Startups

Building better startups with responsible AI

Comment

Digital generated image of split net/turbulence structure of artificial intelligence brain on purple surface.
Image Credits: Andriy Onufriyenko (opens in a new window) / Getty Images

Tom Zick

Contributor

Tom Zick is a researcher in AI ethics at the Berkman Klein Center for Internet and Society at Harvard University, where she is also a J.D. candidate. She holds a Ph.D. from UC Berkeley and was previously a fellow at Bloomberg Beta and the City of Boston.

Founders tend to think responsible AI practices are challenging to implement and may slow the progress of their business. They often jump to mature examples like Salesforce’s Office of Ethical and Humane Use and think that the only way to avoid creating a harmful product is building a big team. The truth is much simpler.

I set out to learn how founders were thinking about responsible AI practices on the ground by speaking with a handful of successful early-stage founders and found many of them were implementing responsible AI practices.

Only they didn’t call it that. They just call it “good business.”

It turns out, simple practices that make business sense and result in better products will go a long way toward reducing the risk of unforeseen societal harms. These practices rely on the insight that people, not data, are at the heart of deploying an AI solution successfully. If you account for the reality that humans are always in the loop, you can build a better business, more responsibly.

Think of AI as a bureaucracy. Like a bureaucracy, AI relies on having some general policy to follow (“the model”) that makes reasonable decisions in most cases. However, this general policy can never account for all possible scenarios a bureaucracy will need to handle — much like an AI model cannot be trained to anticipate every possible input.

When these general policies (or models) fail, those who are already marginalized are disproportionately impacted (a classic algorithmic example is of Somali immigrants being tagged for fraud because of their atypical community shopping habits).

Bureaucracies work to solve this problem with “street-level bureaucrats” like judges, DMV agents and even teachers, who can handle unique cases or decide not to enforce the policy. For example, teachers can waive a course prerequisite given extenuating circumstances, or judges can be more or less lenient in sentencing.

If any AI will inevitably fail, then — like with a bureaucracy — we must keep humans in the loop and design with them in mind. As one founder told me, “If I were a Martian coming to Earth for the first time, I would think: Humans are processing machines — I should use them.”

Whether the humans are operators augmenting the AI system by stepping in when it’s uncertain, or users choosing whether to reject, accept or manipulate a model outcome, these people determine how well any AI-based solution will work in the real world.

Here are five practical suggestions that founders of AI companies shared for keeping, and even harnessing, humans in the loop to build a more responsible AI that’s also good for business:

Introduce only as little AI as you need

Today, many companies plan to launch some services with an end-to-end AI-driven process. When those processes struggle to function under a wide range of use cases, the people who are most harmed tend to be those already marginalized.

In trying to diagnose failures, founders subtract one component at a time, still hoping to automate as much as possible. They should consider the opposite: introducing one AI component at a time.

Many processes are — even with all the wonders of AI — still just less expensive and more reliable to run with humans in the loop. If you build an end-to-end system with many components coming online at once, you may find it hard to identify which are best suited to AI.

Many founders we spoke with view AI as a way to delegate the most time-consuming, low-stakes tasks in their system away from humans, and they started with all human-run systems to identify what these important-to-automate tasks were.

This “AI second” approach also enables founders to enter fields where data is not immediately available. The people who operate parts of a system also create the very data you’ll need to automate those tasks. One founder told us that, without the advice to introduce AI gradually, and only when it was demonstrably more accurate than an operator, they would have never gotten off the ground.

Create some friction

Many founders believe that to be successful, a product must run out of the box, with as little user input as possible.

Because AI is typically used to automate part of an existing workflow — complete with associated preconceptions on how much to trust that workflow output — a perfectly seamless approach can be catastrophic.

For example, when an ACLU audit showed that Amazon’s facial recognition tool would misidentify 28 members of Congress (a disproportionately large fraction of whom were Black) as criminals, lax default settings were at the heart of the problem. The accuracy threshold out of the box was set to only 80%, clearly the wrong setting if a user takes a positive result at face value.

Motivating users to engage with a product’s strengths and weaknesses before deploying it can offset the potential for harmful assumption mismatches. It can also make customers happier with eventual product performance.

One founder we spoke with found that customers ultimately used their product more effectively if the customer had to customize it before use. He views this as a dominant component of a “design-first” approach and found it helped users play to the strengths of the product on a context-specific basis. While this approach required more upfront time to get going, it ended up translating into revenue gains for customers.

Give context, not answers

Many AI-based solutions focus on providing an output recommendation. Once these recommendations are made, they have to be acted on by humans.

Without context, poor recommendations could be blindly followed, causing downstream harm. Similarly, great recommendations could be rejected if the humans in the loop do not trust the system and lack context.

Rather than delegating decisions away from users, consider giving them the tools to make decisions. This approach harnesses the power of humans in the loop to identify problematic model outputs while securing the user buy-in necessary for a successful product.

One founder shared that when their AI made direct recommendations, users didn’t trust it. Their customers were happy with the accuracy that their model predictions turned out to have, but individual users just ignored the recommendations. Then they nixed the recommendation feature and instead used their model to augment the resources that could inform a user’s decision (e.g., this procedure is like these five past procedures and here is what worked). This led to increased adoption rates and revenue.

Consider your not-users and not-buyers

It is a known problem in enterprise tech that products can easily serve the CEO and not the end users. This is even more problematic in the AI space, where a solution is often part of a greater system that interfaces with a few direct users and many more indirect ones.

Take, for example, the controversy that arose when Starbucks began using automated scheduling software to assign shifts. The scheduler optimized for efficiency, completely disregarding working conditions. After a successful labor petition and a high-profile New York Times article, the baristas’ input was taken under consideration, improving morale and productivity.

Instead of taking a customer literally on what they ask you to solve, consider mapping out all of the stakeholders involved and understanding their needs before you decide what your AI will help optimize. That way, you will avoid inadvertently making a product that is needlessly harmful and possibly find an even better business opportunity.

One founder we spoke with took this approach to heart, camping out next to their users to understand their needs before deciding what to optimize their product for. They followed this up by meeting with both customers and union representatives to figure out how to make a product that worked for both.

While customers originally wanted a product that would allow each user to take on a greater workload, these conversations revealed an opportunity to unlock savings for their customers by optimizing the existing workload.

This insight allowed the founder to develop a product that empowered the humans in the loop and saved management more money than the solution they thought they wanted would have.

Be clear on what’s AI theater

If you limit the degree to which you hype up what your AI can do, you can both avoid irresponsible consequences and sell your product more effectively.

Yes, the hype around AI helps sell products. However, knowing how to keep those buzzwords from getting in the way of precision is crucial. While talking up the autonomous capabilities of your product might be good for sales, it can backfire if you apply that rhetoric indiscriminately.

For example, one of the founders we spoke to found that playing up the power of their AI also increased their customers’ privacy concerns. This concern persisted even when the founders explained that the portions of the product in question did not rely on data, but rather on human judgment.

Language choice can help align expectations and build trust in a product. Rather than using the language of autonomy with their users, some of the founders we talked to found that words like “augment” and “assist” were more likely to inspire adoption. This “AI as a tool” framing was also less likely to engender the blind trust that can lead to bad outcomes down the line. Being clear can both dissuade overconfidence in AI and help you sell.

These are some practical lessons learned by real founders for mitigating the risk of unforeseen harms from AI and creating more successful products built for the long term. We also believe there’s an opportunity for new startups to build services that help make it easier to create ethical AI that’s also good for business. So here are a couple of requests for startups:

  • Engage humans in the loop: We need startups that solve the “human in the loop” attention problem. Delegating to humans requires making sure those humans notice when an AI is uncertain so that they can meaningfully intervene. If an AI is correct 95% of the time, research shows that people get complacent and are unlikely to catch the 5% of instances the AI gets wrong. The solution requires more than just technology; much like social media was more of a psychological innovation than a technical one, we think startups in this space can (and should) emerge from social insights.
  • Standard compliance for responsible AI: There’s opportunity for startups that consolidate existing standards around responsible AI and measure compliance. Publication of AI standards has been on the rise in the past two years as public pressure on AI regulation has been increasing. A recent survey showed 84% of Americans think AI should be carefully managed and rate this as a top priority. Companies want to signal they are taking this seriously and showing they are following standards put forth by IEEE, CSET and others would be useful. Meanwhile, the current draft of the EU’s expansive AI Act (AIA) strongly emphasizes industry standards. If the AIA passes, compliance will become a necessity. Given the market that formed around GDPR compliance, we think this is a space to watch.

Whether you’re trying one of these tips or starting one of these companies, simple, responsible AI practices can let you unlock immense business opportunities. To avoid creating a harmful product, you need to be thoughtful in your deployment of AI.

Luckily, this thoughtfulness will pay dividends when it comes to the long-term success of your business.

More TechCrunch

Featured Article

Amazon buys Indian video streaming service MX Player

Amazon has agreed to acquire Indian video streaming service MX Player from the local media powerhouse Times Internet, the latest step by the e-commerce giant to make its services and brand popular in smaller cities and towns in the key overseas market.  The two firms reached a definitive agreement for…

21 mins ago
Amazon buys Indian video streaming service MX Player

Dealt is now building a service platform for retailers instead of end customers.

Dealt turns retailers into service providers and proves that pivots sometimes work

Snowflake is the latest company in a string of high-profile security incidents and sizable data breaches caused by the lack of MFA.

Hundreds of Snowflake customer passwords found online are linked to info-stealing malware

The buy will benefit ChromeOS, Google’s lightweight Linux-based operating system, by giving ChromeOS users greater access to Windows apps “without the hassle of complex installations or updates.”

Google acquires Cameyo to bring Windows apps to ChromeOS

Mistral is no doubt looking to grow revenue as it faces considerable — and growing — competition in the generative AI space.

Mistral launches new services and SDK to let customers fine-tune its models

The warning for the Ai Pin was issued “out of an abundance of caution,” according to Humane.

Humane urges customers to stop using charging case, citing battery fire concerns

The keynote will be focused on Apple’s software offerings and the developers that power them, including the latest versions of iOS, iPadOS, macOS, tvOS, visionOS and watchOS.

Watch Apple kick off WWDC 2024 right here

As WWDC 2024 nears, all sorts of rumors and leaks have emerged about what iOS 18 and its AI-powered apps and features have in store.

What to expect from Apple’s AI-powered iOS 18 at WWDC 2024

Welcome to Elon Musk’s X. The social network formerly known as Twitter where the rules are made up and the check marks don’t matter. Or do they? The Tesla and…

Elon Musk’s X: A complete timeline of what Twitter has become

TechCrunch has kept readers informed regarding Fearless Fund’s courtroom battle to provide business grants to Black women. Today, we are happy to announce that Fearless Fund CEO and co-founder Arian…

Fearless Fund’s Arian Simone coming to Disrupt 2024

Bridgy Fed is one of the efforts aimed at connecting the fediverse with the web, Bluesky and, perhaps later, other networks like Nostr.

Bluesky and Mastodon users can now talk to each other with Bridgy Fed

Zoox, Amazon’s self-driving unit, is bringing its autonomous vehicles to more cities.  The self-driving technology company announced Wednesday plans to begin testing in Austin and Miami this summer. The two…

Zoox to test self-driving cars in Austin and Miami 

Called Stable Audio Open, the generative model takes a text description and outputs a recording up to 47 seconds in length.

Stability AI releases a sound generator

It’s not just instant-delivery startups that are struggling. Oda, the Norway-based online supermarket delivery startup, has confirmed layoffs of 150 jobs as it drastically scales back its expansion ambitions to…

SoftBank-backed grocery startup Oda lays off 150, resets focus on Norway and Sweden

Newsletter platform Substack is introducing the ability for writers to send videos to their subscribers via Chat, its private community feature, the company announced on Wednesday. The rollout of video…

Substack brings video to its Chat feature

Hiya, folks, and welcome to TechCrunch’s inaugural AI newsletter. It’s truly a thrill to type those words — this one’s been long in the making, and we’re excited to finally…

This Week in AI: Ex-OpenAI staff call for safety and transparency

Ms. Rachel isn’t a household name, but if you spend a lot of time with toddlers, she might as well be a rockstar. She’s like Steve from Blues Clues for…

Cameo fumbles on Ms. Rachel fundraiser as fans receive credits instead of videos  

Cartwheel helps animators go from zero to basic movement, so creating a scene or character with elementary motions like taking a step, swatting a fly or sitting down is easier.

Cartwheel generates 3D animations from scratch to power up creators

The new tool, which is set to arrive in Wix’s app builder tool this week, guides users through a chatbot-like interface to understand the goals, intent and aesthetic of their…

Wix’s new tool taps AI to generate smartphone apps

ClickUp Knowledge Management combines a new wiki-like editor and with a new AI system that can also bring in data from Google Drive, Dropbox, Confluence, Figma and other sources.

ClickUp wants to take on Notion and Confluence with its new AI-based Knowledge Base

New York City, home to over 60,000 gig delivery workers, has been cracking down on cheap, uncertified e-bikes that have resulted in battery fires across the city.  Some e-bike providers…

Whizz wants to own the delivery e-bike subscription space, starting with NYC

This is the last major step before Starliner can be certified as an operational crew system, and the first Starliner mission is expected to launch in 2025. 

Boeing’s Starliner astronaut capsule is en route to the ISS 

TechCrunch Disrupt 2024 in San Francisco is the must-attend event for startup founders aiming to make their mark in the tech world. This year, founders have three exciting ways to…

Three ways founders can shine at TechCrunch Disrupt 2024

Google’s newest startup program, announced on Wednesday, aims to bring AI technology to the public sector. The newly launched “Google for Startups AI Academy: American Infrastructure” will offer participants hands-on…

Google’s new startup program focuses on bringing AI to public infrastructure

eBay’s newest AI feature allows sellers to replace image backgrounds with AI-generated backdrops. The tool is now available for iOS users in the U.S., U.K., and Germany. It’ll gradually roll…

eBay debuts AI-powered background tool to enhance product images

If you’re anything like me, you’ve tried every to-do list app and productivity system, only to find yourself giving up sooner rather than later because managing your productivity system becomes…

Hoop uses AI to automatically manage your to-do list

Asana is using its work graph to train LLMs with the goal of creating AI assistants that work alongside human employees in company workflows.

Asana introduces ‘AI teammates’ designed to work alongside human employees

Taloflow, an early stage startup changing the way companies evaluate and select software, has raised $1.3M in a seed round.

Taloflow puts AI to work on software vendor selection to reduce costs and save time

The startup is hoping its durable filters can make metals refining and battery recycling more efficient, too.

SiTration uses silicon wafers to reclaim critical minerals from mining waste

Spun out of Bosch, Dive wants to change how manufacturers use computer simulations by both using modern mathematical approaches and cloud computing.

Dive goes cloud-native for its computational fluid dynamics simulation service