Startups

Building better startups with responsible AI

Comment

Digital generated image of split net/turbulence structure of artificial intelligence brain on purple surface.
Image Credits: Andriy Onufriyenko (opens in a new window) / Getty Images

Tom Zick

Contributor

Tom Zick is a researcher in AI ethics at the Berkman Klein Center for Internet and Society at Harvard University, where she is also a J.D. candidate. She holds a Ph.D. from UC Berkeley and was previously a fellow at Bloomberg Beta and the City of Boston.

Founders tend to think responsible AI practices are challenging to implement and may slow the progress of their business. They often jump to mature examples like Salesforce’s Office of Ethical and Humane Use and think that the only way to avoid creating a harmful product is building a big team. The truth is much simpler.

I set out to learn how founders were thinking about responsible AI practices on the ground by speaking with a handful of successful early-stage founders and found many of them were implementing responsible AI practices.

Only they didn’t call it that. They just call it “good business.”

It turns out, simple practices that make business sense and result in better products will go a long way toward reducing the risk of unforeseen societal harms. These practices rely on the insight that people, not data, are at the heart of deploying an AI solution successfully. If you account for the reality that humans are always in the loop, you can build a better business, more responsibly.

Think of AI as a bureaucracy. Like a bureaucracy, AI relies on having some general policy to follow (“the model”) that makes reasonable decisions in most cases. However, this general policy can never account for all possible scenarios a bureaucracy will need to handle — much like an AI model cannot be trained to anticipate every possible input.

When these general policies (or models) fail, those who are already marginalized are disproportionately impacted (a classic algorithmic example is of Somali immigrants being tagged for fraud because of their atypical community shopping habits).

Bureaucracies work to solve this problem with “street-level bureaucrats” like judges, DMV agents and even teachers, who can handle unique cases or decide not to enforce the policy. For example, teachers can waive a course prerequisite given extenuating circumstances, or judges can be more or less lenient in sentencing.

If any AI will inevitably fail, then — like with a bureaucracy — we must keep humans in the loop and design with them in mind. As one founder told me, “If I were a Martian coming to Earth for the first time, I would think: Humans are processing machines — I should use them.”

Whether the humans are operators augmenting the AI system by stepping in when it’s uncertain, or users choosing whether to reject, accept or manipulate a model outcome, these people determine how well any AI-based solution will work in the real world.

Here are five practical suggestions that founders of AI companies shared for keeping, and even harnessing, humans in the loop to build a more responsible AI that’s also good for business:

Introduce only as little AI as you need

Today, many companies plan to launch some services with an end-to-end AI-driven process. When those processes struggle to function under a wide range of use cases, the people who are most harmed tend to be those already marginalized.

In trying to diagnose failures, founders subtract one component at a time, still hoping to automate as much as possible. They should consider the opposite: introducing one AI component at a time.

Many processes are — even with all the wonders of AI — still just less expensive and more reliable to run with humans in the loop. If you build an end-to-end system with many components coming online at once, you may find it hard to identify which are best suited to AI.

Many founders we spoke with view AI as a way to delegate the most time-consuming, low-stakes tasks in their system away from humans, and they started with all human-run systems to identify what these important-to-automate tasks were.

This “AI second” approach also enables founders to enter fields where data is not immediately available. The people who operate parts of a system also create the very data you’ll need to automate those tasks. One founder told us that, without the advice to introduce AI gradually, and only when it was demonstrably more accurate than an operator, they would have never gotten off the ground.

Create some friction

Many founders believe that to be successful, a product must run out of the box, with as little user input as possible.

Because AI is typically used to automate part of an existing workflow — complete with associated preconceptions on how much to trust that workflow output — a perfectly seamless approach can be catastrophic.

For example, when an ACLU audit showed that Amazon’s facial recognition tool would misidentify 28 members of Congress (a disproportionately large fraction of whom were Black) as criminals, lax default settings were at the heart of the problem. The accuracy threshold out of the box was set to only 80%, clearly the wrong setting if a user takes a positive result at face value.

Motivating users to engage with a product’s strengths and weaknesses before deploying it can offset the potential for harmful assumption mismatches. It can also make customers happier with eventual product performance.

One founder we spoke with found that customers ultimately used their product more effectively if the customer had to customize it before use. He views this as a dominant component of a “design-first” approach and found it helped users play to the strengths of the product on a context-specific basis. While this approach required more upfront time to get going, it ended up translating into revenue gains for customers.

Give context, not answers

Many AI-based solutions focus on providing an output recommendation. Once these recommendations are made, they have to be acted on by humans.

Without context, poor recommendations could be blindly followed, causing downstream harm. Similarly, great recommendations could be rejected if the humans in the loop do not trust the system and lack context.

Rather than delegating decisions away from users, consider giving them the tools to make decisions. This approach harnesses the power of humans in the loop to identify problematic model outputs while securing the user buy-in necessary for a successful product.

One founder shared that when their AI made direct recommendations, users didn’t trust it. Their customers were happy with the accuracy that their model predictions turned out to have, but individual users just ignored the recommendations. Then they nixed the recommendation feature and instead used their model to augment the resources that could inform a user’s decision (e.g., this procedure is like these five past procedures and here is what worked). This led to increased adoption rates and revenue.

Consider your not-users and not-buyers

It is a known problem in enterprise tech that products can easily serve the CEO and not the end users. This is even more problematic in the AI space, where a solution is often part of a greater system that interfaces with a few direct users and many more indirect ones.

Take, for example, the controversy that arose when Starbucks began using automated scheduling software to assign shifts. The scheduler optimized for efficiency, completely disregarding working conditions. After a successful labor petition and a high-profile New York Times article, the baristas’ input was taken under consideration, improving morale and productivity.

Instead of taking a customer literally on what they ask you to solve, consider mapping out all of the stakeholders involved and understanding their needs before you decide what your AI will help optimize. That way, you will avoid inadvertently making a product that is needlessly harmful and possibly find an even better business opportunity.

One founder we spoke with took this approach to heart, camping out next to their users to understand their needs before deciding what to optimize their product for. They followed this up by meeting with both customers and union representatives to figure out how to make a product that worked for both.

While customers originally wanted a product that would allow each user to take on a greater workload, these conversations revealed an opportunity to unlock savings for their customers by optimizing the existing workload.

This insight allowed the founder to develop a product that empowered the humans in the loop and saved management more money than the solution they thought they wanted would have.

Be clear on what’s AI theater

If you limit the degree to which you hype up what your AI can do, you can both avoid irresponsible consequences and sell your product more effectively.

Yes, the hype around AI helps sell products. However, knowing how to keep those buzzwords from getting in the way of precision is crucial. While talking up the autonomous capabilities of your product might be good for sales, it can backfire if you apply that rhetoric indiscriminately.

For example, one of the founders we spoke to found that playing up the power of their AI also increased their customers’ privacy concerns. This concern persisted even when the founders explained that the portions of the product in question did not rely on data, but rather on human judgment.

Language choice can help align expectations and build trust in a product. Rather than using the language of autonomy with their users, some of the founders we talked to found that words like “augment” and “assist” were more likely to inspire adoption. This “AI as a tool” framing was also less likely to engender the blind trust that can lead to bad outcomes down the line. Being clear can both dissuade overconfidence in AI and help you sell.

These are some practical lessons learned by real founders for mitigating the risk of unforeseen harms from AI and creating more successful products built for the long term. We also believe there’s an opportunity for new startups to build services that help make it easier to create ethical AI that’s also good for business. So here are a couple of requests for startups:

  • Engage humans in the loop: We need startups that solve the “human in the loop” attention problem. Delegating to humans requires making sure those humans notice when an AI is uncertain so that they can meaningfully intervene. If an AI is correct 95% of the time, research shows that people get complacent and are unlikely to catch the 5% of instances the AI gets wrong. The solution requires more than just technology; much like social media was more of a psychological innovation than a technical one, we think startups in this space can (and should) emerge from social insights.
  • Standard compliance for responsible AI: There’s opportunity for startups that consolidate existing standards around responsible AI and measure compliance. Publication of AI standards has been on the rise in the past two years as public pressure on AI regulation has been increasing. A recent survey showed 84% of Americans think AI should be carefully managed and rate this as a top priority. Companies want to signal they are taking this seriously and showing they are following standards put forth by IEEE, CSET and others would be useful. Meanwhile, the current draft of the EU’s expansive AI Act (AIA) strongly emphasizes industry standards. If the AIA passes, compliance will become a necessity. Given the market that formed around GDPR compliance, we think this is a space to watch.

Whether you’re trying one of these tips or starting one of these companies, simple, responsible AI practices can let you unlock immense business opportunities. To avoid creating a harmful product, you need to be thoughtful in your deployment of AI.

Luckily, this thoughtfulness will pay dividends when it comes to the long-term success of your business.

More TechCrunch

India’s mobile payments regulator is likely to extend the deadline for imposing market share caps on the popular UPI payments rail by one to two years, sources familiar with the…

India weighs delaying caps on UPI market share in win for PhonePe, Google Pay

Line Man Wongnai, an on-demand food delivery service in Thailand, is considering an initial public offering on a Thai exchange or the U.S. in 2025.

Thai food delivery app Line Man Wongnai weighs IPO in Thailand, US in 2025

The problem is not the media, but the message.

Apple’s ‘Crush’ ad is disgusting

Ever wonder why conversational AI like ChatGPT says “Sorry, I can’t do that” or some other polite refusal? OpenAI is offering a limited look at the reasoning behind its own…

OpenAI offers a peek behind the curtain of its AI’s secret instructions

The federal government agency responsible for granting patents and trademarks is alerting thousands of filers whose private addresses were exposed following a second data spill in as many years. The…

US Patent and Trademark Office confirms another leak of filers’ address data

As part of an investigation into people involved in the pro-independence movement in Catalonia, the Spanish police obtained information from the encrypted services Wire and Proton, which helped the authorities…

Encrypted services Apple, Proton and Wire helped Spanish police identify activist

Match Group, the company that owns several dating apps, including Tinder and Hinge, released its first-quarter earnings report on Tuesday, which shows that Tinder’s paying user base has decreased for…

Match looks to Hinge as Tinder fails

Private social networking is making a comeback. Gratitude Plus, a startup that aims to shift social media in a more positive direction, is expanding its wellness-focused, personal reflections journal to…

Gratitude Plus makes social networking positive, private and personal

With venture totals slipping year-over-year in key markets like the United States, and concern that venture firms themselves are struggling to raise more capital, founders might be worried. After all,…

Can AI help founders fundraise more quickly and easily?

Google has found a way to bring a variation of its clever “Circle to Search” gesture to iPhone users. The new interaction, launched in January, allows Android users to search…

Google brings a variation on ‘Circle to Search’ to iPhone users

A new sculpture going live on Wednesday in the Flatiron South Public Plaza in New York is not your typical artwork. It combines technology, sociology, anthropology and art to let…

Always-on video portal lets people in NYC and Dublin interact in real time

Apple’s iPad event had a lot to like. New iPads with new chips and new sizes, a new Apple Pencil, and even some software updates. If you are a big…

TechCrunch Minute: When did iPads get as expensive as MacBooks?

Autonomous, AI-based players are coming to a gaming experience near you, and a new startup, Altera, is joining the fray to build this new guard of AI agents. The company announced…

Bye-bye bots: Altera’s game-playing AI agents get backing from Eric Schmidt

Google DeepMind has taken the wraps off a new version of AlphaFold, their transformative machine learning model that predicts the shape and behavior of proteins. AlphaFold 3 is not only…

Google DeepMind debuts huge AlphaFold update and free proteomics-as-a-service web app

Uber plans to deliver more perks to Uber One members, like member-exclusive events, in a bid to gain more revenue through subscriptions.  “You will see more member-exclusives coming up where…

Uber promises member exclusives as Uber One passes $1B run-rate

We’ve all seen them. The inspector with a clipboard, walking around a building, ticking off the last time the fire extinguishers were checked, or if all the lights are working.…

Checkfirst raises $1.5M pre-seed to apply AI to remote inspections and audits

Close to a decade ago, brothers Aviv and Matteo Shapira co-founded a company, Replay, that created a video format for 360-degree replays — the sorts of replays that have become…

Controversial drone company Xtend leans into defense with new $40 million round

Usually, when something starts to rot, it gets pitched in the trash. But Joanne Rodriguez wants to turn the concept of rot on its head by growing fungus on trash…

Mycocycle uses mushrooms to upcycle old tires and construction waste

Monzo has raised another £150 million ($190 million), as the challenger bank looks to expand its presence internationally — particularly in the U.S. The new round comes just two months…

UK challenger bank Monzo nabs another $190M as US expansion beckons

iRobot has announced the successor to longtime CEO, Colin Angle. Gary Cohen, who previous held chief executive role at Timex and Qualitor Automotive, will be heading up the company, marking a major…

iRobot names former Timex head Gary Cohen as CEO

Reddit — now a publicly-traded company with more scrutiny on revenue growth — is putting a big focus on boosting its international audience, starting with francophones. In their first-ever earnings…

Reddit tests automatic, whole-site translation into French using LLM-based AI

Mushrooms continue to be a big area for alternative proteins. Canada-based Maia Farms recently raised $1.7 million to develop a blend of mushroom and plant-based protein using biomass fermentation. There’s…

Meati Foods bites into another $100M amid growth to 7,000 retail locations

Cleaning the outside of buildings is a dirty job, and it’s also dangerous. Lucid Bots came on the scene in 2018 with its Sherpa line of drones to clean windows…

Lucid Bots secures $9M for drones to clean more than your windows

High interest rates and financial pressures make it more important than ever for finance teams to have a better handle on their cash flow, and several startups are hoping to…

Israeli startup Panax raises a $10M Series A for its AI-driven cash flow management platform

The European Union has deepened the investigation of Elon Musk-owned social network, X, that it opened back in December under the bloc’s online governance and content moderation rulebook, the Digital Services Act…

EU grills Elon Musk’s X about content moderation and deepfake risks

For the founders of Atlan, a data governance startup, data has always been at the heart of what they do, even before they launched the company. In fact, co-founders Prukalpa…

Atlan scores $105M for its data control plane, as LLMs boost importance of data

It is estimated that about 2 billion people, especially those in lower and middle-income countries, lack access to quality and affordable essential medicines. The situation is exacerbated by low-quality or even killer…

Axmed raises $2M from Founderful to streamline drug supply chains in underserved markets

For decades, the Global Positioning System (GPS) has maintained a de facto monopoly on positioning, navigation and timing, because it’s cheap and already integrated into billions of devices around the…

Xona Space Systems closes $19M Series A to build out ultra-accurate GPS alternative

Bankruptcy lawyers representing customers impacted by the dramatic crash of cryptocurrency exchange FTX 17 months ago say that the vast majority of victims will receive their money back — plus interest. The…

FTX crypto fraud victims to get their money back — plus interest

On Wednesday, Google launched its digital wallet in India with local integrations, nearly two years after the app was relaunched as a digital wallet platform in the U.S. As TechCrunch exclusively reported last month,…

Google Wallet is now available in India