AI

‘Embarrassing and wrong’: Google admits it lost control of image-generating AI

Comment

Image Credits: Adobe Firefly generative AI / composite by TechCrunch

Google has apologized (or come very close to apologizing) for another embarrassing AI blunder this week, an image-generating model that injected diversity into pictures with a farcical disregard for historical context. While the underlying issue is perfectly understandable, Google blames the model for “becoming” oversensitive. But the model didn’t make itself, guys.

The AI system in question is Gemini, the company’s flagship conversational AI platform, which when asked calls out to a version of the Imagen 2 model to create images on demand.

Recently, however, people found that asking it to generate imagery of certain historical circumstances or people produced laughable results. For instance, the Founding Fathers, who we know to be white slave owners, were rendered as a multi-cultural group, including people of color.

This embarrassing and easily replicated issue was quickly lampooned by commentators online. It was also, predictably, roped into the ongoing debate about diversity, equity, and inclusion (currently at a reputational local minimum), and seized by pundits as evidence of the woke mind virus further penetrating the already liberal tech sector.

Image Credits: An image generated by Twitter user Patrick Ganley.

It’s DEI gone mad, shouted conspicuously concerned citizens. This is Biden’s America! Google is an “ideological echo chamber,” a stalking horse for the left! (The left, it must be said, was also suitably perturbed by this weird phenomenon.)

But as anyone with any familiarity with the tech could tell you, and as Google explains in its rather abject little apology-adjacent post today, this problem was the result of a quite reasonable workaround for systemic bias in training data.

Say you want to use Gemini to create a marketing campaign, and you ask it to generate 10 pictures of “a person walking a dog in a park.” Because you don’t specify the type of person, dog, or park, it’s dealer’s choice — the generative model will put out what it is most familiar with. And in many cases, that is a product not of reality, but of the training data, which can have all kinds of biases baked in.

What kinds of people, and for that matter dogs and parks, are most common in the thousands of relevant images the model has ingested? The fact is that white people are over-represented in a lot of these image collections (stock imagery, rights-free photography, etc.), and as a result the model will default to white people in a lot of cases if you don’t specify.

That’s just an artifact of the training data, but as Google points out, “because our users come from all over the world, we want it to work well for everyone. If you ask for a picture of football players, or someone walking a dog, you may want to receive a range of people. You probably don’t just want to only receive images of people of just one type of ethnicity (or any other characteristic).”

Illustration of a group of people recently laid off and holding boxes.
Imagine asking for an image like this — what if it was all one type of person? Bad outcome! Image Credits: Getty Images / victorikart

Nothing wrong with getting a picture of a white guy walking a golden retriever in a suburban park. But if you ask for 10, and they’re all white guys walking goldens in suburban parks? And you live in Morocco, where the people, dogs, and parks all look different? That’s simply not a desirable outcome. If someone doesn’t specify a characteristic, the model should opt for variety, not homogeneity, despite how its training data might bias it.

This is a common problem across all kinds of generative media. And there’s no simple solution. But in cases that are especially common, sensitive, or both, companies like Google, OpenAI, Anthropic, and so on invisibly include extra instructions for the model.

I can’t stress enough how commonplace this kind of implicit instruction is. The entire LLM ecosystem is built on implicit instructions — system prompts, as they are sometimes called, where things like “be concise,” “don’t swear,” and other guidelines are given to the model before every conversation. When you ask for a joke, you don’t get a racist joke — because despite the model having ingested thousands of them, it has also been trained, like most of us, not to tell those. This isn’t a secret agenda (though it could do with more transparency), it’s infrastructure.

Where Google’s model went wrong was that it failed to have implicit instructions for situations where historical context was important. So while a prompt like “a person walking a dog in a park” is improved by the silent addition of “the person is of a random gender and ethnicity” or whatever they put, “the U.S. Founding Fathers signing the Constitution” is definitely not improved by the same.

As the Google SVP Prabhakar Raghavan put it:

First, our tuning to ensure that Gemini showed a range of people failed to account for cases that should clearly not show a range. And second, over time, the model became way more cautious than we intended and refused to answer certain prompts entirely — wrongly interpreting some very anodyne prompts as sensitive.

These two things led the model to overcompensate in some cases, and be over-conservative in others, leading to images that were embarrassing and wrong.

I know how hard it is to say “sorry” sometimes, so I forgive Raghavan for stopping just short of it. More important is some interesting language in there: “The model became way more cautious than we intended.”

Now, how would a model “become” anything? It’s software. Someone — Google engineers in their thousands — built it, tested it, iterated on it. Someone wrote the implicit instructions that improved some answers and caused others to fail hilariously. When this one failed, if someone could have inspected the full prompt, they likely would have found the thing Google’s team did wrong.

Google blames the model for “becoming” something it wasn’t “intended” to be. But they made the model! It’s like they broke a glass, and rather than saying “we dropped it,” they say “it fell.” (I’ve done this.)

Mistakes by these models are inevitable, certainly. They hallucinate, they reflect biases, they behave in unexpected ways. But the responsibility for those mistakes does not belong to the models — it belongs to the people who made them. Today that’s Google. Tomorrow it’ll be OpenAI. The next day, and probably for a few months straight, it’ll be X.AI.

These companies have a strong interest in convincing you that AI is making its own mistakes. Don’t let them.

More TechCrunch

According to a recent Dealroom report on the Spanish tech ecosystem, the combined enterprise value of Spanish startups surpassed €100 billion in 2023. In the latest confirmation of this upward trend, Madrid-based…

Spain’s exposure to climate change helps Madrid-based VC Seaya close €300M climate tech fund

Forestay, an emerging VC based out of Geneva, Switzerland, has been busy. This week it closed its second fund, Forestay Capital II, at a hard cap of $220 million. The…

Forestay, Europe’s newest $220M growth-stage VC fund, will focus on AI

Threads, Meta’s alternative to Twitter, just celebrated its first birthday. After launching on July 5 last year, the social network has reached 175 million monthly active users — that’s a…

A year later, what Threads could learn from other social networks

J2 Ventures, a firm led mostly by U.S. military veterans, announced on Thursday that it has raised a $150 million second fund. The Boston-based firm invests in startups whose products…

J2 Ventures, focused on military healthcare, grabs $150M for its second fund

HealthEquity said in an 8-K filing with the SEC that it detected “anomalous behavior by a personal use device belonging to a business partner.”

HealthEquity says data breach is an ‘isolated incident’

Roll20 said that on June 29 it had detected that a “bad actor” gained access to an account on the company’s administrative website for one hour.

Roll20, an online tabletop role-playing game platform, discloses data breach

Fisker has a willing buyer for its remaining inventory of all-electric Ocean SUVs, and has asked the Delaware Bankruptcy Court judge overseeing its Chapter 11 case to approve the sale.…

Fisker asks bankruptcy court to sell its EVs at average of $14,000 each

Teddy Solomon just moved to a new house in Palo Alto, so he turned to the Stanford community on Fizz to furnish his room. “Every time I show up to…

Fizz, the anonymous Gen Z social app, adds a marketplace for college students

With increasing competition for what is, essentially, still a small number of hard tech and deep tech deals, Sidney Scott realized it would be a challenge for smaller funds like…

Why deep tech VC Driving Forces is shutting down

A guide to turn off reactions on your iPhone and Mac so you don’t get surprised by effects during work video calls.

How to turn off those silly video call reactions on iPhone and Mac

Amazon has decided to discontinue its Astro for Business device, a security robot for small- and medium-sized businesses, just seven months after launch.  In an email sent to customers and…

Amazon retires its Astro for Business security robot after only 7 months

Hiya, folks, and welcome to TechCrunch’s regular AI newsletter. This week in AI, the U.S. Supreme Court struck down “Chevron deference,” a 40-year-old ruling on federal agencies’ power that required…

This Week in AI: With Chevron’s demise, AI regulation seems dead in the water

Noplace had already gone viral ahead of its public launch because of its feature that allows users to express themselves by customizing the colors of their profile.

noplace, a mashup of Twitter and Myspace for Gen Z, hits No. 1 on the App Store

Cloudflare analyzed AI bot and crawler traffic to fine-tune automatic bot detection models.

Cloudflare launches a tool to combat AI bots

Twilio says “threat actors were able to identify” phone numbers of people who use the two-factor app Authy.

Twilio says hackers identified cell phone numbers of two-factor app Authy users

The news brings closure to more than two years of volleying back and forth between some of the biggest names in additive manufacturing.

Nano Dimension is buying Desktop Metal

Planning to attend TechCrunch Disrupt 2024 with your team? Maximize your team-building time and your company’s impact across the entire conference when you bring your team. Groups of 4 to…

Groups save big at TechCrunch Disrupt 2024

As more music streaming apps and creation tools emerge to compete for users’ attention, social music-sharing app Popster is getting two new features to grow its user base: an AI…

Music video-sharing app Popster uses generative AI and lets artists remix videos

Meta’s Threads now has more than 175 million monthly active users, Mark Zuckerberg announced on Wednesday. The announcement comes two days away from Threads’ first anniversary. Zuckerberg revealed back in…

Threads nears its one-year anniversary with more than 175M monthly active users

Cartken and its diminutive sidewalk delivery robots first rolled into the world with a narrow charter: carrying everything from burritos and bento boxes to pizza and pad thai that last…

From burritos to biotech: How robotics startup Cartken found its AV niche

Ashwin Nandakumar and Ashwin Jainarayanan were working on their doctorates at adjacent departments in Oxford, but they didn’t know each other. Nandakumar, who was studying oncology, one day stumbled across…

Granza Bio grabs $7M seed from Felicis and YC to advance delivery of cancer treatments

LG has acquired an 80% stake in Athom, a Dutch smart home company and maker of the Homey smart home hub. According to LG’s announcement, it will purchase the remaining…

LG acquires smart home platform Athom to bring third-party connectivity to its ThinQ ecosytem

CoinDCX, India’s leading cryptocurrency exchange, is expanding internationally through the acquisition of BitOasis, a digital asset platform in the Middle East and North Africa, the companies said Wednesday. The Bengaluru-based…

CoinDCX acquires BitOasis in international expansion push

Collaborative document features are being made available inside Proton Drive, further extending the company’s trademark pitch of robust security.

In a major update, Proton adds privacy-safe document collaboration to Drive, its freemium E2EE cloud storage service

Telegram launched a digital currency called Stars for in-app use last month. Now, the company is expanding its use cases to paid content. The chat app is also allowing channels…

Telegram lets creators share paid content to channels

For the past couple of years, innovation has been accelerating in new materials development. And a new French startup called Altrove plans to play a role in this innovation cycle.…

Altrove uses AI models and lab automation to create new materials

The Indian social media platform Koo, which positioned itself as a competitor to Elon Musk’s X, is ceasing operations after its last-resort acquisition talks with Dailyhunt collapsed. Despite securing over…

Indian social network Koo is shutting down as buyout talks collapse

Apiday leverages AI to save time for its customers. But like legacy consultants, it also offers human expertise.

Europe is still serious about ESG, and Apiday is helping companies comply

Google totally dodges the question of how much energy is AI is using — perhaps because the answer is “way more than we’d care to say.”

Google’s environmental report pointedly avoids AI’s actual energy cost

SpaceX’s ambitious plans to launch its Starship mega-rocket up to 44 times per year from NASA’s Kennedy Space Center are causing a stir among some of its competitors. Late last…

SpaceX wants to launch up to 120 times a year from Florida — and competitors aren’t happy about it