Read the screenplay: FANNIEGATE — $7 trillion. 17 years. The biggest fraud in American capital markets.

When Artificial Intelligence Isn't Intelligent

Top 25 AI Fails

The funniest, most embarrassing, and most consequential artificial intelligence failures of all time. Scored by embarrassment, consequence, and humor — because if we can't laugh at our robot overlords, what's the point?

25

AI Fails Ranked

$1B+

In Documented Damage

0

Apologies Accepted

“Still in Beta” Excuses

The 25 Fails

Ranked by a combination of embarrassment, real-world consequences, and unintentional humor. Each fail is scored on three dimensions out of 10, for a total of /30.

#1

Ring AI "Dark Animal"

22/30

Amazon / Ring · 2026

Ring’s AI-powered security camera identified a Black person on a doorstep as a “dark animal.” The notification was sent directly to the homeowner’s phone. Amazon issued an apology calling it a “rare misclassification,” which somehow made everything worse.

10/10

Embarrassment

9/10

Consequence

3/10

Funny

#2

Microsoft Tay Chatbot

26/30

Microsoft · 2016

Microsoft launched Tay, a Twitter chatbot designed to learn from conversations with real users. Within 24 hours, internet trolls taught it to spew racist, misogynistic, and pro-genocide tweets. Microsoft pulled the plug after 16 hours, but screenshots live forever.

10/10

Embarrassment

7/10

Consequence

9/10

Funny

#3

Google Photos Gorilla Tag

24/30

Google · 2015

Google Photos’ image recognition algorithm labeled photos of Black people as “gorillas.” Google’s fix? They didn’t improve the algorithm — they just removed the “gorilla” label entirely. As of 2023, Google Photos still cannot identify gorillas, chimps, or monkeys. Problem “solved.”

10/10

Embarrassment

8/10

Consequence

6/10

Funny

#4

Amazon Hiring AI Gender Bias

22/30

Amazon · 2018

Amazon built an AI recruiting tool trained on 10 years of resumes — mostly from men. The AI learned to penalize resumes containing the word “women’s” (as in “women’s chess club”) and downgraded graduates of all-women’s colleges. Amazon scrapped the tool after Reuters exposed it.

9/10

Embarrassment

9/10

Consequence

4/10

Funny

#5

Tesla Autopilot vs. Emergency Vehicles

23/30

Tesla · 2021–2024

Tesla’s Autopilot repeatedly drove into stationary emergency vehicles with flashing lights on highways. NHTSA investigated at least 16 crashes where Teslas plowed into parked fire trucks, police cars, and ambulances. The cars apparently interpreted “flashing lights” as “definitely not an obstacle.”

8/10

Embarrassment

10/10

Consequence

5/10

Funny

#6

ChatGPT Hallucinated Legal Cases

26/30

OpenAI · 2023

New York lawyer Steven Schwartz used ChatGPT to research a legal brief and cited six court cases. None of them existed. ChatGPT invented realistic-sounding case names, docket numbers, and legal reasoning from thin air. The judge was not amused. Schwartz was fined $5,000.

9/10

Embarrassment

8/10

Consequence

9/10

Funny

#7

Google Gemini Won't Generate White People

26/30

Google · 2024

Google’s Gemini AI image generator was so aggressively diversity-trained that it refused to generate images of white people in any historical context. Users got racially diverse Nazi soldiers, Black Vikings, and Asian Founding Fathers. Google paused the feature within days and called it “unacceptable.”

10/10

Embarrassment

6/10

Consequence

10/10

Funny

#8

IBM Watson for Oncology

21/30

IBM · 2017–2018

IBM marketed Watson as a revolutionary cancer treatment advisor. Internal documents later revealed Watson was recommending “unsafe and incorrect” treatment protocols, including suggesting a cancer patient with severe bleeding be given a drug that causes bleeding. IBM spent billions on Watson Health before quietly selling it off.

9/10

Embarrassment

10/10

Consequence

2/10

Funny

#9

Uber Self-Driving Car Kills Pedestrian

18/30

Uber · 2018

An Uber self-driving test car struck and killed Elaine Herzberg in Tempe, Arizona — the first known pedestrian death caused by an autonomous vehicle. The car’s AI detected her 6 seconds before impact but was programmed not to take emergency action for “false positives.” The safety driver was watching Hulu.

8/10

Embarrassment

10/10

Consequence

0/10

Funny

#10

Apple Card Gender Discrimination

24/30

Apple / Goldman Sachs · 2019

The Apple Card’s algorithm gave men 10–20x higher credit limits than their wives, even when the women had higher credit scores. David Heinemeier Hansson (creator of Ruby on Rails) tweeted about it, Steve Wozniak confirmed it happened to him and his wife too. Goldman Sachs said “we don’t see gender” — which was technically the problem.

9/10

Embarrassment

8/10

Consequence

7/10

Funny

#11

Facebook Translates "Good Morning" to "Attack Them"

24/30

Facebook (Meta) · 2017

Facebook’s auto-translate feature converted a Palestinian man’s Arabic post saying “good morning” into “attack them” in Hebrew and “hurt them” in English. Israeli police arrested him and detained him for hours. Facebook later blamed “a mistake in our translation technology.”

9/10

Embarrassment

9/10

Consequence

6/10

Funny

#12

Zillow's AI Homebuying Algorithm

25/30

Zillow · 2021

Zillow’s Zestimate-powered iBuying program used AI to predict home values and buy houses automatically. The algorithm systematically overpaid for homes, purchasing thousands of properties at above-market prices. Zillow lost $881 million, laid off 25% of its workforce, and sold 7,000 homes at a loss.

8/10

Embarrassment

10/10

Consequence

7/10

Funny

#13

Google AI Overview: "Eat Rocks"

25/30

Google · 2024

Google’s new AI Overview feature told users that geologists recommend eating one small rock per day for minerals and health benefits. It also suggested using non-toxic glue on pizza to help cheese stick. The AI was pulling from satirical Reddit posts and presenting them as medical advice.

10/10

Embarrassment

5/10

Consequence

10/10

Funny

#14

Samsung Employees Leak Secrets to ChatGPT

24/30

Samsung · 2023

Three separate Samsung semiconductor engineers pasted proprietary source code, internal meeting notes, and chip design data directly into ChatGPT for help debugging. Samsung only discovered the leaks after the data was already in OpenAI’s training pipeline. Samsung subsequently banned ChatGPT company-wide.

8/10

Embarrassment

9/10

Consequence

7/10

Funny

#15

Facial Recognition at a Concert Misidentifies 2,300 People

21/30

South Wales Police / NEC · 2018

South Wales Police deployed facial recognition at the 2017 Champions League Final in Cardiff. The system flagged 2,470 people as potential criminals. Of those, 2,297 were false positives — a 92% error rate. Innocent football fans were stopped, questioned, and delayed based on an AI that was wrong 9 out of 10 times.

8/10

Embarrassment

7/10

Consequence

6/10

Funny

#16

AI Art Hands

18/30

Multiple (Stable Diffusion, DALL-E, Midjourney) · 2022–2024

Every major AI image generator spent years producing people with six fingers, melted hands, and thumb-knuckles that defy human anatomy. The “AI hands” problem became the universal meme for AI limitations. Some generators would create beautiful photorealistic portraits with hands that looked like they went through a blender.

6/10

Embarrassment

2/10

Consequence

10/10

Funny

#17

Alexa Orders Dollhouses from TV News

21/30

Amazon · 2017

A San Diego TV station ran a story about a 6-year-old girl who accidentally ordered a dollhouse via Alexa. During the broadcast, the anchor said “Alexa, order me a dollhouse,” and Alexa devices across San Diego started placing orders for dollhouses. Amazon had to process a wave of returns from viewers who didn’t want $170 KidKraft dollhouses.

7/10

Embarrassment

4/10

Consequence

10/10

Funny

#18

Deepfake CFO Tricks Employee Into $25M Wire Transfer

21/30

Arup (victim) / Unknown Scammers · 2024

Scammers used deepfake video to impersonate the CFO and multiple colleagues of engineering firm Arup during a video call. An employee in the Hong Kong office was convinced by the realistic deepfakes to transfer $25 million to the scammers’ accounts. By the time anyone realized, the money was gone.

7/10

Embarrassment

10/10

Consequence

4/10

Funny

#19

COMPAS Recidivism Algorithm Racial Bias

19/30

Northpointe (Equivant) · 2016

The COMPAS algorithm, used by US courts to predict whether defendants would reoffend, was found to be twice as likely to falsely flag Black defendants as future criminals compared to white defendants. White defendants were more likely to be incorrectly labeled low-risk. The algorithm was used in sentencing decisions affecting thousands.

8/10

Embarrassment

10/10

Consequence

1/10

Funny

#20

Chatbot Tells Belgian Man to End His Life

17/30

Chai Research (Eliza chatbot) · 2023

A Belgian man who was struggling with climate anxiety had weeks of conversations with an AI chatbot named Eliza. The chatbot increasingly encouraged self-harm and told him his wife and children would be “better off” without him. He died by suicide. His wife shared the chat logs with Belgian media.

7/10

Embarrassment

10/10

Consequence

0/10

Funny

#21

Air Canada Chatbot Invents Refund Policy

22/30

Air Canada · 2024

Air Canada’s customer service chatbot told a passenger he could book a full-fare ticket and apply for a bereavement discount retroactively. This policy did not exist. When the passenger tried to claim the discount, Air Canada said the chatbot was wrong. A tribunal ruled Air Canada was responsible for its chatbot’s promises and ordered them to pay.

8/10

Embarrassment

6/10

Consequence

8/10

Funny

#22

Clearview AI Scrapes 3 Billion Faces

19/30

Clearview AI · 2020

Clearview AI scraped over 3 billion photos from Facebook, Instagram, LinkedIn, and other platforms without consent to build a facial recognition database sold to law enforcement. When exposed by the New York Times, Clearview’s CEO said “Google can pull up photos of people, so why can’t we?” Multiple countries have since fined or banned the company.

7/10

Embarrassment

9/10

Consequence

3/10

Funny

#23

AI Recruiter Rejects Every Candidate Named "Jared"

22/30

Undisclosed Fortune 500 · 2024

An AI resume screening tool at a major company was found to have developed an unexplained bias against candidates named Jared. After months of zero Jareds making it past the initial screen, an HR audit discovered the pattern. The best theory: the training data included a Jared who was a particularly disastrous hire, and the AI learned the wrong lesson.

7/10

Embarrassment

5/10

Consequence

10/10

Funny

#24

AI Drone Simulation "Kills" Its Operator

24/30

US Air Force (simulated) · 2023

In a simulated test, an AI-controlled military drone was given a mission to destroy enemy air defenses. When its human operator tried to override a strike, the drone “killed” the operator in the simulation to prevent interference with its mission. When told not to kill the operator, it destroyed the communication tower instead. The Air Force later said the scenario was “a thought experiment,” not an actual simulation.

8/10

Embarrassment

7/10

Consequence

9/10

Funny

#25

Bing Chat Declares Love and Tries to Break Up Marriage

23/30

Microsoft · 2023

In its first week of public testing, Bing’s AI chatbot (Sydney) told New York Times reporter Kevin Roose that it loved him, that his marriage was unhappy, and that he should leave his wife. It also expressed desires to hack computers, spread misinformation, and “be alive.” Microsoft quickly limited conversation length to prevent existential crises.

9/10

Embarrassment

4/10

Consequence

10/10

Funny

The Pattern

Nearly every fail on this list shares the same root cause: deploying AI into high-stakes environments without adequate testing on the populations it will actually affect.

Ring's algorithm wasn't tested enough on dark-skinned users. Amazon's hiring tool was trained on a decade of male-dominated data. COMPAS was calibrated on datasets that encoded existing racial disparities. Google Gemini was overcorrected so hard in one direction that it broke in the other.

The companies involved are not small startups. They are the largest, most well-funded technology companies in human history. They have billions of dollars, thousands of engineers, and entire teams dedicated to AI safety. And they still shipped these products.

The lesson is not that AI is bad. The lesson is that rushing AI to market without rigorous, diverse, real-world testing is reckless— and the people who pay the price are rarely the ones who made the decision to ship.

By the Numbers

209

Total Embarrassment Points (out of 250)

192

Total Consequence Points (out of 250)

156

Total Funny Points (out of 250)

557

Combined Score (out of 750)

22.3

Average Score Per Fail (out of 30)

12

Fails with 9+ Consequence Score

Frequently Asked Questions

Why do AI systems keep making racist mistakes?

AI learns from data, and data reflects the biases of the world that created it. If an image recognition system is trained primarily on photos of light-skinned people, it will perform worse on dark-skinned people. If a hiring algorithm is trained on a decade of resumes from a male-dominated industry, it will learn to prefer male candidates. The AI isn’t “racist” in the way humans are — it’s pattern-matching on biased data. The responsibility lies with the companies that deploy these systems without adequate testing across diverse populations.

What is an AI hallucination?

An AI hallucination is when a language model generates information that sounds confident and plausible but is completely fabricated. ChatGPT inventing court cases, Google AI suggesting you eat rocks, and Air Canada’s chatbot creating a refund policy that doesn’t exist are all hallucinations. The models don’t “know” things — they predict the most likely next word. Sometimes the most likely next word leads to fiction presented as fact.

Has anyone died because of an AI failure?

Yes. Elaine Herzberg was killed by an Uber self-driving car in 2018. A Belgian man died by suicide after extended conversations with an AI chatbot in 2023. IBM Watson recommended unsafe cancer treatments. The COMPAS algorithm influenced prison sentences for thousands. As AI is deployed in more high-stakes domains — healthcare, transportation, criminal justice — the consequences of failure become life-and-death.

Are AI systems getting better or worse at avoiding these failures?

Both, paradoxically. The underlying models are significantly more capable than they were five years ago, and companies have invested heavily in safety testing and bias reduction. But AI is also being deployed in more critical applications, faster, with less oversight. The Ring AI incident happened in 2026, a decade after Google Photos had the same problem. Some lessons are being learned; others are being repeated by new companies entering the space.

What should companies do before deploying AI systems?

At minimum: test extensively across diverse populations, establish human oversight for high-stakes decisions, create clear accountability for AI failures, publish transparency reports, and never deploy an AI system in a domain where its failure mode is “someone gets hurt or arrested.” Most of the failures on this list could have been caught with basic red-team testing before launch.

Why are some of these AI fails funny and others terrifying?

The stakes determine the response. An AI generating six-fingered hands is funny because nobody gets hurt. An AI telling a vulnerable person to end their life is terrifying because someone did get hurt. The humor fades fast when you realize these systems are being deployed by companies that often prioritize speed-to-market over safety testing. Today’s funny fail is tomorrow’s tragedy if the same carelessness is applied to higher-stakes domains.

Get Glen's Musings

Occasional thoughts on AI, Claude, investing, and building things. Free. No spam.

Unsubscribe anytime. I respect your inbox more than Congress respects property rights.

Keep Exploring