AI Mistakes Are Very Different from Human Mistakes

Humans make mistakes all the time. All of us do, every day, in tasks both new and routine. Some of our mistakes are minor and some are catastrophic. Mistakes can break trust with our friends, lose the confidence of our bosses, and sometimes be the difference between life and death.

Over the millennia, we have created security systems to deal with the sorts of mistakes humans commonly make. These days, casinos rotate their dealers regularly, because they make mistakes if they do the same task for too long. Hospital personnel write on limbs before surgery so that doctors operate on the correct body part, and they count surgical instruments to make sure none were left inside the body. From copyediting to double-entry bookkeeping to appellate courts, we humans have gotten really good at correcting human mistakes.

Humanity is now rapidly integrating a wholly different kind of mistake-maker into society: AI. Technologies like large language models (LLMs) can perform many cognitive tasks traditionally fulfilled by humans, but they make plenty of mistakes. It seems ridiculous when chatbots tell you to eat rocks or add glue to pizza. But it’s not the frequency or severity of AI systems’ mistakes that differentiates them from human mistakes. It’s their weirdness. AI systems do not make mistakes in the same ways that humans do.

Much of the friction—and risk—associated with our use of AI arise from that difference. We need to invent new security systems that adapt to these differences and prevent harm from AI mistakes.

Human Mistakes vs AI Mistakes

Life experience makes it fairly easy for each of us to guess when and where humans will make mistakes. Human errors tend to come at the edges of someone’s knowledge: Most of us would make mistakes solving calculus problems. We expect human mistakes to be clustered: A single calculus mistake is likely to be accompanied by others. We expect mistakes to wax and wane, predictably depending on factors such as fatigue and distraction. And mistakes are often accompanied by ignorance: Someone who makes calculus mistakes is also likely to respond “I don’t know” to calculus-related questions.

To the extent that AI systems make these human-like mistakes, we can bring all of our mistake-correcting systems to bear on their output. But the current crop of AI models—particularly LLMs—make mistakes differently.

AI errors come at seemingly random times, without any clustering around particular topics. LLM mistakes tend to be more evenly distributed through the knowledge space. A model might be equally likely to make a mistake on a calculus question as it is to propose that cabbages eat goats.

And AI mistakes aren’t accompanied by ignorance. A LLM will be just as confident when saying something completely wrong—and obviously so, to a human—as it will be when saying something true. The seemingly random inconsistency of LLMs makes it hard to trust their reasoning in complex, multi-step problems. If you want to use an AI model to help with a business problem, it’s not enough to see that it understands what factors make a product profitable; you need to be sure it won’t forget what money is.

How to Deal with AI Mistakes

This situation indicates two possible areas of research. The first is to engineer LLMs that make more human-like mistakes. The second is to build new mistake-correcting systems that deal with the specific sorts of mistakes that LLMs tend to make.

We already have some tools to lead LLMs to act in more human-like ways. Many of these arise from the field of “alignment” research, which aims to make models act in accordance with the goals and motivations of their human developers. One example is the technique that was arguably responsible for the breakthrough success of ChatGPT: reinforcement learning with human feedback. In this method, an AI model is (figuratively) rewarded for producing responses that get a thumbs-up from human evaluators. Similar approaches could be used to induce AI systems to make more human-like mistakes, particularly by penalizing them more for mistakes that are less intelligible.

When it comes to catching AI mistakes, some of the systems that we use to prevent human mistakes will help. To an extent, forcing LLMs to double-check their own work can help prevent errors. But LLMs can also confabulate seemingly plausible, but truly ridiculous, explanations for their flights from reason.

Other mistake mitigation systems for AI are unlike anything we use for humans. Because machines can’t get fatigued or frustrated in the way that humans do, it can help to ask an LLM the same question repeatedly in slightly different ways and then synthesize its multiple responses. Humans won’t put up with that kind of annoying repetition, but machines will.

Understanding Similarities and Differences

Researchers are still struggling to understand where LLM mistakes diverge from human ones. Some of the weirdness of AI is actually more human-like than it first appears. Small changes to a query to an LLM can result in wildly different responses, a problem known as prompt sensitivity. But, as any survey researcher can tell you, humans behave this way, too. The phrasing of a question in an opinion poll can have drastic impacts on the answers.

LLMs also seem to have a bias towards repeating the words that were most common in their training data; for example, guessing familiar place names like “America” even when asked about more exotic locations. Perhaps this is an example of the human “availability heuristic” manifesting in LLMs, with machines spitting out the first thing that comes to mind rather than reasoning through the question. And like humans, perhaps, some LLMs seem to get distracted in the middle of long documents; they’re better able to remember facts from the beginning and end. There is already progress on improving this error mode, as researchers have found that LLMs trained on more examples of retrieving information from long texts seem to do better at retrieving information uniformly.

In some cases, what’s bizarre about LLMs is that they act more like humans than we think they should. For example, some researchers have tested the hypothesis that LLMs perform better when offered a cash reward or threatened with death. It also turns out that some of the best ways to “jailbreak” LLMs (getting them to disobey their creators’ explicit instructions) look a lot like the kinds of social engineering tricks that humans use on each other: for example, pretending to be someone else or saying that the request is just a joke. But other effective jailbreaking techniques are things no human would ever fall for. One group found that if they used ASCII art (constructions of symbols that look like words or pictures) to pose dangerous questions, like how to build a bomb, the LLM would answer them willingly.

Humans may occasionally make seemingly random, incomprehensible, and inconsistent mistakes, but such occurrences are rare and often indicative of more serious problems. We also tend not to put people exhibiting these behaviors in decision-making positions. Likewise, we should confine AI decision-making systems to applications that suit their actual abilities—while keeping the potential ramifications of their mistakes firmly in mind.

This essay was written with Nathan E. Sanders, and originally appeared in IEEE Spectrum.

EDITED TO ADD (1/24): Slashdot thread.

Posted on January 21, 2025 at 7:02 AM11 Comments

Comments

Matt January 21, 2025 11:54 AM

“Technologies like large language models (LLMs) can perform many cognitive tasks”

No, they can’t perform ANY cognitive tasks. They do not cogitate. They do not think and are not capable of reasoning. They are nothing more than word-prediction engines. (This is not the same as saying they are useless.)

You should know better than that, Bruce.

RealFakeNews January 21, 2025 12:35 PM

Part of the problem is AI can’t fundamentally differentiate a fact from something it just made up. It can check cabbages and goats are related via some probability, but it can’t check that a cabbage doesn’t eat goats because it can’t use the lack of data to verify if that is correct.

Ray Dillinger January 21, 2025 3:15 PM

Before anyone bases a business or career plan on this information I feel the need to inject one caveat.

It’s still the ‘wild west’ in Neural-network AI right now. We are discovering new things just by looking over the next hill. Somebody has a random idea on a Thursday afternoon, whomps up demo code over the weekend, trains or adapts a trained model over the course of the next week or two, analyzes the results, discovers that it helps with a problem likely while causing other problems in some degree, and then either (A) files a patent, (B) writes an academic paper, or (C) quietly puts the code up on github for somebody else to discover on a different random Thursday afternoon.

Someone else, somewhere else, on a different random Thursday afternoon, finds an idea they haven’t tried yet on Arxiv or in Github, spends a couple of months and half-a-million dollars worth of electricity training a full-size model, makes more discoveries about the tradeoffs involved or fills in gaps that the original author missed, and the process repeats.

And this happens somewhere, literally every Thursday afternoon. The models trained this year do not resemble the models trained two years past. They make mistakes, but they make different mistakes, in different proportions, with different efficiency tradeoffs. The models trained two years from now will not resemble the models trained this year. The mistakes they make will again be different mistakes, with different efficiency tradeoffs, in different proportions.

I see anti-AI apps that supposedly “detect the use of generative AI” that detect the use of the kind of generative AI that was already obsolete when the apps were released. I see anti-AI apps that “poison” artwork for the purposes of using it as training data, in a way that would definitely mess up a convolution kernel in a convolutional neural network – which used to be part of the training process for generative art AI back in the days of Deep Dream but aren’t any more.

So that’s the caveat. Before you base a lot of effort on this, understand that it’s a moving target. By the time you get sighted in on adapting to this generation’s mistakes, the state of the art will be somewhere else making different kinds of mistakes.

The other thing you should realize is that AI researchers everywhere are going to gratefully and appreciatively read every article they can get their hands on about what kind of mistakes their creations are making, as part of their iterative efforts to improve them. The more attention any type of mistake gets, the more attention correcting that mistake will get. And AI researchers don’t consider this to be an arms race; they (I ought to say “we” technically but I worked a long time in security and it still feels pretentious) will be genuinely happy that you are helping them out.

Obsolete Human January 21, 2025 5:13 PM

Part of the reason LLM’s make very odd errors is that an LLM has never experienced the world first hand. They get their knowledge through training data- it’s all 2nd hand references to the physical world.

We know that eating rocks is a bad idea, partly because most of us have put a rock or pebble in our mouth as a kid. Same too for poetry- a lot of a good rhyme is how it actually sounds. A text based LLM, build on training text from people, has never heard a sound, ever. Of course you get strange attempts at rhymes.

All this to say, once we have AI’s that are able to experience the world on their own (ala Optimus or other robotic body), there will be a giant next leap. They will likely be boosted through AI’s freely adding to training data through own experiences in their free time. Today an LLM is running only when you send a prompt- then it stops. Humans are constantly thinking, and inspiration happens between our “prompts” as we freely think and play.

LLM’s may have plateau’ed for now. The next steps will usher in a new exponential wave of growth.

Chris January 21, 2025 6:01 PM

„ Humans may occasionally make seemingly random, incomprehensible, and inconsistent mistakes, but such occurrences are rare and often indicative of more serious problems. We also tend not to put people exhibiting these behaviors in decision-making positions.“

Do we?

ResearcherZero January 22, 2025 3:05 AM

We like to hope we do not place people exhibiting these behaviors in decision-making positions, but we do not always test for it and sometimes ignore the results when we do.

If you fail your police psyche evaluation because you are diagnosed as a dangerous sociopath, don’t worry, it’s quite likely the results might be ignored.

Clive Robinson January 22, 2025 7:01 AM

@ Ray Dillinger,

With regards,

“It’s still the ‘wild west’ in Neural-network AI right now.”

And according to some like Gartner more like the OK Corral,

https://www.theregister.com/2025/01/22/business_value_genai_elusive/

Which is what I’ve cautioned about for some time now.

With regards,

“LLM’s may have plateau’ed for now. The next steps will usher in a new exponential wave of growth.”

The first part as with all AI and robotic systems I’ve had any involvement with since the 1980’s is certainly true, but so far the second part has never yet happened. And in all honesty the old “it’s just ten to twenty years away” of “experts” has been true for getting on for seven decades and I can not see that trend changing in my lifetime.

What however is true is the small AI systems like “fuzzy logic” and “expert systems” from the 1980’s do not go away, they are still in use in control systems for “automation systems” “under computer control”. It’s hard for some to think of vehicle breaking systems and engine control being “robots” but by early definitions they are and the results of earlier AI migrate into them not as as a tusnami, tidal, or exponential wave, but more as a gentle swell of a rising tide that lifts all boats.

Have a look under the hood of that robot doing “back flip and somersaults” and you will find the derivations of both fuzzy logic and expert systems, doing there thing. Likewise the power control chain in those “light railways” that can run without drivers so smoothly. Those old AI are there because of not just the smoothness of the ride, humans crave, but they also “increase efficiency” and more importantly they “smooth the load” on the power network easing the expense on cabling and generator sets.

Is there a place for the latest LLM and ML AI systems in our future?

Yes, but importantly only with changes to both the models and underlying hardware. When they run on what we would think of as “Single Board Computers”(SBCs) and “System On a Chip”(SOC) microcontrollers as they eventually will they will be useful in all the places most don’t see computers in use. Like in factory and industrial plant control equipment and construction equipment. Saving just a few percent of increasingly expensive power, reducing damage costs due to inertia, and upping process flow thus productivity rates.

Will they cause changes in the job market?

Yes without a doubt, all new technology has and will continue to do so But not as such in the way some are doomsaying it will. Those jobs that will get hit and hit hard are the 20-40% of current jobs that are not really jobs but low end “makework”.

Will that matter to employment figures?

Not really if you think only “detrimentally”, because “makework” has always and will always exist, it will just be a bit more sophisticated as time progresses. But also new work will come into existence as a “natural consequence” that due to the low grade AI assisting can now be done by those that would previously have done what we see as soul destroying or dead end “drudge makework”. But more than half of those new jobs actually won’t be “makework” as such so overall even the current AI LLM and ML systems when shrunk down to run efficiently on small hardware will eventually create increasingly worthwhile jobs and increase productivity and efficiency.

Just look on it as “technology evolution in progress”, what works will survive, what does not will become “tomorrows lessons for the future”.

Which leaves one question that people will have reason to worry about,

Will the LLM and ML AI systems change society?

To which the answer is certainly yes, and we already know that.

What people are realy worried about is,

Will the KLM and ML AI systems change society for better or worse?

That is will it be “good or bad AI”.

The answer boils down to a very human issue which I mention from time to time,

“The independent observers view of the intent of the directing mind.”

Which these days unfortunately boils down to two things,

1, Who is paying?
2, What is their bias?

And thirdly and perhaps most importantly,

3, Can they hide their bias or responsability behind AI?

That is using AI as a “cut out” or “arms length system” to absolve them of any or all responsibility, much as the current “The Computer Says” meme that is so prevalent shows.

We know that there are many politicians and similar just lining up to do this, with “public money”. And we also know there are many people who will take that money quite happily and do what they know is not just wrong but actually illegal. The UK Post Office Horizon Project is just one of many examples. Previous examples have been the various RoboDebt and Tax Reclaim systems and likewise testing during lockdown. It appears at times that where ever you look these days you will find this sort of unlawful behaviour, so we can almost take it as a given it will happen with future LLM and ML systems.

Noman January 30, 2025 5:11 AM

Something else that may be a contributing issue is that the output is generally given with 100% certainty. No “I think” or “it might be” and the end user believes it must be so. As with trying to solve a complex problem, getting it wrong, and looking up the solution in the back of a study guide. Which has an incorrect solution. That must be so. Can you even realize the solution is incorrect or are you just fucked in a small way?
I was once taking an 8-hour exam for a professional license and found a problem that was incorrectly stated and therefore had no proper answer. Should I grind the gears and waste time trying to solve it to the detriment of my score, and possibly not get the license, or move on. Or recognize I had an NG and quickly move on to the next problem. Interestingly many of the multiple choice answers offered wrong solutions based on common computational or conceptual errors so one might think their answers were right because choice B matched my answer to 3 decimal places even though choice B was wrong.
I propose Odysseus’ theorem akin to garbage in, garbage out. “If the answers are based on a database with a random probability of being correct, you have a random probability of surviving your chosen path”. Someone out there state that more elegantly please.

Rontea January 30, 2025 1:43 PM

“random, incomprehensible, inconsistent ”
“Man, Machine, Magic” -shadowrun

@Chris
“Do we?
I don’t think there will be any progress if we do.

@Clive
“what works will survive, what does not will become “tomorrows lessons for the future”
Without language these lessons are unavailable and we cannot learn from the mistakes of our ancestors.

@Noman
“Someone out there state that more elegantly please”

In the forest of choices, we wander alone,
Where paths intertwine, like roots overgrown.
Each step that we take, a whisper of fate,
Guiding us gently towards an unknown gate.

The answers we seek in shadows reside,
With mysteries hidden, where secrets abide.
A tapestry woven with threads of chance,
In the dance of destiny, we bravely advance.

The path may be dark, with turns unforeseen,
Yet hope is the light that keeps our hearts keen.
For in every decision, a new world takes flight,
With dreams to discover and stars to ignite.

In the randomness lies a delicate grace,
Where courage and choice find their rightful place.
Embrace the uncertainty, trust in your way,
For the journey's the treasure, come what may.

In our increasingly AI-driven world, differentiating between human mistakes and artificial intelligence mistakes is crucial. Understanding these distinctions is vital to developing more reliable AI technologies and ensuring their safe and beneficial integration into human activities.

Clive Robinson January 30, 2025 7:40 PM

@ Rontea,

With regards “mistakes” and,

“Without language these lessons are unavailable and we cannot learn from the mistakes of our ancestors.”

Nor can we learn from their successes unless they get “passed down” in some way.

The word “mistake” is also quite nuanced around human processes that in theory are rather more than just acting from “rote example” or “established actions” that have been “passed down”.

To,

“Make a mistake, requires the agent be capable of not making mistakes”.

Yes it sounds like a circular argument but it’s actually not. Most humans take actions based on experience both directly and indirectly related. When they don’t have sufficient “rote examples” that are directly relevant they either simply make a “gut call” / “guess” or by way of sufficient awareness, and experience, reason the issue through based on what they can interpret from information that is known to them they can put into the current situational context.

Current LLM’s even with ML are,

1, Not aware in any meaningful way.
2, Can not interpret actions from what is in the data set to what is not in the data set.
3, Can not reason.
4, Have no understanding of situational context.

Thus they will provide unwanted answers… But they are not answers most would call mistakes. Because the LLM provides it’s output from a “probability curve” where the point is selected by a a stochastic / random process. So in effect the LLM output is,

“As the fall of the dice, or the turn of a card.”

Clive Robinson January 30, 2025 7:57 PM

@ Bruce, ALL,

It appears that even the self proclaimed servant of the Christian God in the North West hemisphere has an opinion on AI.

From the Vatican,

https://www.vatican.va/roman_curia/congregations/cfaith/documents/rc_ddf_doc_20250128_antiqua-et-nova_en.html

Make of it what you will, it appears that,

“Every man and his dog, has an opinion.”

These days… And almost the only thing they mostly have in common, is that they are probably wrong, such is,

“The nature of the beast.”

We call mankind. We are all just observers making calls of “good or bad” based on insufficient information, our positions within society, and the morales and mores we have.

Leave a comment

Blog moderation policy

Login

Allowed HTML <a href="URL"> • <em> <cite> <i> • <strong> <b> • <sub> <sup> • <ul> <ol> <li> • <blockquote> <pre> Markdown Extra syntax via https://michelf.ca/projects/php-markdown/extra/

Sidebar photo of Bruce Schneier by Joe MacInnis.