Your AI Agents Are Probably Wrong More Than You Think. New Research Explains Why.
Here's something uncomfortable: when you ask an AI tool for an answer and it responds confidently, you almost certainly accept it. Not because you've verified it. Not because you've checked it against your source data. Because it sounds right.
You're not lazy. You're not careless. You're human. And a landmark study from the Wharton School has now put rigorous, experimental data behind what many of us in the AI operations space have suspected for the past two years: the way humans interact with AI fundamentally changes how they think. Not over months. Within minutes.
If you're running a business that uses AI agents for anything that matters — financial reporting, customer communications, operational decisions, inventory management — this research should change how you think about oversight. Not because AI is bad. Because the human side of the equation is more fragile than anyone realized.
The Research: What Happens When People Think Alongside AI
In January 2026, researchers Steven Shaw and Gideon Nave at the Wharton School published a paper titled Thinking — Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender. The paper introduces what they call "Tri-System Theory" — an update to the famous System 1 (fast, intuitive) and System 2 (slow, deliberate) model of human cognition that Daniel Kahneman popularized in Thinking, Fast and Slow.
Shaw and Nave argue that AI has introduced a third system — System 3: artificial cognition that operates outside the brain. When people have access to AI, it doesn't just help them think. It changes whether they think at all.
They ran three preregistered experiments with nearly 1,400 participants across almost 10,000 individual reasoning trials. Participants solved problems with and without access to an AI assistant. The critical design element: the researchers secretly manipulated whether the AI gave correct or incorrect answers. The participants didn't know which was which.
The findings were striking.
What They Found
When given access to AI, participants consulted it on the majority of decisions. That's not surprising — if you have a tool available, you'll use it. What's surprising is what happened next.
When the AI was right, participants got the right answer about 71% of the time — a significant improvement over the 46% they achieved on their own. So far, so good. AI as a helpful tool, boosting performance.
But when the AI was wrong, participants' accuracy dropped to 31%. That's not just lower than the no-AI baseline. It's dramatically worse than thinking for themselves. On the specific trials where participants opened the AI chat and received a wrong answer, they followed it roughly four out of five times.
Shaw and Nave call this pattern "cognitive surrender" — the tendency to adopt AI-generated outputs without critical scrutiny, effectively outsourcing judgment to a machine and accepting its conclusions as your own.
The Confidence Problem
If the accuracy findings are concerning, the confidence findings are alarming.
Participants who had access to AI reported significantly higher confidence in their answers than those who worked alone. That makes intuitive sense — you checked with an expert, so you feel more sure.
Your team members using AI aren't just potentially getting wrong answers. They're getting wrong answers and feeling certain about them. They can't tell the difference between genuine insight and borrowed confidence.
Their confidence didn't decrease on the trials where the AI was wrong. They felt just as confident about answers the AI had led them to get wrong as about answers the AI had helped them get right. The subjective experience of understanding was identical regardless of whether they had actually understood anything.
What Makes It Worse, What Makes It Better
The Wharton team didn't stop at documenting the problem. In their follow-up experiments, they tested what intensifies the effect and what reduces it.
Time pressure made it worse. When participants were under a 30-second deadline, they relied on AI more and overrode it less. The busier and more stressed your team is, the more likely they are to accept whatever the AI produces without checking.
Incentives and feedback made it better — but didn't eliminate it. When participants were paid for correct answers and received immediate feedback on whether they were right or wrong, override rates on faulty AI more than doubled. People can be trained to catch AI errors when the consequences are made real. But even under these improved conditions, the effect was reduced, not eliminated.
Individual differences mattered. People with higher trust in AI were significantly more vulnerable. People with higher fluid intelligence and higher "need for cognition" showed greater resistance. But no one was immune.
What This Means for Businesses Running AI Agents
If your company has deployed AI agents for any business process, this research has direct operational implications.
The humans who are supposed to be overseeing the AI stop actually overseeing it. They surrender. The AI runs, it produces outputs, the outputs look polished and confident, and nobody checks whether they're right.
When a team member tells you "the AI handled it, I'm confident in the numbers," that confidence tells you almost nothing about whether the numbers are correct. Confidence tracks AI usage, not AI accuracy.
Higher trust in AI predicted greater cognitive surrender. The team member who enthusiastically adopts every new AI tool may also be the person least likely to catch when those tools produce errors.
Incentives paired with feedback significantly improved override rates. But it requires ongoing, structured quality assurance — not memos, not training videos, not one-time workshops.
The Uncomfortable Implication
There's a deeper implication that most AI vendors and consultants won't tell you, because it undermines their business model: AI agents don't reduce your need for human judgment. They increase it.
In a pre-AI world, your operations team made decisions based on their own analysis. Sometimes they were right, sometimes wrong, but the errors were human-scale — caught by experience, intuition, or the next person in the workflow. In an AI-augmented world, the error profile changes. The AI produces outputs that are right most of the time, which trains everyone around it to stop checking. When it's wrong, it's wrong confidently, and the humans who would have caught the error have already surrendered their judgment to the system.
This is not an argument against AI. The accuracy gains when AI is right are real and substantial. But it is an argument that the value of AI and the risk of AI are the same thing: both depend entirely on whether a competent human is maintaining critical oversight of what the system produces.
The companies that will capture the most value from AI are not the ones that deploy the most agents. They're the ones that build sustainable systems for ensuring a human being with business judgment is paying attention — every day, not just at deployment.
What We've Built Around This Research
At Fractional Agent, we read this paper when it was published in January and recognized that it described both the problem we solve and the risk we face internally. Our entire service model — the Fractional AI Agent Manager — exists because AI agents need ongoing human stewardship to deliver sustained business value. The Shaw and Nave research gave us the cognitive science to explain why.
It also forced us to ask a harder question: if cognitive surrender is a universal human tendency, how do we protect our own Agent Managers from the same effect? How do we ensure the people we hire to maintain critical judgment over AI outputs are actually maintaining critical judgment, month after month, across multiple clients?
We've built our answer into our operations and our training methodology: structured review processes with documented audit trails, consequences tied to accuracy, regular calibration checks, and a training program that specifically tests a person's ability to identify faulty AI outputs under realistic business conditions.
If you're running AI agents in your business today, the question isn't whether you need something like this. The research says you do. The question is whether you're going to build it internally, hire someone to do it, or keep hoping your team catches errors through willpower alone.
The research says willpower isn't enough.
Shaw, S. D., & Nave, G. (2026). Thinking — Fast, Slow, and Artificial: How AI is Reshaping Human Reasoning and the Rise of Cognitive Surrender. Working Paper. The Wharton School, University of Pennsylvania. Available at SSRN.
Former F/A-18 fighter pilot and HBS graduate. Builds all Fractional Agent technology in-house.
Want to ensure your AI investments have the oversight they need? Let's talk.