Where the line between thinking and mimicking blurs

Artificial intelligence often gets sold as being on the cusp of human-like thinking. Tech headlines claim the latest large language models can reason, solve tricky problems, and edge closer to real understanding. But a new paper from Apple’s AI research team throws cold water on those claims. Titled ‘The Illusion of Thinking’, the paper argues that much of what passes for AI “reasoning” is actually just smoke and mirrors.

Coming from Apple, a company renowned for guarded research the critique isn’t just another academic paper. At a moment when AI reasoning is being hyped up like never before, it serves as a pointed reminder:

Don’t mistake AI’s articulate responses for genuine understanding, its underlying thought processes are often more limited than they seem.

Here’s a rapid breakdown of Apple’s key insights:

data x ai icon

Even the “smart” models miss the basics: Apple shows that reasoning-enabled models often second-guess themselves out of the right answer. Simpler models actually outperform them on straightforward tasks; more compute doesn’t mean more accuracy

Hard problems break them: As tasks get more complex, the apparent reasoning falls apart. In Apple’s tests, performance flatlined once a certain threshold was crossed

Icon

Reasoning is mostly surface-level: Apple’s researchers say there’s no real logic at play. Models aren’t developing deep strategies – they’re mimicking patterns that look like thought

Icon

Blind trust is risky: If businesses hand over critical decisions to models that only appear to reason, the consequences could be serious. Confidently wrong answers are often more dangerous than unclear ones.

Reasoning – rehearsed not real

To test how well today’s “reasoning” models actually think, Apple built controlled puzzle environments – like the Tower of Hanoi – with rising complexity. Rather than relying on benchmark questions the models might’ve already seen, they focused on unseen problems that let them observe not just whether a model got the answer right, but how it tried to get there.

The results were striking. On simple puzzles, reasoning-enabled models often made things worse. They’d find the correct answer quickly, then carry on “thinking” until they talked themselves into a wrong one. In contrast, standard models got it right by sticking with their first guess.

At moderate complexity, the reasoning models performed better. They broke problems into parts and used longer step-by-step outputs to reach accurate answers – this is where the so-called chain-of-thought approach (where models break down problems into sequential steps) made sense.

But that advantage didn’t last. Once puzzles became properly difficult, both standard and reasoning models hit a wall. Performance dropped to zero. The reasoning models generated longer answers, but not better ones. They didn’t adapt. They didn’t generalise. They just collapsed.

The models didn’t discover useful strategies or underlying algorithms. Despite having tools for reflection, they failed to build consistent approaches across similar problems. They weren’t reasoning – they were essentially mimicking structure without grasping what it meant.

As Apple put it, these systems “fail to use explicit algorithms and reason inconsistently across puzzles.” In other words, they imitate the appearance of logic, but not the logic itself.

Calling time on the reasoning hype

Apple’s paper challenges a growing narrative in AI – that new models don’t just produce text but can plan, solve, and “think” like humans. Firms like OpenAI, Google, and Anthropic are selling that story hard – Sam Altman has made repeated claims about GPT-4’s improved reasoning abilities.

But Apple’s findings throw doubt on all of that with a clear message: stacking more layers, throwing in more tokens, or asking a model to “think harder” doesn’t get you to true intelligence. If someone’s betting on AGI showing up next year, they may want to adjust their expectations. Critics like cognitive scientist Gary Marcus – long sceptic of large language models – were quick to applaud the paper. Gary has often said that today’s models are just “stochastic parrots,” repeating patterns they’ve seen without understanding them. Apple’s work offers evidence for that view. He even called it a “knockout blow” to the idea that scale alone will solve intelligence.

And he’s got a point. These systems can sound smart – even persuasive – while getting everything wrong. That’s fine if you’re summarising emails. But in medicine, law, or finance? The illusion becomes dangerous.

What this means for businesses

Apple’s findings are not merely academic; they deliver a crucial reality check for businesses grappling with AI adoption. In an era of burgeoning ‘AI co-pilots’ and decision-support tools, the pressure to delegate complex tasks to AI is immense, yet Apple’s research strongly suggests this is often premature and risky.

In high-stakes areas like finance, healthcare, or legal advice, overestimating AI’s abilities could be costly. An AI-generated strategy that looks convincing but rests on flawed logic might not just fail but could also cause real harm.

So what’s the takeaway for businesses?

Treat AI as a tool, not a decision-maker. Use it to assist, not to replace human judgement – especially for complex decisions

Icon

Keep humans in the loop. Particularly in areas where the cost of being wrong is high

blog

Push vendors for clarity. If a product claims to reason, ask to see evidence from tough, novel problems – the kind Apple used

technology partnerships icon

Start with narrow use cases. Look for areas where pattern recognition works well: document summaries, information extraction, or repetitive tasks. Avoid broad planning or abstract analysis unless you’ve got rigorous oversight

Icon shield with a tick

Stress-test your systems. Just like banks test financial stability under strain, test AI under difficult problem scenarios to see where it breaks

Keep it useful. Keep it grounded.

The gap between what current AI models appear to do and what they actually understand remains wide. For decision-makers, developers, and policy strategists, the lesson is simple: don’t confuse articulate output with actual intelligence. These models are still fragile, still shallow, and still prone to confident failure – presenting flawed logic with the same conviction as correct answers, which can lead down costly or even harmful paths.

That doesn’t mean progress isn’t happening. Instead, it highlights that scaling current approaches appears insufficient to bridge the fundamental gap between pattern recognition and genuine understanding. The road ahead likely requires truly novel architectures, perhaps blending symbolic AI or classical logic with modern learning techniques.

In the meantime, the best way to approach AI is clear-eyed and grounded. Let it help where it can. Don’t trust it where it can’t. And above all, remember: just because something sounds smart doesn’t mean it is.

Latest AI blogs

Talk to us