Apple’s “Illusion of Thinking” and What It Really Tells Us About Large Reasoning Models
Published on 6/8/2025
- Artificial Intelligence
- Machine Learning
- LRMs
Apple’s “Illusion of Thinking” and What It Really Tells Us About Large Reasoning Models
TL;DR: Apple stress-tests large reasoning models on puzzles with rising complexity and shows where they still snap.
With a provocative title and a clean experimental setup, Apple’s new paper, “The Illusion of Thinking,” takes aim at a central question in AI: are today’s models truly reasoning, or just mimicking it? Beneath the surface, it’s less a takedown and more a stress test, revealing where even the strongest models start to break.
What the authors actually did
- Built four simulator-based puzzles (Tower of Hanoi, River Crossing, Checker Jumping, Blocks World) where complexity can be dialed up in clean, quantifiable steps.
- Ran mainstream “Large Reasoning Models” (LRMs) alongside their plain-LLM counterparts under identical token budgets.
- Measured final accuracy, pass-at-k, and even the length of each chain-of-thought to see where things break down.
The headline result: both vanilla LLMs and “thinking” variants hit an abrupt wall once compositional depth crosses a modest threshold. LRMs do a bit better in the middle zone, then collapse along with their shorter-and-shorter reasoning traces.
Pattern matching vs. thinking
Today’s generative models are sophisticated statistical engines, not cognitive agents. Apple’s curves reinforce that view. The models exploit patterns in familiar territory, but when a task demands an extended, multi-step plan, they run out of statistical shortcuts.
Yet calling that just pattern matching undersells the breakthrough. GPT-class models, Gemini, Claude, and Apple’s own work have pushed language understanding, code generation, and tool integration to levels that were unthinkable five years ago. Society is moving faster precisely because these “pattern matchers” keep finding useful shortcuts.
Why one-shot evaluation feels limited
A single prompt with a small token cap is rarely how we build production systems:
- Tools and retrieval: External search, structured knowledge, and calculators add trusted facts.
- Agentic workflows: Planners decompose big goals into bite-sized calls.
- Memory: Long-lived vectors or databases preserve state across turns.
- Multi-agent collaboration: Specialist agents critique, verify, or vote, offsetting the blind spots of any single model.
Apple’s paper freezes all that complexity and asks, “How far can one model go alone?” That is a valid baseline, but it is not a verdict on the wider ecosystem.
Why the study still matters
- Sets a higher bar for benchmarks. Simple pass/fail tasks hide scaling cliffs; complexity-controlled suites expose them.
- Keeps AGI hype in check. When accuracy slumps to zero beyond modest depth, we are reminded there is still ground to cover.
- Informs model-tool design. If LRMs shorten their thought chains right before failure, future systems might detect that signature and hand control to a symbolic planner or external solver.
Take-away for practitioners and execs
Treat chain-of-thought as one component in a larger reasoning stack. Measure complexity alongside accuracy. And remember: real-world AI success almost always involves the model plus tooling, retrieval, and iterative control.
Apple’s paper doesn’t say LRMs are useless. It says they remain remarkable, but not magical. That sober message is exactly what our field needs while we keep pushing the frontier.
Have you seen similar failure points when reasoning depth increases? I’d love to hear real-world parallels.