AI Reasoning Debate Rages: Rebuttal Reframes Apple's 'Illusion' as Test Flaw

Apple's 'Illusion of Thinking' ignites debate: Are AI models hitting limits, or are the tests themselves flawed?

July 4, 2025

AI Reasoning Debate Rages: Rebuttal Reframes Apple's 'Illusion' as Test Flaw
A recent paper from Apple researchers, provocatively titled "The Illusion of Thinking," has sent ripples through the artificial intelligence community, suggesting that the industry's most advanced reasoning models have fundamental limitations. The study contends that despite their sophistication, models designed for complex thought processes face a "complete accuracy collapse" when problems reach a certain level of difficulty. However, these claims are now facing significant pushback from a wave of critiques and at least one formal rebuttal, which, while confirming some of Apple's findings, challenge the paper's most dramatic conclusions and point to potential flaws in its experimental design. The ensuing debate is casting a fresh light on how we measure and understand machine "reasoning" and the true nature of progress toward more general AI.
Apple's research presented a sobering analysis of so-called Large Reasoning Models (LRMs), such as variants of Claude, DeepSeek, and OpenAI's "o" series, which are designed to "think" by generating step-by-step rationales before delivering an answer.[1][2] To test their capabilities, Apple's team used a series of controllable puzzles like the Tower of Hanoi, allowing them to precisely ramp up the complexity.[3][4] Their findings outlined three distinct performance regimes.[5] In low-complexity scenarios, standard large language models (LLMs) sometimes outperformed their LRM counterparts, which had a tendency to "overthink" simple problems.[5][6] At medium complexity, the LRMs held a clear advantage, their "chain-of-thought" processes proving effective.[3][5] But at high complexity, the researchers observed a total breakdown, with the performance of all models collapsing to zero.[3][7] Perhaps the most startling claim was that as problems neared this collapse point, the models paradoxically began to reduce their reasoning effort, essentially "giving up" despite having a sufficient computational budget.[1][7][8] The paper concluded that current models lack generalizable problem-solving strategies and may be hitting a "fundamental scaling limitation."[9]
The response to Apple's paper was swift, with many researchers and AI practitioners launching pointed critiques. A central argument against Apple's conclusions focuses on the methodology. Critics have pointed out that the performance collapse observed by the Apple team may not be a failure of reasoning but a consequence of technical constraints, such as the models' maximum output token limits.[10][11] For a puzzle like the Tower of Hanoi, the number of required moves grows exponentially with complexity; the point of "collapse" identified by Apple for an eight-disk puzzle coincides almost exactly with where the full solution would exceed the token capacity of the tested models.[10] This suggests the models may not be giving up on reasoning, but are simply hitting an architectural wall.
Further scrutiny has been applied to the specific tasks chosen for the evaluation. In what some have called a "damning" oversight, it was discovered that for the River Crossing puzzle, the researchers tested models on scenarios that are mathematically impossible to solve.[10] The study included tests with six or more "actors" and a boat that could only carry three, a configuration for which no solution exists when there are more than five actors.[10] Other critiques argue that using abstract puzzles, many of which have well-known algorithmic solutions likely present in the training data, is not representative of real-world reasoning tasks.[5][4][12] This has led to accusations that the study's design was flawed, potentially engineered to show failure rather than to genuinely probe the limits of AI cognition.[11] Some commentators also noted the paper was released as a preprint without peer review, and its timing, just before Apple's major developer conference, has fueled speculation that it may have been a strategic move to cast doubt on competitors' advancements while Apple plays catch-up in the AI race.[12][11]
A formal rebuttal paper, titled "Comment on The Illusion of Thinking," encapsulates many of these criticisms, arguing that the failures Apple identified are not intrinsic to the models but are instead artifacts of poor evaluation design.[10] The authors of the rebuttal and other critics contend that by not allowing the models to use code—a primary tool for solving complex logical problems—and by imposing restrictive parameters, the study failed to test the models under realistic conditions.[11] While a replication study confirmed that models do struggle and their performance breaks down, it reframes the cause. It challenges Apple's central conclusion of a fundamental reasoning limit, suggesting instead that the "illusion" may lie in the experimental setup itself. This more nuanced view proposes that while today's models have clear limitations, particularly in areas like deep abstract reasoning, the picture is not as bleak as Apple's paper suggests.[13] Both the original paper and its critiques may be partially correct: models do fail, but the reasons for that failure are more complex than initially presented.[13]
Ultimately, the debate sparked by "The Illusion of Thinking" serves as a crucial check on the AI industry's breathless narrative of exponential progress. It underscores the difficulty of accurately benchmarking true reasoning abilities and highlights the gap between mimicking human thought processes and possessing genuine, generalizable problem-solving skills.[9][6] Apple's research, despite its potential flaws, has forced a critical conversation about the limitations of current architectures and the need for more robust and transparent evaluation methods.[1][6] While the industry may not agree on whether current models are truly "thinking," there is a growing consensus that the path to more powerful and reliable AI will require more than just scaling up existing models; it may demand a fundamental rethink of their core design.[9][5]

Sources
Share this article