A recent paper by Apple researchers, titled "The Illusion of Thinking," has sparked debate within the AI community by suggesting that Large Reasoning Models (LRMs), despite generating detailed "thinking traces," struggle with increased problem complexity and may not truly reason as perceived. The paper highlights that as tasks become more complex, LRMs' performance deteriorates, with models often abandoning complex problems and overthinking simpler ones. However, a counter-paper, "The Illusion of the Illusion of Thinking," argues that these observed limitations result from experimental flaws such as output token constraints and misinterpretation of unsolvable tasks, rather than inherent reasoning deficits. This rebuttal suggests that with corrected evaluations, models can handle tasks previously deemed failures. The debate underscores the importance of carefully designed evaluations that distinguish genuine cognitive limitations from engineering constraints, and highlights ongoing philosophical and practical discussions around AI's reasoning capabilities and the industry's strategic positioning in AI advancements.