Recent discussions in the AI community, highlighted by researcher Rohan Paul, indicate a significant conceptual shift in understanding how Large Language Models (LLMs) perform reasoning. This emerging perspective moves away from viewing LLM operations as a series of disconnected token-level predictions towards a more integrated understanding of continuous semantic movement within a reasoning space. The change suggests a fundamental re-evaluation of how LLM reasoning should be analyzed and developed.Traditionally, LLMs have been understood at the "token-level," where the model progresses step-by-step through words, making discrete choices in a vast, isolated space of possible tokens. As Rohan Paul stated in a recent tweet, "At the token-level, the model is seen as moving step by step through words, like hopping between isolated points in a huge space. It treats each token choice as a separate move in a giant sequence." This approach, while effective for next-token prediction, has limitations in fully capturing complex thought processes.However, a new "reasoning-level" viewpoint proposes that the space of all token sequences is so dense it can be treated as continuous. This allows for a more fluid interpretation of LLM processes, akin to navigating smoothly through a meaning space rather than making abrupt jumps between words. Research initiatives like the "CoT-Space" framework from Renmin University of China and Meta AI's "Large Concept Model (LCM)" and "Coconut (Chain of Continuous Thought)" exemplify this shift, exploring how LLMs can operate and reason in continuous latent spaces.This paradigm shift suggests that "reasoning should not be studied as disconnected token jumps but as movement in a continuous semantic space," as emphasized by Paul. This continuous perspective is believed to be more natural for analyzing and enhancing how LLM reasoning actually works, potentially leading to more human-like cognitive abilities. Such advancements aim to overcome the computational inefficiencies and limitations of purely discrete token processing, fostering improved abstract reasoning and generalization in AI models.