The field of large language models (LLMs) is moving towards developing more efficient and adaptive reasoning strategies. Recent research has focused on improving the computational efficiency of LLMs, particularly in real-time applications, by dynamically adjusting their reasoning depth and budget. This has led to the development of novel frameworks and methods that enable LLMs to optimize their reasoning processes, balancing efficiency and accuracy without compromising performance. Noteworthy papers in this regard include Aware First, Think Less, which introduces the Dynamic Reasoning-Boundary Self-Awareness Framework (DR. SAF) to improve the efficiency of LLMs, and Think in Blocks, which proposes an adaptive reasoning framework that enables LLMs to dynamically adjust the length of their reasoning processes based on task complexity. Additionally, papers such as Exploring Efficiency Frontiers of Thinking Budget in Medical Reasoning and OptimalThinkingBench have made significant contributions to the field by evaluating the efficiency of LLMs in medical reasoning tasks and introducing a unified benchmark to evaluate overthinking and underthinking in LLMs, respectively.