The field of large language models (LLMs) is moving towards more efficient and adaptive reasoning methods. Recent research has focused on improving the accuracy and interpretability of LLMs while reducing their computational costs and latency. One notable direction is the development of methods that enable LLMs to reason in a more human-like way, by generating explicit step-by-step rationales for their decisions. However, this approach can be computationally expensive and may not always be necessary for simpler tasks. To address this, researchers have proposed various techniques for adaptive reasoning, which allow LLMs to adjust their reasoning depth and complexity based on the task at hand. These techniques include difficulty-adaptive reasoning, latent reasoning, and compressed knowledge distillation. Noteworthy papers in this area include Dual-Head Reasoning Distillation, which improves classifier accuracy with train-time-only reasoning, and MARCOS, which models reasoning as a hidden Markov chain of continuous thoughts. Overall, the field is moving towards more efficient, adaptive, and interpretable reasoning methods for LLMs, with potential applications in a wide range of areas, including natural language processing, computer vision, and decision-making.