The field of large language models (LLMs) is rapidly advancing, with a focus on improving reasoning capabilities. Recent developments have led to the creation of novel benchmarks and evaluation frameworks, such as those assessing abstract reasoning, symbolic mathematics, and data-flow analysis. These advancements highlight the current limitations of LLMs in terms of true understanding and generalization, but also demonstrate the potential for significant improvements through innovative approaches like computational thinking and exchange of perspective prompting. Noteworthy papers include ASyMOB, which introduces a novel assessment framework for symbolic mathematics, and TimeHC-RL, which enhances LLMs' social intelligence through temporal-aware hierarchical cognitive reinforcement learning. Overall, the field is moving towards a deeper understanding of LLM reasoning capabilities and the development of more effective evaluation methods.