The field of artificial intelligence is witnessing significant developments in reinforcement learning and task planning. Researchers are exploring new methods to improve the efficiency and effectiveness of reinforcement learning agents in complex, long-horizon tasks. One notable direction is the use of hierarchical task-planning methods, which decompose complex goals into manageable subgoals and enable agents to reason and act at multiple levels of abstraction. Another area of focus is the development of novel memory mechanisms that allow agents to learn from experience and adapt to changing environments. These advancements have the potential to enable more efficient and effective decision-making in a wide range of applications, from robotics to natural language processing. Noteworthy papers in this area include: ReAcTree, which proposes a hierarchical task-planning method that achieves state-of-the-art performance on several benchmarks. EvoMem, which introduces a dual-evolving memory mechanism that improves multi-agent planning in several domains. Tree Training, which accelerates agentic LLM training via shared prefix reuse, reducing training time by up to 3.9x.