The field of reinforcement learning is moving towards the development of more efficient, safe, and interpretable agents. Recent research has focused on combining model-free and model-based approaches to improve sample efficiency and safety. Model-based methods, such as model predictive control, can leverage prior system knowledge to inform and constrain agent decisions, while model-free methods can help remedy model mismatch. Additionally, there is a growing interest in using expert demonstrations and offline data to guide learning and improve decision-making. Noteworthy papers in this area include: On the Fundamental Limitations of Dual Static CVaR Decompositions in Markov Decision Processes, which highlights the limitations of dual CVaR decompositions in MDPs. Mixture of Autoencoder Experts Guidance using Unlabeled and Incomplete Data for Exploration in Reinforcement Learning, which proposes a framework for utilizing expert demonstrations to guide exploration in RL. Kernel Based Maximum Entropy Inverse Reinforcement Learning for Mean-Field Games, which introduces a kernel-based approach to inverse reinforcement learning for mean-field games.