The field of reinforcement learning and Markov decision processes is moving towards developing more robust and efficient algorithms. Recent research has focused on improving the robustness of policies under uncertainty and adversity, with a particular emphasis on developing algorithms that can handle epistemic uncertainty in environment dynamics. Another key area of research is the development of more efficient algorithms for solving Markov decision processes, including the use of homomorphic mappings and adaptive low-rank structures. Additionally, there is a growing interest in developing algorithms that can learn to guide planning in partially observable Markov decision processes. Noteworthy papers include: ADARL, which proposes a bi-level optimization framework that improves robustness by aligning policy complexity with the intrinsic dimension of the task. Pruning Cannot Hurt Robustness, which develops the first theoretical framework for certified robustness under pruning in state-adversarial Markov decision processes. GammaZero, which introduces an action-centric graph representation framework for learning to guide planning in partially observable Markov decision processes.