The field of distributed learning is moving towards more scalable and efficient methods, with a focus on reducing communication overhead and improving convergence rates. Recent works have explored various techniques, including relaxed global communication, alternating low-rank updates, and cyclical updates, to achieve these goals. These innovations have the potential to improve the performance of large language models and other applications. Noteworthy papers include Pier, which proposes an efficient and scalable optimizer for large language model pretraining, and ADF-LoRA, which introduces a decentralized federated learning method with improved stability and convergence. Additionally, CycleSL presents a novel split learning framework that enhances scalability and performance, while Mitigating Participation Imbalance Bias in Asynchronous Federated Learning provides a theoretical analysis and proposes methods to address heterogeneity amplification in asynchronous federated learning.