The field of music and audio technologies is moving towards more intuitive and interactive systems, enabling non-experts to engage with music creation and audio processing. Researchers are exploring new interfaces and models that allow for embodied and explainable interactions, making it easier for users to generate and manipulate music and audio. Noteworthy papers in this area include: SonicMaster, which introduces a unified generative model for music restoration and mastering with text-based control. Live Music Models, which presents a new class of generative models for music that produce a continuous stream of music in real-time with synchronized user control. TofuML, which employs a physical and spatial interface to make machine learning concepts more accessible and engaging for non-expert users. These innovative approaches are advancing the field and opening up new possibilities for music and audio creation, processing, and interaction.