The field of recommendation systems is witnessing a significant shift towards incorporating multimodal data and techniques to improve performance and address long-standing challenges. Researchers are exploring the potential of multimodal learning to combine collaborative and side information, narrowing the modality gap and providing accurate recommendations even in missing modality scenarios. Notable papers in this area include:
- Understanding Embedding Scaling in Collaborative Filtering, which discovers two novel phenomena when scaling collaborative filtering models and provides a theoretical analysis of the noise robustness of these models.
- Multimodal-enhanced Federated Recommendation, which proposes a novel multimodal fusion mechanism in federated recommendation settings, enabling fine-grained knowledge sharing among similar users while retaining individual preferences.