Advances in Multimodal Learning and Counterfactual Explanations

The field of multimodal learning is rapidly evolving, with a focus on developing more robust and effective methods for text-to-video retrieval, video recommendation, and counterfactual explanations. Recent research has highlighted the importance of capturing intricate interactions between visual and textual modalities, as well as the need for more realistic and actionable counterfactual explanations. The use of large language models and multimodal large language models has shown promise in enhancing video recommendations and generating more accurate counterfactual explanations. Noteworthy papers in this area include: Adversarial Video Promotion Against Text-to-Video Retrieval, which pioneers a new attack against text-to-video retrieval models. Describe What You See with Multimodal Large Language Models to Enhance Video Recommendations, which introduces a simple framework for injecting high-level semantics into video recommendation pipelines. RealAC: A Domain-Agnostic Framework for Realistic and Actionable Counterfactual Explanations, which presents a domain-agnostic framework for generating realistic and actionable counterfactual explanations.

Sources

Adversarial Video Promotion Against Text-to-Video Retrieval

Mitigating Popularity Bias in Counterfactual Explanations using Large Language Models

AI Blob! LLM-Driven Recontextualization of Italian Television Archives

Describe What You See with Multimodal Large Language Models to Enhance Video Recommendations

RealAC: A Domain-Agnostic Framework for Realistic and Actionable Counterfactual Explanations

Built with on top of