The field of artificial intelligence is moving towards increased transparency and explainability, with a focus on multimodal learning and human-AI collaboration. Recent developments have highlighted the importance of designing and evaluating models that genuinely integrate visual and textual cues, rather than relying on single-modality signals. This shift is driven by the need for more accurate and trustworthy decision-making in high-stakes applications, such as healthcare and finance. Noteworthy papers in this area include ReVise, which introduces a visual analytic workflow for incremental recourse planning, and MultiSHAP, a model-agnostic interpretability framework for explaining cross-modal interactions in multimodal AI models. Additionally, papers like Your Model Is Unfair, Are You Even Aware? and On the Risk of Misleading Reports: Diagnosing Textual Biases in Multimodal Clinical AI have shed light on the importance of addressing bias and misinformation in AI systems.
Advances in Explainable AI and Multimodal Learning
Sources
StackLiverNet: A Novel Stacked Ensemble Model for Accurate and Interpretable Liver Disease Detection
Your Model Is Unfair, Are You Even Aware? Inverse Relationship Between Comprehension and Trust in Explainability Visualizations of Biased ML Models