The field of Artificial Intelligence (AI) is moving towards increased transparency and explainability, with a focus on developing techniques that can provide insights into the decision-making processes of AI models. Recent research has highlighted the importance of explainability in various applications, including language grounding, autonomous vehicles, and healthcare. The development of counterfactual explanations, model-agnostic approaches, and world models are some of the innovative techniques being explored to address the black box problem in AI. These techniques aim to provide meaningful explanations that can help practitioners and engineers improve model performance, identify biases, and enhance trust in AI systems.
Noteworthy papers in this area include:
- Why Are You Wrong? Counterfactual Explanations for Language Grounding with 3D Objects, which presents a method for generating counterfactual examples to explain incorrect predictions.
- Realistic Counterfactual Explanations for Machine Learning-Controlled Mobile Robots using 2D LiDAR, which introduces a novel approach for generating realistic counterfactual explanations in mobile robotics.
- Explainable Reinforcement Learning Agents Using World Models, which proposes a technique for using world models to generate explanations for model-based deep RL agents.