The field of artificial intelligence is moving towards developing more fair and unbiased systems. Recent research has focused on identifying and mitigating biases in language models, retrieval-augmented generation, and other AI systems. Studies have shown that biases can be introduced through various means, including data poisoning, prompt injection, and unfairness in tool selection. To address these issues, researchers have proposed novel debiasing methods, such as BiasUnlearn, Open-DeBias, and BiasFreeBench, which have demonstrated promising results in reducing biases while preserving language modeling capabilities. Noteworthy papers include Open-DeBias, which introduces a comprehensive benchmark for evaluating biases across a wide range of categories and subgroups, and BiasUnlearn, which proposes a novel model debiasing framework that achieves targeted debiasing via dual-pathway unlearning mechanisms.
Advances in Mitigating Bias in AI Systems
Sources
Your RAG is Unfair: Exposing Fairness Vulnerabilities in Retrieval-Augmented Generation via Backdoor Attacks
Assessing Algorithmic Bias in Language-Based Depression Detection: A Comparison of DNN and LLM Approaches