The field of natural language processing is moving towards a deeper understanding of the biases present in large language models. Recent studies have highlighted the need for more nuanced and effective methods to mitigate these biases, which can perpetuate harmful stereotypes and reinforce existing social inequalities. The use of prompting as a fairness intervention has shown promise, but its effects can be highly model-specific and may not always lead to more diverse and representative outputs. Furthermore, the evaluation of bias in language models is a complex task, and the choice of metrics and testing designs can significantly impact the results.
Noteworthy papers in this area include: Prompting Away Stereotypes, which introduces a pilot benchmark for assessing representational societal bias in text-to-image models and highlights the potential of prompting as a fairness intervention. Bias after Prompting, which investigates the transfer of biases from pre-trained models to adapted models and finds that popular prompt-based mitigation methods do not consistently prevent biases from transferring. Measuring Bias or Measuring the Task, which examines how signaling the evaluative purpose of a task impacts measured gender bias in language models and highlights the brittleness of bias evaluations.