Advancements in Large Language Models and Speech Recognition

The field of natural language processing is witnessing significant developments, particularly in the areas of large language models (LLMs) and speech recognition. Researchers are exploring new evaluation frameworks and benchmarks to assess the performance of LLMs, such as dynamic and adversarial evaluation environments, to better understand their capabilities and limitations. Additionally, there is a growing focus on improving the robustness and reliability of speech recognition systems, including the development of new architectures and techniques to mitigate hallucination errors. Noteworthy papers in this area include: Evaluating from Benign to Dynamic Adversarial: A Squid Game for Large Language Models, which introduces a dynamic and adversarial evaluation environment to assess LLMs. Listen Like a Teacher: Mitigating Whisper Hallucinations using Adaptive Layer Attention and Knowledge Distillation, which presents a two-stage architecture to reduce hallucinations in Whisper-style ASR systems. AfriSpeech-MultiBench: A Verticalized Multidomain Multicountry Benchmark Suite for African Accented English ASR, which provides a comprehensive benchmark suite for evaluating ASR systems on African accented English.

Sources

Evaluating from Benign to Dynamic Adversarial: A Squid Game for Large Language Models

Learn to Select: Exploring Label Distribution Divergence for In-Context Demonstration Selection in Text Classification

M-DAIGT: A Shared Task on Multi-Domain Detection of AI-Generated Text

Listen Like a Teacher: Mitigating Whisper Hallucinations using Adaptive Layer Attention and Knowledge Distillation

AfriSpeech-MultiBench: A Verticalized Multidomain Multicountry Benchmark Suite for African Accented English ASR

Fine-tuning Pre-trained Audio Models for COVID-19 Detection: A Technical Report

Liars' Bench: Evaluating Lie Detectors for Language Models

Classification of worldwide news articles by perceived quality, 2018-2024

Built with on top of