The field of natural language processing is witnessing significant developments, particularly in the areas of large language models (LLMs) and speech recognition. Researchers are exploring new evaluation frameworks and benchmarks to assess the performance of LLMs, such as dynamic and adversarial evaluation environments, to better understand their capabilities and limitations. Additionally, there is a growing focus on improving the robustness and reliability of speech recognition systems, including the development of new architectures and techniques to mitigate hallucination errors. Noteworthy papers in this area include: Evaluating from Benign to Dynamic Adversarial: A Squid Game for Large Language Models, which introduces a dynamic and adversarial evaluation environment to assess LLMs. Listen Like a Teacher: Mitigating Whisper Hallucinations using Adaptive Layer Attention and Knowledge Distillation, which presents a two-stage architecture to reduce hallucinations in Whisper-style ASR systems. AfriSpeech-MultiBench: A Verticalized Multidomain Multicountry Benchmark Suite for African Accented English ASR, which provides a comprehensive benchmark suite for evaluating ASR systems on African accented English.
Advancements in Large Language Models and Speech Recognition
Sources
Learn to Select: Exploring Label Distribution Divergence for In-Context Demonstration Selection in Text Classification
Listen Like a Teacher: Mitigating Whisper Hallucinations using Adaptive Layer Attention and Knowledge Distillation