Advances in Multilingual Language Models

The field of natural language processing is witnessing significant advancements in multilingual language models, with a growing focus on improving cross-lingual transfer, cultural alignment, and sociolinguistic diversity. Researchers are exploring innovative approaches to bridge the performance gap between high-resource and low-resource languages, including fine-tuning on synthetic code-switched text, using word association learning, and developing culturally grounded evaluation frameworks. These efforts aim to enhance the fairness and robustness of large language models in multilingual contexts, enabling more effective communication and understanding across linguistic and cultural boundaries. Noteworthy papers include: When Does Language Transfer Help, which investigates the effectiveness of sequential fine-tuning for cross-lingual euphemism detection. ALIGN, which introduces a cost-efficient approach to modeling and aligning culture in large language models using word association learning. Long Chain-of-Thought Reasoning Across Languages, which presents a systematic study of long chain-of-thought generation across multiple languages.

Sources

Beyond the Rosetta Stone: Unification Forces in Generalization Dynamics

When Does Language Transfer Help? Sequential Fine-Tuning for Cross-Lingual Euphemism Detection

SEA-BED: Southeast Asia Embedding Benchmark

The Cultural Gene of Large Language Models: A Study on the Impact of Cross-Corpus Training on Model Values and Biases

LoraxBench: A Multitask, Multilingual Benchmark Suite for 20 Indonesian Languages

Breaking Language Barriers: Equitable Performance in Multilingual Language Models

ALIGN: Word Association Learning for Cross-Cultural Generalization in Large Language Models

Benchmarking Sociolinguistic Diversity in Swahili NLP: A Taxonomy-Guided Approach

Evaluating Multilingual and Code-Switched Alignment in LLMs via Synthetic Natural Language Inference

Long Chain-of-Thought Reasoning Across Languages

Bridging the Culture Gap: A Framework for LLM-Driven Socio-Cultural Localization of Math Word Problems in Low-Resource Languages

WangchanThaiInstruct: An instruction-following Dataset for Culture-Aware, Multitask, and Multi-domain Evaluation in Thai

UniCoM: A Universal Code-Switching Speech Generator

Built with on top of