Advances in Large Language Model Alignment and Optimization

The field of natural language processing is moving towards more effective and efficient alignment of large language models (LLMs) with human preferences. Recent developments have focused on improving the quality and diversity of training data, as well as developing new methods for optimizing LLMs. One key area of research is the use of multi-objective alignment, which allows LLMs to balance multiple competing objectives, such as informativeness and conciseness. Another important area is the development of more effective and efficient methods for generating and selecting high-quality training data, including the use of synthetic data and data selection principles. Additionally, researchers are exploring new approaches to prompt engineering, including the use of controlled natural language and inference-aware prompt optimization. Noteworthy papers in this area include: Pareto Multi-Objective Alignment for Language Models, which proposes a principled and computationally efficient algorithm for multi-objective alignment. Beyond Single: A Data Selection Principle for LLM Alignment via Fine-Grained Preference Signals, which introduces a novel data selection principle that advocates for selecting a subset of high-consensus data for efficient training. Inference-Aware Prompt Optimization for Aligning Black-Box Large Language Models, which introduces a unified framework that jointly optimizes the prompt and inference scale, while being aware of the inference budget and different task objectives.

Sources

Annotating Errors in English Learners' Written Language Production: Advancing Automated Written Feedback Systems

When Prompt Engineering Meets Software Engineering: CNL-P as Natural and Robust "APIs'' for Human-AI Interaction

A Stable and Principled Loss Function for Direct Language Model Alignment

Beyond Single: A Data Selection Principle for LLM Alignment via Fine-Grained Preference Signals

Pareto Multi-Objective Alignment for Language Models

Enhancing Small LLM Alignment through Margin-Based Objective Modifications under Resource Constraints

Beyond Ordinal Preferences: Why Alignment Needs Cardinal Human Feedback

APIO: Automatic Prompt Induction and Optimization for Grammatical Error Correction and Text Simplification

SYNAPSE-G: Bridging Large Language Models and Graph Learning for Rare Event Classification

AutoGeTS: Knowledge-based Automated Generation of Text Synthetics for Improving Text Classification

Inference-Aware Prompt Optimization for Aligning Black-Box Large Language Models

Inductive Bias Extraction and Matching for LLM Prompts

MCP2OSC: Parametric Control by Natural Language

Diversity First, Quality Later: A Two-Stage Assumption for Language Model Alignment

Built with on top of