Advancements in Retrieval-Augmented Generation

The field of retrieval-augmented generation (RAG) is rapidly evolving, with a focus on improving the adaptability and reasoning capabilities of RAG systems. Recent developments have centered around incorporating human feedback, decoupling semantic matching from contextual assembly, and enhancing the composability and scalability of retrieval systems. Notably, researchers are exploring novel approaches to feedback-driven adaptation, iterative retrieval, and multi-step reasoning, which have shown promising results in various benchmarks.

Some noteworthy papers in this area include: DPRM, which introduces a dual implicit process reward model for multi-hop question answering, achieving up to 16.6% improvement on Hit@1. Think Before You Retrieve, which proposes a training framework that enables compact models to perform iterative retrieval through learned search strategies, outperforming retrievers up to 200-400x larger on five of six benchmarks. Structured RAG, which constructs a structured representation of the corpus and translates natural-language queries into formal queries, substantially outperforming common RAG systems and long-context LLMs on aggregative queries.

Sources

DMA: Online RAG Alignment with Human Feedback

Search Is Not Retrieval: Decoupling Semantic Matching from Contextual Assembly in RAG

A Preliminary Study of RAG for Taiwanese Historical Archives

Think Before You Retrieve: Learning Test-Time Adaptive Search with Small Language Models

DPRM: A Dual Implicit Process Reward Model in Multi-Hop Question Answering

Structured RAG for Answering Aggregative Questions

Efficient Model-Agnostic Continual Learning for Next POI Recommendation

Thinking Forward and Backward: Multi-Objective Reinforcement Learning for Retrieval-Augmented Reasoning

Built with on top of