The field of retrieval-augmented generation (RAG) is rapidly evolving, with a focus on improving the adaptability and reasoning capabilities of RAG systems. Recent developments have centered around incorporating human feedback, decoupling semantic matching from contextual assembly, and enhancing the composability and scalability of retrieval systems. Notably, researchers are exploring novel approaches to feedback-driven adaptation, iterative retrieval, and multi-step reasoning, which have shown promising results in various benchmarks.
Some noteworthy papers in this area include: DPRM, which introduces a dual implicit process reward model for multi-hop question answering, achieving up to 16.6% improvement on Hit@1. Think Before You Retrieve, which proposes a training framework that enables compact models to perform iterative retrieval through learned search strategies, outperforming retrievers up to 200-400x larger on five of six benchmarks. Structured RAG, which constructs a structured representation of the corpus and translates natural-language queries into formal queries, substantially outperforming common RAG systems and long-context LLMs on aggregative queries.