RA-DIT: Retrieval-Augmented Dual Instruction Tuning
https://papers.cool/arxiv/2310.01352
Authors: Xi Victoria Lin ; Xilun Chen ; Mingda Chen ; Weijia Shi ; Maria Lomeli ; Rich James ; Pedro Rodriguez ; Jacob Kahn ; Gergely Szilvasy ; Mike Lewis ; Luke Zettlemoyer ; Scott Yih
Summary: Retrieval-augmented language models (RALMs) improve performance by accessing long-tail and up-to-date knowledge from external data stores, but are challenging to build. Existing approaches require either expensive retrieval-specific modifications to LM pre-training or use post-hoc integration of the data store that leads to suboptimal performance. We introduce Retrieval-Augmented Dual Instruction Tuning (RA-DIT), a lightweight fine-tuning methodology that provides a third option by retrofitting any LLM with retrieval capabilities. Our approach operates in two distinct fine-tuning steps: (1) one updates a pre-trained LM to better use retrieved information, while (2) the other updates the retriever to return more relevant results, as preferred by the LM. By fine-tuning over tasks that require both knowledge utilization and contextual awareness, we demonstrate that each stage yields significant performance improvements, and using both leads to additional gains. Our best model, RA-DIT 65B, achieves state-of-the-art performance across a range of knowledge-intensive zero- and few-shot learning benchmarks, significantly outperforming existing in-context RALM approaches by up to +8.9% in 0-shot setting and +1.4% in 5-shot setting on average.
A: 这篇论文试图解决的问题是如何有效地构建和改进检索增强型语言模型(Retrieval-Augmented Language Models, RALMs)。具体来说,论文针对以下挑战:
总的来说,这篇论文的目标是提出一种有效的方法来增强大型语言模型的检索能力,使其能够在各种知识密集型任务中更好地利用外部知识,同时避免现有方法的高计算成本和性能限制。
A: 这篇论文提到了多个与检索增强型语言模型(RALMs)相关的研究,以下是一些关键的相关研究: