Demonstrate-Search-Predict: Composing retrieval and language models for knowledge-intensive NLP
https://papers.cool/arxiv/2212.14024
Authors: Omar Khattab ; Keshav Santhanam ; Xiang Lisa Li ; David Hall ; Percy Liang ; Christopher Potts ; Matei Zaharia
Summary: Retrieval-augmented in-context learning has emerged as a powerful approach for addressing knowledge-intensive tasks using frozen language models (LM) and retrieval models (RM). Existing work has combined these in simple "retrieve-then-read" pipelines in which the RM retrieves passages that are inserted into the LM prompt. To begin to fully realize the potential of frozen LMs and RMs, we propose Demonstrate-Search-Predict (DSP), a framework that relies on passing natural language texts in sophisticated pipelines between an LM and an RM. DSP can express high-level programs that bootstrap pipeline-aware demonstrations, search for relevant passages, and generate grounded predictions, systematically breaking down problems into small transformations that the LM and RM can handle more reliably. We have written novel DSP programs for answering questions in open-domain, multi-hop, and conversational settings, establishing in early evaluations new state-of-the-art in-context learning results and delivering 37-120%, 8-39%, and 80-290% relative gains against the vanilla LM (GPT-3.5), a standard retrieve-then-read pipeline, and a contemporaneous self-ask pipeline, respectively. We release DSP at https://github.com/stanfordnlp/dsp
A: 这篇论文提出了一个名为DEMONSTRATE–SEARCH–PREDICT (DSP) 的框架,旨在解决以下问题:
总的来说,DSP框架试图通过更智能地结合检索和语言模型,提高知识密集型自然语言处理(NLP)任务的性能,同时简化模型开发过程。
A: 这篇论文提到了多项相关研究,以下是一些主要的相关工作: