A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts
https://papers.cool/arxiv/2402.09727
Authors: Kuang-Huei Lee ; Xinyun Chen ; Hiroki Furuta ; John Canny ; Ian Fischer
Summary:
Current Large Language Models (LLMs) are not only limited to some maximum context length, but also are not able to robustly consume long inputs. To address these limitations, we propose ReadAgent, an LLM agent system that increases effective context length up to 20x in our experiments. Inspired by how humans interactively read long documents, we implement ReadAgent as a simple prompting system that uses the advanced language capabilities of LLMs to (1) decide what content to store together in a memory episode, (2) compress those memory episodes into short episodic memories called gist memories, and (3) take actions to look up passages in the original text if ReadAgent needs to remind itself of relevant details to complete a task. We evaluate ReadAgent against baselines using retrieval methods, using the original long contexts, and using the gist memories. These evaluations are performed on three long-document reading comprehension tasks: QuALITY, NarrativeQA, and QMSum. ReadAgent outperforms the baselines on all three tasks while extending the effective context window by 3-20x.

Q: 这篇论文试图解决什么问题?
A: 这篇论文试图解决大型语言模型(LLMs)在处理长文本时面临的两个主要问题:
为了克服这些限制,论文提出了一个名为ReadAgent的LLM代理系统。ReadAgent的设计灵感来源于人类如何交互式地阅读长文档,它通过以下三个主要步骤来增加有效上下文长度:
通过这种方法,ReadAgent能够在三个长文档阅读理解任务(QuALITY, NarrativeQA, 和 QMSum)上超越基线模型,同时显著扩展了有效上下文窗口,最高可达20倍。
Q: 有哪些相关研究?
A: 这篇论文提到了多个与长上下文LLMs相关的研究方向,包括: