BriefGPT.xyz
Nov, 2024
解耦大语言模型中的记忆与推理能力
Disentangling Memory and Reasoning Ability in Large Language Models
HTML
PDF
Mingyu Jin, Weidi Luo, Sitao Cheng, Xinyi Wang, Wenyue Hua...
TL;DR
本文解决了现有大语言模型(LLM)推理过程中知识检索与推理步骤不明确的问题。我们提出了一种新推理范式,将复杂推理过程分解为记忆回忆和推理两种独立的操作,从而提高了模型性能和推理过程的可解释性。实验结果表明,这种分解有助于减少模型错误和知识遗忘的现象,提高了在关键领域中的可靠性。
Abstract
Large Language Models
(LLMs) have demonstrated strong performance in handling complex tasks requiring both extensive knowledge and
Reasoning
abilities. However, the existing LLM inference pipeline operates as an
→