BriefGPT.xyz
Aug, 2024
提升长上下文语言模型的多跳推理能力
Making Long-Context Language Models Better Multi-Hop Reasoners
HTML
PDF
Yanyang Li, Shuo Liang, Michael R. Lyu, Liwei Wang
TL;DR
本研究解决了当前长上下文语言模型在多跳推理中表现不足的问题,尤其是在嘈杂上下文中。我们提出了“推理与归因”的新方法,通过促使模型在推理过程中为每个断言提供归因,从而增强推理能力。实验结果表明,经过微调的模型在多跳推理基准测试上取得了与现有先进模型接近的竞争性表现。
Abstract
Recent advancements in long-context modeling have enhanced language models (LMs) for complex tasks across multiple
NLP applications
. Despite this progress, we find that these models struggle with
Multi-hop reasoning
→