BriefGPT.xyz
Feb, 2025
大型语言模型如何在上下文中进行两跳推理?
How Do LLMs Perform Two-Hop Reasoning in Context?
HTML
PDF
Tianyu Guo, Hanlin Zhu, Ruiqi Zhang, Jiantao Jiao, Song Mei...
TL;DR
本研究聚焦于大型语言模型(LLMs)在面对干扰前提时进行两跳推理的能力,揭示了它们从随机猜测到精准推理的学习机制。通过训练三层变换器并进行逆向工程,研究发现模型在初始阶段受到干扰影响,最终则能自如忽略干扰,实现高准确率。这一发现为理解LLMs的推理过程及其训练动态提供了新视角。
Abstract
"Socrates is human. All humans are mortal. Therefore, Socrates is mortal." This classical example demonstrates
Two-hop reasoning
, where a conclusion logically follows from two connected premises. While
Transformer
→