BriefGPT.xyz
Jul, 2024
LLM内部状态揭示面对查询时的幻觉风险
LLM Internal States Reveal Hallucination Risk Faced With a Query
HTML
PDF
Ziwei Ji, Delong Chen, Etsuko Ishii, Samuel Cahyawijaya, Yejin Bang...
TL;DR
研究大型语言模型的幻觉问题如何通过自我感知估计幻觉风险,通过对神经元、激活层和令牌的研究,提出了一种探测器来实现语言模型的自我评估,试验结果表明其幻觉估计准确率达到84.32%。
Abstract
The
hallucination problem
of
large language models
(LLMs) significantly limits their reliability and trustworthiness. Humans have a
self-awarenes
→