BriefGPT.xyz
Oct, 2024
长文本生成中的大型语言模型原子校准
Atomic Calibration of LLMs in Long-Form Generations
HTML
PDF
Caiqi Zhang, Ruihan Yang, Zhisong Zhang, Xinting Huang, Sen Yang...
TL;DR
该研究解决了大型语言模型(LLMs)在长文本生成中常见的幻觉问题,并提出了一种新的原子校准方法,能够以细粒度评估事实准确性。实验结果表明,原子校准不仅适用于长文本生成,还能提升整体校准效果,揭示了模型信心水平的动态变化。
Abstract
Large Language Models
(LLMs) often suffer from hallucinations, posing significant challenges for real-world applications.
Confidence Calibration
, which estimates the underlying uncertainty of model predictions, i
→