BriefGPT.xyz
Oct, 2024
渐进学习:利用部分掌握知识优化大型语言模型的微调
Gradual Learning: Optimizing Fine-Tuning with Partially Mastered Knowledge in Large Language Models
HTML
PDF
Bozhou Li, Hao Liang, Yang Li, Fangcheng Fu, Hongzhi Yin...
TL;DR
本文针对大型语言模型在微调阶段无法有效学习新知识的问题提出解决方案。通过实验,我们提出了一种两阶段的微调策略,该策略能够在保留已掌握知识的同时,显著提高模型在特定任务上的准确性和新知识的掌握率,最高提高了24%。
Abstract
During the pretraining phase,
Large Language Models
(LLMs) acquire vast amounts of knowledge from extensive text corpora. Nevertheless, in later stages such as
Fine-Tuning
and inference, the model may encounter k
→