BriefGPT.xyz
Jan, 2023
在保留通用知识的同时调整语言模型
Adapting a Language Model While Preserving its General Knowledge
HTML
PDF
Zixuan Ke, Yijia Shao, Haowei Lin, Hu Xu, Lei Shu...
TL;DR
本文提出了一种新的领域适应预训练方法,通过软遮罩注意力头并对一般和完全表示进行对比学习,以更智能的方式对 LM 中的知识进行适应,实验结果表明了该方法的有效性。
Abstract
domain-adaptive pre-training
(or DA-training for short), also known as post-training, aims to train a pre-trained general-purpose
language model
(LM) using an
→