BriefGPT.xyz
Oct, 2022
PATS:针对预训练语言模型的敏感度感知噪声学习
PATS: Sensitivity-aware Noisy Learning for Pretrained Language Models
HTML
PDF
Yupeng Zhang, Hongzhi Zhang, Sirui Wang, Wei Wu, Zhoujun Li
TL;DR
本文提出了一种嘈杂训练机制PAT(根据敏感性的扰动),通过让一些不敏感的参数添加嘈杂值,以激活他们的下游任务贡献,从而提高预训练语言模型(PLMs)的微调性能,并在GLUE基准测试中进行了广泛的实验,证明了该方法的有效性。
Abstract
A wide range of
nlp
tasks benefit from the
fine-tuning
of
pretrained language models
(PLMs). However, a number of redundant parameters whi
→