BriefGPT.xyz
May, 2021
Transformer 模型微调和组合之间的相互作用
On the Interplay Between Fine-tuning and Composition in Transformers
HTML
PDF
Lang Yu, Allyson Ettinger
TL;DR
研究发现,预训练的转换语言模型在很多NLP任务上表现出色;然而,这些模型在短语级别的表示中,除了词汇内容外,缺乏复杂的组合短语信息,进一步的fine-tuning只能在情感任务中局部提高性能,而在重新释义任务中则由于数据集中可能存在干扰信号的原因而不能提供改进。
Abstract
pre-trained transformer language models
have shown remarkable performance on a variety of NLP tasks. However, recent research has suggested that
phrase-level representations
in these models reflect heavy influenc
→