BriefGPT.xyz
May, 2020
双向编码器句法结构提取预训练
Syntactic Structure Distillation Pretraining For Bidirectional Encoders
HTML
PDF
Adhiguna Kuncoro, Lingpeng Kong, Daniel Fried, Dani Yogatama, Laura Rimell...
TL;DR
本文研究了文本表示学习器在句法表现上的优越性和在自然语言句法中的应用,提出了将句法模型的预测注入BERT的知识蒸馏策略,并表明在一系列结构预测任务中,这种方法能够显著地减少相对误差。
Abstract
textual representation learners
trained on large amounts of data have achieved notable success on downstream tasks; intriguingly, they have also performed well on challenging tests of
syntactic competence
. Given
→