BriefGPT.xyz
Jan, 2020
预训练语言模型的可比性
On the comparability of Pre-trained Language Models
HTML
PDF
Matthias Aßenmacher, Christian Heumann
TL;DR
该文章介绍了近期在自然语言处理中提高模型性能所用的三种方法:更加复杂的语言模型、更大的语料库和并行计算,总结了一些最近两年新增的大型预训练语言模型,并探讨了新架构与资源所带来的影响,以此为起点提供了一些方便可复制的研究结果。
Abstract
Recent developments in
unsupervised representation learning
have successfully established the concept of
transfer learning
in NLP. Mainly three forces are driving the improvements in this area of research: More e
→