BriefGPT.xyz
Jan, 2019
跨语言语言模型预训练
Cross-lingual Language Model Pretraining
HTML
PDF
Guillaume Lample, Alexis Conneau
TL;DR
本文提出了两种跨语言学习模型的方法(XLMs):一种是仅依赖于单语数据的无监督方式, 另一种是利用新的跨语言模型目标并使用平行数据的有监督方式。通过这些方法在跨语言分类、无监督和有监督机器翻译中取得了最先进的结果。
Abstract
Recent studies have demonstrated the efficiency of
generative pretraining
for English natural language understanding. In this work, we extend this approach to multiple languages and show the effectiveness of
cross-lingu
→