BriefGPT.xyz
Apr, 2020
将多语言BERT扩展到低资源语言
Extending Multilingual BERT to Low-Resource Languages
HTML
PDF
Zihan Wang, Karthikeyan K, Stephen Mayhew, Dan Roth
TL;DR
本研究提出一种简单而有效的方法,扩展多语言BERT(E-BERT),使其可以为任何新语言提供帮助,并在27种语言上进行命名实体识别(NER)实验,结果表明我们的方法对已包含在M-BERT中的语言平均F1值提高了6%,对新语言提高了23%的F1值。
Abstract
multilingual bert
(M-BERT) has been a huge success in both supervised and zero-shot
cross-lingual transfer learning
. However, this success has focused only on the top 104 languages in Wikipedia that it was traine
→