BriefGPT.xyz
Jun, 2023
跨语言转移的高效语言特定模型精简
Distilling Efficient Language-Specific Models for Cross-Lingual Transfer
HTML
PDF
Alan Ansell, Edoardo Maria Ponti, Anna Korhonen, Ivan Vulić
TL;DR
提出从大规模多语言Transformer中提取压缩的、语言特定的模型的方法,通过二阶段稀疏精调生成双语模型,从而在目标语言性能方面表现出最小的降级而且速度更快。
Abstract
Massively
multilingual transformers
(MMTs), such as mBERT and XLM-R, are widely used for
cross-lingual transfer learning
. While these are pretrained to represent hundreds of languages, end users of NLP systems ar
→