BriefGPT.xyz
Jan, 2023
XLM-V: 克服多语言掩码语言模型中的词汇瓶颈
XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
HTML
PDF
Davis Liang, Hila Gonen, Yuning Mao, Rui Hou, Naman Goyal...
TL;DR
本文介绍了一种新方法,通过降低语言之间的词汇共享并分配单独语言的词汇容量,实现非常大的多语言词汇量的扩展,用于构建 XLM-V 多语言模型,其表现优于 XLM-R 模型。
Abstract
Large multilingual language models typically rely on a single
vocabulary
shared across 100+ languages. As these models have increased in parameter count and depth,
vocabulary
size has remained largely unchanged.
→