BriefGPT.xyz
May, 2022
超越静态模型和测试集:在任务和语言间评测预训练模型的潜力
Beyond Static Models and Test Sets: Benchmarking the Potential of Pre-trained Models Across Tasks and Languages
HTML
PDF
Kabir Ahuja, Sandipan Dandapat, Sunayana Sitaram, Monojit Choudhury
TL;DR
本文提出了一种利用语言数据和语言类型学特征来预测跨语种语言模型性能的方法,以此取代传统基于翻译的方法评估系统,该方法表现良好并且能够可靠地估计模型在不同语言上的表现。
Abstract
Although recent
massively multilingual language models
(MMLMs) like mBERT and XLMR support around 100 languages, most existing
multilingual nlp benchmarks
provide evaluation data in only a handful of these langua
→