BriefGPT.xyz
Apr, 2021
针对大规模多语言ASR的端到端模型扩展
Scaling End-to-End Models for Large-Scale Multilingual ASR
HTML
PDF
Bo Li, Ruoming Pang, Tara N. Sainath, Anmol Gulati, Yu Zhang...
TL;DR
本篇研究旨在研究15种任务下,训练语言模型的容量瓶颈,并发现通过增加模型参数,添加深度,以及使用大型编码器等方法可以增加训练效率。
Abstract
Building
asr models
across many language families is a challenging
multi-task learning
problem due to large language variations and heavily unbalanced data. Existing work has shown positive transfer from high res
→