Ke Hu, Tara N. Sainath, Bo Li, Nan Du, Yanping Huang...
TL;DR利用单一多语言语言模型(LM)来进行多语言浅层融合任务,并将其应用于最先进的端到端模型,相对于类似推理期间的密集 LM ,GLaM 可将英语长尾测试集的 WER 降低4.4 %,平均相对 WER 降低3.85%,并且最高降低10%。
Abstract
While large language models (LLM) have made impressive progress in natural language processing, it remains unclear how to utilize them in improving automatic speech recognition (ASR). In this work, we propose to