BriefGPT.xyz
Jul, 2022
基于预训练语音模型的低资源语音识别改进:继续预训练与半监督训练
Improving Low-Resource Speech Recognition with Pretrained Speech Models: Continued Pretraining vs. Semi-Supervised Training
HTML
PDF
Mitchell DeHaven, Jayadev Billa
TL;DR
本文通过在多种低资源语言的未标注同语言音频数据上使用XLSR-53预训练模型进行持续预训练来改进自动语音识别的性能,结果表明持续预训练在单词错误率方面与半监督训练的效果相当且更加高效。
Abstract
self-supervised transformer
based models, such as wav2vec 2.0 and HuBERT, have produced significant improvements over existing approaches to
automatic speech recognition
(ASR). This is evident in the performance
→