BriefGPT.xyz
Jun, 2022
联合编码器-解码器自监督预训练用于ASR
Joint Encoder-Decoder Self-Supervised Pre-training for ASR
HTML
PDF
Arunkumar A, Umesh S
TL;DR
本文提出了一种新的自监督学习范式,利用解码器的威力提高语音识别下游任务的性能。HuBERT框架用于计算编码器的传统掩蔽预测损失,同时在框架中引入了解码器和目标准备策略。最终,我们使用一个多任务SSL设置,其同时优化编码器和解码器损失,实现了ASR表现的25%相对改进。
Abstract
self-supervised learning
(SSL) has shown tremendous success in various speech-related downstream tasks, including
automatic speech recognition
(ASR). The output embeddings of the SSL model are treated as powerful
→