BriefGPT.xyz
Oct, 2021
SLAM: 通过语音-文本联合预训练实现语音和语言建模的统一编码器
SLAM: A Unified Encoder for Speech and Language Modeling via Speech-Text Joint Pre-Training
HTML
PDF
Ankur Bapna, Yu-an Chung, Nan Wu, Anmol Gulati, Ye Jia...
TL;DR
将无监督预训练应用于语言理解,在语音和文本之间建立单一模型,包括BERT目标和w2v-BERT目标以及其他预训练技术改进,同时在GLUE任务中也取得了不俗的竞争力。
Abstract
unsupervised pre-training
is now the predominant approach for both text and speech understanding.
self-attention models
pre-trained on large amounts of unannotated data have been hugely successful when fine-tuned
→