BriefGPT.xyz
Aug, 2021
LICHEE:基于多粒度分词的语言模型预训练优化
LICHEE: Improving Language Model Pre-training with Multi-grained Tokenization
HTML
PDF
Weidong Guo, Mingjun Zhao, Lusheng Zhang, Di Niu, Jinwen Luo...
TL;DR
本文提出一种名为LICHEE的简单而有效的预训练方法,旨在高效地融合输入文本的多粒度信息,以提高各种预先训练的语言模型的表示能力。实验结果表明,我们的方法在广泛的NLU任务中取得了全面的改进,并且我们的最佳集成模型在CLUE基准大赛上达到了最先进的性能。
Abstract
Language model
pre-training
based on large corpora has achieved tremendous success in terms of constructing enriched contextual representations and has led to significant performance gains on a diverse range of Natural Language Understanding (
→