BriefGPT.xyz
Jun, 2019
使用整词掩码的中文BERT预训练
Pre-Training with Whole Word Masking for Chinese BERT
HTML
PDF
Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang...
TL;DR
本文旨在介绍中文BERT预训练语言模型的整词Masking策略以及一系列简单但有效的中文预训练语言模型,其中包括MacBERT,提高了RoBERTa的性能,并通过十个中文NLP任务的广泛实验证明了MacBERT在很多NLP任务中可以达到最先进的性能。
Abstract
Bidirectional Encoder Representations from Transformers (
bert
) has shown marvelous improvements across various NLP tasks. Recently, an upgraded version of
bert
has been released with
→