BriefGPT.xyz
May, 2021
HerBERT:基于Transformer的波兰语预训练语言模型的高效实现
HerBERT: Efficiently Pretrained Transformer-based Language Model for Polish
HTML
PDF
Robert Mroczkowski, Piotr Rybak, Alina Wróblewska, Ireneusz Gawlik
TL;DR
本文介绍了第一项针对波兰语的去构建BERT的训练过程的消融研究,其中探讨了跨语言训练和其他因素,最终实现了一个波兰语BERT模型HerBERT,并在多个下游任务中达到了最佳结果。
Abstract
bert-based models
are currently used for solving nearly all
natural language processing
(NLP) tasks and most often achieve state-of-the-art results. Therefore, the NLP community conducts extensive research on und
→