BriefGPT.xyz
Sep, 2021
自适应分词的语言模型高效域适应
Efficient Domain Adaptation of Language Models via Adaptive Tokenization
HTML
PDF
Vin Sachidananda, Jason S. Kessler, Yi-an Lai
TL;DR
通过适应标记器,将预训练语言模型转移到新领域,提供了与基于专门领域的预训练方法相当的性能,同时生成的模型更小且训练和推断时间更短。
Abstract
Contextual embedding-based language models trained on large data sets, such as BERT and RoBERTa, provide strong performance across a wide range of tasks and are ubiquitous in modern
nlp
. It has been observed that
fine-t
→