BriefGPT.xyz
Apr, 2020
利用预训练语言模型生成通用文本嵌入以实现可扩展推断
General Purpose Text Embeddings from Pre-trained Language Models for Scalable Inference
HTML
PDF
Jingfei Du, Myle Ott, Haoran Li, Xing Zhou, Veselin Stoyanov
TL;DR
研究使用共享文本编码器实现多任务推理以及使用二进制量化减少数据存储大小的方法,证明预训练的编码器在多个任务上表现的泛化性好。
Abstract
The state of the art on many
nlp
tasks is currently achieved by large
pre-trained language models
, which require a considerable amount of computation. We explore a setting where many different predictions are mad
→