BriefGPT.xyz
Dec, 2020
让预训练语言模型成为更好的少样本学习者
Making Pre-trained Language Models Better Few-shot Learners
HTML
PDF
Tianyu Gao, Adam Fisch, Danqi Chen
TL;DR
LM-BFF提出了一种改进的面向小型语言模型的少样本 fine-tuning 方法以提升在多种NLP任务上的性能。通过与传统的 fine-tuning 方法相比,LM-BFF组合的技术在低资源环境下具有显著改进,最高可达30%,平均提高11%。
Abstract
The recent GPT-3 model (Brown et al., 2020) achieves remarkable few-shot
performance
solely by leveraging a natural-language prompt and a few task demonstrations as input context. Inspired by their findings, we study
fe
→