BriefGPT.xyz
Jun, 2023
使您的预训练模型可逆:从参数到内存高效微调
Make Your Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning
HTML
PDF
Baohao Liao, Shaomu Tan, Christof Monz
TL;DR
本文提出了一种内存高效的微调方法(MEFT),通过在预训练语言模型中插入适配器以保留PLM的起点并使其可逆,同时将激活内存降低到84%的完全微调水平,并在GLUE基准测试中实现与完全微调相同的分数。
Abstract
Parameter-efficient
fine-tuning
(PEFT) of
pre-trained language models
(PLMs) has emerged as a highly successful approach, with training only a small number of parameters without sacrificing performance and becomi
→