BriefGPT.xyz
Jul, 2024
增加模型容量的免费方式:参数高效微调的简单策略
Increasing Model Capacity for Free: A Simple Strategy for Parameter Efficient Fine-tuning
HTML
PDF
Haobo Song, Hao Zhao, Soumajit Majumder, Tao Lin
TL;DR
通过低秩更新和并行权重模块来增强模型容量,提出了一种名为CapaBoost的简单且有效的策略,可无需添加参数,整合至各种现有的参数高效微调方法中,实验证明其在多个下游任务中均能显著提升性能。
Abstract
fine-tuning
large pre-trained foundation models
, such as the 175B GPT-3, has attracted more attention for downstream tasks recently. While parameter-efficient
→