BriefGPT.xyz
Feb, 2024
规模化大型语言模型微调的差分隐私零阶方法
Differentially Private Zeroth-Order Methods for Scalable Large Language Model Finetuning
HTML
PDF
Z Liu, J Lou, W Bao, Z Qin, K Ren
TL;DR
本文研究了差分隐私零阶方法在预训练语言模型中的潜力,通过近似梯度避免了 SGD 的可扩展性瓶颈,并提出了动态调度超参数的阶段性差分隐私零阶方法和减少可训练参数的数据无关剪枝技术,从理论和实证分析了这两种方法的效果。
Abstract
finetuning
on task-specific datasets is a widely-embraced paradigm of harnessing the powerful capability of pretrained LLMs for various downstream tasks. Due to the popularity of LLMs
finetuning
and its accompany
→