BriefGPT.xyz
Jan, 2025
基于互信息上界的LoRA缩放法则
The Scaling Law for LoRA Base on Mutual Information Upper Bound
HTML
PDF
Jing Zhang, Hui Gao, Peng Zhang, Shuzhen Sun, Chang Yang...
TL;DR
本文研究了模型性能、参数和数据复杂性之间的关系,特别是在大型模型的LoRA微调过程中。提出了一种基于互信息上界的内部度量,能够更准确地捕捉微调过程中新旧知识之间的依赖关系。实验结果表明,该指标在评估LoRA微调的缩放法则方面优于传统的外部度量方法。
Abstract
LoRA
(Low-Rank Adaptation) is a widely used
Model Fine-Tuning
method. In fine-tuning, the law among model performance, model parameters, and data complexity has been a focal issue in the field. Existing methods o
→