BriefGPT.xyz
Apr, 2024
MixLoRA:基于LoRA的专家混合大语言模型微调增强
MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA based Mixture of Experts
HTML
PDF
Dengchun Li, Yingzi Ma, Naizheng Wang, Zhiyuan Cheng, Lei Duan...
TL;DR
提出了一种基于LoRA的资源高效稀疏MoE模型构建方法,名为MixLoRA,能够在消费级GPU上实现多个专家模型的并行微调,减少了GPU内存消耗41%和训练过程中的延迟17%。
Abstract
large language models
(LLMs) have showcased exceptional performance across a wide array of Natural Language Processing (NLP) tasks.
fine-tuning techniques
are commonly utilized to tailor pre-trained models to spe
→