BriefGPT.xyz
Dec, 2023
ASPEN:使用单个GPU实现高吞吐量的大型语言模型LoRA微调
ASPEN: High-Throughput LoRA Fine-Tuning of Large Language Models with a Single GPU
HTML
PDF
Zhengmao Ye, Dengchun Li, Jingqi Tian, Tingfeng Lan, Jie Zuo...
TL;DR
ASPEN是一个高吞吐量的fine-tuning框架,通过LoRA方法在单个GPU上高效地训练多个任务,使用共享的预训练模型和自适应调度,可以显著节省GPU内存并提高训练吞吐量,同时减少工作轮转时间和训练延迟。
Abstract
Transformer-based
large language models
(LLMs) have demonstrated outstanding performance across diverse domains, particularly when fine-turned for specific domains. Recent studies suggest that the resources required for
→