BriefGPT.xyz
Oct, 2024
LLaVA-KD:多模态大语言模型知识蒸馏框架
LLaVA-KD: A Framework of Distilling Multimodal Large Language Models
HTML
PDF
Yuxuan Cai, Jiangning Zhang, Haoyang He, Xinwei He, Ao Tong...
TL;DR
本研究解决了多模态大语言模型(MLLM)在资源受限环境中应用受限的问题,通过引入一种新的LLaVA-KD框架,成功实现了从大规模模型(l-MLLM)向小规模模型(s-MLLM)的知识转移。研究表明,该方法在保持小模型架构不变的情况下,显著提升了性能,具有重要的应用潜力。
Abstract
The success of
Large Language Models
(LLM) has led researchers to explore
Multimodal
Large Language Models
(MLLM) for unified visual and l
→