BriefGPT.xyz
Oct, 2024
LLM-Rank:一种图论方法用于剪枝大型语言模型
LLM-Rank: A Graph Theoretical Approach to Pruning Large Language Models
HTML
PDF
David Hoffmann, Kailash Budhathoki, Matthaeus Kleindessner
TL;DR
本研究解决了大型语言模型在部署时面临的尺寸和成本问题,提出了一种利用图论中心性度量的创新剪枝方法。该方法通过创建加权有向无环图并应用加权PageRank中心性度量计算节点重要性,从而实现更高的精度保留,MLPRank和LLMRank表现出比传统方法更优的性能。
Abstract
The evolving capabilities of
Large Language Models
are accompanied by growing sizes and deployment costs, necessitating effective inference optimisation techniques. We propose a novel
Pruning
method utilising
→