BriefGPT.xyz
Oct, 2021
基于Hessian感知显著性的全局Vision Transformer压缩
NViT: Vision Transformer Compression and Parameter Redistribution
HTML
PDF
Huanrui Yang, Hongxu Yin, Pavlo Molchanov, Hai Li, Jan Kautz
TL;DR
本研究提出了一种称为NViT的基于Hessian的全局结构裁剪方法,能够比以往更高效地利用ViT模型的参数,使得NViT-Base在ImageNet-1K数据集上具备了比DeiT-Base更高的准确率、更低的FLOPs和参数数量以及更快的运行速度。
Abstract
transformers
yield state-of-the-art results across many tasks. However, they still impose huge computational costs during inference. We apply global,
structural pruning
with latency-aware regularization on all pa
→