BriefGPT.xyz
Jul, 2021
通过细粒度流形蒸馏学习高效视觉Transformer
Efficient Vision Transformers via Fine-Grained Manifold Distillation
HTML
PDF
Ding Jia, Kai Han, Yunhe Wang, Yehui Tang, Jianyuan Guo...
TL;DR
本文提出了一种基于细粒度流形知识蒸馏的方法,旨在减少以往视觉变换器的计算量,并在ImageNet-1k分类基准测试中实现了76.5%的高水平准确性。
Abstract
This paper studies the model compression problem of vision
transformers
. Benefit from the self-attention module, transformer architectures have shown extraordinary performance on many
computer vision
tasks. Altho
→