BriefGPT.xyz
Oct, 2023
通过掩码微调来弥合标记剪枝和完全预训练之间的差距
Bridging The Gaps Between Token Pruning and Full Pre-training via Masked Fine-tuning
HTML
PDF
Fengyuan Shi, Limin Wang
TL;DR
采用蒙版微调为动态视觉转换器的静态预训练基础模型提供更好的初始化,以提高准确性并增强其对遮挡的鲁棒性和对信息丢失的抵抗能力。
Abstract
Despite the success of
transformers
on various computer vision tasks, they suffer from excessive memory and computational cost. Some works present dynamic vision
transformers
to accelerate inference by pruning re
→