BriefGPT.xyz
Jun, 2023
ShiftAddViT:基于混杂乘法原语的高效视觉Transformer
ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer
HTML
PDF
Haoran You, Huihong Shi, Yipin Guo, Yingyan, Lin
TL;DR
本文提出了一种名为ShiftAddViT的卷积神经网络,用于多种视觉任务,该网络通过将注意力和多层感知机重新参数化为位移和加法,从而实现在GPU上进行端到端推理加速,并在一定程度上提高了训练和推断效率。
Abstract
vision transformers
(ViTs) have shown impressive performance and have become a unified backbone for multiple vision tasks. But both attention and multi-layer perceptions (
mlps
) in ViTs are not efficient enough du
→