BriefGPT.xyz
Apr, 2023
Slide-Transformer: 具有局部自注意力的分层视觉变压器
Slide-Transformer: Hierarchical Vision Transformer with Local Self-Attention
HTML
PDF
Xuran Pan, Tianzhu Ye, Zhuofan Xia, Shiji Song, Gao Huang
TL;DR
本文提出一种新颖的本地自注意力模块Slide Attention,使用深度卷积和变形平移技术实现高效、灵活和通用的局部特征学习,适用于各种高级视觉Transformer模型,并在多项基准测试中实现了持续的性能改进。
Abstract
self-attention
mechanism has been a key factor in the recent progress of
vision transformer
(ViT), which enables adaptive feature extraction from global contexts. However, existing
→