Semantic Scholar Open Access 2021 1287 sitasi

Twins: Revisiting the Design of Spatial Attention in Vision Transformers

Xiangxiang Chu Zhi Tian Yuqing Wang Bo Zhang Haibing Ren +3 lainnya

Abstrak

Very recently, a variety of vision transformer architectures for dense prediction tasks have been proposed and they show that the design of spatial attention is critical to their success in these tasks. In this work, we revisit the design of the spatial attention and demonstrate that a carefully-devised yet simple spatial attention mechanism performs favourably against the state-of-the-art schemes. As a result, we propose two vision transformer architectures, namely, Twins-PCPVT and Twins-SVT. Our proposed architectures are highly-efficient and easy to implement, only involving matrix multiplications that are highly optimized in modern deep learning frameworks. More importantly, the proposed architectures achieve excellent performance on a wide range of visual tasks, including image level classification as well as dense detection and segmentation. The simplicity and strong performance suggest that our proposed architectures may serve as stronger backbones for many vision tasks. Our code is released at https://github.com/Meituan-AutoML/Twins .

Topik & Kata Kunci

Penulis (8)

X

Xiangxiang Chu

Z

Zhi Tian

Y

Yuqing Wang

B

Bo Zhang

H

Haibing Ren

X

Xiaolin Wei

H

Huaxia Xia

C

Chunhua Shen

Format Sitasi

Chu, X., Tian, Z., Wang, Y., Zhang, B., Ren, H., Wei, X. et al. (2021). Twins: Revisiting the Design of Spatial Attention in Vision Transformers. https://www.semanticscholar.org/paper/6709d5583f658f589ae6a2184805933aceb18849

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2021
Bahasa
en
Total Sitasi
1287×
Sumber Database
Semantic Scholar
Akses
Open Access ✓