SA-Net: Shuffle Attention for Deep Convolutional Neural Networks

Attention mechanisms, which enable a neural network to accurately focus on all the relevant elements of the input, have become an essential component to improve the performance of deep neural networks. There are mainly two attention mechanisms widely used in computer vision studies, spatial attentio...

Full description

Saved in:
Bibliographic Details
Published inProceedings of the ... IEEE International Conference on Acoustics, Speech and Signal Processing (1998) pp. 2235 - 2239
Main Authors Zhang, Qing-Long, Yang, Yu-Bin
Format Conference Proceeding
LanguageEnglish
Published IEEE 06.06.2021
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Attention mechanisms, which enable a neural network to accurately focus on all the relevant elements of the input, have become an essential component to improve the performance of deep neural networks. There are mainly two attention mechanisms widely used in computer vision studies, spatial attention and channel attention, which aim to capture the pixel-level pairwise relationship and channel dependency, respectively. Although fusing them together may achieve better performance than their individual implementations, it will inevitably increase the computational overhead. In this paper, we propose an efficient Shuffle Attention (SA) module to address this issue, which adopts Shuffle Units to combine two types of attention mechanisms effectively. Specifically, SA first groups channel dimensions into multiple sub-features before processing them in parallel. Then, for each sub-feature, SA utilizes a Shuffle Unit to depict feature dependencies in both spatial and channel dimensions. After that, all sub-features are aggregated and a "channel shuffle" operator is adopted to enable information communication between different sub-features. The proposed SA module is efficient yet effective, e.g., the parameters and computations of SA against the backbone ResNet50 are 300 vs. 25.56M and 2.76e-3 GFLOPs vs. 4.12 GFLOPs, respectively, and the performance boost is more than 1.34% in terms of Top-1 accuracy. Extensive experimental results on common-used benchmarks, including ImageNet-1k for classification, MS COCO for object detection, and instance segmentation, demonstrate that the proposed SA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity.
AbstractList Attention mechanisms, which enable a neural network to accurately focus on all the relevant elements of the input, have become an essential component to improve the performance of deep neural networks. There are mainly two attention mechanisms widely used in computer vision studies, spatial attention and channel attention, which aim to capture the pixel-level pairwise relationship and channel dependency, respectively. Although fusing them together may achieve better performance than their individual implementations, it will inevitably increase the computational overhead. In this paper, we propose an efficient Shuffle Attention (SA) module to address this issue, which adopts Shuffle Units to combine two types of attention mechanisms effectively. Specifically, SA first groups channel dimensions into multiple sub-features before processing them in parallel. Then, for each sub-feature, SA utilizes a Shuffle Unit to depict feature dependencies in both spatial and channel dimensions. After that, all sub-features are aggregated and a "channel shuffle" operator is adopted to enable information communication between different sub-features. The proposed SA module is efficient yet effective, e.g., the parameters and computations of SA against the backbone ResNet50 are 300 vs. 25.56M and 2.76e-3 GFLOPs vs. 4.12 GFLOPs, respectively, and the performance boost is more than 1.34% in terms of Top-1 accuracy. Extensive experimental results on common-used benchmarks, including ImageNet-1k for classification, MS COCO for object detection, and instance segmentation, demonstrate that the proposed SA outperforms the current SOTA methods significantly by achieving higher accuracy while having lower model complexity.
Author Zhang, Qing-Long
Yang, Yu-Bin
Author_xml – sequence: 1
  givenname: Qing-Long
  surname: Zhang
  fullname: Zhang, Qing-Long
  organization: Nanjing University,State Key Laboratory for Novel Software Technology
– sequence: 2
  givenname: Yu-Bin
  surname: Yang
  fullname: Yang, Yu-Bin
  organization: Nanjing University,State Key Laboratory for Novel Software Technology
BookMark eNotj8tOwzAURA0Cibb0C9j4B1LutZM4ZkUUnlJVkAISu8qJr0UgxFXigPh7AnQ1mrMYnZmzo853xBhHWCGCPr8v8rJ8lFqJbCVA4ErHGCdpdsCWWmU4YVQpJMkhmwmpdIQaXk7YfBjeACBTcTZjl2UebShc8PJ1dK4lnodAXWh8x53v-RXRjhe--_Tt-AtNyzc09n8Rvnz_PpyyY2fagZb7XLDnm-un4i5aP9xOfuuoESBD5CTWAHUl06qWzlnEhASSAS0rk9hJlVxSW-UMWSkFQT3VWFTKKatdbOWCnf3vNkS03fXNh-m_t_vD8geTa037
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/ICASSP39728.2021.9414568
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Engineering
EISBN 9781728176055
1728176050
EISSN 2379-190X
EndPage 2239
ExternalDocumentID 9414568
Genre orig-research
GroupedDBID 23M
6IE
6IF
6IH
6IK
6IL
6IM
6IN
AAJGR
AAWTH
ABLEC
ACGFS
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
IJVOP
IPLJI
M43
OCL
RIE
RIL
RIO
RNS
ID FETCH-LOGICAL-i203t-f31c00cb36bc3ffd115e21ea093ba5d728ef5cd7faed332e0cf5c42b7f7d9f4d3
IEDL.DBID RIE
IngestDate Wed Aug 27 02:39:02 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i203t-f31c00cb36bc3ffd115e21ea093ba5d728ef5cd7faed332e0cf5c42b7f7d9f4d3
PageCount 5
ParticipantIDs ieee_primary_9414568
PublicationCentury 2000
PublicationDate 2021-June-6
PublicationDateYYYYMMDD 2021-06-06
PublicationDate_xml – month: 06
  year: 2021
  text: 2021-June-6
  day: 06
PublicationDecade 2020
PublicationTitle Proceedings of the ... IEEE International Conference on Acoustics, Speech and Signal Processing (1998)
PublicationTitleAbbrev ICASSP
PublicationYear 2021
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0008748
Score 2.6564417
Snippet Attention mechanisms, which enable a neural network to accurately focus on all the relevant elements of the input, have become an essential component to...
SourceID ieee
SourceType Publisher
StartPage 2235
SubjectTerms channel attention
channel shuffle
Computer architecture
Computer vision
Conferences
grouped features
Image segmentation
Neural networks
Object detection
Signal processing
spatial attention
Title SA-Net: Shuffle Attention for Deep Convolutional Neural Networks
URI https://ieeexplore.ieee.org/document/9414568
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV07T8MwELZKJ1h4tIi3PDDiNIljJ2GiKlQFqVWlUKlb5cdZIKq2goSBX4-dpOUhBibHkWJbd5G_s_3dZ4QubQQumEolCaQUJKKak1TzgGgmYxEqEJS55OThiA8m0cOUTRvoapMLAwAl-Qw891ie5eulKtxWWSeNAov3yRbasgu3KldrM-smcZSsmTp-2rnvdbNsbME2dPytMPDqb39colJiSH8XDde9V9SRF6_Ipac-fgkz_nd4e6j9la2Hxxsc2kcNWBygnW9Cgy10k3XJCPJrnD0VxswBd_O84jliG7TiW4AVtq291_-hmGMn2lEWJUv8rY0m_bvH3oDUdyeQ59CnOTE0UL6vJOVSUWO0DfwgDED4KZWCaWsWMEzp2AjQlIbgK1uNQhmbWKcm0vQQNRfLBRwhzAKjQWouueGR9WviMF445beEycSIY9RytpitKnmMWW2Gk79fn6Jt54-SbcXPUDN_LeDc4nouL0qHfgIrWaQl
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3LTsJAFJ0gLtSNDzC-nYVLW9pOpw9XEpSAAiEpJOzIPO5EIwGirQu_3pm24CMuXM1Mk2mbe5uem_accxG60hU4oyLmlss5s3wiAyuWgWtJykPmCWCEGnFyfxB0xv7DhE4q6HqthQGAnHwGtpnm__LlQmTmU1kj9l2N99EG2tS4T91CrbV-70ahH624Ok7c6LaaSTLUcOsZBpfn2uXuH21UchRp76L-6voFeeTFzlJui49f1oz_vcE9VP_S6-HhGon2UQXmB2jnm9VgDd0mTWsA6Q1OnjKlZoCbaVowHbEuW_EdwBLrs72XTyKbYWPbkQ85T_ytjsbt-1GrY5XdE6xnzyGppYgrHEdwEnBBlJK69APPBebEhDMqdVhAUSFDxUAS4oEj9NL3eKhCGStfkkNUnS_mcIQwdZUELgMeqMDXmY0MyjPj_RZRHil2jGomFtNlYZAxLcNw8vfhS7TVGfV701538HiKtk1ucu5VcIaq6WsG5xrlU36RJ_cT7GCnbg
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+of+the+...+IEEE+International+Conference+on+Acoustics%2C+Speech+and+Signal+Processing+%281998%29&rft.atitle=SA-Net%3A+Shuffle+Attention+for+Deep+Convolutional+Neural+Networks&rft.au=Zhang%2C+Qing-Long&rft.au=Yang%2C+Yu-Bin&rft.date=2021-06-06&rft.pub=IEEE&rft.eissn=2379-190X&rft.spage=2235&rft.epage=2239&rft_id=info:doi/10.1109%2FICASSP39728.2021.9414568&rft.externalDocID=9414568