Squeeze-and-Excitation Networks

Type: Article

Publication Date: 2019-05-02

Citations: 11224

DOI: https://doi.org/10.1109/tpami.2019.2913372

Abstract

The central building block of convolutional neural networks (CNNs) is the convolution operator, which enables networks to construct informative features by fusing both spatial and channel-wise information within local receptive fields at each layer. A broad range of prior research has investigated the spatial component of this relationship, seeking to strengthen the representational power of a CNN by enhancing the quality of spatial encodings throughout its feature hierarchy. In this work, we focus instead on the channel relationship and propose a novel architectural unit, which we term the "Squeeze-and-Excitation" (SE) block, that adaptively recalibrates channel-wise feature responses by explicitly modelling interdependencies between channels. We show that these blocks can be stacked together to form SENet architectures that generalise extremely effectively across different datasets. We further demonstrate that SE blocks bring significant improvements in performance for existing state-of-the-art CNNs at slight additional computational cost. Squeeze-and-Excitation Networks formed the foundation of our ILSVRC 2017 classification submission which won first place and reduced the top-5 error to 2.251 percent, surpassing the winning entry of 2016 by a relative improvement of -25 percent. Models and code are available at https://github.com/hujie-frank/SENet.

Locations

  • IEEE Transactions on Pattern Analysis and Machine Intelligence - View
  • arXiv (Cornell University) - View - PDF
  • PubMed - View

Similar Works

Action Title Year Authors
+ PDF Chat Squeeze-and-Excitation Networks 2018 Jie Hu
Li Shen
Gang Sun
+ Squeeze aggregated excitation network 2023 N. Mahendran
+ SENetV2: Aggregated dense layer for channelwise and global representations 2023 M. S. Narayanan
+ Variations of Squeeze and Excitation networks 2023 Mahendran
+ Channel Locality Block: A Variant of Squeeze-and-Excitation 2019 Huayu Li
+ PDF Chat ME-CapsNet: A Multi-Enhanced Capsule Networks with Routing Mechanism 2022 Jerrin Bright
Suryaprakash Rajkumar
Arockia Selvakumar Arockia Doss
+ Gather-Excite: Exploiting Feature Context in Convolutional Neural Networks 2018 Jie Hu
Li Shen
Samuel Albanie
Gang Sun
Andrea Vedaldi
+ Gather-Excite: Exploiting Feature Context in Convolutional Neural Networks 2018 Jie Hu
Li Shen
Samuel Albanie
Gang Sun
Andrea Vedaldi
+ ME-CapsNet: A Multi-Enhanced Capsule Networks with Routing Mechanism 2022 Jerrin Bright
Suryaprakash Rajkumar
Arockia Selvakumar Arockia Doss
+ Recalibrating Fully Convolutional Networks with Spatial and Channel 'Squeeze & Excitation' Blocks 2018 Abhijit Guha Roy
Nassir Navab
Christian Wachinger
+ Recalibrating Fully Convolutional Networks with Spatial and Channel 'Squeeze & Excitation' Blocks 2018 Abhijit Guha Roy
Nassir Navab
Christian Wachinger
+ Concurrent Spatial and Channel Squeeze & Excitation in Fully Convolutional Networks 2018 Abhijit Guha Roy
Nassir Navab
Christian Wachinger
+ Concurrent Spatial and Channel Squeeze & Excitation in Fully Convolutional Networks 2018 Abhijit Guha Roy
Nassir Navab
Christian Wachinger
+ Tiled Squeeze-and-Excite: Channel Attention With Local Spatial Context 2021 Niv Vosco
Alon Shenkler
Mark Grobman
+ Tiled Squeeze-and-Excite: Channel Attention With Local Spatial Context 2021 Niv Vosco
Alon Shenkler
Mark Grobman
+ Residual Squeeze VGG16 2017 Hussam Qassim
David Feinzimer
Abhishek Verma
+ Residual Squeeze VGG16 2017 Hussam Qassim
David Feinzimer
Abhishek Verma
+ Network compression and faster inference using spatial basis filters 2021 Roy Miles
Krystian Mikolajczyk
+ RePr: Improved Training of Convolutional Filters 2018 Aaditya Prakash
James A. Storer
Dinei Florêncio
Cha Zhang
+ LeanConvNets: Low-Cost Yet Effective Convolutional Neural Networks 2020 Jonathan Ephrath
Moshe Eliasof
Lars Ruthotto
Eldad Haber
Eran Treister

Works That Cite This (3002)

Action Title Year Authors
+ Generative Residual Attention Network for Disease Detection 2021 Euyoung Kim
Soochahn Lee
Kyoung Mu Lee
+ Training BatchNorm Only in Neural Architecture Search and Beyond 2021 Yichen Zhu
Jie Du
Yuqin Zhu
Yi Wang
Zhicai Ou
Feifei Feng
Jian Tang
+ PDF Chat SCPM-Net: An anchor-free 3D lung nodule detection network using sphere representation and center points matching 2021 Xiangde Luo
Tao Song
Guotai Wang
Jieneng Chen
Yinan Chen
Kang Li
Dimitris Metaxas
Shaoting Zhang
+ PDF Chat Face Super-Resolution Guided by 3D Facial Priors 2020 Xiaobin Hu
Wenqi Ren
John LaMaster
Xiaochun Cao
Xiaoming Li
Zechao Li
Bjoern Menze
Wei Liu
+ PDF Chat Lightweight Feature Fusion Network for Single Image Super-Resolution 2019 Wenming Yang
Wei Wang
Xuechen Zhang
Shuifa Sun
Qingmin Liao
+ PDF Chat Multiple sclerosis lesion activity segmentation with attention-guided two-path CNNs 2020 Nils Gessert
Julia Krüger
Roland Opfer
Ann‐Christin Ostwaldt
Praveena Manogaran
Hagen H. Kitzler
Sven Schippling
Alexander Schlaefer
+ PDF Chat End-to-end learning for simultaneously generating decision map and multi-focus image fusion result 2021 Boyuan Ma
Xiang Yin
Di Wu
Haokai Shen
Xiaojuan Ban
Yu Wang
+ PDF Chat Sequential vessel segmentation via deep channel attention network 2020 Dongdong Hao
Song Ding
Linwei Qiu
Yisong Lv
Baowei Fei
Yueqi Zhu
Binjie Qin
+ PDF Chat Integrating Frequency Translational Invariance in TDNNs and Frequency Positional Information in 2D ResNets to Enhance Speaker Verification 2021 Jenthe Thienpondt
Brecht Desplanques
Kris Demuynck
+ PDF Chat Attention-guided chained context aggregation for semantic segmentation 2021 Quan Tang
Fagui Liu
Tong Zhang
Jun Jiang
Yu Zhang