Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

Type: Article

Publication Date: 2015-12-01

Citations: 16897

DOI: https://doi.org/10.1109/iccv.2015.123

Download PDF

Abstract

Rectified activation units (rectifiers) are essential for state-of-the-art neural networks. In this work, we study rectifier neural networks for image classification from two aspects. First, we propose a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU improves model fitting with nearly zero extra computational cost and little overfitting risk. Second, we derive a robust initialization method that particularly considers the rectifier nonlinearities. This method enables us to train extremely deep rectified models directly from scratch and to investigate deeper or wider network architectures. Based on the learnable activation and advanced initialization, we achieve 4.94% top-5 test error on the ImageNet 2012 classification dataset. This is a 26% relative improvement over the ILSVRC 2014 winner (GoogLeNet, 6.66% [33]). To our knowledge, our result is the first to surpass the reported human-level performance (5.1%, [26]) on this dataset.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Flexible Rectified Linear Units for Improving Convolutional Neural Networks 2017 Suo Qiu
Bolun Cai
+ FReLU: Flexible Rectified Linear Units for Improving Convolutional Neural Networks 2017 Suo Qiu
Xiangmin Xu
Bolun Cai
+ Piecewise Linear Units Improve Deep Neural Networks 2021 Jordan Inturrisi
Suiyang Khoo
Abbas Z. Kouzani
Riccardo M. Pagliarella
+ Empirical Evaluation of Rectified Activations in Convolutional Network 2015 Bing Xu
Naiyan Wang
Tianqi Chen
Mu Li
+ Improving Deep Neural Network with Multiple Parametric Exponential Linear Units 2016 Yang Li
Chunxiao Fan
Yong Li
Qiong Wu
Yue Ming
+ Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) 2015 Djork-Arné Clevert
Thomas Unterthiner
Sepp Hochreiter
+ Effects of the Nonlinearity in Activation Functions on the Performance of Deep Learning Models 2020 Nalinda Kulathunga
N. R. Ranasinghe
D. Vrinceanu
Zackary Kinsman
Lei Huang
Yunjiao Wang
+ ResNet Sparsifier: Learning Strict Identity Mappings in Deep Residual Networks 2018 Xin Yu
Zhiding Yu
Srikumar Ramalingam
+ Dynamic ReLU 2020 Yinpeng Chen
Xiyang Dai
Mengchen Liu
Dongdong Chen
Lu Yuan
Zicheng Liu
+ Deep Learning using Rectified Linear Units (ReLU) 2018 Abien Fred Agarap
+ Natural-Logarithm-Rectified Activation Function in Convolutional Neural Networks 2019 Yang Liu
Jianpeng Zhang
Chao Gao
QU Jing-hua
Lixin Ji
+ PDF Chat Natural-Logarithm-Rectified Activation Function in Convolutional Neural Networks 2019 Yang Liu
Jianpeng Zhang
Chao Gao
QU Jing-hua
Lixin Ji
+ Deep Global-Connected Net With The Generalized Multi-Piecewise ReLU Activation in Deep Learning 2018 Zhi Chen
Pin–Han Ho
+ Learning specialized activation functions with the Piecewise Linear Unit 2021 Yucong Zhou
Zezhou Zhu
Zhao Zhong
+ PDF Chat Deep Learning with S-Shaped Rectified Linear Activation Units 2016 Xiaojie Jin
Chunyan Xu
Jiashi Feng
Yunchao Wei
Junjun Xiong
Shuicheng Yan
+ Adjustable Bounded Rectifiers: Towards Deep Binary Representations 2015 Zhirong Wu
Dahua Lin
Xiaoou Tang
+ Overcoming Overfitting and Large Weight Update Problem in Linear Rectifiers: Thresholded Exponential Rectified Linear Units. 2020 Vijay Pandey
+ PDF Chat Learning specialized activation functions with the Piecewise Linear Unit 2021 Yucong Zhou
Zezhou Zhu
Zhao Zhong
+ Parametric Variational Linear Units (PVLUs) in Deep Convolutional Networks 2021 Aarush Gupta
Shikhar Ahuja
+ Overcoming Overfitting and Large Weight Update Problem in Linear Rectifiers: Thresholded Exponential Rectified Linear Units 2020 Vijay Pandey