ReBNet: Residual Binarized Neural Network

Type: Article

Publication Date: 2018-04-01

Citations: 119

DOI: https://doi.org/10.1109/fccm.2018.00018

Download PDF

Abstract

This paper proposes ReBNet, an end-to-end framework for training reconfigurable binary neural networks on software and developing efficient accelerators for execution on FPGA. Binary neural networks offer an intriguing opportunity for deploying large-scale deep learning models on resource-constrained devices. Binarization reduces the memory footprint and replaces the power-hungry matrix-multiplication with light-weight XnorPopcount operations. However, binary networks suffer from a degraded accuracy compared to their fixed-point counterparts. We show that the state-of-the-art methods for optimizing binary networks accuracy, significantly increase the implementation cost and complexity. To compensate for the degraded accuracy while adhering to the simplicity of binary networks, we devise the first reconfigurable scheme that can adjust the classification accuracy based on the application. Our proposition improves the classification accuracy by representing features with multiple levels of residual binarization. Unlike previous methods, our approach does not exacerbate the area cost of the hardware accelerator. Instead, it provides a tradeoff between throughput and accuracy while the area overhead of multi-level binarization is negligible.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ ReBNet: Residual Binarized Neural Network 2017 Mohammad Ghasemzadeh
Mohammad Samragh
Farinaz Koushanfar
+ PDF Chat LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference 2020 Erwei Wang
James J. Davis
Peter Y. K. Cheung
George A. Constantinides
+ Scaling Binarized Neural Networks on Reconfigurable Logic 2017 Nicholas J. Fraser
Yaman Umuroglu
Giulio Gambardella
Michaela Blott
Philip H. W. Leong
Magnus Jahre
Kees Vissers
+ PDF Chat Scaling Binarized Neural Networks on Reconfigurable Logic 2017 Nicholas J. Fraser
Yaman Umuroglu
Giulio Gambardella
Michaela Blott
Philip H. W. Leong
Magnus Jahre
Kees Vissers
+ FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with Fractional Activations 2020 Yichi Zhang
Junhao Pan
Xinheng Liu
Hongzheng Chen
Deming Chen
Zhiru Zhang
+ PDF Chat FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with Fractional Activations 2020 Yichi Zhang
Junhao Pan
Xinheng Liu
Hongzheng Chen
Deming Chen
Zhiru Zhang
+ FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with Fractional Activations 2020 Yichi Zhang
Junhao Pan
Xinheng Liu
Hongzheng Chen
Deming Chen
Zhiru Zhang
+ BoolNet: Minimizing The Energy Consumption of Binary Neural Networks 2021 Nianhui Guo
Joseph Bethge
Haojin Yang
Kai Zhong
Xuefei Ning
Christoph Meinel
Yu Wang
+ PDF Chat BinArray: A Scalable Hardware Accelerator for Binary Approximated CNNs 2021 Mario Fischer
Juergen Wassner
+ LUTNet: Rethinking Inference in FPGA Soft Logic 2019 Erwei Wang
James J. Davis
Peter Y. K. Cheung
George A. Constantinides
+ PDF Chat LUTNet: Rethinking Inference in FPGA Soft Logic 2019 Erwei Wang
James J. Davis
Peter Y. K. Cheung
George A. Constantinides
+ How to train accurate BNNs for embedded systems? 2022 Floran de Putter
Henk Corporaal
+ Towards Fast and Energy-Efficient Binarized Neural Network Inference on FPGA 2018 Cheng Fu
Shilin Zhu
Hao Su
Ching‐En Lee
Jishen Zhao
+ Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural Networks 2017 Hokchhay Tann
Soheil Hashemi
R. Iris Bahar
Sherief Reda
+ Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural Networks 2017 Hokchhay Tann
Soheil Hashemi
Iris Bahar
Sherief Reda
+ PDF Chat Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural Networks 2017 Hokchhay Tann
Soheil Hashemi
R. Iris Bahar
Sherief Reda
+ PANTHER: A Programmable Architecture for Neural Network Training Harnessing Energy-Efficient ReRAM 2020 Aayush Ankit
Izzat El Hajj
Sai Rahul Chalamalasetti
Sapan Agarwal
Matthew Marinella
Martin FoltĂ­n
John Paul Strachan
Dejan Milojičić
Wen‐mei Hwu
Kaushik Roy
+ Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic 2018 Michaela Blott
Thomas B. Preußer
Nicholas J. Fraser
Giulio Gambardella
Kenneth M. O'Brien
Yaman Umuroglu
Miriam Leeser
+ Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic 2018 Michaela Blott
Thomas B. Preußer
Nicholas C. Fraser
Giulio Gambardella
Kenneth M. O'Brien
Yaman Umuroglu
Miriam Leeser
+ PANTHER: A Programmable Architecture for Neural Network Training Harnessing Energy-efficient ReRAM 2019 Aayush Ankit
Izzat El Hajj
Sai Rahul Chalamalasetti
Sapan Agarwal
Matthew Marinella
Martin FoltĂ­n
John Paul Strachan
Dejan Milojičić
Wen‐mei Hwu
Kaushik Roy

Works Cited by This (18)

Action Title Year Authors
+ PDF Chat Efficient and accurate approximations of nonlinear convolutional networks 2015 Xiangyu Zhang
Jianhua Zou
Ming Xiang
Kaiming He
Jian Sun
+ Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding 2015 Song Han
Huizi Mao
William J. Dally
+ PDF Chat Quantized Convolutional Neural Networks for Mobile Devices 2016 Jiaxiang Wu
Cong Leng
Yuhang Wang
Qinghao Hu
Jian Cheng
+ PDF Chat XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks 2016 Mohammad Rastegari
Vicente Ordóùez
Joseph Redmon
Ali Farhadi
+ Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1 2016 Matthieu Courbariaux
Itay Hubara
Daniel Soudry
Ran El‐Yaniv
Yoshua Bengio
+ DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients 2016 Shuchang Zhou
Yuxin Wu
Zekun Ni
Xinyu Zhou
He Wen
Yuheng Zou
+ Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations 2016 Itay Hubara
Matthieu Courbariaux
Daniel Soudry
Ran El‐Yaniv
Yoshua Bengio
+ Ternary Neural Networks with Fine-Grained Quantization 2017 Naveen Mellempudi
Abhisek Kundu
Dheevatsa Mudigere
Dipankar Das
Bharat Kaul
Pradeep Dubey
+ MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications 2017 Andrew Howard
Menglong Zhu
Bo Chen
Dmitry Kalenichenko
Weijun Wang
Tobias Weyand
Marco Andreetto
Hartwig Adam
+ PDF Chat DeepRebirth: Accelerating Deep Neural Network Execution on Mobile Devices 2018 Dawei Li
Xiaolong Wang
Deguang Kong