Energy-Efficient Architecture for FPGA-based Deep Convolutional Neural Networks with Binary Weights

Yunzhi Duan, Shuai Li, Ruipeng Zhang, Qi Wang, Jienan Chen, Gerald E. Sobelman

Research output: Chapter in Book/Report/Conference proceedingConference contribution

8 Scopus citations

Abstract

This paper presents an energy-efficient, deep parallel Convolutional Neural Network (CNN) accelerator. By adopting a recently proposed binary weight method, the CNN computations are converted into multiplication-free processing. To allow parallel accessing and storing of data, we use two RAM banks, where each bank is composed of NRAM blocks corresponding to N-parallel processing. We also design a reconfigurable CNN computing unit in a divide-and-reuse to support a variable-size convolutional filter. Compared with full-precision computing on the MNIST and CIFAR-10 classification tasks, the inference Top-1 accuracy of the binary weight CNN has dropped by 1.21% and 1.34%, respectively. The hardware implementation results show that the proposed design can achieve 2100 GOPs with a 4.6 millisecond processing latency. The deep parallel accelerator exhibits 3X energy efficiency compared to a GPU-based design.

Original languageEnglish (US)
Title of host publication2018 IEEE 23rd International Conference on Digital Signal Processing, DSP 2018
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781538668115
DOIs
StatePublished - Jul 2 2018
Event23rd IEEE International Conference on Digital Signal Processing, DSP 2018 - Shanghai, China
Duration: Nov 19 2018Nov 21 2018

Publication series

NameInternational Conference on Digital Signal Processing, DSP
Volume2018-November

Conference

Conference23rd IEEE International Conference on Digital Signal Processing, DSP 2018
Country/TerritoryChina
CityShanghai
Period11/19/1811/21/18

Bibliographical note

Publisher Copyright:
© 2018 IEEE.

Keywords

  • Convolutional Neural Network (CNN)
  • deep neural network
  • energy efficiency
  • parallel implementation

Fingerprint

Dive into the research topics of 'Energy-Efficient Architecture for FPGA-based Deep Convolutional Neural Networks with Binary Weights'. Together they form a unique fingerprint.

Cite this