Search-free Accelerator for Sparse Convolutional Neural Networks

  • Bosheng Liu
  • , Xiaoming Chen*
  • , Yinhe Han
  • , Ying Wang
  • , Jiajun Li
  • , Haobo Xu
  • , Xiaowei Li
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Sparsification is an efficient solution to reduce the demand of on-chip memory space for deep convolutional neural networks (CNNs). Most of state-of-the-art CNN accelerators can deliver high throughput for sparse CNNs by searching pairs of nonzero weights and activations, and then sending them to processing elements (PEs) for multiplication-accumulation (MAC) operations. However, their PE scales are difficult to be increased for superior and efficient computing because of the significant internal interconnect and memory bandwidth consumption. To deal with this dilemma, we propose a sparsity-aware architecture, called Swan, which frees the search process for sparse CNNs under limited interconnect and bandwidth resources. The architecture comprises two parts: A MAC unit that can free the search operation for the sparsity-aware MAC calculation, and a systolic compressive dataflow that well suits the MAC architecture and greatly reuses inputs for interconnect and bandwidth saving. With the proposed architecture, only one column of the PEs needs to load/store data while all PEs can operate in full scale. Evaluation results based on a place-and-route process show that the proposed design, in a compact factor of 4096 PEs, 4.9TOP/s peak performance, and 2.97W power running at 600MHz, achieves 1.5-2.1× speedup and 6.0-9.1× higher energy efficiency than state-of-the-art CNN accelerators with the same PE scale.

Original languageEnglish
Title of host publicationASP-DAC 2020 - 25th Asia and South Pacific Design Automation Conference, Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages524-529
Number of pages6
ISBN (Electronic)9781728141237
DOIs
StatePublished - Jan 2020
Externally publishedYes
Event25th Asia and South Pacific Design Automation Conference, ASP-DAC 2020 - Beijing, China
Duration: 13 Jan 202016 Jan 2020

Publication series

NameProceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC
Volume2020-January

Conference

Conference25th Asia and South Pacific Design Automation Conference, ASP-DAC 2020
Country/TerritoryChina
CityBeijing
Period13/01/2016/01/20

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 7 - Affordable and Clean Energy
    SDG 7 Affordable and Clean Energy

Keywords

  • Sparse convolution neural networks
  • internal interconnect
  • memory bandwidth.
  • sparsity-aware CNN accelerator

Fingerprint

Dive into the research topics of 'Search-free Accelerator for Sparse Convolutional Neural Networks'. Together they form a unique fingerprint.

Cite this