TY - GEN
T1 - A Survey
T2 - 2021 IEEE High Performance Extreme Computing Conference, HPEC 2021
AU - Geng, Tong
AU - Wu, Chunshu
AU - Tan, Cheng
AU - Xie, Chenhao
AU - Guo, Anqi
AU - Haghi, Pouya
AU - He, Sarah Yuan
AU - Li, Jiajia
AU - Herbordt, Martin
AU - Li, Ang
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - In the last decade, Artificial Intelligence (AI) through Deep Neural Networks (DNNs) has penetrated virtually every aspect of science, technology, and business. Many types of DNNs have been and continue to be developed, including Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Graph Neural Networks (GNNs). The overall problem for all of these Neural Networks (NNs) is that their target applications generally pose stringent constraints on latency and throughput, while also having strict accuracy requirements. There have been many previous efforts in creating hardware to accelerate NNs. The problem designers face is that optimal NN models typically have significant irregularities, making them hardware-unfriendly. In this paper, we first define the problems in NN acceleration by characterizing common irregularities in NN processing into 4 types; then we summarize the existing works that handle the four types of irregularities efficiently using hardware, especially FPGAs; finally, we provide a new vision of next-generation FPGA-based NN acceleration: That the emerging heterogeneity in the next-generation FPGAs is the key to achieving higher performance.
AB - In the last decade, Artificial Intelligence (AI) through Deep Neural Networks (DNNs) has penetrated virtually every aspect of science, technology, and business. Many types of DNNs have been and continue to be developed, including Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Graph Neural Networks (GNNs). The overall problem for all of these Neural Networks (NNs) is that their target applications generally pose stringent constraints on latency and throughput, while also having strict accuracy requirements. There have been many previous efforts in creating hardware to accelerate NNs. The problem designers face is that optimal NN models typically have significant irregularities, making them hardware-unfriendly. In this paper, we first define the problems in NN acceleration by characterizing common irregularities in NN processing into 4 types; then we summarize the existing works that handle the four types of irregularities efficiently using hardware, especially FPGAs; finally, we provide a new vision of next-generation FPGA-based NN acceleration: That the emerging heterogeneity in the next-generation FPGAs is the key to achieving higher performance.
UR - https://www.scopus.com/pages/publications/85123496359
U2 - 10.1109/HPEC49654.2021.9622877
DO - 10.1109/HPEC49654.2021.9622877
M3 - 会议稿件
AN - SCOPUS:85123496359
T3 - 2021 IEEE High Performance Extreme Computing Conference, HPEC 2021
BT - 2021 IEEE High Performance Extreme Computing Conference, HPEC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 20 September 2021 through 24 September 2021
ER -