In the last decade, Artificial Intelligence (AI) through Deep Neural Networks (DNNs) has penetrated virtually every aspect of science, technology, and business. Many types of DNNs have been and continue to be developed, including Convolutional Neural Networks (CNNs), Recurrent Neural Net- works (RNNs), and Graph Neural Networks (GNNs). The overall problem for all of these Neural Networks (NNs) is that their target applications generally pose stringent constraints on latency and throughput, while also having strict accuracy requirements. There have been many previous efforts in creating hardware to accelerate NNs. The problem designers face is that optimal NN models typically have significant irregularities, making them hardware-unfriendly. In this paper, we first define the problems in NN acceleration by characterizing common irregularities in NN processing into 4 types; then we summarize the existing works that handle the four types of irregularities efficiently using hardware, especially FPGAs; finally, we provide a new vision of next-generation FPGA-based NN acceleration: that the emerging heterogeneity in the next-generation FPGAs is the key to achieving higher performance.
Published: March 16, 2022
Citation
Geng T., C. Wu, C. Tan, C. Xie, A. Guo, P. Haghi, and S. He, et al. 2021.A Survey: Handling Irregularities in Neural Network Acceleration with FPGAs. In IEEE High Performance Extreme Computing Conference (HPEC 2021), September 20-24, 2021, Virtual, Online, 1-8. Piscataway, New Jersey:IEEE.PNNL-SA-165315.doi:10.1109/HPEC49654.2021.9622877