skip to main content
10.1145/3531437.3539715acmconferencesArticle/Chapter ViewAbstractPublication PagesislpedConference Proceedingsconference-collections
research-article
Open access

Sparse Periodic Systolic Dataflow for Lowering Latency and Power Dissipation of Convolutional Neural Network Accelerators

Published: 01 August 2022 Publication History

Abstract

This paper introduces the sparse periodic systolic (SPS) dataflow, which advances the state-of-the-art hardware accelerator for supporting lightweight neural networks. Specifically, the SPS dataflow enables a novel hardware design approach unlocked by an emergent pruning scheme, periodic pattern-based sparsity (PPS). By exploiting the regularity of PPS, our sparsity-aware compiler optimally reorders the weights and uses a simple indexing unit in hardware to create matches between the weights and activations. Through the compiler-hardware codesign, SPS dataflow enjoys higher degrees of parallelism while being free of the high indexing overhead and without model accuracy loss. Evaluated on popular benchmarks such as VGG and ResNet, the SPS dataflow and accompanying neural network compiler outperform prior work in convolutional neural network (CNN) accelerator designs targeting FPGA devices. Against other sparsity-supporting weight storage formats, SPS results in 4.49 × energy efficiency gain while lowering storage requirements by 3.67 × for total weight storage (non-pruned weights plus indexing) and 22,044 × for indexing memory.

References

[1]
John Cheng, Max Grossman, and Ty McKercher. 2014. Professional CUDA c programming. John Wiley & Sons.
[2]
Dave 2021. Hardware Acceleration of Sparse and Irregular Tensor Computations of ML Models: A Survey and Insights. Proc. IEEE 109, 10 (2021), 1706–1752.
[3]
Xiaohan Ding 2019. Global Sparse Momentum SGD for Pruning Very Deep Neural Networks. In Neurips. 6379–6391.
[4]
Song Han 2015. Learning both Weights and Connections for Efficient Neural Networks. CoRR abs/1506.02626(2015). arxiv:1506.02626
[5]
Geoffrey E. Hinton, Oriol Vinyals, and Jeffrey Dean. 2015. Distilling the Knowledge in a Neural Network. CoRR abs/1503.02531(2015). arxiv:1503.02531
[6]
Alex Krizhevsky 2012. ImageNet Classification with Deep Convolutional Neural Networks. In Neurips. 1106–1114.
[7]
Souvik Kundu 2020. Pre-Defined Sparsity for Low-Complexity Convolutional Neural Networks. IEEE Trans. Computers 69 (2020), 1045–1058.
[8]
Mahdi Nazemi 2021. NullaNet Tiny: Ultra-low-latency DNN Inference Through Fixed-function Combinational Logic. In 29th IEEE FCCM.
[9]
Wei Niu 2020. PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with Pattern-based Weight Pruning. In ASPLOS ’20. ACM, 907–922.
[10]
Wei Wen 2016. Learning Structured Sparsity in Deep Neural Networks. In Neurips 2016, December 5-10, 2016, Barcelona, Spain, Daniel D. Leeet al. (Eds.). 2074–2082.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
ISLPED '22: Proceedings of the ACM/IEEE International Symposium on Low Power Electronics and Design
August 2022
192 pages
ISBN:9781450393546
DOI:10.1145/3531437
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 August 2022

Check for updates

Author Tags

  1. CNN Acceleration
  2. Deep Learning
  3. FPGA
  4. Pattern-based Pruning

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ISLPED '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 398 of 1,159 submissions, 34%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)207
  • Downloads (Last 6 weeks)16
Reflects downloads up to 28 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media