[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3566097.3567872acmconferencesArticle/Chapter ViewAbstractPublication PagesaspdacConference Proceedingsconference-collections
research-article

BARVINN: Arbitrary Precision DNN Accelerator Controlled by a RISC-V CPU

Published: 31 January 2023 Publication History

Abstract

We present a DNN accelerator that allows inference at arbitrary precision with dedicated processing elements that are configurable at the bit level. Our DNN accelerator has 8 Processing Elements controlled by a RISC-V controller with a combined 8.2 TMACs of computational power when implemented with the recent Alveo U250 FPGA platform. We develop a code generator tool that ingests CNN models in ONNX format and generates an executable command stream for the RISC-V controller. We demonstrate the scalable throughput of our accelerator by running different DNN kernels and models when different quantization levels are selected. Compared to other low precision accelerators, our accelerator provides run time programmability without hardware reconfiguration and can accelerate DNNs with multiple quantization levels, regardless of the target FPGA size. BARVINN is an open source project and it is available at https://github.com/hossein1387/BARVINN.

References

[1]
2021. FINN Dataflow Accelerator Examples. (2021). https://github.com/Xilinx/finn-examples
[2]
2021. ONNX Model Zoo. (2021). https://github.com/onnx/models
[3]
2021. Open Neural Network Exchange. (2021). https://onnx.ai/
[4]
MohammadHossein AskariHemmat, Olexa Bilaniuk, Sean Wagner, Yvon Savaria, and Jean-Pierre David. 2021. RISC-V Barrel Processor for Deep Neural Network Acceleration. In 2021 IEEE International Symposium on Circuits and Systems (ISCAS). 1--5.
[5]
Olexa Bilaniuk, Sean Wagner, Yvon Savaria, and Jean-Pierre David. 2019. Bit-Slicing FPGA Accelerator for Quantized Neural Networks. In 2019 IEEE International Symposium on Circuits and Systems (ISCAS). 1--5.
[6]
Michaela Blott, Thomas B. Preußer, Nicholas J. Fraser, Giulio Gambardella, Kenneth O'brien, Yaman Umuroglu, Miriam Leeser, and Kees Vissers. 2018. FINN-R: An End-to-End Deep-Learning Framework for Fast Exploration of Quantized Neural Networks. ACM Trans. Reconfigurable Technol. Syst. 11, 3, Article 16 (dec 2018), 23 pages.
[7]
Adrian Bulat and Georgios Tzimiropoulos. 2021. Bit-Mixer: Mixed-Precision Networks With Runtime Bit-Width Selection. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). 5188--5197.
[8]
Meghan Cowan, Thierry Moreau, Tianqi Chen, James Bornholt, and Luis Ceze. 2020. Automatic Generation of High-Performance Quantized Machine Learning Kernels. Association for Computing Machinery, New York, NY, USA, 305--316.
[9]
Steven K. Esser, Jeffrey L. McKinstry, Deepika Bablani, Rathinakumar Appuswamy, and Dharmendra S. Modha. 2020. Learned Step Size Quantization. In International Conference on Learning Representations.
[10]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11]
Mark Horowitz. 2014. Computing's Energy Problem (and what we can do about it). Interational Solid State Circuits Conference (2014).
[12]
Itay Hubara et al. 2018. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations. JMLR 18, 187 (2018), 1--30.
[13]
B. Ham J. Lee, D. Kim. 2021. Network Quantization with Element-wise Gradient Scaling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[14]
Guilin Li et al. 2020. Residual Distillation: Towards Portable Deep Neural Networks without Shortcuts. In Advances in Neural Information Processing Systems, Vol. 33. 8935--8946.
[15]
Nicolas Limare. 2014. Floating-Point Math Speed vs Precision. (2014). http://nicolas.limare.net/pro/notes/2014/12/16_math_speed/
[16]
Paulius Micikevicius et al. 2017. Mixed precision training. arXiv preprint arXiv:1710.03740 (2017).
[17]
Sungju Ryu, Hyungjun Kim, Wooseok Yi, and Jae-Joon Kim. 2019. Bitblade: Area and energy-efficient precision-scalable neural network accelerator with bitwise summation. In Proceedings of the 56th Annual Design Automation Conference 2019. 1--6.
[18]
Sayeh Sharify, Alberto Delmas Lascorz, Kevin Siu, Patrick Judd, and Andreas Moshovos. 2018. Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks. In 2018 55th ACM/ESDA/IEEE Design Automation Conference (DAC). 1--6.
[19]
Hardik Sharma et al. 2018. Bit fusion: Bit-level dynamically composable architecture for accelerating deep neural network. In 2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA). IEEE, 764--775.
[20]
Mengshu Sun et al. 2022. FILM-QNN: Efficient FPGA Acceleration of Deep Neural Networks with Intra-Layer, Mixed-Precision Quantization. In Proceedings of the 2022 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays. 134--145.
[21]
Stefan Uhlich et al. 2020. Mixed Precision DNNs: All you need is a good parametrization. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26--30, 2020.
[22]
Yaman Umuroglu et al. 2017. FINN: A framework for fast, scalable binarized neural network inference. In Proceedings of the 2017 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays. ACM, 65--74.
[23]
Kuan Wang, Zhijian Liu, Yujun Lin, Ji Lin, and Song Han. 2019. HAQ: Hardware-Aware Automated Quantization With Mixed Precision. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 8604--8612.
[24]
Haichao Yu, Haoxiang Li, Humphrey Shi, Thomas S. Huang, and Gang Hua. 2021. Any-Precision Deep Neural Networks. In Thirty-Fifth AAAI Conference on Artificial Intelligence. AAAI Press, 10763--10771. https://ojs.aaai.org/index.php/AAAI/article/view/17286
[25]
Xiaofan Zhang et al. 2018. DNNBuilder: an Automated Tool for Building High-Performance DNN Hardware Accelerators for FPGAs. In (ICCAD).

Cited By

View all
  • (2023)Quark: An Integer RISC-V Vector Processor for Sub-Byte Quantized DNN Inference2023 IEEE International Symposium on Circuits and Systems (ISCAS)10.1109/ISCAS46773.2023.10181985(1-5)Online publication date: 21-May-2023
  • (2023)Sustainable Computing Through Open Standard ISAs: Leveraging Tailor-Fit Hardware Designs for Circular EconomiesProduction at the Leading Edge of Technology10.1007/978-3-031-47394-4_46(469-480)Online publication date: 18-Nov-2023

Index Terms

  1. BARVINN: Arbitrary Precision DNN Accelerator Controlled by a RISC-V CPU
      Index terms have been assigned to the content through auto-classification.

      Recommendations

      Comments

      Please enable JavaScript to view thecomments powered by Disqus.

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      ASPDAC '23: Proceedings of the 28th Asia and South Pacific Design Automation Conference
      January 2023
      807 pages
      ISBN:9781450397834
      DOI:10.1145/3566097
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      In-Cooperation

      • IPSJ
      • IEEE CAS
      • IEEE CEDA
      • IEICE

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 31 January 2023

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. FPGA
      2. hardware acceleration
      3. low-precision
      4. neural networks

      Qualifiers

      • Research-article

      Funding Sources

      Conference

      ASPDAC '23
      Sponsor:

      Acceptance Rates

      ASPDAC '23 Paper Acceptance Rate 102 of 328 submissions, 31%;
      Overall Acceptance Rate 466 of 1,454 submissions, 32%

      Upcoming Conference

      ASPDAC '25

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)139
      • Downloads (Last 6 weeks)21
      Reflects downloads up to 11 Dec 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2023)Quark: An Integer RISC-V Vector Processor for Sub-Byte Quantized DNN Inference2023 IEEE International Symposium on Circuits and Systems (ISCAS)10.1109/ISCAS46773.2023.10181985(1-5)Online publication date: 21-May-2023
      • (2023)Sustainable Computing Through Open Standard ISAs: Leveraging Tailor-Fit Hardware Designs for Circular EconomiesProduction at the Leading Edge of Technology10.1007/978-3-031-47394-4_46(469-480)Online publication date: 18-Nov-2023

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media