[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3649329.3656514acmconferencesArticle/Chapter ViewAbstractPublication PagesdacConference Proceedingsconference-collections
research-article
Open access

Energy-efficient SNN Architecture using 3nm FinFET Multiport SRAM-based CIM with Online Learning

Published: 07 November 2024 Publication History

Abstract

Current Artificial Intelligence (AI) computation systems face challenges, primarily from the memory-wall issue, limiting overall system-level performance, especially for Edge devices with constrained battery budgets, such as smartphones, wearables, and Internet-of-Things sensor systems. In this paper, we propose a new SRAM-based Compute-In-Memory (CIM) accelerator optimized for Spiking Neural Networks (SNNs) Inference. Our proposed architecture employs a multiport SRAM design with multiple decoupled Read ports to enhance the throughput and Transposable Read-Write ports to facilitate online learning. Furthermore, we develop an Arbiter circuit for efficient data-processing and port allocations during the computation. Results for a 128×128 array in 3nm FinFET technology demonstrate a 3.1× improvement in speed and a 2.2× enhancement in energy efficiency with our proposed multiport SRAM design compared to the traditional single-port design. At system-level, a throughput of 44 MInf/s at 607 pJ/Inf and 29mW is achieved.

References

[1]
Y. Lee, P. Tsung, and M. Wu, "Techology trend of edge AI," in VLSI-DAT 2018, pp. 1--2.
[2]
H. Fujiwara et al. "A 5-nm 254-TOPS/W 221-TOPS/mm2 Fully-Digital Computingin-Memory Macro Supporting Wide-Range Dynamic-Voltage-Frequency Scaling and Simultaneous MAC and Write Operations," in ISSCC 2022, pp 1--3.
[3]
Y. Chih et al. "An 89TOPS/W and 16.3TOPS/mm2 All-Digital SRAM-Based Full-Precision Compute-In Memory Macro in 22nm for Machine-Learning Edge Applications," in ISSCC 2021, pp. 252--254.
[4]
D. Wanq et al. "All-Digital Full-Precision In-SRAM Computing with Reduction Tree for Energy-Efficient MAC Operations," in ICTA 2022, pp. 150--151.
[5]
D. Wang et al. "DIMC: 2219TOPS/W 2569F2/b Digital In-Memory Computing Macro in 28nm Based on Approximate Arithmetic Hardware," in ISSCC 2022, pp. 266--268.
[6]
D. Wang et al. "Always-On, Sub-300-nW, Event-Driven Spiking Neural Network based on Spike-Driven Clock-Generation and Clock- and Power-Gating for an Ultra-Low-Power Intelligent Device," in A-SSCC 2020, pp. 1--4.
[7]
J. Seo et al. "A 45nm CMOS neuromorphic chip with a scalable architecture for learning in networks of spiking neurons," in CICC 2011, pp. 1--3.
[8]
D. Kim et al. "MONETA: A Processing-In-Memory-Based Hardware Platform for the Hybrid Convolutional Spiking Neural Network With Online Learning," in Frontiers in Neuroscience 16, 2022.
[9]
G. Chen et al. "A 4096-Neuron 1M-Synapse 3.8-pJ/SOP Spiking Neural Network With On-Chip STDP Learning and Sparse Weights in 10-nm FinFET CMOS," in JSSC 54.4, 2019, pp. 992--1002.
[10]
J. Kim et al. "Efficient Synapse Memory Structure for Reconfigurable Digital Neuromorphic Hardware," in Frontiers in Neuroscience 12, 2018.
[11]
J. Wang et al. "A Compute SRAM with Bit-Serial Integer/Floating-Point Operations for Programmable In-Memory Vector Acceleration," in ISSCC 2019, pp. 224--226.
[12]
S. K. Bose and A. Basu, "A 389TOPS/W, 1262fps at 1Meps Region Proposal Integrated Circuit for Neuromorphic Vision Sensors in 65nm CMOS," in A-SSCC 2021, pp. 1--3.
[13]
Z. Lin et al. "Two-Direction In-Memory Computing Based on 10T SRAM With Horizontal and Vertical Decoupled Read Ports," in JSSC 56.9, 2021, pp. 2832--2944.
[14]
W. Maass, "Networks of spiking neurons: The third generation of neural network models," in Neural Networks 10.9, 1997, pp. 1659--1671.
[15]
H. Kim, H. Oh, and J. Kim, "Energy-efficient XNOR-free In-Memory BNN Accelerator with Input Distribution Regularization," in ICCAD 2020, pp. 1--9.
[16]
A. Yousefzadeh et al. "On practical issues for stochastic STDP hardware with 1-bit synaptic weights," in Frontiers in Neuroscience 12, 2018.
[17]
A. Yousefzadeh et al. "Hardware implementation of convolutional STDP for on-line visual feature learning," in ISCAS 2017, pp. 1--4.
[18]
A. Rostami et al. "E-prop on SpiNNaker 2: Exploring online learning in spiking RNNs on neuromorphic hardware," in Frontiers in Neuroscience 16, 2022.
[19]
H.-H. Liu et al. "Extended Methodology to Determine SRAM Write Margin in Resistance-Dominated Technology Node," in TED 69.6, 2022, pp. 3113--3117.
[20]
H.-H. Liu et al. "DTCO of sequential and monolithic CFET SRAM," in SPIE 12495, 2023.
[21]
H.-H. Liu et al. "CFET SRAM DTCO, Interconnect Guideline, and Benchmark for CMOS Scaling," in TED 70.3, 2023, pp. 883--890.

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
DAC '24: Proceedings of the 61st ACM/IEEE Design Automation Conference
June 2024
2159 pages
ISBN:9798400706011
DOI:10.1145/3649329
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

In-Cooperation

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 07 November 2024

Check for updates

Qualifiers

  • Research-article

Conference

DAC '24
Sponsor:
DAC '24: 61st ACM/IEEE Design Automation Conference
June 23 - 27, 2024
CA, San Francisco, USA

Acceptance Rates

Overall Acceptance Rate 1,770 of 5,499 submissions, 32%

Upcoming Conference

DAC '25
62nd ACM/IEEE Design Automation Conference
June 22 - 26, 2025
San Francisco , CA , USA

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 231
    Total Downloads
  • Downloads (Last 12 months)231
  • Downloads (Last 6 weeks)75
Reflects downloads up to 01 Mar 2025

Other Metrics

Citations

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media