Hérissé et al., 2024 - Google Patents
An Incremental Time-Domain Mixed-Signal Matrix-Vector-Multiplication Technique for Low-Power Edge-AIHérissé et al., 2024
View PDF- Document ID
- 5484727620504997897
- Author
- Hérissé K
- Larras B
- Stefanelli B
- Kaiser A
- Frappé A
- Publication year
- Publication venue
- IEEE Transactions on Circuits and Systems I: Regular Papers
External Links
Snippet
This paper proposes a time-domain mixed-signal computing architecture for Matrix-Vector Multiplication suited for embedded in-memory computing applications. The system leverages the low data rate of sensors' data in embedded AI applications to target an energy …
- 238000000034 method 0 title description 19
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
- G06N3/0635—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/58—Random or pseudo-random number generators
-
- H—ELECTRICITY
- H03—BASIC ELECTRONIC CIRCUITRY
- H03M—CODING; DECODING; CODE CONVERSION IN GENERAL
- H03M1/00—Analogue/digital conversion; Digital/analogue conversion
- H03M1/66—Digital/analogue converters
- H03M1/74—Simultaneous conversion
- H03M1/80—Simultaneous conversion using weighted impedances
- H03M1/802—Simultaneous conversion using weighted impedances using capacitors, e.g. neuron-mos transistors, charge coupled devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11C—STATIC STORES
- G11C11/00—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor
- G11C11/21—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using electric elements
- G11C11/34—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using electric elements using semiconductor devices
- G11C11/40—Digital stores characterised by the use of particular electric or magnetic storage elements; Storage elements therefor using electric elements using semiconductor devices using transistors
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Xue et al. | Embedded 1-Mb ReRAM-based computing-in-memory macro with multibit input and weight for CNN-based AI edge processors | |
Valavi et al. | A 64-tile 2.4-Mb in-memory-computing CNN accelerator employing charge-domain compute | |
Long et al. | ReRAM-based processing-in-memory architecture for recurrent neural network acceleration | |
Si et al. | A local computing cell and 6T SRAM-based computing-in-memory macro with 8-b MAC operation for edge AI chips | |
Daniels et al. | Energy-efficient stochastic computing with superparamagnetic tunnel junctions | |
Mochida et al. | A 4M synapses integrated analog ReRAM based 66.5 TOPS/W neural-network processor with cell current controlled writing and flexible network architecture | |
Hung et al. | A four-megabit compute-in-memory macro with eight-bit precision based on CMOS and resistive random-access memory for AI edge devices | |
Lee et al. | A charge-domain scalable-weight in-memory computing macro with dual-SRAM architecture for precision-scalable DNN accelerators | |
Sayal et al. | A 12.08-TOPS/W all-digital time-domain CNN engine using bi-directional memory delay lines for energy efficient edge computing | |
Kang et al. | A 19.4-nJ/decision, 364-K decisions/s, in-memory random forest multi-class inference accelerator | |
Kadetotad et al. | Parallel architecture with resistive crosspoint array for dictionary learning acceleration | |
Su et al. | Two-way transpose multibit 6T SRAM computing-in-memory macro for inference-training AI edge chips | |
Correll et al. | A fully integrated reprogrammable CMOS-RRAM compute-in-memory coprocessor for neuromorphic applications | |
Yang et al. | An in-memory-computing charge-domain ternary CNN classifier | |
Nasrin et al. | Mf-net: Compute-in-memory sram for multibit precision inference using memory-immersed data conversion and multiplication-free operators | |
Song et al. | A 28 nm 16 kb bit-scalable charge-domain transpose 6T SRAM in-memory computing macro | |
Sayal et al. | COMPAC: Compressed time-domain, pooling-aware convolution CNN engine with reduced data movement for energy-efficient AI computing | |
TWI803889B (en) | Computing device and computing method | |
Sharma et al. | A reconfigurable 16Kb AND8T SRAM macro with improved linearity for multibit compute-in memory of artificial intelligence edge devices | |
Lee et al. | A 17.5-fJ/bit energy-efficient analog SRAM for mixed-signal processing | |
Singh et al. | Srif: Scalable and reliable integrate and fire circuit adc for memristor-based cim architectures | |
Yamaguchi et al. | An energy-efficient time-domain analog CMOS BinaryConnect neural network processor based on a pulse-width modulation approach | |
Cheon et al. | A 2941-TOPS/W charge-domain 10T SRAM compute-in-memory for ternary neural network | |
Sehgal et al. | A bit-serial, compute-in-SRAM design featuring hybrid-integrating ADCs and input dependent binary scaled precharge eliminating DACs for energy-efficient DNN inference | |
Yeo et al. | A power and area efficient CMOS stochastic neuron for neural networks employing resistive crossbar array |