Meng et al., 2021 - Google Patents
Structured pruning of RRAM crossbars for efficient in-memory computing acceleration of deep neural networksMeng et al., 2021
View PDF- Document ID
- 4526518917197681484
- Author
- Meng J
- Yang L
- Peng X
- Yu S
- Fan D
- Seo J
- Publication year
- Publication venue
- IEEE Transactions on Circuits and Systems II: Express Briefs
External Links
Snippet
The high computational complexity and a large number of parameters of deep neural networks (DNNs) become the most intensive burden of deep learning hardware design, limiting efficient storage and deployment. With the advantage of high-density storage, non …
- 230000001537 neural 0 title abstract description 9
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
- G06N3/0635—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
- G06F7/48—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
- G06F7/52—Multiplying; Dividing
- G06F7/523—Multiplying only
- G06F7/53—Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/16—Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/58—Random or pseudo-random number generators
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Meng et al. | Structured pruning of RRAM crossbars for efficient in-memory computing acceleration of deep neural networks | |
Liu et al. | Parallelizing SRAM arrays with customized bit-cell for binary neural networks | |
Cai et al. | Low bit-width convolutional neural network on RRAM | |
Chou et al. | Cascade: Connecting rrams to extend analog dataflow in an end-to-end in-memory processing paradigm | |
Wan et al. | A compute-in-memory chip based on resistive random-access memory | |
Zhu et al. | A configurable multi-precision CNN computing framework based on single bit RRAM | |
US10346347B2 (en) | Field-programmable crossbar array for reconfigurable computing | |
Yuan et al. | Forms: Fine-grained polarized reram-based in-situ computation for mixed-signal dnn accelerator | |
Sun et al. | Fully parallel RRAM synaptic array for implementing binary neural network with (+ 1,− 1) weights and (+ 1, 0) neurons | |
Tang et al. | Binary convolutional neural network on RRAM | |
Wang et al. | SNrram: An efficient sparse neural network computation architecture based on resistive random-access memory | |
Lin et al. | Learning the sparsity for ReRAM: Mapping and pruning sparse neural network for ReRAM based accelerator | |
Gonugondla et al. | Fundamental limits on the precision of in-memory architectures | |
Sun et al. | Computing-in-memory with SRAM and RRAM for binary neural networks | |
Kim et al. | Input-splitting of large neural networks for power-efficient accelerator with resistive crossbar memory array | |
Tu et al. | SDP: Co-designing algorithm, dataflow, and architecture for in-SRAM sparse NN acceleration | |
Azamat et al. | Quarry: Quantization-based ADC reduction for ReRAM-based deep neural network accelerators | |
Meng et al. | Exploring compute-in-memory architecture granularity for structured pruning of neural networks | |
Qu et al. | ASBP: Automatic structured bit-pruning for RRAM-based NN accelerator | |
Rhe et al. | VWC-SDK: Convolutional weight mapping using shifted and duplicated kernel with variable windows and channels | |
Chang et al. | E-UPQ: Energy-aware unified pruning-quantization framework for CIM architecture | |
Liu et al. | Era-bs: Boosting the efficiency of reram-based pim accelerator with fine-grained bit-level sparsity | |
Yang et al. | Apq: Automated dnn pruning and quantization for reram-based accelerators | |
Liu et al. | An energy-efficient mixed-bit CNN accelerator with column parallel readout for ReRAM-based in-memory computing | |
Kim et al. | SNPU: An energy-efficient spike domain deep-neural-network processor with two-step spike encoding and shift-and-accumulation unit |