[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

Meng et al., 2021 - Google Patents

Structured pruning of RRAM crossbars for efficient in-memory computing acceleration of deep neural networks

Meng et al., 2021

View PDF
Document ID
4526518917197681484
Author
Meng J
Yang L
Peng X
Yu S
Fan D
Seo J
Publication year
Publication venue
IEEE Transactions on Circuits and Systems II: Express Briefs

External Links

Snippet

The high computational complexity and a large number of parameters of deep neural networks (DNNs) become the most intensive burden of deep learning hardware design, limiting efficient storage and deployment. With the advantage of high-density storage, non …
Continue reading at ieeexplore.ieee.org (PDF) (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
    • G06N3/0635Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/52Multiplying; Dividing
    • G06F7/523Multiplying only
    • G06F7/53Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • G06N3/04Architectures, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/12Computer systems based on biological models using genetic models
    • G06N3/126Genetic algorithms, i.e. information processing using digital simulations of the genetic system
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/58Random or pseudo-random number generators

Similar Documents

Publication Publication Date Title
Meng et al. Structured pruning of RRAM crossbars for efficient in-memory computing acceleration of deep neural networks
Liu et al. Parallelizing SRAM arrays with customized bit-cell for binary neural networks
Cai et al. Low bit-width convolutional neural network on RRAM
Chou et al. Cascade: Connecting rrams to extend analog dataflow in an end-to-end in-memory processing paradigm
Wan et al. A compute-in-memory chip based on resistive random-access memory
Zhu et al. A configurable multi-precision CNN computing framework based on single bit RRAM
US10346347B2 (en) Field-programmable crossbar array for reconfigurable computing
Yuan et al. Forms: Fine-grained polarized reram-based in-situ computation for mixed-signal dnn accelerator
Sun et al. Fully parallel RRAM synaptic array for implementing binary neural network with (+ 1,− 1) weights and (+ 1, 0) neurons
Tang et al. Binary convolutional neural network on RRAM
Wang et al. SNrram: An efficient sparse neural network computation architecture based on resistive random-access memory
Lin et al. Learning the sparsity for ReRAM: Mapping and pruning sparse neural network for ReRAM based accelerator
Gonugondla et al. Fundamental limits on the precision of in-memory architectures
Sun et al. Computing-in-memory with SRAM and RRAM for binary neural networks
Kim et al. Input-splitting of large neural networks for power-efficient accelerator with resistive crossbar memory array
Tu et al. SDP: Co-designing algorithm, dataflow, and architecture for in-SRAM sparse NN acceleration
Azamat et al. Quarry: Quantization-based ADC reduction for ReRAM-based deep neural network accelerators
Meng et al. Exploring compute-in-memory architecture granularity for structured pruning of neural networks
Qu et al. ASBP: Automatic structured bit-pruning for RRAM-based NN accelerator
Rhe et al. VWC-SDK: Convolutional weight mapping using shifted and duplicated kernel with variable windows and channels
Chang et al. E-UPQ: Energy-aware unified pruning-quantization framework for CIM architecture
Liu et al. Era-bs: Boosting the efficiency of reram-based pim accelerator with fine-grained bit-level sparsity
Yang et al. Apq: Automated dnn pruning and quantization for reram-based accelerators
Liu et al. An energy-efficient mixed-bit CNN accelerator with column parallel readout for ReRAM-based in-memory computing
Kim et al. SNPU: An energy-efficient spike domain deep-neural-network processor with two-step spike encoding and shift-and-accumulation unit