[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

Crafton et al., 2020 - Google Patents

Breaking barriers: Maximizing array utilization for compute in-memory fabrics

Crafton et al., 2020

View PDF
Document ID
1375611465552224606
Author
Crafton B
Spetalnick S
Murali G
Krishna T
Lim S
Raychowdhury A
Publication year
Publication venue
2020 IFIP/IEEE 28th International Conference on Very Large Scale Integration (VLSI-SOC)

External Links

Snippet

Compute in-memory (CIM) is a promising technique that minimizes data transport, the primary performance bottleneck and energy cost of most data intensive applications. This has found wide-spread adoption in accelerating neural networks for machine learning …
Continue reading at arxiv.org (PDF) (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers
    • G06F15/80Architectures of general purpose stored programme computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
    • G06F15/8007Architectures of general purpose stored programme computers comprising an array of processing units with common control, e.g. single instruction multiple data processors single instruction multiple data [SIMD] multiprocessors
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a programme unit and a register, e.g. for a simultaneous processing of several programmes
    • G06F15/163Interprocessor communication
    • G06F15/173Interprocessor communication using an interconnection network, e.g. matrix, shuffle, pyramid, star, snowflake
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers
    • G06F15/78Architectures of general purpose stored programme computers comprising a single central processing unit
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for programme control, e.g. control unit
    • G06F9/06Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/52Multiplying; Dividing
    • G06F7/523Multiplying only
    • G06F7/53Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Error detection; Error correction; Monitoring responding to the occurence of a fault, e.g. fault tolerance
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/12Computer systems based on biological models using genetic models
    • G06N3/126Genetic algorithms, i.e. information processing using digital simulations of the genetic system

Similar Documents

Publication Publication Date Title
Hu et al. A survey on convolutional neural network accelerators: GPU, FPGA and ASIC
Wang et al. FPDeep: Scalable acceleration of CNN training on deeply-pipelined FPGA clusters
Joardar et al. AccuReD: High accuracy training of CNNs on ReRAM/GPU heterogeneous 3-D architecture
CN107203807B (en) On-chip cache bandwidth balancing method, system and device of neural network accelerator
Yang et al. PIMGCN: A ReRAM-based PIM design for graph convolutional network acceleration
Acer et al. Improving performance of sparse matrix dense matrix multiplication on large-scale parallel systems
Arka et al. ReGraphX: NoC-enabled 3D heterogeneous ReRAM architecture for training graph neural networks
CN112119459A (en) Memory arrangement for tensor data
Feng et al. Cosparse: A software and hardware reconfigurable spmv framework for graph analytics
Catthoor et al. Very large-scale neuromorphic systems for biological signal processing
Arka et al. DARe: DropLayer-aware manycore ReRAM architecture for training graph neural networks
Crafton et al. Breaking barriers: Maximizing array utilization for compute in-memory fabrics
Sun et al. Multi-node acceleration for large-scale GCNs
Joardar et al. Heterogeneous manycore architectures enabled by processing-in-memory for deep learning: From CNNs to GNNs:(ICCAD special session paper)
Akbari et al. A high-performance network-on-chip topology for neuromorphic architectures
Zhang et al. Simeuro: A hybrid CPU-GPU parallel simulator for neuromorphic computing chips
Wang et al. SPCIM: Sparsity-Balanced Practical CIM Accelerator With Optimized Spatial-Temporal Multi-Macro Utilization
Ravichandiran et al. A review of 3D-dynamic random-access memory based near-memory computation
Ascia et al. Networks-on-chip based deep neural networks accelerators for iot edge devices
Venkateswaran et al. Memory in processor: A novel design paradigm for supercomputing architectures
Zhou et al. Pim-dl: Boosting dnn inference on digital processing in-memory architectures via data layout optimizations
Wang et al. Benchmarking DNN Mapping Methods for the In-Memory Computing Accelerators
Joshi et al. Neuromorphic event-driven multi-scale synaptic connectivity and plasticity
Crafton et al. Statistical Array Allocation and Partitioning for Compute In-Memory Fabrics
Liu et al. Regularizing sparse and imbalanced communications for voxel-based brain simulations on supercomputers