[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

Kim et al., 2024 - Google Patents

BS2: Bit-Serial Architecture Exploiting Weight Bit Sparsity for Efficient Deep Learning Acceleration

Kim et al., 2024

Document ID
17844911520676885994
Author
Kim E
Lee S
Kim C
Lim H
Nam J
Sim J
Publication year
Publication venue
2024 21st International SoC Design Conference (ISOCC)

External Links

Snippet

Most of weights in deep learning models are small, thus they show high bit sparsity in MSBs. Based on this observation, we propose a bit-serial processing architecture (BS2) that exploits such bit sparsity to maximize computing efficiency. In this architecture, a bit feed …
Continue reading at ieeexplore.ieee.org (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/52Multiplying; Dividing
    • G06F7/523Multiplying only
    • G06F7/53Multiplying only in parallel-parallel fashion, i.e. both operands being entered in parallel
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/38Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
    • G06F7/48Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
    • G06F7/50Adding; Subtracting
    • G06F7/505Adding; Subtracting in bit-parallel fashion, i.e. having a different digit-handling circuit for each denomination
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for programme control, e.g. control unit
    • G06F9/06Arrangements for programme control, e.g. control unit using stored programme, i.e. using internal store of processing equipment to receive and retain programme
    • G06F9/30Arrangements for executing machine-instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/30007Arrangements for executing specific machine instructions to perform operations on data operands
    • G06F9/30021Compare instructions, e.g. Greater-Than, Equal-To, MINMAX
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/12Computer systems based on biological models using genetic models
    • G06N3/126Genetic algorithms, i.e. information processing using digital simulations of the genetic system
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/22Arrangements for sorting or merging computer data on continuous record carriers, e.g. tape, drum, disc
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • G06N99/005Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F2207/00Indexing scheme relating to methods or arrangements for processing data by operating upon the order or content of the data handled
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/50Computer-aided design
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/76Architectures of general purpose stored programme computers
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRICAL DIGITAL DATA PROCESSING
    • G06F19/00Digital computing or data processing equipment or methods, specially adapted for specific applications
    • G06F19/10Bioinformatics, i.e. methods or systems for genetic or protein-related data processing in computational molecular biology

Similar Documents

Publication Publication Date Title
US20240112029A1 (en) Acceleration of model/weight programming in memristor crossbar arrays
Salamat et al. Rnsnet: In-memory neural network acceleration using residue number system
Sim et al. Scalable stochastic-computing accelerator for convolutional neural networks
CN112068798B (en) Method and device for realizing importance ordering of network nodes
US10824394B2 (en) Concurrent multi-bit adder
US20220188604A1 (en) Method and Apparatus for Performing a Neural Network Operation
KR102409615B1 (en) Method for min-max computation in associative memory
Cai et al. Training low bitwidth convolutional neural network on RRAM
Shukla et al. Mc-cim: Compute-in-memory with monte-carlo dropouts for bayesian edge intelligence
Stevens et al. GNNerator: A hardware/software framework for accelerating graph neural networks
Kang et al. S-FLASH: A NAND flash-based deep neural network accelerator exploiting bit-level sparsity
Geng et al. CQNN: a CGRA-based QNN framework
Karavay et al. Qubit fault detection in SoC logic
Shivanandamurthy et al. Atria: A bit-parallel stochastic arithmetic based accelerator for in-dram cnn processing
Kim et al. BS2: Bit-Serial Architecture Exploiting Weight Bit Sparsity for Efficient Deep Learning Acceleration
US20230244901A1 (en) Compute-in-memory sram using memory-immersed data conversion and multiplication-free operators
JP2022074442A (en) Arithmetic device and arithmetic method
Block et al. A hardware acceleration of a phylogenetic tree reconstruction with maximum parsimony algorithm using FPGA
Kang et al. An energy-efficient programmable mixed-signal accelerator for machine learning algorithms
Klhufek et al. Exploring Quantization and Mapping Synergy in Hardware-Aware Deep Neural Network Accelerators
Zhu et al. iMAT: Energy-Efficient In-Memory Acceleration for Ternary Neural Networks With Sparse Dot Product
US20240143541A1 (en) Compute in-memory architecture for continuous on-chip learning
CN110989971B (en) System and method for energy-saving data processing
CN110765413A (en) Matrix summation structure and neural network computing platform
US12032959B2 (en) Non-volatile memory die with latch-based multiply-accumulate components