Sasaki et al., 2019 - Google Patents
Post training weight compression with distribution-based filter-wise quantization stepSasaki et al., 2019
- Document ID
- 6419765911415733700
- Author
- Sasaki S
- Maki A
- Miyashita D
- Deguchi J
- Publication year
- Publication venue
- 2019 IEEE Symposium in Low-Power and High-Speed Chips (COOL CHIPS)
External Links
Snippet
Quantization of models with lower bit precision is a promising method to develop lower- power and smaller-area neural network hardware. However, 4-or lower bit quantization usually requires additional retraining with labeled dataset for backpropagation to improve …
- 238000007906 compression 0 title description 2
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
- G06F7/48—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/30286—Information retrieval; Database structures therefor; File system structures therefor in structured data stores
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114140353B (en) | Swin-Transformer image denoising method and system based on channel attention | |
CN113159173B (en) | Convolutional neural network model compression method combining pruning and knowledge distillation | |
CN110517329B (en) | A deep learning image compression method based on semantic analysis | |
CN106991646B (en) | Image super-resolution method based on dense connection network | |
CN111079781A (en) | A lightweight convolutional neural network image recognition method based on low-rank and sparse decomposition | |
CN110222821A (en) | Convolutional neural networks low-bit width quantization method based on weight distribution | |
CN111147862B (en) | End-to-end image compression method based on target coding | |
Khashman et al. | Image compression using neural networks and Haar wavelet | |
CN112016674A (en) | Knowledge distillation-based convolutional neural network quantification method | |
Laha et al. | Design of vector quantizer for image compression using self-organizing feature map and surface fitting | |
KR20210125425A (en) | System and method of training GAN for real-world super resolution with unknown degradations | |
Sasaki et al. | Post training weight compression with distribution-based filter-wise quantization step | |
CN112734867A (en) | Multispectral image compression method and system based on space spectrum feature separation and extraction | |
Yang et al. | JPEG steganalysis with combined dense connected CNNs and SCA-GFR | |
CN118038269A (en) | Hyperspectral image classification method based on spectral Transformer self-supervised learning algorithm model | |
CN112686384A (en) | Bit-width-adaptive neural network quantization method and device | |
Wu et al. | Fedcomp: A federated learning compression framework for resource-constrained edge computing devices | |
Ando et al. | Dither nn: An accurate neural network with dithering for low bit-precision hardware | |
TW202004568A (en) | Full exponential operation method applied to deep neural network, computer apparatus, and computer-readable recording medium reducing the operation complexity and circuit complexity, increasing the operation speed of the deep neural network and reducing the occupation of memory space. | |
CN108390871A (en) | A kind of radar data compression method based on the prediction of autoregression model frame | |
Khashman et al. | Neural networks arbitration for optimum DCT image compression | |
Zhao et al. | Learned image compression using adaptive block-wise encoding and reconstruction network | |
WO2022247368A1 (en) | Methods, systems, and mediafor low-bit neural networks using bit shift operations | |
Gafour et al. | Genetic fractal image compression | |
Hirose et al. | Quantization error-based regularization for hardware-aware neural network training |