Zero and Narrow-Width Value-Aware Compression for Quantized Convolutional Neural Networks
Abstract
References
Index Terms
- Zero and Narrow-Width Value-Aware Compression for Quantized Convolutional Neural Networks
Recommendations
Simulating quantized inference on convolutional neural networks
AbstractMobile and embedded applications of convolutional neural networks (CNNs) use quantization to reduce model size and increase computational efficiency. However, working with quantized networks often implies using non-standard training ...
Highlights- Simulation of fixed-point quantization inference in convolutional neural networks.
Lossless Text Compression Using Recurrent Neural Networks
AbstractLossless Data compression is the process of reducing the size or the number of bits required to represent data, and Arithmetic coding is one of the popular lossless text compression techniques. This project focuses on lossless data compression ...
MXQN:Mixed quantization for reducing bit-width of weights and activations in deep convolutional neural networks
AbstractQuantization, which involves bit-width reduction, is considered as one of the most effective approaches to rapidly and energy-efficiently deploy deep convolutional neural networks (DCNNs) on resource-constrained embedded hardware. However, bit-...
Comments
Please enable JavaScript to view thecomments powered by Disqus.Information & Contributors
Information
Published In
Publisher
IEEE Computer Society
United States
Publication History
Qualifiers
- Research-article
Contributors
Other Metrics
Bibliometrics & Citations
Bibliometrics
Article Metrics
- 0Total Citations
- 0Total Downloads
- Downloads (Last 12 months)0
- Downloads (Last 6 weeks)0