Authors:
Mariana Oliveira Prazeres
1
;
2
;
Xinlin Li
2
;
Adam Oberman
1
and
Vahid Partovi Nia
2
Affiliations:
1
Department of Mathematics and Statistics, McGill University, Montreal, Canada
;
2
Huawei Noah’s Ark, Montreal, Canada
Keyword(s):
Neural Network Compression, Hardware-aware Architectures.
Abstract:
In order to deploy deep neural networks on edge devices, compressed (resource efficient) networks need to be developed. While established compression methods, such as quantization, pruning, and architecture search are designed for conventional hardware, further gains are possible if compressed architectures are coupled with novel hardware designs. In this work, we propose EuclidNet, a compressed network designed to be implemented on hardware which replaces multiplication, wx, with squared difference (x − w)2. EuclidNet allows for a low precision hardware implementation which is about twice as efficient (in term of logic gate counts) as the comparable conventional hardware, with acceptably small loss of accuracy. Moreover, the network can be trained and quantized using standard methods, without requiring additional training time. Codes and pre-trained models are available.