Abstract
With the advent of deep learning applications on edge devices, researchers actively try to optimize deep learning model deployment on low-power and restricted memory devices. There are established compression methods such as quantization, pruning, and architecture search that leverage commodity hardware. Apart from conventional compression algorithms, one may redesign the operations of deep learning models, leading to more efficient hardware implementation. To this end, we propose EuclidNet, an efficient computing method designed to be implemented on hardware that replaces multiplication, with squared difference. We show that EuclidNet is aligned with matrix multiplication and can be used as a measure of similarity in the case of convolutional layers. Furthermore, we show that under various transformations and noise scenarios, EuclidNet exhibits the same performance compared to the deep learning models designed with multiplication operations.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Benmeziane H, Maghraoui KE, Ouarnoughi H, Niar S, Wistuba M, Wang N. A comprehensive survey on hardware-aware neural architecture search. 2021 arXiv preprint arXiv:2101.09336.
Cheng Y, Wang D, Zhou P, Zhang T. A survey of model compression and acceleration for deep neural networks. 2017 arXiv preprint arXiv:1710.09282.
Hennessy JL, Patterson DA. Computer architecture: a quantitative approach. London: Elsevier; 2011.
Chen H, Wang Y, Xu C, Shi B, Xu C, Tian Q, Xu C. Addernet: Do we really need multiplications in deep learning? In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020 pp. 1468–1477.
de Dinechin F, Pasca B. Large multipliers with fewer dsp blocks. In: 2009 International Conference on Field Programmable Logic and Applications, 2009 pp. 250–255. IEEE.
Baluja S, Marwood D, Covell M, Johnston N. No multiplication? no floating point? no problem! training networks for efficient inference. 2018 arXiv preprint arXiv:1809.09244.
Covell M, Marwood D, Baluja S, Johnston N. Table-based neural units: Fully quantizing networks for multiply-free inference. 2019 arXiv preprint arXiv:1906.04798.
Razlighi MS, Imani M, Koushanfar F, Looknn Rosing T. Neural network with no multiplication. In: Design, Automation & Test in Europe Conference & Exhibition. IEEE; 2017. p. 1775–80.
Frankle J, Carbin M. The lottery ticket hypothesis: Finding sparse, trainable neural networks. 2018 arXiv preprint arXiv:1803.03635.
Cheng Y, Wang D, Zhou P, Zhang T. Model compression and acceleration for deep neural networks: The principles, progress, and challenges. IEEE Sig Process Mag. 2018;35(1):126–36.
Wu C-J, Brooks D, Chen K, Chen D, Choudhury S, Dukhan M, Hazelwood K, Isaac E, Jia Y, Jia B, et al. Machine learning at facebook: Understanding inference at the edge. In: 2019 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2019 , pp. 331–344. IEEE.
Reed R. Pruning algorithms-a survey. IEEE Transact Neural Net. 1993;4(5):740–7.
Guo, Y. A survey on methods and theories of quantized neural networks. 2018 arXiv preprint arXiv:1808.04752.
Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network. 2015 arXiv preprint arXiv:1503.02531.
Iandola FN, Han S, Moskewicz MW, Ashraf K, Dally WJ, Keutzer K. Squeezenet: Alexnet-level accuracy with 50x fewer parameters and< 0.5 mb model size. 2016 arXiv preprint arXiv:1602.07360
Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. 2017 arXiv preprint arXiv:1704.04861
Zhang X, Zhou X, Lin M, Sun J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018 pp. 6848–6856.
Tan M, Le QV. Efficientnet: Rethinking model scaling for convolutional neural networks. 2019 arXiv preprint arXiv:1905.11946.
Hubara I, Courbariaux M, Soudry D, El-Yaniv R, Bengio Y. Binarized neural networks. In: Lee D, Sugiyama M, Luxburg U, Guyon I, Garnett R, editors. Advances in neural information processing systems, vol. 29. Curran Associates Inc; 2016.
Xu Y, Xu C, Chen X, Zhang W, Xu C, Wang Y. Kernel based progressive distillation for adder neural networks. 2020 arXiv preprint arXiv:2009.13044
Dogaru R, Chua LO. The comparative synapse: a multiplication free approach to neuro-fuzzy classifiers. IEEE Transact Circuits Syst. 1999;46(11):1366–71.
Akbaş CE, Bozkurt A, çetin AE, çetin-Atalay R, Üner A. Multiplication-free neural networks. In: 2015 23nd Signal Processing and Communications Applications Conference (SIU), 2015 pp. 2416–2418. Doi: https://doi.org/10.1109/SIU.2015.7130369.
Mallah M. Multiplication free neural networks. PhD thesis, Bilkent University, 2018.
Afrasiyabi A, Badawi D, Nasir B, Yildi O, Vural FTY, Çetin AE. Non-euclidean vector product for neural networks. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018, pp. 6862–6866. IEEE
Badawi D, Akhan E, Mallah M, Üner A, Çetin-Atalay R, Çetin AE. Multiplication free neural network for cancer stem cell detection in h-and-e stained liver images. In: Compressive Sensing VI: From Diverse Modalities to Big Data Analytics, 2017, vol. 10211, pp. 102110C1–102110C9. International Society for Optics and Photonics. https://doi.org/10.1117/12.2262338
Pan H, Badawi D, Zhang X, Cetin AE. Additive neural network for forest fire detection. Sig, Image Video Process. 2019;14:675–82.
You H, Chen X, Zhang Y, Li C, Li S, Liu Z, Wang Z, Lin Y. Shiftaddnet: A hardware-inspired deep network. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (eds.) Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, Virtual.
Limonova E, Alfonso D, Nikolaev D, Arlazarov VV. Resnet-like architecture with low hardware requirements. 2020 arXiv preprint arXiv:2009.07190.
Limonova E, Matveev D, Nikolaev D, Arlazarov VV. Bipolar morphological neural networks: convolution without multiplication. In: Twelfth International Conference on Machine Vision (ICMV 2019), 2020, vol. 11433, p. 114333. International Society for Optics and Photonics.
Mondal R, Santra S, Mukherjee S, Chanda B. Morphological Network: How Far Can We Go with Morphological Neurons? 2022. arXiv:1901.00109
Davidson JL, Ritter GX. Theory of morphological neural networks. In: Digital Optical Computing II. UK: International Society for Optics and Photonics; 1990. p. 709–17.
Ritter GX, Sussner P. An introduction to morphological neural networks. In: Proceedings of 13th International Conference on Pattern Recognition, 1996, vol. 4, pp. 709–717. IEEE.
Weimerskirch A, Paar C. Generalizations of the karatsuba algorithm for efficient implementations. Cryptology ePrint Archive, Paper 2006/224. https://eprint.iacr.org/2006/224
Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp. 234–241. Springer
He K, Gkioxari G, Dollár P, Girshick R. Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2961–2969.
Chen L-C, Papandreou G, Schroff F, Adam H. Rethinking atrous convolution for semantic image segmentation. 2017 arXiv preprint arXiv:1706.05587.
NVIDIA Vingelmann P, Fitzek FHP. CUDA, Accessed: 10 Feb 89, 2020. https://developer.nvidia.com/cuda-toolkit.
Allgower EL, Georg K. Introduct Num Cont Methods. New York: SIAM; 2003.
Chow J, Udpa L, Udpa S. Homotopy continuation methods for neural networks. In: 1991., IEEE International Sympoisum on Circuits and Systems, 1991, pp. 2483–2486. IEEE.
Pathak HN, Paffenroth R. Parameter continuation methods for the optimization of deep neural networks. In: 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA), 2019, pp. 1637–1643. IEEE.
Cao Z, Long M, Wang J, Yu PS. Hashnet: Deep learning to hash by continuation. In: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 5608–5617.
Mobahi H. Training recurrent neural networks by diffusion. 2016, arXiv preprint arXiv:1601.04114.
Farhadi F, Partovi Nia V, Lodi A. Activation adaptation in neural networks. In: Proceedings of the 9th International Conference on Pattern Recognition Applications and Methods - Volume 1: ICPRAM, 2020.
Gulcehre C, Moczulski M, Visin F, Bengio Y. Mollifying networks. 2016 arXiv preprint arXiv:1608.04980.
Chen Q, Hao W. A homotopy training algorithm for fully connected neural networks. 2019. Proc R Soc A. https://doi.org/10.1098/rspa.2019.0662
Bengio Y, Louradour J, Collobert R, Weston J. Curriculum learning. In: Proceedings of the 26th Annual International Conference on Machine Learning, 2009, pp. 41–48.
Yim J, Joo D, Bae J, Kim J. A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 4133–4141.
Krizhevsky A, Hinton G, et al. Learning multiple layers of features from tiny images. 2009. Technical Report. https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf
He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. 2015, CoRR abs/1512.03385 (2015).
Wu H, Judd P, Zhang X, Isaev M, Micikevicius P. Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation, 2020. arXiv:2004.09602
Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: A large-scale hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 2009, pp. 248–255. https://doi.org/10.1109/CVPR.2009.5206848
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
All authors declare that they have no conflicts of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This article is part of the topical collection “Advances on Pattern Recognition Applications and Methods 2022” guest edited by Ana Fred, Maria De Marsico and Gabriella Sanniti di Baja.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Li, X., Parazeres, M., Oberman, A. et al. EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models. SN COMPUT. SCI. 4, 507 (2023). https://doi.org/10.1007/s42979-023-01921-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s42979-023-01921-y