Abstract
In this work, we propose an improved weight-constrained neural network training algorithm, named iWCNN. The proposed algorithm exploits the numerical efficiency of the L-BFGS matrices together with a gradient-projection strategy for handling the bounds on the weights. Additionally, an attractive property of iWCNN is that it utilizes a new scaling factor for defining the initial Hessian approximation used in the L-BFGS formula. Since the L-BFGS Hessian approximation is defined utilizing a small number of correction vector pairs, our motivation is to further exploit them in order to increase the efficiency of the training algorithm and the convergence rate of the minimization process. The preliminary numerical experiments provide empirical evidence that the proposed training algorithm accelerates the training process.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Al-Baali M (1998) Numerical experience with a class of self-scaling quasi-Newton algorithms. J Optim Theory Appl 96(3):533–553
Awan SM, Aslam M, Khan ZA, Saeed H (2014) An efficient model based on artificial bee colony optimization algorithm with neural networks for electric load forecasting. Neural Comput Appl 25(7–8):1967–1978
Barzilai J, Borwein JM (1988) Two-point step size gradient methods. IMA J Numer Anal 8(1):141–148
Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP (2002) SMOTE: synthetic minority over-sampling technique. J Artif Intell Res 16:321–357
Chen W, Wang Z, Zhou J (2014) Large-scale L-BFGS using MapReduce. In: Advances in neural information processing systems, pp 1332–1340
Cui K, Qin X (2018) Virtual reality research of the dynamic characteristics of soft soil under metro vibration loads based on BP neural networks. Neural Comput Appl 29(5):1233–1242
Demertzis K, Iliadis L (2015) Intelligent bio-inspired detection of food borne pathogen by DNA barcodes: the case of invasive fish species Lagocephalus sceleratus. In: International conference on engineering applications of neural networks. Springer, pp 89–99
Dolan E, Moré JJ (2002) Benchmarking optimization software with performance profiles. Math Program 91:201–213
Dua D, Taniskidou EK (2017) UCI machine learning repository
Erzin Y, Gul TO (2014) The use of neural networks for the prediction of the settlement of one-way footings on cohesionless soils based on standard penetration test. Neural Comput Appl 24(3–4):891–900
Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2414–2423
Horton P, Nakai K (1997) Better prediction of protein cellular localization sites with the \(k\)-nearest neighbors classifier. In: Intelligent systems in molecular biology, pp 368–383
Iliadis L, Mansfield SD, Avramidis S, El-Kassaby YA (2013) Predicting Douglas-fir wood density by artificial neural networks (ANN) based on progeny testing information. Holzforschung 67(7):771–777
Iliadis L, Margaritis K, Maglogiannis I (2017) Timely advances in evolving neural-based systems special issue. Evol Syst 8(1):1–2
Jia F, Lei Y, Guo L, Lin J, Xing S (2018) A neural network constructed by deep learning technique and its application to intelligent fault diagnosis of machines. Neurocomputing 272:619–628
Kayaer K, Yıldırım T (2003) Medical diagnosis on pima Indian diabetes using general regression neural networks. In: Proceedings of the international conference on artificial neural networks and neural information processing, pp 181–184
Kostić S, Vasović D (2015) Prediction model for compressive strength of basic concrete mixture using artificial neural networks. Neural Comput Appl 26(5):1005–1024
Li F, Zhang X, Zhang X, Du C, Xu Y, Tian YC (2018) Cost-sensitive and hybrid-attribute measure multi-decision tree over imbalanced data sets. Inf Sci 422:242–256
Liang P, Labedan B, Riley M (2002) Physiological genomics of Escherichia coli protein families. Physiol Genom 9:15–26
Liu DC, Nocedal J (1989) On the limited memory BFGS method for large scale optimization. Math Program 45(1–3):503–528
Livieris IE (2018) Improving the classification efficiency of an ANN utilizing a new training methodology. Informatics 6(1):1–17
Livieris IE (2019) Forecasting economy-related data utilizing constrained recurrent neural networks. Algorithms 12:85
Livieris IE, Pintelas P (2012) An improved spectral conjugate gradient neural network training algorithm. Int J Artif Intell Tools 21(1):1250009
Maren AJ, Harston CT, Pap RM (2014) Handbook of neural computing applications. Academic Press, Cambridge
Morales JL, Nocedal J (2011) Remark on “Algorithm 778: L-BFGS-B: Fortran subroutines for large-scale bound constrained optimization”. ACM Trans Math Softw (TOMS) 38(1):7
Moré JJ, Thuente DJ (1994) Line search algorithms with guaranteed sufficient decrease. ACM Trans Math Softw (TOMS) 20(3):286–307
Nguyen D, Widrow B (1990) Improving the learning speed of 2-layer neural network by choosing initial values of adaptive weights. Biol Cybern 59:71–113
Nocedal J, Wright S (2006) Numerical optimization. Springer, Berlin
Oren SS, Luenberger DG (1974) Self-scaling variable metric (ssvm) algorithms: part I: criteria and sufficient conditions for scaling a class of algorithms. Manag Sci 20(5):845–862
Shanno DF, Phua KH (1978) Matrix conditioning and nonlinear optimization. Math Program 14(1):149–160
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15(1):1929–1958
Yu J, Wang S, Xi L (2008) Evolving artificial neural networks using an improved PSO and DPSO. Neurocomputing 71:1054–1060
Zhou B, Gao L, Dai YH (2006) Gradient methods with adaptive step-sizes. Comput Optim Appl 35(1):69–86
Zhu C, Byrd RH, Lu P, Nocedal J (1997) Algorithm 778: L-BFGS-B: Fortran subroutines for large-scale bound-constrained optimization. ACM Trans Math Softw (TOMS) 23(4):550–560
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declared no potential conflict of interest with respect to the research, authorship and/or publication of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Livieris, I.E., Pintelas, P. An improved weight-constrained neural network training algorithm. Neural Comput & Applic 32, 4177–4185 (2020). https://doi.org/10.1007/s00521-019-04342-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-019-04342-2