[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/3224207.3224219acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicdpaConference Proceedingsconference-collections
research-article

Research Summary of Convolution Neural Network in Image Recognition

Published: 12 May 2018 Publication History

Abstract

Convolution neural networks is a model of deep learning and it is popular in image recognition, object detection and speech recognition. This paper studied Convolution neural networks in detail. Firstly this paper introduced the generation and development of convolution neural networks and illustrated its advantages in image recognition tasks. Then, it summarized the classic structure of convolution neural networks. Next, this paper stated the study trends and summarized five aspects: appropriately simplified networks, reducing over-fitting, increasing gradient signal, deeper networks and randomization. Finally, this paper discussed the problems existed in convolution neural networks and looked forward to the development trends

References

[1]
Hubel, D.H. & Wiesel, T.N. Receptive fields and functional architecture of monkey striate cortex. J. Physiol. (Lond.) 195, 215--243.
[2]
Fukushima, K. Biol. Cybernetics (1980) 36: 193.
[3]
Y. LeCun and Y. Bengio: Convolutional Networks for Images, Speech, and Time-Series, in Arbib, M. A. (Eds), The Handbook of Brain Theory and Neural Networks, MIT Press, 1995.
[4]
Lécun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition{J}. Proceedings of the IEEE, 1998, 86(11): 2278--2324.
[5]
Krizhevsky A, Sutskever I, Hinton G E. ImageNet Classification with Deep Convolutional Neural Networks{J}. Advances in Neural Information Processing Systems, 2012, 25(2):2012.
[6]
Nair, Vinod, and G. E. Hinton. "Rectified Linear Units Improve Restricted Boltzmann Machines Vinod Nair." International Conference on Machine Learning 2010: 807--814.
[7]
Lin M, Chen Q, Yan S. Network In Network{J}. Computer Science, 2014.
[8]
Springenberg J T, Dosovitskiy A, Brox T, et al. Striving for Simplicity: The All Convolutional Net{J}. Eprint Arxiv, 2014.
[9]
Szegedy C, Liu W, Jia Y, et al. Going deeper with convolutions{J}. 2014: 1--9.
[10]
He K, Zhang X, Ren S, et al. Deep Residual Learning for Image Recognition{J}. Computer Science, 2015.
[11]
Srivastava R K, Greff K, Schmidhuber J. Highway Networks{J}. Computer Science, 2015.
[12]
Huang G, Liu Z, Weinberger K Q. Densely Connected Convolutional Networks{J}. 2016.
[13]
Huang G, Sun Y, Liu Z, et al. Deep Networks with Stochastic Depth{J}. 2016.
[14]
Yann LeCun, Yoshua Bengio & Geoffrey Hinton. Deep learning.
[15]
Smith L N, Topin N. Deep Convolutional Neural Network Design Patterns{J}. 2016.
[16]
He K, Zhang X, Ren S, et al. Deep Residual Learning for Image Recognition{J}. Computer Science, 2015.
[17]
Nagi J, Ducatelle F, Caro G A D, et al. Max-pooling convolutional neural networks for vision-based hand gesture recognition{C}// IEEE International Conference on Signal and Image Processing Applications. IEEE, 2011: 342--347.
[18]
Lecun Y, Kavukcuoglu K, Farabet C. Convolutional networks and applications in vision{C}// IEEE International Symposium on Circuits and Systems. IEEE, 2010: 253--256.
[19]
Jarrett K, Kavukcuoglu K, Ranzato M, et al. What is the best multi-stage architecture for object recognition?{J}. 2009, 30(2): 2146--2153.
[20]
Hawkins D M. The problem of overfitting.{J}. ChemInform, 2004, 35(19): 1--12.
[21]
Hariharan B, Arbeláez P, Girshick R, et al. Hypercolumns for object segmentation and fine-grained localization{J}. 2015: 447--456.
[22]
Hinton G E, Srivastava N, Krizhevsky A, et al. Improving neural networks by preventing co-adaptation of feature detectors{J}. Computer Science, 2012, 3(4):págs. 212--223.
[23]
Srivastava N, Hinton G, Krizhevsky A, et al. Dropout: a simple way to prevent neural networks from overfitting{J}. Journal of Machine Learning Research, 2014, 15(1): 1929--1958.
[24]
Wan L, Zeiler M, Zhang S, et al. Regularization of neural networks using dropconnect{C}// International Conference on Machine Learning. 2013: 1058--1066.
[25]
Hinton G E, Dayan P, Frey B J, et al. The "wake-sleep" algorithm for unsupervised neural networks.{J}. Science, 1995, 268(5214): 1158--61.
[26]
Schölkopf, B, Platt, J, Hofmann, T. Greedy Layer-Wise Training of Deep Networks{J}. Advances in Neural Information Processing Systems, 2007, 19: 153--160.
[27]
Erhan D, Bengio Y, Courville A, et al. Why Does Unsupervised Pre-training Help Deep Learning?{J}. Journal of Machine Learning Research, 2010, 11(3): 625--660.
[28]
Erhan D, Manzagol P A, Bengio Y, et al. The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training.{J}. Immunology of Fungal Infections, 2009, 5: 153--160.
[29]
Erhan D. Understanding Deep Architectures and the Effect of Unsupervised Pre-training{J}. 2011.
[30]
Wu Y N. Data Augmentation{J}. Computer Vision, 2014: 165--166.
[31]
Dyk D A V, Meng X L. The Art of Data Augmentation{J}. Journal of Computational & Graphical Statistics, 2012, 10(1): 1--50.
[32]
Howard A G. Some Improvements on Deep Convolutional Neural Network Based Image Classification{J}. Computer Science, 2013.
[33]
Wong S C, Gatt A, Stamatescu V, et al. Understanding data augmentation for classification: when to warp?{J}. 2016.
[34]
Raiko T, Valpola H, Lecun Y. Deep learning made easier by linear transformations in perceptrons{J}. 2012, 22: 924--932.
[35]
He K, Zhang X, Ren S, et al. Identity Mappings in Deep Residual Networks{J}. 2016.
[36]
Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition{J}. Computer Science, 2014.
[37]
Larsson G, Maire M, Shakhnarovich G. FractalNet: Ultra-Deep Neural Networks without Residuals{J}. 2016.
[38]
A. L. Maas, A. Y. Hannun. And Y. Bengio. Rectifier nonlinearities improve neural network acoustic models, in ICML 2103.
[39]
He K, Zhang X, Ren S, et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification{J}. 2015: 1026--1034.
[40]
Xu B, Wang N, Chen T, et al. Empirical Evaluation of Rectified Activations in Convolutional Network{J}. Computer Science, 2015.
[41]
Agostinelli F, Hoffman M, Sadowski P, et al. Learning Activation Functions to Improve Deep Neural Networks{J}. Computer Science, 2014.
[42]
Yu D, Wang H, Chen P, et al. Mixed Pooling for Convolutional Neural Networks{C}// the 9th international conference on rough sets and knowledge technology. 2014: 364--375.
[43]
Zeiler M D, Fergus R. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks{J}. Computer Science, 2013.
[44]
Goodfellow I J, Wardefarley D, Mirza M, et al. Maxout Networks{J}. Computer Science, 2013: 1319--1327.

Cited By

View all
  • (2021)Recognition Method of Tunnel Lining Defects Based on Deep LearningWireless Communications and Mobile Computing10.1155/2021/90701822021(1-12)Online publication date: 30-Sep-2021
  • (2021)A Comparative Study of Different Machine Learning Algorithms on Bitcoin Value Prediction2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT)10.1109/ICAECT49130.2021.9392629(1-7)Online publication date: 19-Feb-2021

Index Terms

  1. Research Summary of Convolution Neural Network in Image Recognition

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICDPA 2018: Proceedings of the International Conference on Data Processing and Applications
    May 2018
    73 pages
    ISBN:9781450364188
    DOI:10.1145/3224207
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    In-Cooperation

    • Peking University: Peking University
    • Guangdong University of Technology: Guangdong University of Technology

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 12 May 2018

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Convolution neural networks
    2. deep learning
    3. image recognition

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    ICDPA 2018

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)12
    • Downloads (Last 6 weeks)2
    Reflects downloads up to 19 Dec 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2021)Recognition Method of Tunnel Lining Defects Based on Deep LearningWireless Communications and Mobile Computing10.1155/2021/90701822021(1-12)Online publication date: 30-Sep-2021
    • (2021)A Comparative Study of Different Machine Learning Algorithms on Bitcoin Value Prediction2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT)10.1109/ICAECT49130.2021.9392629(1-7)Online publication date: 19-Feb-2021

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media