KR20170081887A - Method of determining the final answer by convolution of the artificial neural networks - Google Patents
Method of determining the final answer by convolution of the artificial neural networks Download PDFInfo
- Publication number
- KR20170081887A KR20170081887A KR1020160000880A KR20160000880A KR20170081887A KR 20170081887 A KR20170081887 A KR 20170081887A KR 1020160000880 A KR1020160000880 A KR 1020160000880A KR 20160000880 A KR20160000880 A KR 20160000880A KR 20170081887 A KR20170081887 A KR 20170081887A
- Authority
- KR
- South Korea
- Prior art keywords
- correct answer
- artificial neural
- neural network
- sub
- final
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
A method of determining a final correct answer using convolution in an artificial neural network according to an embodiment of the present invention includes the steps of receiving vectors having an arbitrary dimension, Dividing a plurality of candidate answers into sub-regions of a predetermined size, learning an artificial neural network for each of the partial regions, determining a plurality of correct answers from the vectors obtained as learning results, and And selecting a final correct answer from the plurality of correct answers.
Description
The present invention relates to an artificial neural network technique, and more particularly, to a method for determining a final correct answer using convolution in an artificial neural network.
In general, the Artificial Neural Network in machine learning and cognitive science is a statistical learning algorithm inspired by biological neural networks (the animal's central nervous system, especially the brain). An artificial neural network refers to the entire model that has artificial neurons (nodes) that form a network of synapses by changing the binding strength of synapses through learning.
This artificial neural network technique can improve the performance of classification problem by using convolution. Specifically, the input total observation information is divided into sub-regions. At this time, the sub-regions are allowed to overlap and set to include all or part of the entire observation information. Each sub-region is learned by constructing a separate neural network, forming a convolutional layer.
In each neural network of learned sub-regions, an output vector is generated, which is changed to a representative value or a representative vector by a pooling technique. The input vector of the next layer is generated by generating a new vector following the modified representative vectors. In this way, the convolutional layer can represent upper patterns as a combination of sub-patterns that appear locally in the lower layer.
In contrast to conventional artificial neural networks or kernel machines iteratively transforming input values to ensure non-linearity, learning through this convolutional layer directly implements the abstraction process from lower to higher patterns It is possible to do it. This is not only convenient for visualizing the neural network configuration, but it can also mitigate the data sparseness that occurs when input values are given in very large dimensions.
In the conventional technique, after several stages of abstraction, classification is performed through a fixed decision layer using only a simple logistic regression layer at the final stage. This method is based on the assumption that the output information learned at the top layer immediately before the decision is pattern-related information that directly affects the problem to be solved.
However, if there is no sufficient knowledge or empirical knowledge about the types of patterns and search spaces on the top layer, data sparseness due to large dimensions also occurs in the top layer.
Also, in the process of applying the convolutional layer to the decision layer to obtain the data sparseness mitigation capability provided by the convolutional layer, the pooling technique for the existing representative vector generation is directly used for decision making. .
For example, existing pooling techniques can be divided into max pooling and average pooling. Both of these methods are for designating representative values of the patterns of the respective sub-regions, and the discrete information required in the decision step is lost.
It is an object of the present invention to provide a method for determining a final correct answer of an artificial neural network using convolution for dimension reduction of top-layer patterns and data sparseness reduction.
According to an aspect of the present invention, there is provided a method for determining a final correct answer using convolution in an artificial neural network, comprising: receiving vectors having an arbitrary dimension; inputting the input vectors into a sub- regions, learning the artificial neural network for each of the partial regions, determining a plurality of correct answers from the vectors obtained as a result of the learning, and extracting final correct answers from the plurality of correct answers .
According to the embodiment of the present invention, by using the convolution method in the final correct answer determination process of the artificial neural network, data sparseness due to dimensional reduction is alleviated by learning a wide range of inputs with a separate neural network having a low dimension.
In addition, according to the embodiment of the present invention, an artificial neural network is constructed for sub-regions of input in the step of finally determining the correct answer in the artificial neural network, the correct answer candidates are directly determined at the output of each neural network, By determining the final candidate (final correct answer) from the candidates, it is possible to make error correction in a way that generates one correct answer from the entire input, and even if the correct answer decision from the local patterns is not correct, The artificial neural network can mitigate the difficulty of learning the values that are not the maximum value in the probability vector for determining the correct answer, thereby estimating the best correct answer prediction result through the patterns that can be observed even in the case of the prediction failure .
In addition, according to the embodiment of the present invention, it is possible to learn not only the performance improvement in the final layer but also the most abstract patterns independently appearing in position, thereby reducing the number of stacking required or reducing data, Can improve performance.
BRIEF DESCRIPTION OF THE DRAWINGS FIG. 1 is a block diagram of a final correct answer determining apparatus using convolution of an artificial neural network according to an embodiment of the present invention; FIG.
BACKGROUND OF THE INVENTION Field of the Invention [0001] The present invention relates to an artificial neural network,
3 is a flowchart of a method for determining a final correct answer using convolution in an artificial neural network according to an embodiment of the present invention.
BRIEF DESCRIPTION OF THE DRAWINGS The advantages and features of the present invention and the manner of achieving them will become apparent with reference to the embodiments described in detail below with reference to the accompanying drawings. The present invention may, however, be embodied in many different forms and should not be construed as being limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. And is provided to fully convey the scope of the invention to those skilled in the art, and the present invention is defined by the claims. It is to be understood that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. In the present specification, the singular form includes plural forms unless otherwise specified in the specification. It is noted that " comprises, " or "comprising," as used herein, means the presence or absence of one or more other components, steps, operations, and / Do not exclude the addition.
Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings. In the drawings, the same reference numerals are given to the same or similar components, and in the following description of the present invention, Detailed explanations of the detailed description will be omitted when the gist of the present invention can be obscured.
1 is a structural diagram of a final correct answer determining apparatus using convolution in an artificial neural network according to an embodiment of the present invention.
1, an
Hereinafter, the operation of the final correct
The
The
In FIG. 2, sub-regions having three element values can be consistently generated. For example, in FIG. 2, the first to
The
Here, the artificial neural network can be configured in various ways. For example, the linear combination of the element values of the sub-region input vector may be performed by stacking or repeating several layers of artificial neural networks that yield results through various activation functions such as sigmoid function or tanh function recurrent network).
In addition, the
The
At this time, the
The selecting
For example, in order to select a final correct answer from the
As described above, according to the embodiment of the present invention, by using the convolution method in the final correct answer determination process of the artificial neural network, it is possible to learn a wide range of inputs by using a separate neural network having a low dimension so that data sparseness do.
In addition, according to the embodiment of the present invention, an artificial neural network is constructed for sub-regions of input in the step of finally determining the correct answer in the artificial neural network, the correct answer candidates are directly determined at the output of each neural network, By determining the final candidate (final correct answer) from the candidates, it is possible to make error correction in a way that generates one correct answer from the entire input, and even if the correct answer decision from the local patterns is not correct, The artificial neural network can mitigate the difficulty of learning the values that are not the maximum value in the probability vector for determining the correct answer, thereby estimating the best correct answer prediction result through the patterns that can be observed even in the case of the prediction failure .
In addition, according to the embodiment of the present invention, it is possible to learn not only the performance improvement in the final layer but also the most abstract patterns independently appearing in position, thereby reducing the number of stacking required or reducing data, Can improve performance.
FIG. 3 is a flowchart illustrating a final correct answer determination method using convolution in an artificial neural network according to an embodiment of the present invention.
Hereinafter, unless otherwise noted, it is assumed that the apparatus is operated by the final correct
First, a vector having an arbitrary dimension composed of discrete or continuous values is input (S310). For example, the input vector may be a six-dimensional vector (first to
In operation S320, the input vector is divided into sub-regions of a predetermined size and generated (S320). Here, the predetermined size may be preset by the developer or user of the artificial neural network in advance. In FIG. 2, it can be divided into
Thereafter, the artificial
Here, the artificial neural network can be configured in various ways. For example, the linear combination of the element values of the sub-region input vector may be performed by stacking or repeating several layers of artificial neural networks that yield results through various activation functions such as sigmoid function or tanh function recurrent network).
At this time, the dimension of the output vector for each sub-region and the range of the output value can be determined. Here, the dimension may be the number of all classes to be finally determined through the artificial neural network. For example, a neural network that follows the form of an artificial neural network may be formed by inputting the sub-regions generated in step S320 for each element of the output vector and learning for each sub-region.
The
At this time, it is possible to generate (determine) correct answer candidates (final correct answer value for each partial area) by the number of the sub-regions generated in step S320. For example, as four
The final correct answer 51 is selected from the
For example, in order to select the final correct answer from the correct candidates, the frequency with which each candidate candidate appears as a candidate, the probability that the neural network of each sub-region generates a candidate, the prior probability of each sub- Prior probabilities are used to select the final correct answer, or the ensemble model is used to select the final correct answer through widely known classifiers.
As described above, according to the embodiment of the present invention, by using the convolution method in the final correct answer determination process of the artificial neural network, it is possible to learn a wide range of inputs by using a separate neural network having a low dimension so that data sparseness do.
In addition, according to the embodiment of the present invention, an artificial neural network is constructed for sub-regions of input in the step of finally determining the correct answer in the artificial neural network, the correct answer candidates are directly determined at the output of each neural network, By determining the final candidate (final correct answer) from the candidates, it is possible to make error correction in a way that generates one correct answer from the entire input, and even if the correct answer decision from the local patterns is not correct, The artificial neural network can mitigate the difficulty of learning the values that are not the maximum value in the probability vector for determining the correct answer, thereby estimating the best correct answer prediction result through the patterns that can be observed even in the case of the prediction failure .
In addition, according to the embodiment of the present invention, it is possible to learn not only the performance improvement in the final layer but also the most abstract patterns independently appearing in position, thereby reducing the number of stacking required or reducing data, Can improve performance.
While the present invention has been described in connection with what is presently considered to be practical exemplary embodiments, it is to be understood that the invention is not limited to the disclosed embodiments, but, on the contrary, It is to be understood that the invention may be embodied in other specific forms. It is therefore to be understood that the above-described embodiments are illustrative in all aspects and not restrictive. The scope of the present invention is defined by the appended claims rather than the detailed description, and all changes or modifications derived from the scope of the claims and their equivalents should be construed as being included within the scope of the present invention.
100: Final correct decision device
110: input unit 120:
130: learning unit 140:
150:
Claims (1)
Dividing the input vectors into sub-regions of a predetermined size;
Learning an artificial neural network for each of the partial areas;
Determining a plurality of correct answers candidates from the vectors obtained as the learning result;
Selecting a final correct answer from the plurality of correct answers;
And determining a final correct answer using convolution in the artificial neural network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160000880A KR20170081887A (en) | 2016-01-05 | 2016-01-05 | Method of determining the final answer by convolution of the artificial neural networks |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160000880A KR20170081887A (en) | 2016-01-05 | 2016-01-05 | Method of determining the final answer by convolution of the artificial neural networks |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20170081887A true KR20170081887A (en) | 2017-07-13 |
Family
ID=59352681
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020160000880A KR20170081887A (en) | 2016-01-05 | 2016-01-05 | Method of determining the final answer by convolution of the artificial neural networks |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20170081887A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020001329A1 (en) * | 2018-06-28 | 2020-01-02 | 北京金山安全软件有限公司 | Method and device for input prediction |
KR20200002245A (en) * | 2018-06-29 | 2020-01-08 | 포항공과대학교 산학협력단 | Neural network hardware |
KR20200002248A (en) * | 2018-06-29 | 2020-01-08 | 포항공과대학교 산학협력단 | Neural network hardware |
US11430137B2 (en) | 2018-03-30 | 2022-08-30 | Samsung Electronics Co., Ltd. | Electronic device and control method therefor |
-
2016
- 2016-01-05 KR KR1020160000880A patent/KR20170081887A/en unknown
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11430137B2 (en) | 2018-03-30 | 2022-08-30 | Samsung Electronics Co., Ltd. | Electronic device and control method therefor |
WO2020001329A1 (en) * | 2018-06-28 | 2020-01-02 | 北京金山安全软件有限公司 | Method and device for input prediction |
US11409374B2 (en) | 2018-06-28 | 2022-08-09 | Beijing Kingsoft Internet Security Software Co., Ltd. | Method and device for input prediction |
KR20200002245A (en) * | 2018-06-29 | 2020-01-08 | 포항공과대학교 산학협력단 | Neural network hardware |
KR20200002248A (en) * | 2018-06-29 | 2020-01-08 | 포항공과대학교 산학협력단 | Neural network hardware |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107133934B (en) | Image completion method and device | |
CN112639833B (en) | Self-adaptive neural network | |
EP3767544A1 (en) | Training a generative model and a discriminative model | |
KR102190303B1 (en) | deep-learning model learning apparatus based CNN and method therefor | |
EP4266246A1 (en) | Automated defect classification and detection | |
KR20170081887A (en) | Method of determining the final answer by convolution of the artificial neural networks | |
CN105320945A (en) | Image classification method and apparatus | |
JP6258560B2 (en) | Neural network learning apparatus and learning method | |
JP6977886B2 (en) | Machine learning methods, machine learning devices, and machine learning programs | |
Oh et al. | Learn-explain-reinforce: counterfactual reasoning and its guidance to reinforce an Alzheimer's Disease diagnosis model | |
JP2018116364A (en) | Dictionary generation device, evaluation device, dictionary generation method, evaluation method, and program | |
EP3502978A1 (en) | Meta-learning system | |
US20230419170A1 (en) | System and method for efficient machine learning | |
CN110033009A (en) | The method of image data is handled in connection network | |
Little et al. | Sequence-sensitive exemplar and decision-bound accounts of speeded-classification performance in a modified Garner-tasks paradigm | |
Holzinger et al. | KANDINSKYPatterns--An experimental exploration environment for Pattern Analysis and Machine Intelligence | |
Park et al. | Unraveling the arc puzzle: Mimicking human solutions with object-centric decision transformer | |
Alhroob et al. | Fuzzy min-max classifier based on new membership function for pattern classification: a conceptual solution | |
KR102208688B1 (en) | Apparatus and method for developing object analysis model based on data augmentation | |
Nimma et al. | Advancements in Deep Learning Architectures for Image Recognition and Semantic Segmentation. | |
CN109902729B (en) | Behavior prediction method and device based on sequence state evolution | |
JP2019520656A (en) | System and method for recursive cortical network | |
Volna et al. | Enhanced ensemble-based classifier with boosting for pattern recognition | |
KR20210010769A (en) | Method and apparatus for identifying anomaly area of image using encoder | |
JP7324086B2 (en) | Information processing device, information processing method |