WO2020119169A1 - 计算机可读存储介质、输入数据检查方法以及计算装置 - Google Patents
计算机可读存储介质、输入数据检查方法以及计算装置 Download PDFInfo
- Publication number
- WO2020119169A1 WO2020119169A1 PCT/CN2019/102013 CN2019102013W WO2020119169A1 WO 2020119169 A1 WO2020119169 A1 WO 2020119169A1 CN 2019102013 W CN2019102013 W CN 2019102013W WO 2020119169 A1 WO2020119169 A1 WO 2020119169A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- target
- input data
- computer
- classification model
- Prior art date
Links
- 238000003860 storage Methods 0.000 title claims abstract description 36
- 238000000034 method Methods 0.000 title claims description 76
- 238000013145 classification model Methods 0.000 claims abstract description 112
- 238000004590 computer program Methods 0.000 claims abstract description 11
- 238000007689 inspection Methods 0.000 claims description 131
- 230000005856 abnormality Effects 0.000 claims description 113
- 230000002159 abnormal effect Effects 0.000 claims description 33
- 230000006870 function Effects 0.000 claims description 31
- 238000001514 detection method Methods 0.000 claims description 22
- 230000008569 process Effects 0.000 claims description 20
- 238000013528 artificial neural network Methods 0.000 description 72
- 238000004891 communication Methods 0.000 description 31
- 238000010586 diagram Methods 0.000 description 25
- 238000012545 processing Methods 0.000 description 19
- 238000013527 convolutional neural network Methods 0.000 description 18
- 238000005516 engineering process Methods 0.000 description 8
- 230000003287 optical effect Effects 0.000 description 8
- 238000004422 calculation algorithm Methods 0.000 description 7
- 238000004364 calculation method Methods 0.000 description 6
- 210000002569 neuron Anatomy 0.000 description 6
- 238000005476 soldering Methods 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 5
- 230000002950 deficient Effects 0.000 description 5
- 238000013461 design Methods 0.000 description 5
- 238000005070 sampling Methods 0.000 description 5
- 239000004065 semiconductor Substances 0.000 description 5
- 239000000758 substrate Substances 0.000 description 5
- 230000007547 defect Effects 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 241000282326 Felis catus Species 0.000 description 3
- 230000007423 decrease Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000013507 mapping Methods 0.000 description 3
- 230000002776 aggregation Effects 0.000 description 2
- 238000004220 aggregation Methods 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 230000000670 limiting effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 239000002245 particle Substances 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000000306 recurrent effect Effects 0.000 description 2
- 230000002441 reversible effect Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 238000003466 welding Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000011248 coating agent Substances 0.000 description 1
- 238000000576 coating method Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003278 mimic effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000036961 partial effect Effects 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 230000002829 reductive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
Definitions
- the present disclosure relates to data processing using a computing device, and more particularly, to data processing of artificial intelligence embodied by using a computing device.
- neural network that models human biological nerve cells.
- Neural networks use algorithms that mimic the learning capabilities held by humans.
- the neural network performs the mapping between the input mode and the output mode by learning. Also, based on the learning results, the neural network has the generalization ability to generate relatively correct output for input patterns that are not used for learning.
- the present disclosure is for providing a method of processing data using a computing device.
- An embodiment of the present disclosure for solving the problem discloses a computer-readable storage medium that stores a computer program.
- the work includes: a work for deriving the results of the first check of the input data; The task of classifying the input data by the classification model learned by the learning data with cluster information; and the work of outputting the final inspection result based on at least a part of the classification result of the input data of the classification model.
- the classification model is learned using a learning data set that contains multiple subsets of learning data. It is possible to classify similar data into the same cluster in the learning data contained in the learning data set, and classify non-similar The data is classified into different clusters to learn the classification model.
- the work of deriving the first inspection result of the input data may include processing the input data using a pre-learned inspection model containing more than one network function, thereby, the input data Perform anomaly detection or perform anomaly detection on the input data based on the comparison between the input data and the reference data.
- the classification model may be learned using a learning data set containing a subset of learning data, the learning data set containing learning data labeled with different cluster information.
- the classification model may be learned using a learning data set that includes a subset of learning data, the learning data set including target data, target similar data, and target non-similar data.
- the target data and the target similar data may be data marked with first cluster information, and the target non-similar data may be data marked with second cluster information.
- the target similar data may be data containing types of abnormalities similar to the abnormalities contained in the target data, and the target is not similar
- the data may be data that does not contain abnormalities.
- the target similar data may be an image cropped in a manner that includes at least a part of the abnormality included in the target data, and the target is not Similar data can be images that do not contain anomalies.
- the target data may be an image containing the abnormality at the center of the image.
- the target non-similar data may be an image in which at least a part of the part other than the abnormality of the target data is repeated.
- the target data and the target similar data may be classified into the same cluster, and the target non-similar data may be classified into a cluster different from the cluster to which the target data and the target similar data belong. Way to learn the classification model.
- the target data includes an image related to an abnormal condition that can occur separated from the object of the object
- the target similar data may include an image having at least a part of the target data
- the target data and the target similar data may be data containing false positives.
- outputting the final inspection result based on at least a part of the classification result of the input data of the classification model may include when the classification model classifies the input data as belonging to a specific cluster , Generate the inspection result matching the specific cluster as the second inspection result, and output the final inspection result based on the second inspection result, and, in the case where the classification of the input data fails, convert The output of the first inspection result as a final inspection result.
- the second inspection result may include information about at least one of whether there is an abnormality, the location of the abnormality, the type of the abnormality, and the type of misdetection.
- classifying the input data by using the classification model learned by the learning data marked with cluster information may include: using the pre-learned classification model to process the input data, so that all The work of mapping the features of the input data to the solution space of the pre-learned classification model; and based on the position of the input data on the solution space and based on whether the input data belongs to the solution space One of the more than one clusters to classify the work of the input data.
- the solution space may be composed of more than one dimension, and may contain more than one cluster, and each cluster is based on the characteristics of each target data and the position of the features based on similar data of the target on the solution space. constitute.
- An embodiment of the present disclosure for solving the problem discloses an input data checking method executed in a computing device including more than one processor.
- the method includes the steps of deriving the first inspection result of the input data; the step of classifying the input data by using a classification model learned by the learning data marked with cluster information; and the step of classifying the classification model The step of outputting the final inspection result based on at least a part of the classification results of the input data.
- the computing device includes: more than one processor; and a memory for storing instructions that can be executed in the processor, the processor performs the following tasks: exporting the first inspection result of the input data; by using Classifying the input data by a classification model learned by the learning data marked with cluster information; and outputting a final inspection result based on at least a part of the classification results of the input data of the classification model.
- the present disclosure may provide a method of processing data using a computing device.
- FIG. 1 is a block diagram illustrating a computing device that performs an input data checking method according to an embodiment of the present disclosure.
- FIG. 2a is a diagram illustrating a network function according to an embodiment of the present disclosure.
- FIG. 2b is a schematic diagram illustrating a convolutional neural network according to an embodiment of the present disclosure.
- 3a, 3b, 3c, 3d, 3e, and 3f are diagrams showing learning data according to an embodiment of the present disclosure.
- FIG. 4 is a diagram illustrating a method of training a classification model according to an embodiment of the present disclosure.
- FIG. 5 is a diagram showing the solution space of the classification model according to an embodiment of the present disclosure.
- FIG. 6 is a flowchart of a method for checking input data inspection according to an embodiment of the present disclosure.
- FIG. 7 is a block configuration diagram showing a unit for embodying a method of input data inspection according to an embodiment of the present disclosure.
- FIG. 8 is a block configuration diagram showing a module for embodying an input data checking method according to an embodiment of the present disclosure.
- FIG. 9 is a block diagram showing logic for embodying an input data checking method according to an embodiment of the present disclosure.
- FIG. 10 is a block configuration diagram showing a circuit for embodying an input data checking method according to an embodiment of the present disclosure.
- FIG. 11 is a brief and general diagram showing an exemplary computing environment that can be embodied by an embodiment of the present disclosure.
- component refers to computer-related entities, hardware, firmware, software, software, and a combination of hardware or execution of software.
- the component is a procedure executed on a processor, a processor, an object, an execution thread, a program, and/or a computer, but it is not limited thereto.
- both the application executed on the computing device and the computing device can be components. More than one component may reside in the processor and/or thread of execution.
- One component can be logicalized in one computer.
- a component can be distributed between more than two computers. And, such components can be executed from various computer-readable media having various data structures stored internally.
- a component is based on a signal with more than one data packet (for example, in a logic system, a decentralized system, through data and/or signals from a component interacting with other components, transmitted through a network such as other systems and the Internet Data), to communicate via local and/or remote processing.
- a signal with more than one data packet for example, in a logic system, a decentralized system, through data and/or signals from a component interacting with other components, transmitted through a network such as other systems and the Internet Data
- the terms “include” and/or “include” mean the presence of corresponding features and/or structural elements. However, the terms “include” and/or “include” do not exclude the presence or addition of more than one other feature, structural element, and/or combination. In addition, when there is no special definition or the context does not clearly indicate the singular form, the singular number usually means “one or more” in the scope of the specification and the invention.
- network functions and artificial neural networks and neural networks can be exchanged with each other.
- FIG. 1 is a block diagram illustrating a computing device that performs an input data checking method according to an embodiment of the present disclosure.
- the structure of the computing device 100 shown in FIG. 1 is an illustration briefly shown. In an embodiment of the present disclosure, the computing device 100 may include other structures for performing the embodiments of the present disclosure.
- the computing device 100 may include a processor 110, a memory 120, a communication module 130, and a camera module 140.
- the processor 110 may be composed of more than one core, and may include a central processing unit (CPU) of a computing device, a general graphics processing unit (GPGPU, general purpose graphics processing unit), and a tensor processing unit (TPU, tensor processing unit) ) And other data analysis, deep learning processors.
- the processor 110 reads the computer program stored in the memory 120 to execute the input data checking method according to an embodiment of the present disclosure.
- the processor 110 may perform calculations for neural network learning.
- the processor 110 may perform input data processing for deep learning (DN, deep learning), feature extraction from input data, error calculation, and weighting using backpropagation neural network Calculation of neural network learning such as value update.
- At least one of the CPU, GPGPU, and TPU of the processor 110 may be in the learning of network functions.
- the CPU and GPGPU work together to learn network functions and use the network functions to classify data.
- the processors of a plurality of computing devices are used together to learn network functions and classify data using the network functions.
- the computer program executed by the computing device according to an embodiment of the present disclosure may be a CPU, GPGPU, or TPU executable program.
- the computing device 100 utilizes at least one of CPU, GPGPU, and TPU to decentralize and process network functions. Moreover, in an embodiment of the present disclosure, the computing device 100 decentralizes network functions together with other computing devices for processing.
- the image processed by the network function may be an image stored in the storage medium of the computing device 100, an image captured by the camera module 140 of the computing device 100, and/or a communication module 130, from the image Images transmitted by other computing devices such as databases.
- the image processed by the network function may be an image stored in a computer-readable storage medium (for example, may include flash memory, etc., but the present disclosure is not limited thereto).
- the computing device 100 receives an image file stored in a computer-readable storage medium through an input-output interface (not shown).
- the computing device 100 receives image data and uses pre-learned network functions to check the input data.
- the input data check of the present disclosure may include anomaly detection for determining whether there is an abnormality in the input data.
- the input data inspection of the present disclosure may include misdetection judgment related to abnormality detection and abnormality detection.
- the memory 120 may store a computer program for executing the input data checking method according to an embodiment of the present disclosure.
- the stored computer program is read and driven by the processor 110.
- the communication module 130 transmits and receives an abnormality determination method for performing image data according to an embodiment of the present disclosure to other computing devices, servers, and the like.
- the communication module 130 transmits and receives data for image data and other embodiments of the present disclosure to other computing devices, servers, and the like.
- the communication module 130 receives learning image data in a learning image database or the like.
- the communication module 130 may enable communication between a plurality of computing devices, thereby performing learning dispersion of network functions in each of the plurality of computing devices.
- the communication module 130 enables communication between a plurality of computing devices, so that data classification using network functions can be distributed.
- the camera module 140 captures an inspection object to generate image data in order to execute the image data abnormality determination method according to an embodiment of the present disclosure.
- the computing device 100 may include more than one camera module 140.
- FIG. 2a is a diagram illustrating a network function 200 according to an embodiment of the present disclosure.
- a neural network may be composed of a collection of interconnected computing units generally referred to as "nodes”. Such “nodes” may also be called “neurons”.
- the neuron includes at least one or more nodes.
- the nodes (or neurons) that make up the neural network are connected by more than one "connection unit”.
- connection unit In the neural network, more than one node connected by the connection unit forms the relationship between the input node and the output node.
- the concept of input node and output node is relativity.
- any node that has an output node relationship has a relationship with an input node in the relationship with other nodes, and vice versa.
- the relationship between the input node and the output node is generated with the connection unit as the center.
- At one input node more than one output node is connected through the connection unit, and vice versa.
- the output node determines the value based on the data input to the input node.
- the nodes connecting the input node and the output node with each other may have a weight.
- the weighting value can be changed, and the neural network can be changed by the user or algorithm in order to perform the required function. For example, in the case where one output node is connected to more than one input node through each connection unit, the output node has the value input to the input node connected to the output node and the value corresponding to each input node
- the weighted value set by the connection unit determines the output node value based on the weighted value.
- a neural network more than one node is connected to each other through more than one connection unit, thereby forming an input node and an output node relationship in the neural network.
- the characteristics of the neural network are set according to the number of nodes and connection units and the connection relationship between the nodes and connection units, and the weighted value restored to the connection unit. For example, there are the same number of nodes and connection units, and when there are two neural networks with different weighting values between the connection units, the two neural networks are different.
- the neural network may include more than one node.
- a part of the nodes constituting the neural network form a layer based on the distance from the initial input node.
- the set of nodes with distance n may constitute n layers.
- the distance from the initial input node is determined by the minimum number of connection units that need to be traversed in order to reach the corresponding node from the initial input node.
- the definition of this layer is used to illustrate that within the neural network, the difference in layers can be defined by a method different from that described above.
- the layer of a node can be defined by the distance of the final node.
- the input node is a node in the neural network.
- the input node In the relationship with other nodes, more than one node that directly inputs data without going through the connection unit.
- the nodes of other input nodes connected through the connection unit are not blocked.
- the final output node is one of the nodes in the neural network, and in the relationship with other nodes, there is more than one node that does not have an output node.
- the hidden node is a node that does not constitute the neural network for the first input node and the last output node.
- the number of nodes in the input layer is the same as the number of nodes in the output layer.
- the number of nodes in the input layer is greater than the number of nodes in the output layer, and in the input layer, the number of nodes decreases along hidden nodes.
- the number of nodes in the input layer is greater than the number of nodes in the output layer. In the input layer, the number of nodes increases along hidden nodes.
- the neural network of another embodiment of the present disclosure is a neural network of a combined form of neural networks.
- a deep neural network is a neural network that includes multiple hidden layers in addition to the input layer and the output layer. If the deep neural network is used, the latent structures of the data can be grasped. That is, the underlying structure of photos, texts, videos, voices, and music can be grasped (for example, which objects exist in the photos, what are the contents and feelings of the text, and what are the contents and feelings of the speech, etc.).
- Deep neural networks may include convolutional neural networks (CNN, convolutional neural networks), recurrent neural networks (RNN, recurrent neural networks), autoencoders (autoencoders), generative adversarial networks (GAN, Generative Adversarial Networks), restricted Boltzmann machine (RBM, restricted boltzmann machine), deep belief network (DBN, deep Belief network), Q network, U network, Siam network, etc.
- CNN convolutional neural networks
- RNN recurrent neural networks
- autoencoders autoencoders
- GAN Generative Adversarial Networks
- restricted Boltzmann machine RBM, restricted boltzmann machine
- DBN deep Belief network
- Q network U network
- Siam network etc.
- the description of the deep neural network is an example, and the present disclosure is not limited thereto.
- the network function 200 may further include an auto encoder.
- the automatic encoder may be one type of artificial neural network for outputting output data similar to the input data.
- the autoencoder may include at least one hidden layer, and a singular hidden layer may be disposed between the input and output layers.
- the number of nodes in each layer is reduced to the middle layer called the bottleneck layer (coding) in the number of nodes in the input layer, and then in the bottleneck layer, it is expanded symmetrically with the reduction to the output layer (symmetrical to the input layer) .
- the automatic encoder can perform non-linear dimensionality reduction.
- the number of input layers and output layers corresponds to the number of sensors remaining after the preprocessor of the input data.
- the number of nodes of the hidden layer in the encoder decreases in a direction away from the input layer.
- the number of nodes in the bottleneck layer (the layer with the fewest nodes between the encoder and the decoder) is too small, a sufficient amount of information cannot be transferred, and therefore, it can be maintained above a certain number (for example, half of the input layer) Etc.).
- Neural networks can be learned in at least one of supervised learning, unsupervised learning and semi-supervised learning.
- the learning of the neural network is used to minimize the output error.
- the learning process of the neural network repeatedly input the learning data to the neural network, and calculate the output and target errors of the neural network related to the learning data, along the direction used to reduce the errors, remove the errors of the neural network from the output layer of the neural network Backward propagation in the direction of the input layer to update the weighted value of each node of the neural network.
- learning data marked with correct answers in each learning data ie, marked learning data
- unsupervised learning each learning data is not marked with correct answers.
- the learning data in the case of supervised learning related to data classification is data marked with categories in the learning data.
- the labeled learning data is input to the neural network, and the output (category) of the neural network and the label of the learning data are compared to calculate errors.
- the input learning data in the case of unsupervised learning related to data classification is compared with the neural network output, thereby calculating errors. Calculated errors propagate backward in the neural network (that is, in the direction of the output layer to the input layer). With the reverse propagation, the connection weight of each node of each layer of the neural network will be updated. The updated connection weight value of each node may change according to the learning rate.
- the calculation of the neural network related to the input data and the back propagation of errors can constitute a learning loop (epoch).
- the learning rate can be applied differently according to the number of repetitions of the learning cycle of the neural network. For example, at the beginning of the learning of the neural network, a high learning rate is applied.
- the neural network quickly ensures a specified level of performance to improve efficiency, and at the later stage of learning, a low learning rate is used to improve accuracy.
- the learning data is a partial collection of actual data (ie, data processed by the learned neural network), therefore, there is a learning cycle that reduces errors related to the learning data or adds errors to the actual data .
- overfitting over-learns, resulting in an increase in errors related to actual data.
- one type of overfitting is that after seeing a yellow cat, the neural network that learned the cat cannot recognize cats other than the yellow cat. Overfitting increases the error of the machine running the algorithm.
- a variety of optimization methods can be used. In order to prevent overfitting, increase the learning data, or regularization (regularization), omitting the dropout of a part of the network nodes during the learning process, etc.
- the convolutional neural network (CNN, convolutional neural network) shown in FIG. 2b is a kind of deep neural network, including a neural network with a convolutional layer.
- Convolutional neural networks are a type of multilayer receptors designed with minimal preprocess.
- the convolutional neural network can be composed of one or more convolutional layers and an artificial neural network layer combined with this, and the weighted value and pooling layer are additionally used. Because of this structure, the convolutional neural network can fully use the input data of the two-dimensional structure.
- Convolutional neural networks are used to identify targets in images.
- the convolutional neural network can divide the image data into rows and columns with dimensions to be processed.
- image data encoded in red-green-blue (RGB, red-green-blue)
- RGB red-green-blue
- the image data can be divided into three two-dimensional rows and columns (three-dimensional data array).
- the convolution filter is moved to multiply the convolution filter and the row and column components at various positions of the image to perform the convolution process (input and output of the convolution array).
- the convolution filter may be in the form of n*n rows and columns, and generally, it may be composed of a filter of a fixed form that is smaller than the number of overall pixels of the image. That is, when an m*m image is input to a convolution layer (for example, a convolution layer with a size of n*n of a convolution filter), the rows and convolutions of n*n pixels including each pixel of the image are represented The filter components are multiplied (ie, the components of the rows and columns are multiplied).
- the components matching the convolution filter can be extracted from the image.
- the 3*3 convolution filter used to extract the upper and lower straight line components is composed of [[0, 1, 0], [0, 1, 0], [0, 1, 0]].
- the convolution filter is applied to the input image, and the upper and lower straight line components matching the convolution filter are extracted from the image and output.
- the convolution layer may use a convolution filter in each row and column related to each channel (that is, in the case of R, G, and B coating images, R, G, and B colors).
- the convolutional layer uses a convolution filter on the input image to extract features in the input image that match the convolution filter.
- the filter value of the convolution filter (that is, the value of each component of the row and column) is updated by inverse propagation during the learning process of the convolutional neural network.
- the output of the convolutional layer is connected to the sub-sampling layer, thereby simplifying the output of the convolutional layer to reduce memory usage and calculation values. For example, when the output of the convolutional layer is input to the collection layer having the 2*2 maximum collection filter, each pixel of the image outputs the maximum value included in each patch to compress the image for each 2*2.
- the aggregate output minimum value of the patch, or the average value of the output patch, any aggregation method may be included in the present disclosure.
- the convolutional neural network may include more than one convolutional layer and sub-sampling layer.
- the convolutional neural network repeatedly performs a convolution process and a sub-sampling process (for example, the maximum aggregation, etc.) to extract features from the image. Through repeated convolution and sub-sampling processes, the neural network can extract the global features of the image.
- the output of the convolutional layer or sub-sampling layer can be input to a fully connected layer.
- a fully connected layer is a layer connected to all neurons of a layer adjacent to all neurons of one layer.
- a fully connected layer is a structure in a neural network where all nodes of each layer are connected to all nodes of other layers.
- the neural network may include a deconvolutional neural network (DCNN).
- the deconvolution neural network performs similar work to calculate the convolutional neural network in the reverse direction, and outputs the features extracted from the convolutional neural network as features related to the original data.
- the processor 110 may derive the first inspection result of the input data.
- the first inspection result may include abnormal information related to the input data, and the abnormal information may include arbitrary information related to the abnormality existing in the input data.
- the first inspection result may include the presence or absence of an abnormality, the location of the abnormality in the input data, and the type of abnormality.
- the first inspection result may include information about whether the input data is abnormal, which pixel of the input data is abnormal, and what type of abnormality is included in the input data ( For example, in the case of images related to semiconductors, poor soldering, defective wiring, defective components, etc.).
- the processor 110 may derive the first check result related to whether there is an abnormality in the input data.
- the processor 110 processes the input data using a pre-learned inspection model containing more than one network function, thereby performing an abnormality inspection related to the input data to derive the first inspection result.
- the inspection model is a semi-supervised learning model that judges whether there is an unlearned new pattern (ie, abnormality) in the input data by learning only from normal data, thereby determining whether there is an abnormality in the input data.
- the inspection model is learned from the marked normal data and abnormal data, so that it is possible to determine whether the input data contains abnormal supervised learning models. Checking the model may include any machine learning model that determines whether there is an abnormality in the input data.
- the processor 110 performs abnormality detection related to the input data based on the comparison of the input data and the reference data to derive the first inspection result.
- the processor 110 determines whether there is an abnormality in the input data based on the comparison of the reference data containing the abnormality or the reference data not containing the abnormality and the input data.
- the processor 110 is based on an arbitrary machine vision (machine vision) solution to judge the abnormality in the input data.
- the processor 110 classifies the input data using the classification model learned using the learning data marked with cluster data.
- the classification model uses learning data sets that contain multiple subsets of learning data to learn.
- the learning data contained in the learning data set classifies similar data into the same cluster and classifies dissimilar data into different clusters.
- the learning data subset may contain multiple learning data.
- the learning data subset may contain multiple learning data, and each learning data is marked with cluster information.
- the learning data is a subset of data used in one iteration in the learning of the network function.
- the learning data set is the entire data used for looping in the learning of the network function.
- the learning data subset may contain dissimilar learning data, and the dissimilar learning data contained in the learning data subset may be classified into different clusters through a classification model.
- the learning data subset includes a plurality of learning data, each of the learning data is labeled with cluster information, and the learning data subset may further include learning data labeled with different cluster information.
- the learning data subset may include target data, target similar data, and target non-similar data, and may be labeled with cluster information such as target data and target similar data, and target non-similar data may be marked with different data from target data and target similar data.
- Cluster information The target data and target similar data contained in the learning data subset may be marked with first cluster information, and the target non-similar data may be marked with second cluster information.
- the target similar data is data containing abnormalities
- the target non-similar data is data containing no abnormalities.
- the target non-similar data is data in which at least part of the part other than the abnormality is repeated in the target data.
- the target data is an image
- the target non-similar data is an image of at least a part of the part other than the abnormal part in the image containing the abnormality.
- the target similar data is data including an abnormality similar to the abnormality included in the target data.
- the target similar data is cropped in such a manner as to include at least a part of the abnormality included in the target data.
- the abnormality of the similar type included in the target data may include the abnormality at the same position on the object of the inspection object of the abnormality contained in the target data, and the same type of defect as the object of the inspection object of the abnormality included in the target data (For example, when the object to be inspected is a circuit board and the abnormality included in the target data is poor soldering, the abnormality included in the target similar data is also poor soldering).
- the target similar data may include at least a part of the image of the target data, and the pixel part corresponding to the abnormality in the image of the target data is the data that must be included.
- Target non-similar data are images that do not contain abnormalities.
- the target data is an image that contains abnormality in the center of the image.
- the target non-similar data is an image in which at least a part of the part other than the abnormality of the target data is repeated.
- the abnormal part of the image containing the abnormality and the image not containing the abnormality are different, and other parts may be similar, so that in the case of classification by the usual method, it can be classified as a similar image.
- the abnormal part is 5*5 pixels, and the abnormal part is usually classified at 0.25% of the whole, 99.75% of the two images are similar, Therefore, it is classified as similar by the usual classification method.
- target data, target similar data, and target dissimilar data respectively labeled with cluster information according to an embodiment of the present disclosure to learn a classification model even if a part of the image is different, it is classified as Non-similar data.
- the processor 110 classifies the target data and target similar data into the same cluster, and classifies the target non-similar data into clusters that are different from the cluster to which the target data and target similar data belong.
- the abnormalities that the target data can include, in the case of the target data bit image, at least one of the abnormalities related to the inspection target object included in the image and the abnormalities separated from the inspection target object.
- the abnormalities that the target data may include may be abnormalities that can occur in semiconductor products (eg, defects, etc.) and abnormalities that are not related to the object of the inspection object (eg, images Acquire at least one of the problem of the device, the lens foreign object, etc.). That is, from the viewpoint of abnormality detection, the target data and the target similar data may include false positives. More specifically, when there is no abnormality in the inspection object and there is a situation that can occur that is separate from the inspection object, the target data and the target similar data may contain such a false alarm. Therefore, the classification model classifies abnormal types and false positives into clusters.
- the result of the first inspection may contain a false positive.
- the input data is classified for the second time using the classification model, and the processor 110 uses the classification model to classify the result of the input data.
- the second time In the classification it can be judged whether it is an abnormality related to the actual inspection object or an abnormality not related to the inspection object (ie, a false alarm).
- FIGS. 3a to 3 e are diagrams showing learning data according to an embodiment of the present disclosure.
- the images shown in FIGS. 3 a to 3 e are only examples of learning data, and the present disclosure may include arbitrary learning data and arbitrary images.
- FIG. 3a is a diagram showing a target base image 300 that can be used as learning data.
- the target base image 300 itself can be used as the target image for learning the classification model, and the target base image 300 is cropped for the target image 310.
- the target base image 300 may include a defect 301 abnormality of the inspection object.
- the plurality of target similar images 311, 313 may be images cropped in such a manner as to contain the defective 301 portion of the target image 310.
- the plurality of crop lines 310, 311, and 313 in FIG. 3a are crops of the target image 310 and crops 311 and 313 of the target similar image.
- FIG. 3b shows a target image 310 (target image) and target similar images 311, 313 (target similar image) cropped by the cropping line of FIG. 3a.
- the target image 310 and the target similar images 311, 313 both contain the abnormality 301, and may include other parts of the inspection object.
- FIG. 3c is an illustration showing the target non-similar base image 320.
- the target non-similar base image 320 shown in FIG. 3c itself can be used as a target dissimilar image, a part of which is cropped and used as a target non-similar image.
- the target non-similar image 320 may have resolutions such as the target image 310 and the target similar image 311. The resolutions of the target images, target similar images, and target non-similar images are also examples, and the resolutions of these may be the same or different.
- the target non-similar image 320 may be an image that does not contain abnormalities.
- the target non-similar image 320 is an image that does not contain abnormality in a portion 321 corresponding to the target image.
- FIG. 3d is an illustration showing a subset of learning data used to learn the classification model of an embodiment of the present disclosure.
- the learning data subset may include a target image 310 with an anomaly 301, a target similar image 311, and a target non-similar image 320 without an anomaly 321.
- the target image 310 and the target similar image 311 are marked with the same cluster information in a manner of being classified into the same cluster.
- the target non-similar image 320 is marked with cluster information that is different from the cluster information marked on the target image 310 and the target similar image 311.
- the target image 310 and the target similar image 311 may be marked with first cluster information, and the target non-similar image 320 may be labeled with second cluster information.
- the classification model may classify the images with abnormalities and the images without abnormalities into different clusters.
- the abnormal part is 5*5 pixels, and the abnormal part is usually classified at 0.25% of the whole, 99.75% of the two images are similar, Therefore, it is classified as similar by the usual classification method.
- target data, target similar data, and target dissimilar data respectively labeled with cluster information according to an embodiment of the present disclosure to learn a classification model even if a part of the image is different, it is classified as Non-similar data.
- FIG. 3e is an illustrative diagram showing other learning data subsets for learning the classification model of an embodiment of the present disclosure.
- the target image 330 and the target similar image 333 of FIG. 3e may be images containing false positives. More specifically, the abnormality 331 included in the target image 330 and the target similar image 333 is separated from the inspection target object and an abnormality condition (for example, a lens foreign object, etc.) can occur.
- the target image 330 may be an image related to a normal product or an abnormal image that occurs according to an abnormal condition separated from the normal product. According to the conventional abnormality inspection method, when an abnormal image containing such a false alarm is included, it is determined to be abnormal, and an erroneous result is output.
- the classification model is used to generate clusters related to false alarms using images containing false alarms, whereby it is possible to detect whether the inspection object included in the input image to be inspected actually contains abnormalities
- the object to be inspected does not contain anomalies, but for other reasons, the input data is judged to be anomaly detection.
- the target non-similar image 340 is an image that does not contain abnormalities.
- the classification model may classify images with false alarms and images without false alarms into different clusters.
- FIG. 3f is an illustrative diagram showing another learning data subset for learning the classification model of an embodiment of the present disclosure.
- the target image 330 and the target similar image 333 of FIG. 3f may be images containing false positives. More specifically, the abnormality 331 included in the target image 330 and the target similar image 333 is an abnormal condition (for example, a lens foreign object, etc.) that can occur when separated from the inspection target object.
- the target image 330 may be an image related to a normal product or an abnormal image that occurs according to an abnormal condition separated from the normal product. According to the conventional abnormality inspection method, when an abnormal image containing such a false alarm is included, it is determined to be abnormal, and an erroneous result is output.
- the classification model is used to generate clusters related to false alarms using images containing false alarms, whereby it is possible to detect whether the inspection object included in the input image to be inspected actually contains abnormalities
- the object to be inspected does not contain anomalies, but for other reasons, the input data is judged to be anomaly detection.
- the target non-similar image 310 is an image that includes an abnormality 301 (that is, an abnormality related to the inspection target object), and does not include a false alarm.
- the classification model can classify images with false alarms and images with abnormalities in the actual inspection object (ie, images classified as abnormal in the abnormality inspection ) Is classified into different clusters.
- the target image 330 including a false alarm is added according to the type of false alarm.
- the new misdetection is used as a target image including a false alarm to additionally learn the classification model.
- the classification model of the present disclosure can cluster well-known anomaly types and well-known misdetection types, and can determine which type the input data belongs to.
- FIG. 4 is a diagram illustrating a method of training a classification model according to an embodiment of the present disclosure.
- the classification model of the present disclosure clusters similar data on the solution space 400. More specifically, the classification model includes target data 401 and target similar data 402 in one cluster 410, and target non-similar data 403 is included in a different cluster from target data 401 and target similar data 402.
- each cluster has a prescribed distance margin 420.
- the classification model receives a subset of learning data containing target data 401, target similar data 402, and target non-similar data 403 to match each data with the solution space, and updates the included classification in a clustered manner based on the cluster information marked on the solution space
- the weighted value of more than one network function of the model increases the distance in the solution space between the target data 401 and the target similar data 402 and the target non-similar data 403 in such a manner that the distance in the solution space of the target data 401 and the target similar data 402 becomes shorter.
- the classification model uses a triplet-based cost function to learn.
- the difference between the second distance between the input data is at least the distance margin 420, and the method of learning the classification model includes reducing the first distance to the distance margin Steps below the scale.
- the distance margin 420 is always a positive number.
- the weighted value of more than one network function included in the classification model may be updated, and the weighted value update is performed every iteration or one epoch.
- the classification model is learned by a model based on magnet loss considering not only cluster classification of dissimilar data, but also the semantic relationship between data in a cluster or other clusters.
- the initial distance of the center point difference of each cluster is executed during the learning process.
- the position on the solution space of each data is adjusted based on the similarity to the cluster to which each data belongs and the data within and outside the cluster.
- FIG. 5 is a diagram showing the solution space of the classification model according to an embodiment of the present disclosure.
- FIG. 5 is a diagram showing the solution space 400 of the classification model.
- the solution space 400 shown in FIG. 5 is merely an example, and the classification model may include any number of clusters and any number of data of each cluster.
- the data 431, 433, 441, and 443 in the cluster shown in FIG. 5 are similar data.
- the solution space is composed of a space of more than one dimension, and includes more than one cluster, and each cluster is constructed based on the features based on the respective target data and the positions on the solution space based on the features based on the target-like data.
- the first cluster 430 and the second cluster 440 may be clusters related to dissimilar data.
- the third cluster 450 may be a cluster related to data that is not similar to the first and second clusters.
- the distance 445, 435 between clusters is a scale that shows the difference in data belonging to each cluster.
- the twelfth distance 445 between the first cluster 430 and the second cluster 440 is a measure of the difference between the data belonging to the first cluster 430 and the data belonging to the second cluster 440.
- the thirteenth distance 435 between the first cluster 430 and the second cluster 440 is a measure of the difference between the data belonging to the first cluster 430 and the data belonging to the third cluster 450.
- the data belonging to the first cluster 430 is more dissimilar to the data belonging to the second cluster than the data belonging to the third cluster 450. That is, when the distance between clusters is long, the data belonging to each cluster is more dissimilar, and when the distance between clusters is short, the data belonging to each cluster is more similar.
- the distances 435, 445 between several districts are larger than the radius of the cluster by a predetermined ratio.
- the processor 110 uses the classification model to calculate the input data, thereby classifying the input data based on the position where the feature of the input data matches the solution space of the classification model.
- the processor 110 processes the input data using the pre-learned classification model, thereby matching the characteristics of the input data with the solution space of the pre-learned classification model.
- the processor 110 classifies the input data based on the position on the solution space of the input data, based on which of the one or more clusters on the solution space the input data belongs to.
- the processor 110 determines the final inspection result based on at least a part of the classification result related to the input data of the classification model.
- the processor 110 In the case where the split model classifies the input data as belonging to a specific cluster, the processor 110 generates the inspection result matching the specific cluster as the second inspection result, and determines the final inspection result with the second inspection result. In addition, when the classification of the classification model and the input data fails, the processor 110 determines the first inspection result as the final inspection result, where the second inspection result may include the presence or absence of the abnormality, the location of the abnormality, and the abnormality Information about at least one of the type, misdetection, and misdetection type.
- the pre-learned classification model retains clusters in the solution space according to the type of anomaly. Therefore, the processor 110 uses the classification model to generate the second inspection result related to the input data.
- the processor 110 determines whether there is an abnormality in the input data based on which cluster on the solution space of the classification model the feature of the input data belongs to, and if there is an abnormality, it can be determined whether the abnormality is an abnormality on the inspection object Or an abnormality that has nothing to do with the object of inspection (ie, false detection).
- the processor 110 may generate The data exists as a second inspection result that is an abnormality of substrate soldering failure, and the final inspection result is determined based on the first inspection result and the second inspection result.
- the processor 110 may generate a second corresponding to the input data that is a false detection and there is no abnormal normal data
- the results of the second inspection are based on the results of the first inspection and the second inspection to determine the final inspection results.
- the processor 110 determines that the substrate welding defects are related to the input data The final result of anomaly detection.
- the processor 110 may inform the operator of the content.
- the processor 110 marks the corresponding input data in order to inform the operator of the difference between the first inspection result and the second inspection result.
- the processor 110 ensembles the first inspection result and the second inspection result to generate a final inspection result.
- the ensemble method may use a voting method, a method based on the weighted value of each neural network, and may include any other ensemble method.
- the processor 110 determines that the classification using the input data classification model failed, and generates the first inspection result as the final inspection result. For example, in the case where the input data includes a new abnormality related to the inspection target object with a defective circuit wire or an abnormality related to the inspection target object with a broken lens, the processor 110 (that is, the learned classification model has an unlearned new pattern, etc. ), the input data cannot be classified using the classification model, therefore, the final inspection result related to the corresponding input data is generated based on the first inspection result. In this case, the processor 110 may inform the operator of the content. In order to inform the operator of the content, the processor 110 needs to mark the corresponding input data.
- the processor 110 may generate the corresponding input data as new learning data.
- the processor 110 uses the corresponding input data as the target image, and can generate the corresponding The other part of the anomaly of the input data is used as a new learning data subset with similar data as the target.
- the processor 110 uses the newly generated subset of the learning data to generate clusters on the solution space of the classification model related to the new abnormal pattern and adds the learning classification model.
- the abnormality of the input data is determined by the first and second inspections, and the input data includes abnormalities related to abnormal conditions that occur through factors other than the object of the inspection object (ie, false positives) ), the inspection results can be derived, not for false detections.
- the classification model is learned again by using it, whereby the predetermined inspection performance corresponding to the newly generated misdetection and abnormality can be maintained.
- the first inspection and the second inspection are separated, so that adding a new type of abnormality or a new type of misdetection to the abnormality detection solution does not affect the classification performance related to the previous data. That is, in the conventional solution, in the case of repeated false detections, the solution itself is modified to solve the problem of corresponding false detections, and thus other erroneous operations may occur. In the abnormality detection of an embodiment of the present disclosure, it may be broken Open the possibility of additional error work.
- FIG. 6 is a flowchart of a method for checking input data inspection according to an embodiment of the present disclosure.
- the computing device 100 may derive the first inspection result of the input data (step 610).
- the input data may include image data
- the computing device 100 may determine the first inspection result related to the captured image data or the image data transmitted from an external device.
- the result of the first inspection may contain abnormal information related to the input data.
- the computing device 100 classifies the input data by using a classification model learned using the learning data labeled with cluster information (step S630).
- the classification model uses a learning data set that contains multiple subsets of learning data to learn. In the learning data contained in the learning data set, similar data is classified into the same cluster, and non-similar data is classified into other clusters.
- the computing device 100 outputs a final inspection result based on at least a part of the classification result of the input data of the classification model (step 650).
- the classification model classifies the input data as belonging to a specific cluster
- the computing device 100 generates the inspection result matching the specific cluster as the second inspection result, and determines the final result based on at least a part of the second inspection result test result.
- the computing device 100 determines the first inspection result as the final inspection result.
- FIG. 7 is a block configuration diagram showing a unit for embodying a method of input data inspection according to an embodiment of the present disclosure.
- a method for checking input data may be embodied by the following unit.
- a method for inspecting input data may be embodied by the following units: a unit 710 that derives the first inspection result of the input data; a classification model learned by using learning data marked with cluster information A unit 730 to classify the input data; and a unit 750 to output a final inspection result based on at least a part of the classification result of the input data of the classification model.
- the unit 710 for deriving the first inspection result of the input data may include using a pre-learned inspection model containing more than one network function to The input data is processed to thereby perform abnormality detection on the input data, or a unit that performs abnormality check related to the input data based on the comparison of the input data and reference data.
- a unit 750 for outputting the final inspection result based on at least a part of the classification result of the input data of the classification model may be included in the When the classification model classifies the input data as belonging to a specific cluster, the inspection result matching the specific cluster is generated as the second inspection result, and the final inspection result is output based on the second inspection result, and, When the classification of the input data fails, the first inspection result is output as a unit of the final inspection result.
- the unit 730 for classifying the input data by using a classification model learned using the learning data marked with cluster information may include: The learned classification model processes the input data, thereby mapping the features of the input data to units of the solution space of the pre-learned classification model; and using the input data on the solution space Based on the location, the units of the input data are classified based on whether the input data belongs to one of more than one cluster on the solution space.
- FIG. 8 is a block configuration diagram showing a module for embodying an input data checking method according to an embodiment of the present disclosure.
- a method for checking input data may be embodied by the following modules.
- a method for inspecting input data may be embodied by the following modules: a module 810 for deriving the first inspection result of the input data; a classification model learned by using learning data marked with cluster information A module 830 to classify the input data; and a module 850 to output a final inspection result based on at least a part of the classification result of the input data of the classification model.
- FIG. 9 is a block diagram showing logic for embodying an input data checking method according to an embodiment of the present disclosure.
- a method for checking input data may be embodied by the following logic.
- the method for checking the input data may be embodied by the following logic: Logic 910 that derives the first check result of the input data; by using the classification model learned by the learning data marked with cluster information Logic 930 to classify the input data; and logic 950 to output a final inspection result based on at least a portion of the classification result of the input data of the classification model.
- FIG. 10 is a block configuration diagram showing a circuit for embodying an input data checking method according to an embodiment of the present disclosure.
- a method for checking input data may be embodied by the following unit.
- the method for checking the input data may be embodied by the following circuit: a circuit 1010 that derives the first check result of the input data; a classification model learned by using learning data marked with cluster information A circuit 1030 to classify the input data; and a circuit 1050 to output a final inspection result based on at least a part of the classification result of the input data of the classification model.
- FIG. 11 is a brief and general diagram showing an exemplary computing environment that can be embodied by an embodiment of the present disclosure.
- program modules include routines, programs, components, data structures, etc. that perform specific tasks or embody specific abstract data types.
- the method of the present disclosure is known not only in single-processor or multi-processor computer systems, minicomputers, and mainframe computers, but also in personal computers, handheld computing devices, and microprocessor-based Implement other computer systems including devices such as devices, programmable home appliances, and others (these are connected to more than one related device to work).
- the embodiments described in the present disclosure may be implemented in a decentralized computing environment executed by a remote processing device connected through a communication network through any task.
- program modules can be located in both local and remote storage devices.
- Computers generally include a variety of computer-readable media. Through the computer, accessible media can become computer-readable media, such computer-readable media include volatile and non-volatile media, transitory (transitory) and non-transitory (non-transitory) media, mobile And non-removable media.
- the computer-readable media may include computer-readable storage media and computer-readable transmission media.
- Computer-readable storage media includes volatile and non-volatile media media, transitory and non-transitory media embodied by any method or technology that stores information such as computer-readable instructions, data structures, program modules, or other data, Removable and non-removable media.
- Computer storage media may include RAM, ROM, EEPROM, flash memory or other storage technologies, CD-ROM, digital video disk (DVD, digital) or other optical disk storage devices, magnetic tape, magnetic tape, magnetic disk storage devices or other magnetic storage devices or Any other medium that accesses and stores required information through a computer, but is not limited to this.
- Computer-readable transmission media generally include a medium that embodies computer-readable instructions, data structures, program modules, or other data in a modulated data signal such as a carrier wave or other transmission mechanism and transmits all information. Modulated data signal terminology sets or changes more than one signal in the characteristics of the signal in such a manner as to encode information within the signal.
- computer-readable transmission media include wired media such as a wired network or direct-wired connection, audio, RF, infrared, and other wireless media such as other wireless media. Any combination of such media is also included within the scope of computer-readable transmission media.
- An exemplary environment 1100 embodying multiple aspects of the present disclosure including a computer 1102 is presented, which includes a processing device 1104, a system memory 1106, and a system bus 1108.
- the system bus 1108 connects system components including the system memory 1106 (not limited thereto) to the processing device 1104.
- the processing device 1104 may be any processor among various commonly used processors. Dual-processor and other multi-processor architectures can also be utilized as the processing device 1104.
- the system bus 1108 may be one of several types of bus structures that are connected to each other using an arbitrary local bus in a memory bus, a peripheral device bus, and various common bus architectures.
- the system memory 1106 includes a read-only memory 2110 and a random access memory 2112.
- the basic input/output system (BIOS) is stored in non-volatile memory 2110 such as RROM, EPROM, EEPROM, etc., and includes a basic history of transferring information between structural elements in the computer 1102 when the input/output system is started.
- the random access memory 2112 may include a high-speed random access memory such as a static random access memory for caching data.
- the computer 1102 includes an internal hard disk drive 2114 (HDD) (for example, EIDE, SATA, and the internal hard disk drive 2114 is also used as an external type in a suitable chassis (not shown)), and a magnetic floppy disk drive 2116 (FDD) (for example, Read from the removable floppy disk 2118 or use it to record thereon) and the optical disk drive 1120 (for example, read the CD-RON disk 1122 or read from other high-capacity optical media such as DVD or use it to record thereon).
- the hard disk drive 2114, the magnetic disk drive 2116, and the optical disk drive 1120 may be connected to the system bus 1108 through a hard disk drive interface 1124, a magnetic disk drive interface 1126, and an optical disk drive interface 1128, respectively.
- the interface 1124 for external drive embodiment includes at least one or two of Universal Serial Bus (USB, Universal Serial Bus) and IEEE 1394 interface technologies.
- drives and related computer-readable media provide non-volatile storage of data, data structures, computer-executable instructions, and others.
- the drive and the medium store arbitrary data in an appropriate digital form.
- the description of the computer-readable medium refers to a hard disk drive, a removable magnetic disk, and a portable optical medium such as a CD or DVD, as long as it is a person of ordinary skill in the technical field to which this disclosure belongs, through a zip drive, a magnetic tape,
- Other types of media read by computers, such as flash memory cards, cassette tapes, and others, are also used in the exemplary operating environment, and any such media may contain computer-executable instructions for performing the present disclosure.
- Multiple program modules including the operating system 2130, one or more application programs 2132, other program modules 2134, and program data 2136 may be stored in the drive and the random access memory 2112. All or part of the operating system, applications, modules, and/or data is also cached in the random access memory 2112.
- the present disclosure may be embodied as an operating system or a combination of operating systems available in various commercial aspects.
- the user inputs commands and information to the computer 1102 through more than one wired or wireless input device, such as a pointing device such as a keyboard 2138 and a mouse 1140.
- a pointing device such as a keyboard 2138 and a mouse 1140.
- Other input devices are microphone, IR remote control, joystick, gamepad, stylus, touch screen, etc.
- These and other input devices are connected to the processing device 1104 through an input device interface 1142 connected to the system bus 1108, and connected through other interfaces such as a parallel port, an IEEE 1394 serial port, a game port, a USB port, an IR interface, etc. .
- the display 1144 or other type of display device is also connected to the system bus 1108 through an interface such as a video adapter 1146 or the like. Attached to the display 1144, the computer usually includes speakers, printers, other peripheral output devices (not shown).
- the computer 1102 operates in a networked environment using the logical connection of more than one remote computer, such as a remote computer 1148 via wired and/or wireless communication.
- the remote computer 1148 may be a workstation, a computing device computer, a router, a personal computer, a portable computer, a microprocessor-based entertainment device, a peer device, or other common network nodes.
- the computer 1102 includes many of the described structural elements. For simplicity, only the storage device 1150 is shown.
- the illustrated logical connection includes a short-range communication network 1152 (LAN) and/or a larger network, for example, a wired or wireless connection of a long-range communication network 1154 (WAN).
- LAN short-range communication network
- WAN long-range communication network
- Such short-distance communication and long-distance communication network environments are generally used in offices and companies to make enterprise-wide computer networks such as intranets simple, which are connected to computer networks all over the world, for example, networks.
- the computer 1102 When used in a short-range communication network environment, the computer 1102 is connected to the local network 1152 through a wired and/or wireless communication network interface or adapter 1156.
- the adapter 1156 can be wired or wireless for short-range communication 1152.
- the short-range communication 1152 includes a wireless access point provided for communication with the wireless adapter 1156.
- the computer 1102 may include a modem 1158, connected to a communication computing device on the long-distance communication 1154, or set other units of communication through the long-distance communication 1154 through a network or the like.
- the modem 1158 which may be an internal or external type and a wired or wireless device, is connected to the system bus 1108 through a serial interface 1142.
- the described program module or a part thereof is stored in the remote memory/storage device 1150.
- the illustrated network connection is an exemplary embodiment, and other units for setting a communication connection between computers may also be used.
- the computer 1102 performs any wireless device or individual configured and operated by wireless communication, for example, a printer, a scanner, a desktop computer and/or a portable computer, a personal data assistant (PDA, portable data), a communication satellite, and a wireless detection mark Related to any device or place and telephone communication work.
- PDA personal data assistant
- Wireless fidelity (Wi-Fi, Wireless, Fidelity) can be connected through the network even if it is not wired.
- Wireless fidelity is such a device, for example, the wireless technology of a telephone in which a computer can send and receive data anywhere in the conversation circle of a base station, that is, a computer.
- the wireless fidelity network is reliable and uses IEEE 802.11 (a, b, g, other) wireless technologies to provide high-speed wireless connections.
- IEEE 802.11 a, b, g, other
- wireless fidelity is used in order to connect the computer to the network and the wired network (using IEEE 802.3 or Ethernet).
- the wireless fidelity network operates in the unlicensed 2.4 and 5 GHz radio frequency bands, for example, at a data rate of 11 Mbps (802.11a) or 54 Mbps (802.11b), or in products that include dual bands.
- the various embodiments disclosed herein are embodied by methods, devices, or articles using standard transformation and/or functional technologies.
- article of manufacture includes a computer program, carrier, or media accessed from any computer-readable device.
- computer-readable media include magnetic storage devices (eg, hard disks, floppy disks, magnetic strips, etc.), optical disks (eg, CDs, DVDs, etc.), smart cards, and flash memory devices (eg, EEPROMs, cards, strips, key drives, etc.), But it is not limited to this.
- the various storage media disclosed herein include more than one device and/or other machine-readable media for storing information.
- the specific order or hierarchical structure of the steps in the disclosed process is an example of illustrative proximity. Based on the limited order of design, the specific order or hierarchical structure of the steps of processes within the scope of the present disclosure may be rearranged again.
- the additional method invention claims that the elements of the multiple steps are provided in the order of the samples, and is not limited to the specific order or hierarchical structure disclosed.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (19)
- 一种计算机可读存储介质,存储有计算机程序,其特征在于,所述计算机程序在计算装置的一个以上的处理器中执行的情况下,执行用于对输入数据进行检查的以下的工作,所述工作包括:导出对输入数据的第一次检查结果的工作;通过利用标记有集群信息的学习数据学习的分类模型来对所述输入数据进行分类的工作;以及以对所述分类模型的所述输入数据的分类结果的至少一部分为基础来输出最终检查结果的工作。
- 根据权利要求1所述的计算机可读存储介质,其特征在于,所述分类模型利用包含多个学习数据子集的学习数据集来学习,以在包含在所述学习数据集中的学习数据中将类似的数据分类为相同集群,且将非类似的数据分类为不同集群的方式学习所述分类模型。
- 根据权利要求1所述的计算机可读存储介质,其特征在于,导出对所述输入数据的第一次检查结果的工作,包括利用包含一个以上的网络函数的预先学习的检查模型来对所述输入数据进行处理,由此,对所述输入数据执行异常检测或者以所述输入数据与参考数据的比较为基础来对所述输入数据执行异常检测的工作。
- 根据权利要求1所述的计算机可读存储介质,其特征在于,所述分类模型利用包含学习数据子集的学习数据集来学习,所述学习数据集包含标记有不同集群信息的学习数据。
- 根据权利要求1所述的计算机可读存储介质,其特征在于,所述分类模型利用包含学习数据子集的学习数据集来学习,所述学习数据集包含目标数据、目标类似数据及目标非类似数据。
- 根据权利要求5所述的计算机可读存储介质,其特征在于,所述目标数据和所述目标类似数据为标记有第一集群信息的数据,而且,目标非类似数据为标记有第二集群信息的数据。
- 根据权利要求5所述的计算机可读存储介质,其特征在于,在所述目标数据为包含异常的数据的情况下,所述目标类似数据为包含与所述目标数据所包含的异常类似的类型的异常的数据,而且,所述目标非类似数据为不包含异常的数据。
- 根据权利要求5所述的计算机可读存储介质,其特征在于,在所述目标数据为包含异常的图像的情况下,所述目标类似数据为以包含所述目标数据所包含的异常的至少一部分的方式裁剪的图像,而且,所述目标非类似数据为不包含异常的图像。
- 根据权利要求8所述的计算机可读存储介质,其特征在于,所述目标数据为在图像的中心部包含所述异常的图像。
- 根据权利要求8所述的计算机可读存储介质,其特征在于,所述目标非类似数据为除所述目标数据的异常之外的部分的至少一部分重复的图像。
- 根据权利要求5所述的计算机可读存储介质,其特征在于,以将所述目标数据和所述目标类似数据分类为相同集群,且将所述目标非类似数据分类为与所述目标数据及所述目标类似数据所属的集群不相同的集群的方式学习所述分类模型。
- 根据权利要求5所述的计算机可读存储介质,其特征在于,所述目标数据包含与能够与对象客体分离来发生的异常状况有关的图像,所述目标类似数据包含具有所述目标数据的至少一部分的图像。
- 根据权利要求5所述的计算机可读存储介质,其特征在于,所述目标数据及所述目标类似数据为包含误报的数据。
- 根据权利要求1所述的计算机可读存储介质,其特征在于,以对所述分类模型的所述输入数据的分类结果的至少一部分为基础来输出最终检查结果的工作包括在所述分类模型将所述输入数据分类为属于特定集群的情况下,将与所述特定集群匹配的检查结果生成为第二次检查结果,并以第二次检查结果为基础来输出最终检查结果,而且,在对所述输入数据的分类失败的情况下,将所述第一次检查结果作为最终检查结果来输出的工作。
- 根据权利要求14所述的计算机可读存储介质,其特征在于,所述第二次检查结果包含与是否存在异常、异常的位置、异常的类型及误检类型中的至少一种有关的信息。
- 根据权利要求1所述的计算机可读存储介质,其特征在于,通过利用标记有集群信息的学习数据学习的分类模型来对所述输入数据进行分类的工作包括:利用所述预先学习的分类模型来对所述输入数据进行处理,从而将所述输入数据的特征映射到所述预先学习的分类模型的解空间的工作;以及以所述解空间上的所述输入数据的位置为基础,并基于所述输入数据是否属于 所述解空间上的一个以上的集群中的一个集群来分类所述输入数据的工作。
- 根据权利要求16所述的计算机可读存储介质,其特征在于,所述解空间由一维以上的空间构成,包含一个以上的集群,而且,各个集群以基于各个目标数据的特征及以基于目标类似数据的特征在解空间上的位置为基础构成。
- 一种输入数据检查方法,在包括一个以上的处理器的计算装置中执行,其特征在于,包括:导出对输入数据的第一次检查结果的步骤;通过利用标记有集群信息的学习数据学习的分类模型来对所述输入数据进行分类的步骤;以及以对所述分类模型的所述输入数据的分类结果的至少一部分为基础来输出最终检查结果的步骤。
- 一种计算装置,其特征在于,包括:一个以上的处理器;以及存储器,用于存储能够在所述处理器中执行的指令,所述处理器执行如下工作:导出对输入数据的第一次检查结果;通过利用标记有集群信息的学习数据学习的分类模型来对所述输入数据进行分类;以及以对所述分类模型的所述输入数据的分类结果的至少一部分为基础来输出最终检查结果。
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811528293.5A CN111325224A (zh) | 2018-12-13 | 2018-12-13 | 计算机可读存储介质、输入数据检查方法以及计算装置 |
CN201811528293.5 | 2018-12-13 | ||
KR1020190002307A KR102246085B1 (ko) | 2018-12-13 | 2019-01-08 | 어노말리 디텍션 |
KR10-2019-0002307 | 2019-01-08 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020119169A1 true WO2020119169A1 (zh) | 2020-06-18 |
Family
ID=71075937
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/102013 WO2020119169A1 (zh) | 2018-12-13 | 2019-08-22 | 计算机可读存储介质、输入数据检查方法以及计算装置 |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2020119169A1 (zh) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110295778A1 (en) * | 2010-05-27 | 2011-12-01 | Sony Corporation | Information processing apparatus, information processing method, and program |
CN106294762A (zh) * | 2016-08-11 | 2017-01-04 | 齐鲁工业大学 | 一种基于学习的实体识别方法 |
CN107884475A (zh) * | 2017-10-18 | 2018-04-06 | 常州大学 | 一种基于深度学习神经网络的城市燃气管道故障诊断方法 |
CN108027911A (zh) * | 2015-07-29 | 2018-05-11 | 伊利诺斯工具制品有限公司 | 促进焊接即服务软件的系统及方法 |
CN108985214A (zh) * | 2018-07-09 | 2018-12-11 | 上海斐讯数据通信技术有限公司 | 图像数据的标注方法和装置 |
-
2019
- 2019-08-22 WO PCT/CN2019/102013 patent/WO2020119169A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110295778A1 (en) * | 2010-05-27 | 2011-12-01 | Sony Corporation | Information processing apparatus, information processing method, and program |
CN108027911A (zh) * | 2015-07-29 | 2018-05-11 | 伊利诺斯工具制品有限公司 | 促进焊接即服务软件的系统及方法 |
CN106294762A (zh) * | 2016-08-11 | 2017-01-04 | 齐鲁工业大学 | 一种基于学习的实体识别方法 |
CN107884475A (zh) * | 2017-10-18 | 2018-04-06 | 常州大学 | 一种基于深度学习神经网络的城市燃气管道故障诊断方法 |
CN108985214A (zh) * | 2018-07-09 | 2018-12-11 | 上海斐讯数据通信技术有限公司 | 图像数据的标注方法和装置 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7112501B2 (ja) | ディープラーニングに基づく画像比較装置と方法及びコンピューター可読媒体に保存されたコンピュータープログラム | |
KR102246085B1 (ko) | 어노말리 디텍션 | |
CN107430705B (zh) | 用于重新训练分类器的样本选择 | |
US10275719B2 (en) | Hyper-parameter selection for deep convolutional networks | |
KR20200091808A (ko) | 어노말리 디텍션 | |
US11443514B2 (en) | Recognizing minutes-long activities in videos | |
US11769060B2 (en) | Method for determining optimal anomaly detection model for processing input data | |
KR20200000541A (ko) | 인공 신경망의 학습 방법 | |
KR102078133B1 (ko) | 데이터의 어노말리 감지 방법 | |
US20220269718A1 (en) | Method And Apparatus For Tracking Object | |
KR102402194B1 (ko) | 딥러닝 기반 엔드-투-엔드 o-ring 결함 검출 방법 | |
KR20190041961A (ko) | 딥러닝 기반 이미지 비교 장치, 방법 및 컴퓨터 판독가능매체에 저장된 컴퓨터 프로그램 | |
US11625574B2 (en) | Method for generating abnormal data | |
KR20200010984A (ko) | 어노말리 디텍션 | |
KR20220112164A (ko) | 목표 물체 검출 방법 및 장치 | |
KR20210098420A (ko) | 비정상 데이터 생성 방법 | |
KR20220105102A (ko) | 레이더 영상 처리 방법 및 장치 | |
WO2020119169A1 (zh) | 计算机可读存储介质、输入数据检查方法以及计算装置 | |
KR20200019919A (ko) | 데이터의 어노말리 감지 방법 | |
KR102070730B1 (ko) | 이미지 세그먼테이션 방법 | |
KR102260169B1 (ko) | 데이터 생성 방법 | |
KR102283885B1 (ko) | 이미지 정합 방법 | |
KR102718267B1 (ko) | 웨이퍼맵의 패턴에 기반한 불량 장비 검출 방법, 장치 및 프로그램 | |
KR102588998B1 (ko) | 영상에서 이상행동 검출을 위한 딥러닝 모델을 이용하여 태스크를 수행하는 방법 및 장치 | |
KR20230154601A (ko) | 표의 픽셀 정보를 획득하는 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19894961 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19894961 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 03.08.2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19894961 Country of ref document: EP Kind code of ref document: A1 |