Specific embodiment
Multi-source image target association method and technology scheme proposed by the present invention based on improvement dictionary learning includes following step
It is rapid:
Step 1: data set is collected and pretreatment stage, collects plurality of information resources to the picture number of Same Scene or same target
Handmarking's analysis is carried out according to collection, and to multi-source image data set, stores image data set and handmarking as a result, being formed with mark
Sign multi-source image target association raw data set S;
Step 1.1: mainly including visible images, multispectral image, infrared image, SAR image to plurality of information resources image
In two kinds of images collected extensively, and guarantee between the multi-source image for target association to be for Same Scene, comprising same
A kind of target or corresponding target;
Step 1.2: having label multi-source image target association raw data set S={ (yi,yj',aij),i∈(0,N),j∈
(0, N') }, wherein yiAnd yj' indicate that the data of any two kinds of information sources indicate, aij=0 or 1, it is handmarking as a result, aij=0 table
Show yiAnd yj' it is dereferenced target, aij=1 indicates yiAnd yj' it is associated objects, N and N' are respectively that two kinds of multi-source image collection include
Number of elements.
Step 2: unified wordbook generation phase, using the method for improvement dictionary learning to all figures in raw data set S
As data carry out dictionary learning, the generation common wordbook D=[d of multi-source image1,d2,…dk];
Step 2.1: dictionary learning is the sample Y=[y for common dense expression1,…,yn]∈Rn×N, find one it is excessively complete
Standby wordbook D=[d1,…,dK]∈Rn×KFormer dense sample is expressed as linear group of each element in wordbook by (K > n)
It closes, guarantees the coefficient X=[x of linear combination1,…,xN]∈RK×NIt is sparse as far as possible, and the degree of information loss is controlled, to make to learn
Habit task is simplified, and reduces the complexity of model, is seen shown in formula (1);
Step 2.2: to make the sparse coefficient obtained that there is certain classification capacity, dictionary learning is further improved,
On the basis of its objective function, increase a label to limitation, the objective function for improving dictionary learning is obtained, as shown in formula (2);
Wherein, α and β is hyper parameter, for controlling the weight of error representated by items;Section 3 is to have sparse coefficient
The limit entry of classification capacity, when target is to (yi,yj') label be associated objects, i.e. same category when, Mij=1;When target pair
(yi,yj') label be dereferenced target, i.e., when different classes of, Mij=-1;Other situations, such as (yi,yj') it is same target,
That is when i=j, Mij=0, see shown in formula (3);
N=diag { N1,…,NN, it is diagonal matrix, each diagonal element is the sum of each column element in matrix M,L=N-M;;
Step 2.3: objective function (2) has two parameters of D and X, while being optimized for non-convex optimization problem to parameter D and X,
Need to fix parametric solution another parameter, thus the optimization problem of objective function can be converted to following two formula
(4), the iteration optimization of (5):
Wherein, formula (4) is that fixed wordbook D optimizes sparse coefficient X one by one, and formula (5) is fixed sparse system
Several pairs of wordbooks optimize, and the optimization of formula (5) can use K-SVD progress, update the element and right in wordbook one by one
Answer sparse coefficient;To formula (4), due to there is a presence of L1 norm, objective function be not it is continuously differentiable, searched using characteristic symbol
Rope algorithm is corresponded to the sign vector θ of sparse coefficient by grey iterative generation, converted objective function to standard, unconfined
Double optimization problem, therefore, the gradient of calculation formula (4) are shown in that formula (6) is shown,
Formula (6) gradient is enabled to be equal to zero, can obtain sparse coefficient as a result, sees shown in formula (7),
Step 2.3.1;When generating wordbook to formula (5) using K-SVD algorithm, every wheel iteration only updates a dictionary member
Plain dk, keep other elements constant, find a best dkAnd update corresponding sparse coefficient(For the K of coefficient matrix X
Row, the kth element d with wordbookkIt is corresponding), so that formula (5) is reached minimum;
As shown in formula (8), directly to EkCarry out singular value decomposition (SVD) Lai Gengxin dkWithIt is likely to occur sparse coefficientNot sparse phenomenon, i.e., it is updatedBefore updateThe position of nonzero element is inconsistent, therefore, only retains sparse system
NumberIn nonzero value, EkOnly retain dkWithItem after the product of middle non-zero position is formedIt is rightSingular value decomposition is carried out to obtain
It arrives,And it enablesIt is updated to the first row of U, the product conduct of the first row and Δ (1,1) of VUpdated value,
Complete the update of a dictionary element;
Step 2.3.2;For formula (4), characteristic symbol searching algorithm is dilute to generate by iteration one group of active set of update
Sparse coefficient symbol, the element in active set are the number of sparse coefficient, specifically includes the following steps:
Step 2.3.2.1;Initialization, characteristic symbol vectorWherein θiIt is sparse that ∈ { -1,0,1 } respectively indicates correspondence
Coefficient xiSymbol, active set activeset:={ } be empty set;
Step 2.3.2.2;The sign of each non-zero sparse coefficient and the element of active set are set, is selected conducive to target letter
The sparse coefficient number that number (4) reduces is active set element, it may be assumed that
WhenWhen, if θi=-1, activeset:={ i } ∪
activeset;
WhenWhen, if θi=1, activeset:={ i } ∪
activeset;
Step 2.3.2.3;After generating active set, respectively to being selected in wordbook D, sparse coefficient X and characteristic symbol vector θ
Number element in active set corresponds to generating subsetFormula (4) can be converted into be seen without limitation double optimization problem
Shown in formula (9),
It enables the derivative of formula (9) be equal to 0, can obtain optimal
Step 3: the representative learning stage, using the common wordbook D of multi-source image to multi-source image target association initial data
Multi-source image data in collection S carry out unified rarefaction representation, the sparse coefficient collection X of each multi-source image data are generated, with original number
Multi-source image target association is collectively formed according to the label information of concentration and differentiates data set S';
Step 3.1: using the wordbook D that improvement dictionary learning generates in step 2 to multi-source image target association original number
Rarefaction representation is carried out according to the multi-source image data in collection S, using formula (4), fixed wordbook D carries out one by one sparse coefficient X
Optimization, generates the sparse coefficient of each multi-source image data;
Step 3.2: being collectively formed in conjunction with the sparse coefficient of label information and each multi-source image data in raw data set S
Multi-source image target association differentiates data set S'={ (xi,xj',aij), i ∈ (0, N), j ∈ (0, N') }, wherein xiAnd xj' be
The rarefaction representation of the two kinds of information sources generated in 3.1, aij=0 or 1, it marks for the Central Plains S somebody's work as a result, aij=0 indicates xiWith
xj' it is dereferenced target, aij=1 indicates xiAnd xj' it is associated objects.
Step 4: the classification learning stage establishes multi-source image target association discrimination model, utilizes multilayer neural network model
Data and label in data set S', which are learnt, train, to be differentiated to multi-source image target association, neural network association is obtained and sentences
Other model;
Step 4.1: the network structure of multilayer neural network model uses three layers of full connection structure, and input layer is two any more
The L2 distance of source image data sparse coefficient is shown in that wherein n is characterized dimension shown in formula (10), for 256 dimensions, the second node layer
Number is 512, and third layer node number is 256, and in addition to output layer, each interlayer is all made of dropout strategy to avoid mistake
Fitting;
Step 4.2: defeated since the problem of whether being associated with two targets is converted into binary classification discrimination
Node layer number is set as 1 out, and the activation primitive of the output layer of the multilayer neural network of use uses sigmoid function, sees public affairs
Shown in formula (11), export as the number of 0~1 range, 1 to represent two image datas associated, 0 represent it is uncorrelated between two image datas
Connection is exported closer to 1, and association probability is bigger, conversely, association probability is smaller;The activation primitive of each middle layer of network structure uses
Relu function is shown in shown in formula (12).
sigmoid(wTX+b)=1/1+exp [- (wTx+b)] (11)