CN114913434A - High-resolution remote sensing image change detection method based on global relationship reasoning - Google Patents
High-resolution remote sensing image change detection method based on global relationship reasoning Download PDFInfo
- Publication number
- CN114913434A CN114913434A CN202210622122.9A CN202210622122A CN114913434A CN 114913434 A CN114913434 A CN 114913434A CN 202210622122 A CN202210622122 A CN 202210622122A CN 114913434 A CN114913434 A CN 114913434A
- Authority
- CN
- China
- Prior art keywords
- network
- change detection
- remote sensing
- features
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000008859 change Effects 0.000 title claims abstract description 76
- 238000001514 detection method Methods 0.000 title claims abstract description 76
- 238000000034 method Methods 0.000 claims abstract description 30
- 230000004927 fusion Effects 0.000 claims abstract description 18
- 238000012549 training Methods 0.000 claims description 37
- 230000008569 process Effects 0.000 claims description 20
- 238000012795 verification Methods 0.000 claims description 13
- 230000006870 function Effects 0.000 claims description 9
- 238000007781 pre-processing Methods 0.000 claims description 9
- 238000012360 testing method Methods 0.000 claims description 8
- 239000000203 mixture Substances 0.000 claims description 7
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 6
- 238000011156 evaluation Methods 0.000 claims description 6
- 230000003213 activating effect Effects 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims description 4
- 230000002123 temporal effect Effects 0.000 claims description 4
- 238000000605 extraction Methods 0.000 claims description 3
- 230000003044 adaptive effect Effects 0.000 claims description 2
- 238000011084 recovery Methods 0.000 abstract description 3
- 230000000694 effects Effects 0.000 abstract description 2
- 230000011218 segmentation Effects 0.000 abstract 1
- 230000003287 optical effect Effects 0.000 description 5
- 239000000284 extract Substances 0.000 description 4
- 230000000007 visual effect Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000009499 grossing Methods 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 238000005096 rolling process Methods 0.000 description 2
- 238000004088 simulation Methods 0.000 description 2
- 238000000638 solvent extraction Methods 0.000 description 2
- 235000002566 Capsicum Nutrition 0.000 description 1
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 1
- 239000006002 Pepper Substances 0.000 description 1
- 235000016761 Piper aduncum Nutrition 0.000 description 1
- 235000017804 Piper guineense Nutrition 0.000 description 1
- 244000203593 Piper nigrum Species 0.000 description 1
- 235000008184 Piper nigrum Nutrition 0.000 description 1
- 238000010420 art technique Methods 0.000 description 1
- 230000002238 attenuated effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000007499 fusion processing Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- XULSCZPZVQIMFM-IPZQJPLYSA-N odevixibat Chemical class C12=CC(SC)=C(OCC(=O)N[C@@H](C(=O)N[C@@H](CC)C(O)=O)C=3C=CC(O)=CC=3)C=C2S(=O)(=O)NC(CCCC)(CCCC)CN1C1=CC=CC=C1 XULSCZPZVQIMFM-IPZQJPLYSA-N 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 150000003839 salts Chemical class 0.000 description 1
- 239000002904 solvent Substances 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/13—Satellite images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/467—Encoded features or binary features, e.g. local binary patterns [LBP]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Astronomy & Astrophysics (AREA)
- Remote Sensing (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
A high-resolution remote sensing image change detection method based on global relationship reasoning comprises the following steps: firstly, extracting multi-scale features of a double-time phase remote sensing image through a pre-trained encoder; secondly, a global relationship reasoning module is used for carrying out relationship reasoning between the regions on each scale characteristic; and finally, establishing a multi-scale feature fusion decoder, and generating a final change detection result through a semantic segmentation head. In addition to acquiring local information from the stacked convolutional layers, the invention also fully considers the global semantic information capable of representing the internal relation between the changing objects, and in addition, the method is a network structure of an encoder-decoder, can realize detail information recovery, effectively weakens the interference of background noise and reduces the phenomenon of false detection. The method and the device fully utilize the multi-scale information and the global semantic information of the remote sensing image to generate the division characteristics with the resolution effect, effectively improve the change detection precision and have wide application prospect.
Description
Technical Field
The invention belongs to the technical field of remote sensing image processing, and relates to a high-resolution remote sensing image change detection method based on global relationship reasoning. The invention can be used for detecting the building change in two optical remote sensing images with different time phases, and can be applied to land utilization and coverage, city planning and the like.
Background
The primary objective of the remote sensing image change detection task is to make a decision on the change state of an image pixel or an object by comparing double-temporal or multi-temporal images with time intervals, and can provide an important theoretical basis for upper-layer management and decision of tasks such as urban spatial layout, land utilization detection, water body monitoring, disaster monitoring and the like. In recent years, with the rapid development of satellite sensor technology, the resolution of remote sensing images is developed from the past dozens of meters and hundreds of meters to the present meter level and sub-meter level, and the remote sensing images show the trend of high resolution. The high-resolution remote sensing images bring more technical challenges to earth surface change detection, the high-resolution remote sensing images contain more earth surface information, the texture details of the ground features are finer, the topological relation among the ground features is more complex, and the uncertainty of a change detection result is greatly increased. In addition, different from the imaging mode of a common optical image, the remote sensing image is obtained from a satellite or an unmanned aerial vehicle at a high altitude overlooking the ground, the process is easily influenced by factors such as time, light, environment and the like, interference such as shape distortion, overexposure, shadow and the like can be caused, and the change detection difficulty is further increased.
With the development of big data and computing resources, the deep learning technology is gradually introduced into the field of change detection analysis. The depth network can learn abstract representation of data layer by layer, the nonlinear learning capability of the depth network is easier to handle complex pattern recognition, and an actual solution is provided for high-resolution remote sensing image change detection. In the image domain, the minimum learning unit of the depth network is 2-dimensional convolution. Depending on the translation invariance of convolution and the characteristic of being capable of smoothing noise, the spatial hierarchical structure of the mode can be learned by stacking multiple layers of convolution, and the extracted multi-layer features contain rich local details. Therefore, many researches and designs various convolutional neural networks at home and abroad to meet the task of detecting the change of the remote sensing image.
The paper "Daudtr, Bertrand LS, Alexandre B. full capacitive Signal Networks for Change Detection [ C ]// IEEEInternationality Conference on image Processing, ICIP 2018, Athens, Greece, October 7-10,2018: 4063-4067" by Rodrigo CayeDaudt et al presents 3 Full Convolutional (FC) network structures, FC-EF, FC-Sim-Conc and FC-Sim-Diff, respectively. The FC-EF comprises an encoder and a decoder which can extract 4 scale features, and double time phase images are spliced and input into a network, belonging to a single-input single-output structure; the FC-Sim-Conc comprises a twin encoder and a decoder, the double-temporal images are respectively input into the twin encoder, are subjected to connection processing on the extracted characteristic channels with the same scale and then are input into the decoder, and the FC-Sim-Conc belongs to a double-input single-output structure; the FC-Sim-Diff also comprises a twin encoder and a decoder, wherein the twin encoder is respectively input into the double-phase images, and the difference value of the extracted same-scale features is calculated and then input into the decoder. Further, Rodrigo CayeDaudt et al, in the paper "DaudR, Bertrand LS, Alexandre B, et al, Multitask learning for large-scale session management detection [ J ]. Computer Vision and Image interpretation, 2019,187:102783," improve FC-EF networks by adding a residual block at the end of each layer of the FC-EF encoder to facilitate the training of the network (i.e., FC-EF-Res).
Chen et al, in the paper "Chen H, Shi Z W.A Spatial-Temporal Attention-Based Method and a New data set for Mobile Sensing image Change Detection [ J ]. Remote Sensing,2020,12(10):1662," propose spatio-Temporal Attention Network (STANet), which extracts visual features Based on the ResNet18 convolutional Network and uses Attention to self-adapt the region of interest of the image of interest. Shi et al in the article "ShiQ, Liu M X, Li S C, et al, ADeeply Supervised Attention method-based Network and an Open orthogonal Image data set for Remote Sensing Change Detection [ J ]. IEEEtransactions on geographic and Remote Sensing,2022,60: 1-16" propose a depth Supervised Attention method-based Network (DSAMNet), which also extracts Image features based on the ResNet18 convolutional Network and uses the spatial Attention and channel Attention mechanisms to generate discriminative features.
However, the depth network formed by stacking convolutional layers is limited to fine detail feature extraction in a smaller area by the limitation of the receptive field of the convolutional kernel. And the context connection among the multi-scale features is not tight, and semantic information is not fully utilized. It is known that in the same scene, semantic attributes of variant objects are related, i.e. variant object bodies are consistent. If semantic information can be well utilized, one or a small number of change objects are determined, and the remaining change objects are further distinguished through semantic relations, so that not only can the detection precision be improved, but also the detection efficiency can be improved. In view of the deficiencies of the prior art, a change detection method that comprehensively utilizes global semantic information and local information is needed.
Disclosure of Invention
The optical remote sensing image has wide scene coverage range and complex background, and has the problems of shadow noise, overexposure, cloud layer shielding and the like caused by an imaging mode. Detecting some local changes in complex pairs of optical remote sensing images remains an open and challenging task at present. The invention provides an effective change detection method by considering the internal relation between the global semantic information and the objects of the image aiming at the problem that a plurality of changed objects are difficult to accurately detect only through local detail information in a high-resolution remote sensing image. The invention provides a high-resolution remote sensing image change detection method based on global relationship reasoning, aiming at the defect that the prior art focuses more on local detail information of an optical remote sensing image and does not fully utilize global semantic information to carry out change detection, and the change detection precision can be improved.
In order to solve the problems, the invention adopts the technical scheme that:
a high-resolution remote sensing image change detection method based on global relationship reasoning is disclosed. The method not only obtains local information from the stacked convolution layer, but also fully considers global semantic information which can represent the internal relation between the changing objects, and particularly adopts a global relation reasoning module to model the semantic relation between different objects or regions on the remote sensing image characteristic diagram. In addition, the method is a network structure of the encoder-decoder, can realize detail information recovery, effectively weakens the interference of background noise, and reduces the phenomenon of false detection. The specific high-resolution remote sensing image change detection method based on global relationship reasoning comprises the following steps:
the first step, sample partitioning and data preprocessing.
1.1) sample partitioning. A pair of different times t covering the same earth's surface 1 ,t 2 The high-resolution remote sensing image and the corresponding actual label are taken as a sample, and the collected sample is divided into a training set, a verification set and a test set;
1.2) data preprocessing. The preprocessing process is divided into image cropping and standardization. Taking a sample as an example, the image and the actual label are clipped to 256 × 256, and the clipped image is normalized channel by channel to obtain the processed t 1 Temporal videoAnd t 2 Temporal videoC, H, W indicates the number of channels, height, and width of the video, respectively.
And secondly, constructing a change detection network architecture based on global relationship reasoning.
2.1) constructing an encoder. Twin encoders built on the basis of the ResNet18 network for extraction respectivelyAndobtaining a multi-scale feature by using a feature map on 4 scalesAnd
2.2) constructing a Global relationship Reasoning module (Global learning, GloRe). Multi-scale features extracted from twin encodersAndeach input was to 4 GloRe. Each GloRe is connected with a rolling block with a kernel of 3 multiplied by 3 in series, and the output of the rolling block is enhancedAnd
2.3) constructing a decoder. The main role of the decoder is the fusion of the multi-scale enhancement features and the generation of difference features, with the following sub-steps:
2.3.1) multiscale enhancement feature fusion. The minimum scale features gradually get close to the upper scale features. Namely, it isAndgeneration and fusion of unitsBlend features of equal sizeAndgeneration and fusion of unitsBlend features of equal sizeBy analogy, finally generate andblend features of equal sizeEach fusion unit has the same structure to fuseAndfor example, the fusion process is described as follows:
1) will be characterized byInputting the data into an up-sampling layer using bilinear interpolation, inputting the data into a convolution layer with a kernel of 3 multiplied by 3, and obtaining characteristics by activating a function ReLU layerNamely, it is
2) Will be characterized byInputting the data into a kernel of 3 × 3 convolutional layer, and activating the function ReLU layer to obtain the characteristicsNamely, it is
3) Merging in a channel cascadeAndand fed to a kernel 1 x 1 convolution block to obtain the fusion characteristicsNamely, it is
2.3.2) generating a difference signature. Fused featuresAndmaking difference, inputting the difference into a kernel to obtain difference characteristic F for a 3 multiplied by 3 convolution block diff I.e. by
2.4) construction of Change detection Head. Is formed by connecting an up-sampling layer of bilinear interpolation and a convolution block in series and is used for generating a change detection graphComprising the following substeps:
2.4.1) differential feature F diff Upsampling (scale factor of 2, bilinear interpolationMethod), get the updated feature F by the volume block (Conv 3 × 3+ BN + ReLU) diff(1) ;
2.4.2) feature F diff(1) Upsampling (scaling factor of 2, bilinear interpolation), by means of a convolution block (Conv 3 × 3+ BN + ReLU), to obtain a feature F diff(2) ;
2.4.3) feature F diff(2) Output through convolutional layer (Conv 1 × 1, output channel 1) and Sigmoid layer
And thirdly, constructing a loss function.
The Loss function is constructed in a combined manner of Dice Loss (Dice Loss) and Binary cross entropy Loss (BCE Loss).
And fourthly, training and verifying the network.
4.1) basic setting. Parameters required to initialize the network training process include iteration round (Epoch), batch size (Batchsize), initial Learning Rate (LR). And setting a learning rate updating strategy, such as linear attenuation, exponential attenuation, fixed step attenuation and the like. Updating the weight of the change detection network by using an Adaptive Moment Estimation optimizer (Adam), and setting a first-order Moment attenuation coefficient beta of the change detection network 1 And second moment attenuation coefficient beta 2 。
4.2) network training. The network one-time training process corresponds to one Epoch and comprises the following sub-steps:
4.2.1) inputting a training sample with the size of Batchsize into a change detection network based on global relationship reasoning to obtain a change detection result;
4.2.2) calculating Dice Loss and BCE Loss, transmitting the sum of the Dice Loss and the BCE Loss to an Adam optimizer to update the weight of the change detection network, and repeating the step 4.2.1) until the training of the network on all training samples is completed, so as to obtain the change detection network trained under Epoch once.
4.3) network authentication. And inputting the verification set to a trained network to obtain a change detection result of the verification set, and calculating an evaluation index of the network according to the change detection result and an actual sample label.
And fifthly, repeating the network training and verifying process.
And finishing the Epoch network training, and selecting the optimal verification result as the final network.
Compared with the prior art, the invention has the beneficial effects that:
(1) from the aspects of practicability and operability of the model, the invention constructs an end-to-end remote sensing image change detection network, the network can acquire local information through an encoder and global semantic information through a global relationship reasoning module, the local information and the semantic information are fully utilized to detect the change of the image, and the problems of detection loss and missing detection of a change object caused by low global modeling efficiency of the conventional convolutional neural network are solved, so that the invention has the advantage of high detection precision.
(2) The detection network constructed by the invention uses decoding characteristics instead of coding characteristics to calculate the change information, and can effectively avoid complex background and imaging interference in the image. And in consideration of the complementary action of the multi-scale features, the multi-scale coding features are fused in the decoder, and can guide the detail recovery of the variable object to resist the problem of scale difference of the variable object, so that the identification of the variable object is more accurate.
(3) The encoder constructed by the invention firstly integrates the multi-scale features of each image and then calculates the difference features, but not integrates the multi-scale features after calculating the difference features of each layer, so that more salt and pepper noises can be effectively prevented from being generated, and the encoder has excellent change detection visual effect.
Drawings
FIG. 1 is a flow chart of a high resolution remote sensing image change detection method based on global relationship reasoning;
FIG. 2 is a block diagram of a high resolution remote sensing image change detection network based on global relationship reasoning;
FIG. 3 is a block diagram of a decoder;
FIG. 4 is a graph of the variation of the training loss of the network;
fig. 5 is a change curve of the evaluation index of the network authentication process F1;
FIG. 6(a) is a true tag of the LEVIR test set;
FIG. 6(b) is a graph showing the results of detecting changes in FC-EF;
FIG. 6(c) is a graph showing the results of detecting a change in FC-EF-Conc;
FIG. 6(d) is a graph showing the results of detecting changes in FC-EF-Diff;
FIG. 6(e) is a graph showing the results of detecting changes in FC-EF-Res;
FIG. 6(f) is a graph of change detection results for STANet;
FIG. 6(g) is a graph showing the results of detecting a change in DSAMNet;
FIG. 6(h) is a graph showing the results of change detection according to the present invention.
Detailed Description
In order to make the process problems solved by the present invention, the process schemes adopted and the process effects achieved more clear, the present invention will be further described in detail with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the invention and are not limiting of the invention. It should be further noted that, for the convenience of description, only some but not all of the relevant elements of the present invention are shown in the drawings.
1. Data and operating environment.
This example is illustrated using a LEVIR change detection public data set containing building changes. The hardware platform of the simulation experiment of this example is: CPU is Intel (R) core (TM) i7-8700, main frequency is 3.2GHz, memory is 16GB, GPU is NVIDIA GTX 1070, and memory is 8G. The software platform of the simulation experiment of this example is: python 3.8.
2. And (5) carrying out the steps.
(1) Data acquisition and data pre-processing.
(1a) And (6) acquiring data. The LEVIR remote sensing image collects the changes of various buildings (villas or high-rise apartments) in 20 areas of a plurality of cities in Texas USA in the years 2002-2018. The data set contains 637 pairs of remote sensing images with the resolution of 0.5 m/pixel and corresponding real labels, and has 3 channels of red, green and blue, and the size of 1024 × 1024. The data set was calculated as 7: 1: the scale division of 2 is training set, verification set and test set, respectively including 445, 64 and 128 pairs of remote sensing images.
(1b) And (4) preprocessing data. The preprocessing process is divided into image cropping and standardization. Due to the limitation of the GPU, the experiment cannot be trained directly on the whole image, so all samples are cropped to 256 × 256 size and the cropped image is normalized channel by channel according to equation (1).
Wherein,is t 1 、t 2 One training sample taken at a time of day,calculating the average value of all images under red, green and blue channels at the moment, and calculating the LEVIR data set channel by channel, includingThe variance of all images under red, green and blue channels at the time is calculated
(2) And inputting the LEVIR training data set into a change detection network based on global relationship reasoning for forward propagation.
FIG. 2 is a high-resolution remote sensing image change detection network structure based on global relationship reasoning, provided by the invention, for a pair of preprocessed LEVIR imagesAndfor example, the forward propagation process of the network is as follows:
(2a) the multi-scale features are obtained by an encoder. To speed up training, ResNet18 pre-trained on ImageNet datasets (containing over 120 million images, for a total of 1000 classes) was used for encoding. ResNet18 extracts separatelyAndfeatures on 4 scales, i.e.Andwherein,and
(2b) and acquiring the multi-scale enhanced features through a global relationship reasoning module (GloRe). The essence of GloRe is that local areas on the feature map are treated as nodes, and the relationship between the nodes represents the relationship between the local areas on the feature map. The present example sets the number of nodes (N) for 4 glores to 128, 64, 32, and 16, respectively. Multi-scale features extracted by twin encoders And 3X 3 rolled blocks (Co) by GloRe and corresponding kernel, respectivelynv 3 × 3+ BN + ReLU), output enhancement featuresAnd
(2c) obtaining a difference feature F by a decoder diff . Referring to fig. 3, the decoding process is as follows:
s2c.1 multi-scale enhancement features are obtained through a fusion unit andblend features of equal sizeThe minimum scale features gradually get close to the upper scale features. In particular, the amount of the solvent to be used,andthrough a fusion unit FF 1 Generating featuresAndthrough a fusion unit FF 2 Generating featuresAndthrough a fusion unit FF 3 Generating featuresFusion unit FF 1 、FF 2 And FF 3 Has the advantages ofThe same structure is adopted.
S2c.2 fusion featureAnddifference of (2)Input to kernel for 3 × 3 convolution block to obtain difference features
Wherein L is a loss function, L Dice Is DiceLoss, L BCE Are BCELoss. DiceLoss addresses the problem of class imbalance, and expression is shown in equation (3).
In the formula,and Y ij Respectively representing the predicted value and the real value of the index position of i rows and j columns. ε is a smoothing constant, preventing the denominator to be 0, and is set to 10 in this example -5 . BCELoss is used for calculating the matching degree between the detection result of the predicted change and the real label, and the expression is shown as a formula (4).
(4) And (5) network iterative training and verification.
(4a) And setting parameters. The network training process sets Epoch to 50, Batchsize to 16, and LR to 0.001. The learning rate updating strategy adopts linear attenuation, specifically, the learning rates of the first 25 epochs are kept unchanged, and the learning rates of the last 25 epochs are linearly attenuated to 0. Parameter β of Adam optimizer 1 =0.9,β 2 =0.999。
(4b) And (5) network iterative training. The network one-time training process corresponds to one Epoch and comprises the following sub-steps:
s4b.1, inputting a training sample with the size of Batchsize into a change detection network based on global relationship reasoning for forward propagation to obtain a change detection result;
s4b.2 calculates corresponding loss, updates the network weight by using an Adam optimizer, repeats the step S4b.1 until the training of the network on all training samples is completed, and obtains a change detection network trained under one Epoch.
(4c) And (5) network authentication. And inputting the verification set to the trained network to obtain a change detection result of the verification set, and calculating an evaluation index F1 score (formula (6)) according to the change detection result and the actual sample label. And repeating the network training and verifying process to finish the Epoch network training. From fig. 4, where the training loss decays rapidly and tends to stabilize, and fig. 5, where the example validation set F1 evaluates, the 50 th round of training networks may be determined to be the optimal network.
(5) And (5) testing the network. And inputting the test set into the optimal network, and calculating the following evaluation indexes.
Accuracy (Acc) measures the degree of closeness of the predicted value and the true value, i.e. the proportion of correctly predicted (un) changed categories, and the expression is shown in formula (5).
Wherein, TP, TN, FP and FN are respectively a true positive case, a true negative case, a false positive case and a false negative case.
The F1 score can be regarded as a harmonic mean of the precision and recall, expressed as shown in equation (6).
The Mean Intersection over Union (MIoU) is the ratio of the Intersection and Union of the two sets of the actual value and the predicted value, and the expression is shown in formula (7).
The Kappa Coefficient (KC) measures the consistency of the model prediction result and the actual class, and is different from Acc in that the Kappa Coefficient can score the bias of a sample with unbalanced class, and the expression is shown in a formula (8).
(6) And (5) comparing and analyzing results.
Comparative analyses were performed using 6 prior art techniques and the present invention, all techniques performing in the LEVIR test set as shown in table 1. The invention represents the remote sensing image change detection method based on the global relationship reasoning, and the FC-EF, the FC-Sim-Conc, the FC-Sim-Diff and the FC-EF-Res represent 4 full-convolution change detection networks proposed by Daudt and the like, the STATNet represents a space-time attention network proposed by Chen and the like, and the DSAMNet represents a deep supervision attention measurement network proposed by Shi and the like.
Table 1 performance evaluation table of the present invention and the existing remote sensing image change detection model
The combination of table 1 shows that the invention obtains excellent detection results, and all indexes are improved higher than those of other technologies, which proves that the invention can obtain higher change detection precision. In addition, fig. 6 shows the visual results of the present invention and the prior art on one of the test samples, and it can be seen that the present invention can obtain a cleaner visual effect, and the changed objects have clearer edges and better internal consistency.
Finally, it should be noted that: the above examples are only intended to represent embodiments of the present invention, it is understood that the examples are only intended to illustrate the present invention and not to limit the scope of the present invention, after reading the present invention, it will be apparent to those skilled in the art that several variations and modifications can be made without departing from the spirit of the invention, and various equivalent modifications of the present invention fall within the scope of the invention defined in the appended claims.
Claims (3)
1. A high-resolution remote sensing image change detection method based on global relationship reasoning is characterized by comprising the following steps:
firstly, dividing a sample and preprocessing data;
1.1) dividing samples;
1.2) data preprocessing to obtain processed t 1 Temporal imagesAnd t 2 Temporal videoC, H, W denotes the number of channels, height and width of the image;
secondly, constructing a change detection network architecture based on global relationship reasoning;
2.1) constructing an encoder; twin encoders built on the basis of the ResNet18 network for extraction respectivelyAndobtaining a multi-scale feature by using a feature map on 4 scalesAnd
2.2) constructing a global relationship reasoning module GloRe; multi-scale features to be extracted by twin encodersAnd(*∈{t 1 ,t 2 }), input to 4 GloRe, respectively; each GloRe is connected with a convolution block with a kernel of 3 multiplied by 3 in series to output enhanced featuresAnd
2.3) constructing a decoder for fusion of the multi-scale enhancement features and generating difference features, having the following sub-steps:
2.3.1) multi-scale enhanced feature fusion; gradually closing the minimum scale feature to the upper-level scale feature; namely, it isAndgeneration and fusion of unitsBlend features of equal sizeAndgeneration and fusion of unitsBlend features of equal sizeBy analogy, finally generate andblend features of equal sizeEach fusion unit has the same structure to fuseAndto illustrate, fuseThe process is described as follows:
1) will be characterized byInputting the data into an up-sampling layer using bilinear interpolation, inputting the data into a convolution layer with a kernel of 3 multiplied by 3, and obtaining characteristics by activating a function ReLU layerNamely, it is
2) Will be characterized byInputting the data into a kernel of 3 × 3 convolutional layer, and activating the function ReLU layer to obtain the characteristicsNamely, it is
3) Merging in a channel cascadeAndand fed to a kernel 1 x 1 convolution block to obtain the fusion characteristicsNamely that
2.3.2) generating difference characteristics; fused featuresAndmaking difference, inputting the difference into a kernel to obtain difference characteristic F for a 3 x 3 convolution block diff I.e. by
2.4) constructing a change detection Head; is formed by connecting an up-sampling layer with bilinear interpolation and a convolution block in series and is used for generating a change detection graphComprising the following substeps:
2.4.1) Difference characteristic F diff Upsampling to obtain updated features F by convolution block diff(1) ;
2.4.2) feature F diff(1) Upsampling to obtain feature F by convolution block diff(2) ;
Thirdly, constructing a loss function;
constructing a Loss function in a combined mode of Dice Loss and binary cross entropy Loss BCE Loss;
fourthly, network training and verification;
4.1) basic setting, including initializing parameters required by the network training process, setting a learning rate updating strategy, updating the weight of the change detection network by using an adaptive moment estimation optimizer Adam, and setting a first moment attenuation coefficient beta thereof 1 And second moment attenuation coefficient beta 2 ;
4.2) network training; the network one-time training process corresponds to one Epoch and comprises the following sub-steps:
4.2.1) inputting a training sample with the size of Batchsize into a change detection network based on global relationship reasoning to obtain a change detection result;
4.2.2) calculating Dice Loss and BCE Loss, transmitting the sum of the Dice Loss and the BCE Loss to an Adam optimizer, updating the weight of the change detection network, and repeating the step 4.2.1) until the training of the network on all training samples is completed to obtain the change detection network trained under Epoch once;
4.3) network authentication; inputting the verification set to a trained network to obtain a change detection result of the verification set, and calculating an evaluation index of the network according to the change detection result and an actual sample label;
fifthly, repeating the network training and verifying process;
and finishing the Epoch network training, and selecting the optimal verification result as the final network.
2. The method for detecting the change of the high-resolution remote sensing image based on the global relationship reasoning according to claim 1, wherein in the step 1.1), the sample division specifically comprises the following steps: a pair of different times t covering the same earth's surface 1 ,t 2 And (4) taking the high-resolution remote sensing image and the corresponding actual label as a sample, and dividing the collected sample into a training set, a verification set and a test set.
3. The method for detecting the change of the high-resolution remote sensing image based on the global relationship reasoning according to claim 1, wherein in the step 1.2), the data preprocessing process comprises image cutting and standardization.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210622122.9A CN114913434B (en) | 2022-06-02 | 2022-06-02 | High-resolution remote sensing image change detection method based on global relation reasoning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210622122.9A CN114913434B (en) | 2022-06-02 | 2022-06-02 | High-resolution remote sensing image change detection method based on global relation reasoning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114913434A true CN114913434A (en) | 2022-08-16 |
CN114913434B CN114913434B (en) | 2024-06-11 |
Family
ID=82771619
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210622122.9A Active CN114913434B (en) | 2022-06-02 | 2022-06-02 | High-resolution remote sensing image change detection method based on global relation reasoning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114913434B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115526886A (en) * | 2022-10-26 | 2022-12-27 | 中国铁路设计集团有限公司 | Optical satellite image pixel level change detection method based on multi-scale feature fusion |
CN117319610A (en) * | 2023-11-28 | 2023-12-29 | 松立控股集团股份有限公司 | Smart city road monitoring method based on high-order panoramic camera region enhancement |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109063569A (en) * | 2018-07-04 | 2018-12-21 | 北京航空航天大学 | A kind of semantic class change detecting method based on remote sensing image |
AU2020103901A4 (en) * | 2020-12-04 | 2021-02-11 | Chongqing Normal University | Image Semantic Segmentation Method Based on Deep Full Convolutional Network and Conditional Random Field |
CN113420662A (en) * | 2021-06-23 | 2021-09-21 | 西安电子科技大学 | Remote sensing image change detection method based on twin multi-scale difference feature fusion |
CN113706482A (en) * | 2021-08-16 | 2021-11-26 | 武汉大学 | High-resolution remote sensing image change detection method |
CN113936217A (en) * | 2021-10-25 | 2022-01-14 | 华中师范大学 | Priori semantic knowledge guided high-resolution remote sensing image weakly supervised building change detection method |
-
2022
- 2022-06-02 CN CN202210622122.9A patent/CN114913434B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109063569A (en) * | 2018-07-04 | 2018-12-21 | 北京航空航天大学 | A kind of semantic class change detecting method based on remote sensing image |
AU2020103901A4 (en) * | 2020-12-04 | 2021-02-11 | Chongqing Normal University | Image Semantic Segmentation Method Based on Deep Full Convolutional Network and Conditional Random Field |
CN113420662A (en) * | 2021-06-23 | 2021-09-21 | 西安电子科技大学 | Remote sensing image change detection method based on twin multi-scale difference feature fusion |
CN113706482A (en) * | 2021-08-16 | 2021-11-26 | 武汉大学 | High-resolution remote sensing image change detection method |
CN113936217A (en) * | 2021-10-25 | 2022-01-14 | 华中师范大学 | Priori semantic knowledge guided high-resolution remote sensing image weakly supervised building change detection method |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115526886A (en) * | 2022-10-26 | 2022-12-27 | 中国铁路设计集团有限公司 | Optical satellite image pixel level change detection method based on multi-scale feature fusion |
CN115526886B (en) * | 2022-10-26 | 2023-05-26 | 中国铁路设计集团有限公司 | Optical satellite image pixel level change detection method based on multi-scale feature fusion |
CN117319610A (en) * | 2023-11-28 | 2023-12-29 | 松立控股集团股份有限公司 | Smart city road monitoring method based on high-order panoramic camera region enhancement |
CN117319610B (en) * | 2023-11-28 | 2024-01-30 | 松立控股集团股份有限公司 | Smart city road monitoring method based on high-order panoramic camera region enhancement |
Also Published As
Publication number | Publication date |
---|---|
CN114913434B (en) | 2024-06-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110705457B (en) | Remote sensing image building change detection method | |
CN111539316B (en) | High-resolution remote sensing image change detection method based on dual-attention twin network | |
CN112668494A (en) | Small sample change detection method based on multi-scale feature extraction | |
CN109871875B (en) | Building change detection method based on deep learning | |
Su et al. | DLA-Net: Learning dual local attention features for semantic segmentation of large-scale building facade point clouds | |
CN114296075B (en) | Ground penetrating radar image artificial intelligence identification method and device | |
Chen et al. | Exchange means change: An unsupervised single-temporal change detection framework based on intra-and inter-image patch exchange | |
CN103714148B (en) | SAR image search method based on sparse coding classification | |
CN111339827A (en) | SAR image change detection method based on multi-region convolutional neural network | |
CN111008644B (en) | Ecological change monitoring method based on local dynamic energy function FCN-CRF model | |
Xu et al. | Feature-based constraint deep CNN method for mapping rainfall-induced landslides in remote regions with mountainous terrain: An application to Brazil | |
CN109635726B (en) | Landslide identification method based on combination of symmetric deep network and multi-scale pooling | |
CN114913434A (en) | High-resolution remote sensing image change detection method based on global relationship reasoning | |
CN114241208A (en) | Local and global depth contrast feature learning network construction method for inter-image change detection | |
Li et al. | Incorporating open source data for Bayesian classification of urban land use from VHR stereo images | |
Zhang et al. | CDMamba: Remote Sensing Image Change Detection with Mamba | |
Seydi et al. | BDD-Net+: A building damage detection framework based on modified coat-net | |
CN113988147A (en) | Multi-label classification method and device for remote sensing image scene based on graph network, and multi-label retrieval method and device | |
Liu et al. | Diverse hyperspectral remote sensing image synthesis with diffusion models | |
Wang et al. | CF-GCN: Graph Convolutional Network for Change Detection in Remote Sensing Images | |
CN118661207A (en) | Method for large-scale near real-time flood detection in geographical areas covering urban and rural areas and related computer program product | |
CN117911879B (en) | SAM-fused fine-granularity high-resolution remote sensing image change detection method | |
CN111275680B (en) | SAR image change detection method based on Gabor convolution network | |
Albalooshi et al. | Deep belief active contours (DBAC) with its application to oil spill segmentation from remotely sensed sea surface imagery | |
CN115456957B (en) | Method for detecting change of remote sensing image by full-scale feature aggregation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |