CN115861938B - Unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition - Google Patents
Unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition Download PDFInfo
- Publication number
- CN115861938B CN115861938B CN202310063450.4A CN202310063450A CN115861938B CN 115861938 B CN115861938 B CN 115861938B CN 202310063450 A CN202310063450 A CN 202310063450A CN 115861938 B CN115861938 B CN 115861938B
- Authority
- CN
- China
- Prior art keywords
- unmanned aerial
- aerial vehicle
- feature
- image
- feature map
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 28
- 230000004927 fusion Effects 0.000 claims abstract description 89
- 238000001514 detection method Methods 0.000 claims abstract description 64
- 230000011218 segmentation Effects 0.000 claims abstract description 63
- 238000012549 training Methods 0.000 claims abstract description 10
- 238000010586 diagram Methods 0.000 claims description 106
- 238000012544 monitoring process Methods 0.000 claims description 40
- 238000000605 extraction Methods 0.000 claims description 20
- 230000003287 optical effect Effects 0.000 claims description 6
- 230000002829 reductive effect Effects 0.000 claims description 5
- 238000010276 construction Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 4
- 230000002452 interceptive effect Effects 0.000 claims description 3
- 238000011176 pooling Methods 0.000 description 20
- 230000000694 effects Effects 0.000 description 3
- 238000012938 design process Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Image Analysis (AREA)
Abstract
The invention belongs to the technical field of unmanned aerial vehicle countering, and particularly relates to an unmanned aerial vehicle countering method and system based on unmanned aerial vehicle identification, wherein the method comprises the following steps: acquiring an unmanned aerial vehicle detection data set; constructing a deconvolution fusion network, wherein the deconvolution fusion network is used for outputting a segmentation result graph; training a deconvolution fusion network by adopting an unmanned aerial vehicle detection data set to obtain an unmanned aerial vehicle detection model; obtaining a visible light image of a region to be monitored and a depth image corresponding to the visible light image; inputting the visible light image into an unmanned aerial vehicle detection model, and judging whether the visible light image contains the unmanned aerial vehicle according to the output of the unmanned aerial vehicle detection model; when the visible light image contains the unmanned aerial vehicle, determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result image output by the unmanned aerial vehicle detection model and the corresponding depth image; and transmitting an interference signal to the unmanned aerial vehicle according to the space center point coordinates of the unmanned aerial vehicle. The invention improves the accuracy of unmanned aerial vehicle identification and improves the accuracy of interference.
Description
Technical Field
The invention relates to the technical field of unmanned aerial vehicle countering, in particular to an unmanned aerial vehicle countering method and system based on unmanned aerial vehicle identification.
Background
In recent years, the Chinese unmanned aerial vehicle industry rapidly develops, and the application scene of the unmanned aerial vehicle is widened and deepened continuously. However, the proliferation of the number of unmanned aerial vehicles and the imperfection of a supervision system thereof also cause a series of problems such as personal privacy leakage, secret information leakage and the like, and the social security and military security are seriously threatened. Therefore, achieving the countering of low-altitude unmanned aerial vehicle targets is important.
Compared with countering means such as laser weapons, net capturing and the like, the radio interference can be countered on the premise of not damaging the unmanned aerial vehicle, and the method has the advantages of low implementation difficulty and low cost. The existing radio interference countermeasures mostly rely on manual or photoelectric equipment to find and position the unmanned aerial vehicle target, and the unmanned aerial vehicle is countered through an interference system, so that the countermeasures effect can be influenced and electromagnetic environment pollution can be caused once the unmanned aerial vehicle is found out untimely, erroneously identified or positioned with low precision. Therefore, an automatic unmanned aerial vehicle countering method capable of rapidly and accurately finding and positioning the unmanned aerial vehicle and performing directional interference is needed.
Disclosure of Invention
The invention aims to provide an unmanned aerial vehicle countering method and system based on unmanned aerial vehicle identification, which improve the accuracy of unmanned aerial vehicle identification and the accuracy of interference.
In order to achieve the above object, the present invention provides the following solutions:
an unmanned aerial vehicle countering method based on unmanned aerial vehicle identification, comprising:
acquiring an unmanned aerial vehicle detection data set;
constructing a deconvolution fusion network, wherein the deconvolution fusion network is used for outputting a segmentation result graph;
training the deconvolution fusion network by adopting the unmanned aerial vehicle detection data set to obtain an unmanned aerial vehicle detection model;
obtaining a visible light image of a region to be monitored and a depth image corresponding to the visible light image;
inputting the visible light image into the unmanned aerial vehicle detection model, and judging whether the visible light image contains the unmanned aerial vehicle according to the output of the unmanned aerial vehicle detection model;
when the visible light image contains the unmanned aerial vehicle, determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result image output by the unmanned aerial vehicle detection model and the corresponding depth image;
transmitting an interference signal to the unmanned aerial vehicle according to the space center point coordinates of the unmanned aerial vehicle;
the deconvolution fusion network comprises a feature extraction module, a first deconvolution fusion module, a second deconvolution fusion module, a third deconvolution fusion module, a fourth deconvolution fusion module and a segmentation result diagram determining module, wherein the feature extraction module is used for generating a feature diagram U1, a feature diagram U2, a feature diagram U3, a feature diagram U4 and a feature diagram U5 with sequentially reduced sizes according to an input image; the first deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature map U4 to obtain a feature map U4', carrying out convolution operation with the channel number of 2 on the feature map U5 to obtain a feature map U5', carrying out deconvolution operation on the feature map U5' to obtain a feature map V5 with the same size as the feature map U4', and carrying out pixel-by-pixel addition on the feature map V5 and the feature map U4' to obtain a feature map V4; the second deconvolution fusion module is configured to perform a convolution operation with a channel number of 2 on the feature map U3 to obtain a feature map U3', perform a deconvolution operation on the feature map V4' to obtain a feature map V4 with the same size as the feature map U3', and perform a pixel-by-pixel addition on the feature map V4 and the feature map U3' to obtain a feature map V3'; the third deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature image U2 to obtain a feature image U2', carrying out deconvolution operation on the feature image V3' to obtain a feature image V3 with the same size as the feature image U2', carrying out pixel-by-pixel addition on the feature image V3 and the feature image U2' to obtain a feature image V2', carrying out convolution operation with the channel number of 2 on the feature image U1 to obtain a feature image U1', carrying out deconvolution operation on the feature image V2 'to obtain a feature image V2 with the same size as the feature image U1', and carrying out pixel-by-pixel addition on the feature image V2 and the feature image U1 'to obtain a feature image V1'; the segmentation result diagram determining module is used for deconvoluting the characteristic diagram V1' to obtain a segmentation result diagram V1 with the same size as the input image.
Optionally, when the visible light image includes the unmanned aerial vehicle, determining a spatial center point coordinate of the unmanned aerial vehicle according to the segmentation result graph output by the unmanned aerial vehicle detection model and the corresponding depth image, specifically includes:
when the visible light image contains the unmanned aerial vehicle, calculating centroid coordinates of the unmanned aerial vehicle in a segmentation result diagram output by the unmanned aerial vehicle detection model based on a centroid calculation function;
obtaining a depth value of the centroid coordinate position of the unmanned aerial vehicle on the corresponding depth image;
and determining the space center point coordinates of the unmanned aerial vehicle according to the centroid coordinates of the unmanned aerial vehicle and the depth value.
Optionally, the transmitting an interference signal to the unmanned aerial vehicle according to the spatial center point coordinates of the unmanned aerial vehicle specifically includes:
determining an altitude angle and an azimuth angle of interference signal emission according to the space center point coordinates of the unmanned aerial vehicle;
and controlling the interference equipment to emit interference signals according to the altitude angle and the azimuth angle through a servo system.
Optionally, the method further comprises:
inputting a visible light image acquired in real time into the unmanned aerial vehicle detection model to output a segmentation result diagram, and determining a monitoring result according to the output segmentation result diagram, wherein the monitoring result comprises or does not comprise an unmanned aerial vehicle;
when the monitoring result at the current moment is that the unmanned aerial vehicle is not contained and the monitoring result at the previous moment is that the unmanned aerial vehicle is contained, the interference equipment is controlled to reset through the servo system, and the interference equipment is closed.
Optionally, the feature extraction module is configured to input the input image into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 4*4 to obtain a feature map U1, input the feature map U1 into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 2×2 to obtain a feature map U2, input the feature map U2 into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 2×2 to obtain a feature map U3, and input the feature map U3 into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 2×2 to obtain a feature map U4.
The invention discloses an unmanned aerial vehicle countering system based on unmanned aerial vehicle identification, which comprises the following components:
the data set acquisition module is used for acquiring an unmanned aerial vehicle detection data set;
the deconvolution fusion network construction module is used for constructing a deconvolution fusion network which is used for outputting a segmentation result graph;
the deconvolution fusion network training module is used for training the deconvolution fusion network by adopting the unmanned aerial vehicle detection data set to obtain an unmanned aerial vehicle detection model;
the device comprises an area to be monitored image acquisition module, a display module and a display module, wherein the area to be monitored image acquisition module is used for acquiring a visible light image of an area to be monitored and a depth image corresponding to the visible light image;
the unmanned aerial vehicle detection module is used for inputting the visible light image into the unmanned aerial vehicle detection model, and judging whether the visible light image contains an unmanned aerial vehicle or not according to the output of the unmanned aerial vehicle detection model;
the unmanned aerial vehicle space center point coordinate determining module is used for determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result graph output by the unmanned aerial vehicle detection model and the corresponding depth image when the visible light image contains the unmanned aerial vehicle;
the interference signal transmitting module is used for transmitting an interference signal to the unmanned aerial vehicle according to the space center point coordinates of the unmanned aerial vehicle;
the deconvolution fusion network comprises a feature extraction module, a first deconvolution fusion module, a second deconvolution fusion module, a third deconvolution fusion module, a fourth deconvolution fusion module and a segmentation result diagram determining module, wherein the feature extraction module is used for generating a feature diagram U1, a feature diagram U2, a feature diagram U3, a feature diagram U4 and a feature diagram U5 with sequentially reduced sizes according to an input image; the first deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature map U4 to obtain a feature map U4', carrying out convolution operation with the channel number of 2 on the feature map U5 to obtain a feature map U5', carrying out deconvolution operation on the feature map U5' to obtain a feature map V5 with the same size as the feature map U4', and carrying out pixel-by-pixel addition on the feature map V5 and the feature map U4' to obtain a feature map V4; the second deconvolution fusion module is configured to perform a convolution operation with a channel number of 2 on the feature map U3 to obtain a feature map U3', perform a deconvolution operation on the feature map V4' to obtain a feature map V4 with the same size as the feature map U3', and perform a pixel-by-pixel addition on the feature map V4 and the feature map U3' to obtain a feature map V3'; the third deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature image U2 to obtain a feature image U2', carrying out deconvolution operation on the feature image V3' to obtain a feature image V3 with the same size as the feature image U2', carrying out pixel-by-pixel addition on the feature image V3 and the feature image U2' to obtain a feature image V2', carrying out convolution operation with the channel number of 2 on the feature image U1 to obtain a feature image U1', carrying out deconvolution operation on the feature image V2 'to obtain a feature image V2 with the same size as the feature image U1', and carrying out pixel-by-pixel addition on the feature image V2 and the feature image U1 'to obtain a feature image V1'; the segmentation result diagram determining module is used for deconvoluting the characteristic diagram V1' to obtain a segmentation result diagram V1 with the same size as the input image.
Optionally, the spatial center point coordinate determining module of the unmanned aerial vehicle specifically includes:
the unmanned aerial vehicle detection model comprises a centroid coordinate determining unit, a processing unit and a processing unit, wherein the centroid coordinate determining unit is used for calculating the centroid coordinate of the unmanned aerial vehicle in a segmentation result diagram output by the unmanned aerial vehicle detection model based on a centroid calculation function when the visible light image contains the unmanned aerial vehicle;
the depth value determining unit is used for obtaining the depth value of the centroid coordinate position of the unmanned aerial vehicle on the corresponding depth image;
and the space center point coordinate determining unit of the unmanned aerial vehicle is used for determining the space center point coordinate of the unmanned aerial vehicle according to the centroid coordinate of the unmanned aerial vehicle and the depth value.
Optionally, the interference signal transmitting module specifically includes:
the system comprises an interference signal transmitting altitude and azimuth determining unit, a receiving unit and a receiving unit, wherein the interference signal transmitting altitude and azimuth determining unit is used for determining the interference signal transmitting altitude and azimuth according to the space center point coordinates of the unmanned aerial vehicle;
and the interference signal transmitting unit is used for controlling the interference equipment to transmit interference signals according to the altitude angle and the azimuth angle through a servo system.
Optionally, the system further comprises:
the monitoring result real-time output module is used for inputting the visible light image acquired in real time into the unmanned aerial vehicle detection model to output a segmentation result diagram, and determining a monitoring result according to the output segmentation result diagram, wherein the monitoring result comprises or does not comprise the unmanned aerial vehicle;
and the interference equipment resetting and closing module is used for controlling the interference equipment to be reset and closing the interference equipment through the servo system when the monitoring result at the current moment is that the unmanned aerial vehicle is not contained and the monitoring result at the last moment is that the unmanned aerial vehicle is contained.
Optionally, the feature extraction module is configured to input the input image into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 4*4 to obtain a feature map U1, input the feature map U1 into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 2×2 to obtain a feature map U2, input the feature map U2 into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 2×2 to obtain a feature map U3, and input the feature map U3 into 2 convolution layers with a core 3*3 and 1 pooling layer with a core 2×2 to obtain a feature map U4.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects:
the invention discloses an unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition, wherein a counterconvolution fusion network expands the size of a deep semantic feature map layer by layer through a counterconvolution operation, and ensures high resolution of an output feature map in a mode of gradually fusing with low-layer features, so that end-to-end and accurate pixel level segmentation of an unmanned aerial vehicle target is realized, the accuracy of unmanned aerial vehicle recognition is improved, in addition, the spatial center point coordinate of the unmanned aerial vehicle is determined according to a segmentation result map and a corresponding depth image, an interference signal is transmitted to the unmanned aerial vehicle according to the spatial center point coordinate of the unmanned aerial vehicle, the accuracy of interference signal transmission is improved, and the countering efficiency of the unmanned aerial vehicle is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions of the prior art, the drawings that are needed in the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a schematic flow chart of an unmanned aerial vehicle countering method based on unmanned aerial vehicle identification;
fig. 2 is a schematic flow chart II of an unmanned aerial vehicle countering method based on unmanned aerial vehicle recognition;
FIG. 3 is a schematic diagram of a deconvolution converged network in accordance with the present invention;
fig. 4 is a schematic structural diagram of an unmanned aerial vehicle countering system based on unmanned aerial vehicle recognition.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The invention aims to provide an unmanned aerial vehicle countering method and system based on unmanned aerial vehicle identification, which improve the accuracy of unmanned aerial vehicle identification and the accuracy of interference.
In order that the above-recited objects, features and advantages of the present invention will become more readily apparent, a more particular description of the invention will be rendered by reference to the appended drawings and appended detailed description.
Fig. 1 is a schematic flow diagram of an unmanned aerial vehicle countering method based on unmanned aerial vehicle recognition according to the present invention, fig. 2 is a schematic flow diagram of an unmanned aerial vehicle countering method based on unmanned aerial vehicle recognition according to the present invention, as shown in fig. 1-2, an unmanned aerial vehicle countering method based on unmanned aerial vehicle recognition, including:
step 101: and acquiring a detection data set of the unmanned aerial vehicle.
Step 102: and constructing a deconvolution fusion network, wherein the deconvolution fusion network is used for outputting the segmentation result graph.
As shown in fig. 3, the deconvolution fusion network includes a feature extraction module, a first deconvolution fusion module, a second deconvolution fusion module, a third deconvolution fusion module, a fourth deconvolution fusion module, and a segmentation result diagram determining module, wherein the dashed boxes from top to bottom in fig. 3 are the first deconvolution fusion module, the second deconvolution fusion module, the third deconvolution fusion module, and the fourth deconvolution fusion module in sequence, and the feature extraction module is configured to generate a feature diagram U1, a feature diagram U2, a feature diagram U3, a feature diagram U4, and a feature diagram U5 with sizes decreasing in sequence according to an input image; the first deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature map U4 to obtain a feature map U4', carrying out convolution operation with the channel number of 2 on the feature map U5 to obtain a feature map U5', carrying out deconvolution operation on the feature map U5 'to obtain a feature map V5 with the same size as the feature map U4', and carrying out pixel-by-pixel addition on the feature map V5 and the feature map U4 'to obtain a feature map V4'; the second deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the characteristic diagram U3 to obtain a characteristic diagram U3', carrying out deconvolution operation on the characteristic diagram V4' to obtain a characteristic diagram V4 with the same size as the characteristic diagram U3', and carrying out pixel-by-pixel addition on the characteristic diagram V4 and the characteristic diagram U3' to obtain a characteristic diagram V3'; the third deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the characteristic image U2 to obtain the characteristic image U2', carrying out deconvolution operation on the characteristic image V3' to obtain the characteristic image V3 with the same size as the characteristic image U2', carrying out pixel-by-pixel addition on the characteristic image V3 and the characteristic image U2' to obtain the characteristic image V2', carrying out convolution operation with the channel number of 2 on the characteristic image U1 to obtain the characteristic image U1', carrying out deconvolution operation on the characteristic image V2 'to obtain the characteristic image V2 with the same size as the characteristic image U1', and carrying out pixel-by-pixel addition on the characteristic image V2 and the characteristic image U1 'to obtain the characteristic image V1'; the segmentation result diagram determining module is used for deconvoluting the feature diagram V1' to obtain a segmentation result diagram V1 with the same size as the input image.
The feature extraction module is used for inputting an input image into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 4*4 to obtain a feature map U1, inputting the feature map U1 into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 2 x 2 to obtain a feature map U2, inputting the feature map U2 into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 2 x 2 to obtain a feature map U3, and inputting the feature map U3 into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 2 x 2 to obtain a feature map U4.
The following specifically shows a design process of the deconvolution fusion network, taking an example of inputting a visible light monitoring image (size 1024×1024×3) including a black unmanned aerial vehicle:
(1) First, feature extraction is performed on a monitoring image (input image): inputting the monitoring image into 2 convolution layers with the core 3*3 and 1 pooling layer with the pooling core 4*4 to obtain a feature map U1; inputting the characteristic diagram U1 into 2 convolution layers with the core of 3*3 and 1 pooling layer with the pooling core of 2 x 2 to obtain the characteristic diagram U2; repeating the operation process of U1 to U2 to generate a feature map which is deeper and has richer semantic representation, namely U3 and U4; and then carrying out convolution operation with the number of convolutions of U4 being 4096 and the step length being 2, and obtaining a characteristic diagram U5.
Through layer-by-layer feature extraction, the output feature graphs and corresponding sizes of all layers are U1:256×256×24, U2:128 x 48, U3:64×64×96, U4:32 x 192 and U5:16*16*4096.
(2) Next, constructing a deconvolution fusion module based on the reverse direction of the backbone network:
and respectively carrying out convolution operation with the channel number of 2 on the characteristic diagram U4 and the characteristic diagram U5, so that the channel dimension of the characteristic diagram is converted into 2, and obtaining characteristic diagrams U4 'and U5', wherein the corresponding sizes of the characteristic diagrams are 32 x 2 and 16 x 2. The number of modified channels is 2 here because the task of the semantic segmentation is to identify and separate the unmanned aerial vehicle target from the background, i.e. the class of pixels in the image is both unmanned aerial vehicle and background.
Next, deconvolution is performed on the feature map U5', i.e., the size of the feature map U5' is first enlarged to 32×32×2 by interpolation and zero padding, and then forward convolution of 1*1 is performed. After the deconvolution operation, a feature map V5 with a 2-fold size expansion is obtained. Then, the feature map V5 is added pixel by pixel with U4' of the same size, resulting in a feature map of size 32 x 2. In order to further enhance the fusion effect, the obtained feature map is input into a layer of convolution layer again to obtain a feature map V4', so that the size of the feature map V4' is kept unchanged, but the semantic feature information is enhanced. The above is the construction process of the deconvolution fusion module.
(3) Then inputting the characteristic diagrams V4' and U3' into the deconvolution module to obtain a characteristic diagram V3' with the size of 64 x 2; inputting the obtained characteristic diagrams V3' and U2' into a deconvolution module to obtain a characteristic diagram V2' with the size of 128 x 2; finally, a feature map V1' of size 256×256×2 is obtained based on the obtained feature maps V3', U2' and the deconvolution module.
In order to obtain the exact position of the drone target, pixel-level segmentation of the input map is required. At this time, the size of the feature map V1 'is 1/4 of the input map, so that the deconvolution operation is performed again on V1' to enlarge the size thereof to the original map size, and the final segmentation result map V1 is obtained. The size of V1 is 1024×1024×2, and the value of the first channel indicates the probability that the pixel belongs to the unmanned plane target, and when the pixel value is greater than the target threshold, the pixel belongs to the unmanned plane; the value of the second channel indicates the probability that the pixel belongs to the background, and when the pixel value is greater than the background threshold, it indicates that the pixel is in the background region.
The above is the design process of the deconvolution fusion network. The four deconvolution fusion modules contained in the network can enlarge the size of the deep semantic feature map layer by deconvolution operation, and ensure the high resolution of the output feature map by low-level feature fusion, thereby realizing end-to-end and accurate pixel-level segmentation.
Step 103: and training the deconvolution fusion network by adopting the unmanned aerial vehicle detection data set to obtain an unmanned aerial vehicle detection model.
Step 103 specifically includes: when the deconvolution fusion network is trained by adopting the unmanned aerial vehicle detection data set, the parameter of the whole deconvolution fusion network is updated based on the two classification cross entropy loss functions, and a deconvolution fusion model (unmanned aerial vehicle detection model) is obtained after training is finished.
Step 104: and obtaining a visible light image of the area to be monitored and a depth image corresponding to the visible light image.
Step 104 specifically includes:
and monitoring a security area (an area to be monitored) to be subjected to unmanned aerial vehicle reaction by adopting a depth camera, and shooting at regular time to obtain a visible light image and a corresponding depth picture of the security area.
Step 105: and inputting the visible light image into an unmanned aerial vehicle detection model, and judging whether the visible light image contains the unmanned aerial vehicle according to the output of the unmanned aerial vehicle detection model.
Step 105 specifically includes: inputting the shot visible light image into an unmanned aerial vehicle detection model, when an output segmentation result image V1 contains unmanned aerial vehicle pixels, generating a non-cooperative unmanned aerial vehicle in a current security area, outputting a '1' by an intelligent monitoring module, converting a first channel image in the segmentation result image V1 into a binary image, and outputting the binary image; when all pixel classifications in the output segmentation result diagram V1 are background classifications, the unmanned aerial vehicle target is not found in the security area, and the intelligent monitoring module outputs 0.
The control module obtains the output of the intelligent monitoring module.
Step 106: when the visible light image contains the unmanned aerial vehicle, determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result image output by the unmanned aerial vehicle detection model and the corresponding depth image.
Step 106 specifically includes:
when the visible light image contains the unmanned aerial vehicle (when the output of the intelligent monitoring module is 1), calculating centroid coordinates (a, b) of the unmanned aerial vehicle pixel region in the segmentation result diagram output by the unmanned aerial vehicle detection model based on the centroid calculation function of opencv. The segmentation result graph is a binary graph.
And obtaining a depth value d of the centroid coordinates (a, b) of the unmanned aerial vehicle on the corresponding depth image.
And determining the coordinates (x, y, z) of the spatial center point of the unmanned aerial vehicle according to the centroid coordinates and the depth values of the unmanned aerial vehicle.
And (3) calculating and obtaining the space center point coordinates (x, y, z) of the unmanned aerial vehicle according to the formulas (1) - (3).
In the formulas (1) - (3), a and b are respectively the abscissa and the ordinate of the centroid of the unmanned aerial vehicle, d is the depth value of the centroid of the unmanned aerial vehicle, c_f is the internal parameter of the depth camera, and f x Is the focal length value in the x direction, f y C_f is the focal length value in the y direction x C_f is the offset of the optical axis in the x direction from the center of the projection plane coordinates y Is the offset of the optical axis in the y-direction from the center of the projection plane coordinates.
Step 107: and transmitting an interference signal to the unmanned aerial vehicle according to the space center point coordinates of the unmanned aerial vehicle.
Step 107 specifically includes:
and determining the altitude and azimuth angle of the interference signal emission according to the space center point coordinates of the unmanned aerial vehicle.
The interference device is controlled by a servo system to emit interference signals according to the altitude angle and the azimuth angle.
In order to aim the interfering device at the unmanned aerial vehicle target, the angle of the servo system needs to be adjusted. Therefore, according to the coordinates (x, y, z) of the spatial center point of the unmanned aerial vehicle and the formulas (4) to (5), the motion angles required to be reached by the servo system, namely the altitude angle alpha and the azimuth angle theta, are obtained.
And the control module sends a servo motion instruction and an interference starting instruction according to the output value '1' of the intelligent monitoring module and the calculated motion angle of the servo system.
The unmanned aerial vehicle countering method based on unmanned aerial vehicle identification further comprises the following steps:
and inputting the visible light image acquired in real time into an unmanned aerial vehicle detection model to output a segmentation result diagram, and determining a monitoring result according to the output segmentation result diagram, wherein the monitoring result comprises or does not comprise the unmanned aerial vehicle.
When the monitoring result at the current moment is that the unmanned aerial vehicle is not contained and the monitoring result at the previous moment is that the unmanned aerial vehicle is contained, the interference equipment is controlled to reset and the interference equipment is closed through the servo system.
When the output of the intelligent monitoring module is 0, namely, no unmanned aerial vehicle target appears, the control module judges whether the value at the current moment is the same as the output value of the intelligent monitoring module of the previous frame. If the two types of the wireless communication signals are the same, the unmanned aerial vehicle target is not appeared at the last moment, the wireless interference transmitter (interference equipment) is in a closed state, and the control module does not send any instruction; if the target is different, the unmanned aerial vehicle targets appear at the last moment but do not exist at the current moment (the target is knocked down by countering or the target flies out of the security area), the radio interference transmitter is in an on state, and the control module sends a servo homing instruction and an interference closing instruction.
The interference module comprises interference equipment and a servo system, and after a host of the interference module receives an instruction of the control module, the servo system is controlled to move to a designated angle or azimuth of the control module according to the instruction type; and then the switch of the radio interference transmitter is turned on or off according to the instruction type.
That is, when the unmanned aerial vehicle target appears in the security area, the interference module firstly enables the servo system to move to a designated angle and starts the radio interference transmitter, so that the interference transmitter aims at the azimuth of the unmanned aerial vehicle target to perform interference countermeasures; along with the change of the spatial position of the unmanned aerial vehicle, the angle of the interference transmitter is continuously adjusted according to the new servo motion instruction so as to lead the interference transmitter to be aligned with the unmanned aerial vehicle for accurate countering; and (3) enabling the servo system to return to a position and closing the interference transmitter until the unmanned aerial vehicle is knocked down after losing control, namely, the unmanned aerial vehicle is disappeared in a security area.
In order to accurately identify and position the area of the unmanned aerial vehicle from the background, the invention discloses a deconvolution fusion network, wherein the deconvolution network gradually enlarges the size of a deep semantic feature map through deconvolution operation, and ensures the high resolution of an output feature map in a mode of gradually fusing with low-level features, thereby realizing the end-to-end and accurate pixel level segmentation of the unmanned aerial vehicle target.
According to the invention, the control module and the interference module can calculate and adjust the transmitting direction of the interference device according to the output result of the intelligent monitoring module, so that the directional accurate interference is realized.
Fig. 4 is a schematic structural diagram of an unmanned aerial vehicle countering system based on unmanned aerial vehicle recognition according to the present invention, as shown in fig. 4, an unmanned aerial vehicle countering system based on unmanned aerial vehicle recognition includes:
the data set acquisition module 201 is configured to acquire a detection data set of the unmanned aerial vehicle.
The deconvolution fusion network construction module 202 is configured to construct a deconvolution fusion network, where the deconvolution fusion network is configured to output a segmentation result graph.
The deconvolution fusion network training module 203 is configured to train the deconvolution fusion network by using the unmanned aerial vehicle detection data set, so as to obtain an unmanned aerial vehicle detection model.
The region to be monitored image obtaining module 204 is configured to obtain a visible light image and a depth image corresponding to the visible light image of the region to be monitored.
The unmanned aerial vehicle detection module 205 is configured to input a visible light image into an unmanned aerial vehicle detection model, and determine whether the visible light image includes an unmanned aerial vehicle according to an output of the unmanned aerial vehicle detection model.
The space center point coordinate determining module 206 of the unmanned aerial vehicle is configured to determine, when the visible light image includes the unmanned aerial vehicle, the space center point coordinate of the unmanned aerial vehicle according to the segmentation result map output by the unmanned aerial vehicle detection model and the corresponding depth image.
The interference signal transmitting module 207 is configured to transmit an interference signal to the unmanned aerial vehicle according to the spatial center point coordinates of the unmanned aerial vehicle.
The deconvolution fusion network comprises a feature extraction module, a first deconvolution fusion module, a second deconvolution fusion module, a third deconvolution fusion module, a fourth deconvolution fusion module and a segmentation result diagram determining module, wherein the feature extraction module is used for generating a feature diagram U1, a feature diagram U2, a feature diagram U3, a feature diagram U4 and a feature diagram U5 with sizes reduced in sequence according to an input image; the first deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature map U4 to obtain a feature map U4', carrying out convolution operation with the channel number of 2 on the feature map U5 to obtain a feature map U5', carrying out deconvolution operation on the feature map U5 'to obtain a feature map V5 with the same size as the feature map U4', and carrying out pixel-by-pixel addition on the feature map V5 and the feature map U4 'to obtain a feature map V4'; the second deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the characteristic diagram U3 to obtain a characteristic diagram U3', carrying out deconvolution operation on the characteristic diagram V4' to obtain a characteristic diagram V4 with the same size as the characteristic diagram U3', and carrying out pixel-by-pixel addition on the characteristic diagram V4 and the characteristic diagram U3' to obtain a characteristic diagram V3'; the third deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the characteristic image U2 to obtain the characteristic image U2', carrying out deconvolution operation on the characteristic image V3' to obtain the characteristic image V3 with the same size as the characteristic image U2', carrying out pixel-by-pixel addition on the characteristic image V3 and the characteristic image U2' to obtain the characteristic image V2', carrying out convolution operation with the channel number of 2 on the characteristic image U1 to obtain the characteristic image U1', carrying out deconvolution operation on the characteristic image V2 'to obtain the characteristic image V2 with the same size as the characteristic image U1', and carrying out pixel-by-pixel addition on the characteristic image V2 and the characteristic image U1 'to obtain the characteristic image V1'; the segmentation result diagram determining module is used for deconvoluting the feature diagram V1' to obtain a segmentation result diagram V1 with the same size as the input image.
The space center point coordinate determining module 206 of the unmanned plane specifically includes:
and the centroid coordinate determining unit of the unmanned aerial vehicle is used for calculating the centroid coordinate of the unmanned aerial vehicle in the segmentation result diagram output by the unmanned aerial vehicle detection model based on the centroid calculation function when the visible light image contains the unmanned aerial vehicle.
And the depth value determining unit is used for obtaining the depth value of the centroid coordinate position of the unmanned aerial vehicle on the corresponding depth image.
And the space center point coordinate determining unit of the unmanned aerial vehicle is used for determining the space center point coordinate of the unmanned aerial vehicle according to the centroid coordinate and the depth value of the unmanned aerial vehicle.
The interfering signal transmitting module 207 specifically includes:
and the interference signal transmitting altitude and azimuth angle determining unit is used for determining the altitude and azimuth angle of the interference signal transmitting according to the space center point coordinates of the unmanned aerial vehicle.
And the interference signal transmitting unit is used for controlling the interference equipment to transmit interference signals according to the altitude angle and the azimuth angle through the servo system.
Unmanned aerial vehicle reaction system based on unmanned aerial vehicle discernment still includes:
and the monitoring result real-time output module is used for inputting the visible light image acquired in real time into the unmanned aerial vehicle detection model to output a segmentation result diagram, and determining the monitoring result according to the output segmentation result diagram, wherein the monitoring result comprises or does not comprise the unmanned aerial vehicle.
And the interference equipment resetting and closing module is used for controlling the interference equipment to be reset and closing the interference equipment through the servo system when the monitoring result at the current moment is that the unmanned aerial vehicle is not contained and the monitoring result at the last moment is that the unmanned aerial vehicle is contained.
The feature extraction module is used for inputting an input image into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 4*4 to obtain a feature map U1, inputting the feature map U1 into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 2 x 2 to obtain a feature map U2, inputting the feature map U2 into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 2 x 2 to obtain a feature map U3, and inputting the feature map U3 into 2 convolution layers with the core 3*3 and 1 pooling layers with the core 2 x 2 to obtain a feature map U4.
In the present specification, each embodiment is described in a progressive manner, and each embodiment is mainly described in a different point from other embodiments, and identical and similar parts between the embodiments are all enough to refer to each other. For the system disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant points refer to the description of the method section.
The principles and embodiments of the present invention have been described herein with reference to specific examples, the description of which is intended only to assist in understanding the methods of the present invention and the core ideas thereof; also, it is within the scope of the present invention to be modified by those of ordinary skill in the art in light of the present teachings. In view of the foregoing, this description should not be construed as limiting the invention.
Claims (8)
1. An unmanned aerial vehicle countering method based on unmanned aerial vehicle recognition is characterized by comprising the following steps:
acquiring an unmanned aerial vehicle detection data set;
constructing a deconvolution fusion network, wherein the deconvolution fusion network is used for outputting a segmentation result graph;
training the deconvolution fusion network by adopting the unmanned aerial vehicle detection data set to obtain an unmanned aerial vehicle detection model;
obtaining a visible light image of a region to be monitored and a depth image corresponding to the visible light image;
inputting the visible light image into the unmanned aerial vehicle detection model, and judging whether the visible light image contains the unmanned aerial vehicle according to the output of the unmanned aerial vehicle detection model;
when the visible light image contains the unmanned aerial vehicle, determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result image output by the unmanned aerial vehicle detection model and the corresponding depth image;
transmitting an interference signal to the unmanned aerial vehicle according to the space center point coordinates of the unmanned aerial vehicle;
the deconvolution fusion network comprises a feature extraction module, a first deconvolution fusion module, a second deconvolution fusion module, a third deconvolution fusion module, a fourth deconvolution fusion module and a segmentation result diagram determining module, wherein the feature extraction module is used for generating a feature diagram U1, a feature diagram U2, a feature diagram U3, a feature diagram U4 and a feature diagram U5 with sequentially reduced sizes according to an input image; the first deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature map U4 to obtain a feature map U4', carrying out convolution operation with the channel number of 2 on the feature map U5 to obtain a feature map U5', carrying out deconvolution operation on the feature map U5' to obtain a feature map V5 with the same size as the feature map U4', and carrying out pixel-by-pixel addition on the feature map V5 and the feature map U4' to obtain a feature map V4; the second deconvolution fusion module is configured to perform a convolution operation with a channel number of 2 on the feature map U3 to obtain a feature map U3', perform a deconvolution operation on the feature map V4' to obtain a feature map V4 with the same size as the feature map U3', and perform a pixel-by-pixel addition on the feature map V4 and the feature map U3' to obtain a feature map V3'; the third deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature image U2 to obtain a feature image U2', carrying out deconvolution operation on the feature image V3' to obtain a feature image V3 with the same size as the feature image U2', carrying out pixel-by-pixel addition on the feature image V3 and the feature image U2' to obtain a feature image V2', carrying out convolution operation with the channel number of 2 on the feature image U1 to obtain a feature image U1', carrying out deconvolution operation on the feature image V2 'to obtain a feature image V2 with the same size as the feature image U1', and carrying out pixel-by-pixel addition on the feature image V2 and the feature image U1 'to obtain a feature image V1'; the segmentation result diagram determining module is used for deconvoluting the feature diagram V1' to obtain a segmentation result diagram V1 with the same size as the input image;
when the visible light image contains the unmanned aerial vehicle, determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result image output by the unmanned aerial vehicle detection model and the corresponding depth image, specifically comprising:
when the visible light image contains the unmanned aerial vehicle, calculating centroid coordinates of the unmanned aerial vehicle in a segmentation result diagram output by the unmanned aerial vehicle detection model based on a centroid calculation function;
obtaining a depth value of the centroid coordinate position of the unmanned aerial vehicle on the corresponding depth image;
determining the space center point coordinates of the unmanned aerial vehicle according to the centroid coordinates of the unmanned aerial vehicle and the depth values;
calculating the space center point coordinates of the unmanned aerial vehicle according to the formulas (1) - (3)x,y,z);
wherein ,ais the abscissa of the centroid of the unmanned aerial vehicle,bis the ordinate of the centroid of the unmanned aerial vehicle,dis the depth value of the centroid of the unmanned aerial vehicle,c_fas an internal parameter of the depth camera,f x is thatxThe focal length value in the direction is set,f y is thatyThe focal length value in the direction is set,c_f x is thatxThe offset of the optical axis in the direction from the center of the projection plane coordinates,c_f y is thatyOffset of the optical axis in the direction from the center of the projection plane coordinates.
2. The unmanned aerial vehicle countering method based on unmanned aerial vehicle identification according to claim 1, wherein the transmitting the interference signal to the unmanned aerial vehicle according to the spatial center point coordinates of the unmanned aerial vehicle specifically comprises:
determining an altitude angle and an azimuth angle of interference signal emission according to the space center point coordinates of the unmanned aerial vehicle;
and controlling the interference equipment to emit interference signals according to the altitude angle and the azimuth angle through a servo system.
3. The unmanned aerial vehicle countering method based on unmanned aerial vehicle identification of claim 2, wherein the method further comprises:
inputting a visible light image acquired in real time into the unmanned aerial vehicle detection model to output a segmentation result diagram, and determining a monitoring result according to the output segmentation result diagram, wherein the monitoring result comprises or does not comprise an unmanned aerial vehicle;
when the monitoring result at the current moment is that the unmanned aerial vehicle is not contained and the monitoring result at the previous moment is that the unmanned aerial vehicle is contained, the interference equipment is controlled to reset through the servo system, and the interference equipment is closed.
4. The unmanned aerial vehicle countering method based on unmanned aerial vehicle recognition according to claim 1, wherein the feature extraction module is configured to input an input image into 2 convolutions layers with a core 3*3 and 1 pooled layers with a core 4*4 to obtain a feature map U1, input the feature map U1 into 2 convolutions layers with a core 3*3 and 1 pooled layers with a core 2 x 2 to obtain a feature map U2, input the feature map U2 into 2 convolutions layers with a core 3*3 and 1 pooled layers with a core 2 x 2 to obtain a feature map U3, and input the feature map U3 into 2 convolutions layers with a core 3*3 and 1 pooled layers with a core 2 x 2 to obtain a feature map U4.
5. Unmanned aerial vehicle reaction system based on unmanned aerial vehicle discernment, characterized in that includes:
the data set acquisition module is used for acquiring an unmanned aerial vehicle detection data set;
the deconvolution fusion network construction module is used for constructing a deconvolution fusion network which is used for outputting a segmentation result graph;
the deconvolution fusion network training module is used for training the deconvolution fusion network by adopting the unmanned aerial vehicle detection data set to obtain an unmanned aerial vehicle detection model;
the device comprises an area to be monitored image acquisition module, a display module and a display module, wherein the area to be monitored image acquisition module is used for acquiring a visible light image of an area to be monitored and a depth image corresponding to the visible light image;
the unmanned aerial vehicle detection module is used for inputting the visible light image into the unmanned aerial vehicle detection model, and judging whether the visible light image contains an unmanned aerial vehicle or not according to the output of the unmanned aerial vehicle detection model;
the unmanned aerial vehicle space center point coordinate determining module is used for determining the space center point coordinate of the unmanned aerial vehicle according to the segmentation result graph output by the unmanned aerial vehicle detection model and the corresponding depth image when the visible light image contains the unmanned aerial vehicle;
the interference signal transmitting module is used for transmitting an interference signal to the unmanned aerial vehicle according to the space center point coordinates of the unmanned aerial vehicle;
the deconvolution fusion network comprises a feature extraction module, a first deconvolution fusion module, a second deconvolution fusion module, a third deconvolution fusion module, a fourth deconvolution fusion module and a segmentation result diagram determining module, wherein the feature extraction module is used for generating a feature diagram U1, a feature diagram U2, a feature diagram U3, a feature diagram U4 and a feature diagram U5 with sequentially reduced sizes according to an input image; the first deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature map U4 to obtain a feature map U4', carrying out convolution operation with the channel number of 2 on the feature map U5 to obtain a feature map U5', carrying out deconvolution operation on the feature map U5' to obtain a feature map V5 with the same size as the feature map U4', and carrying out pixel-by-pixel addition on the feature map V5 and the feature map U4' to obtain a feature map V4; the second deconvolution fusion module is configured to perform a convolution operation with a channel number of 2 on the feature map U3 to obtain a feature map U3', perform a deconvolution operation on the feature map V4' to obtain a feature map V4 with the same size as the feature map U3', and perform a pixel-by-pixel addition on the feature map V4 and the feature map U3' to obtain a feature map V3'; the third deconvolution fusion module is used for carrying out convolution operation with the channel number of 2 on the feature image U2 to obtain a feature image U2', carrying out deconvolution operation on the feature image V3' to obtain a feature image V3 with the same size as the feature image U2', carrying out pixel-by-pixel addition on the feature image V3 and the feature image U2' to obtain a feature image V2', carrying out convolution operation with the channel number of 2 on the feature image U1 to obtain a feature image U1', carrying out deconvolution operation on the feature image V2 'to obtain a feature image V2 with the same size as the feature image U1', and carrying out pixel-by-pixel addition on the feature image V2 and the feature image U1 'to obtain a feature image V1'; the segmentation result diagram determining module is used for deconvoluting the feature diagram V1' to obtain a segmentation result diagram V1 with the same size as the input image;
the space center point coordinate determining module of the unmanned aerial vehicle specifically comprises:
the unmanned aerial vehicle detection model comprises a centroid coordinate determining unit, a processing unit and a processing unit, wherein the centroid coordinate determining unit is used for calculating the centroid coordinate of the unmanned aerial vehicle in a segmentation result diagram output by the unmanned aerial vehicle detection model based on a centroid calculation function when the visible light image contains the unmanned aerial vehicle;
the depth value determining unit is used for obtaining the depth value of the centroid coordinate position of the unmanned aerial vehicle on the corresponding depth image;
the space center point coordinate determining unit of the unmanned aerial vehicle is used for determining the space center point coordinate of the unmanned aerial vehicle according to the centroid coordinate of the unmanned aerial vehicle and the depth value;
calculating the space center point coordinates of the unmanned aerial vehicle according to the formulas (1) - (3)x,y,z);
wherein ,ais the abscissa of the centroid of the unmanned aerial vehicle,bis the ordinate of the centroid of the unmanned aerial vehicle,dis the depth value of the centroid of the unmanned aerial vehicle,c_fas an internal parameter of the depth camera,f x is thatxThe focal length value in the direction is set,f y is thatyThe focal length value in the direction is set,c_f x is thatxThe offset of the optical axis in the direction from the center of the projection plane coordinates,c_f y is thatyOffset of the optical axis in the direction from the center of the projection plane coordinates.
6. The unmanned aerial vehicle countering system based on unmanned aerial vehicle recognition according to claim 5, wherein the interfering signal transmitting module specifically comprises:
the system comprises an interference signal transmitting altitude and azimuth determining unit, a receiving unit and a receiving unit, wherein the interference signal transmitting altitude and azimuth determining unit is used for determining the interference signal transmitting altitude and azimuth according to the space center point coordinates of the unmanned aerial vehicle;
and the interference signal transmitting unit is used for controlling the interference equipment to transmit interference signals according to the altitude angle and the azimuth angle through a servo system.
7. The unmanned aerial vehicle countering system based on unmanned aerial vehicle identification of claim 6, wherein the system further comprises:
the monitoring result real-time output module is used for inputting the visible light image acquired in real time into the unmanned aerial vehicle detection model to output a segmentation result diagram, and determining a monitoring result according to the output segmentation result diagram, wherein the monitoring result comprises or does not comprise the unmanned aerial vehicle;
and the interference equipment resetting and closing module is used for controlling the interference equipment to be reset and closing the interference equipment through the servo system when the monitoring result at the current moment is that the unmanned aerial vehicle is not contained and the monitoring result at the last moment is that the unmanned aerial vehicle is contained.
8. The unmanned aerial vehicle countering system based on unmanned aerial vehicle recognition according to claim 5, wherein the feature extraction module is configured to input an input image into 2 convolutions layers with kernels 3*3 and 1 pooled layers with kernels 4*4 to obtain a feature map U1, input the feature map U1 into 2 convolutions layers with kernels 3*3 and 1 pooled layers with kernels 2 x 2 to obtain a feature map U2, input the feature map U2 into 2 convolutions layers with kernels 3*3 and 1 pooled layers with kernels 2 x 2 to obtain a feature map U3, and input the feature map U3 into 2 convolutions layers with kernels 3*3 and 1 pooled layers with kernels 2 x 2 to obtain a feature map U4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310063450.4A CN115861938B (en) | 2023-02-06 | 2023-02-06 | Unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310063450.4A CN115861938B (en) | 2023-02-06 | 2023-02-06 | Unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115861938A CN115861938A (en) | 2023-03-28 |
CN115861938B true CN115861938B (en) | 2023-05-26 |
Family
ID=85657620
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310063450.4A Active CN115861938B (en) | 2023-02-06 | 2023-02-06 | Unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115861938B (en) |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113554656A (en) * | 2021-07-13 | 2021-10-26 | 中国科学院空间应用工程与技术中心 | Optical remote sensing image example segmentation method and device based on graph neural network |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11288517B2 (en) * | 2014-09-30 | 2022-03-29 | PureTech Systems Inc. | System and method for deep learning enhanced object incident detection |
US10776673B2 (en) * | 2019-01-31 | 2020-09-15 | StradVision, Inc. | Learning method and learning device for sensor fusion to integrate information acquired by radar capable of distance estimation and information acquired by camera to thereby improve neural network for supporting autonomous driving, and testing method and testing device using the same |
CN109753903B (en) * | 2019-02-27 | 2020-09-15 | 北航(四川)西部国际创新港科技有限公司 | Unmanned aerial vehicle detection method based on deep learning |
CN111666822A (en) * | 2020-05-13 | 2020-09-15 | 飒铂智能科技有限责任公司 | Low-altitude unmanned aerial vehicle target detection method and system based on deep learning |
CN113255589B (en) * | 2021-06-25 | 2021-10-15 | 北京电信易通信息技术股份有限公司 | Target detection method and system based on multi-convolution fusion network |
CN113822383B (en) * | 2021-11-23 | 2022-03-15 | 北京中超伟业信息安全技术股份有限公司 | Unmanned aerial vehicle detection method and system based on multi-domain attention mechanism |
CN114550016B (en) * | 2022-04-22 | 2022-07-08 | 北京中超伟业信息安全技术股份有限公司 | Unmanned aerial vehicle positioning method and system based on context information perception |
CN115187959B (en) * | 2022-07-14 | 2023-04-14 | 清华大学 | Method and system for landing flying vehicle in mountainous region based on binocular vision |
CN115223067B (en) * | 2022-09-19 | 2022-12-09 | 季华实验室 | Point cloud fusion method, device and equipment applied to unmanned aerial vehicle and storage medium |
-
2023
- 2023-02-06 CN CN202310063450.4A patent/CN115861938B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113554656A (en) * | 2021-07-13 | 2021-10-26 | 中国科学院空间应用工程与技术中心 | Optical remote sensing image example segmentation method and device based on graph neural network |
Also Published As
Publication number | Publication date |
---|---|
CN115861938A (en) | 2023-03-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111507166A (en) | Method and apparatus for learning CNN by using camera and radar together | |
CN109871763B (en) | Specific target tracking method based on YOLO | |
CN108694724A (en) | A kind of long-time method for tracking target | |
CN112085728B (en) | Submarine pipeline and leakage point detection method | |
CN108108697B (en) | Real-time unmanned aerial vehicle video target detection and tracking method | |
CN111507210A (en) | Traffic signal lamp identification method and system, computing device and intelligent vehicle | |
CN113203409B (en) | Method for constructing navigation map of mobile robot in complex indoor environment | |
CN109635661B (en) | Far-field wireless charging receiving target detection method based on convolutional neural network | |
CN111985365A (en) | Straw burning monitoring method and system based on target detection technology | |
CN111260687B (en) | Aerial video target tracking method based on semantic perception network and related filtering | |
CN112207821B (en) | Target searching method of visual robot and robot | |
CN114217303A (en) | Target positioning and tracking method and device, underwater robot and storage medium | |
CN112132900A (en) | Visual repositioning method and system | |
Li et al. | Weak moving object detection in optical remote sensing video with motion-drive fusion network | |
CN115861938B (en) | Unmanned aerial vehicle countering method and system based on unmanned aerial vehicle recognition | |
CN115424072A (en) | Unmanned aerial vehicle defense method based on detection technology | |
CN113391644B (en) | Unmanned aerial vehicle shooting distance semi-automatic optimization method based on image information entropy | |
CN113298177B (en) | Night image coloring method, device, medium and equipment | |
CN112233079B (en) | Method and system for fusing images of multiple sensors | |
Li et al. | Feature point extraction and tracking based on a local adaptive threshold | |
CN112529917A (en) | Three-dimensional target segmentation method, device, equipment and storage medium | |
CN117593620A (en) | Multi-target detection method and device based on fusion of camera and laser radar | |
CN114740878B (en) | Unmanned aerial vehicle flight obstacle detection method based on computer image recognition | |
CN111008555B (en) | Unmanned aerial vehicle image small and weak target enhancement extraction method | |
CN114764822A (en) | Image processing method and device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |