CN111985375B - Visual target tracking self-adaptive template fusion method - Google Patents
Visual target tracking self-adaptive template fusion method Download PDFInfo
- Publication number
- CN111985375B CN111985375B CN202010810873.4A CN202010810873A CN111985375B CN 111985375 B CN111985375 B CN 111985375B CN 202010810873 A CN202010810873 A CN 202010810873A CN 111985375 B CN111985375 B CN 111985375B
- Authority
- CN
- China
- Prior art keywords
- target
- template
- frame
- current frame
- fusion
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000007 visual effect Effects 0.000 title claims abstract description 15
- 238000007500 overflow downdraw method Methods 0.000 title claims abstract description 12
- 230000004044 response Effects 0.000 claims abstract description 31
- 230000004927 fusion Effects 0.000 claims abstract description 27
- 238000000034 method Methods 0.000 claims abstract description 25
- 238000004364 calculation method Methods 0.000 claims description 10
- 230000003044 adaptive effect Effects 0.000 claims description 8
- 238000000605 extraction Methods 0.000 claims description 8
- 238000012549 training Methods 0.000 claims description 7
- 230000006870 function Effects 0.000 claims description 6
- 238000004590 computer program Methods 0.000 claims description 4
- 238000013135 deep learning Methods 0.000 claims description 4
- 238000013461 design Methods 0.000 claims description 3
- 238000010586 diagram Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 claims description 3
- 230000008569 process Effects 0.000 abstract description 11
- 230000000694 effects Effects 0.000 abstract description 3
- 238000001914 filtration Methods 0.000 abstract description 3
- 241001465754 Metazoa Species 0.000 description 7
- 238000002474 experimental method Methods 0.000 description 6
- 238000004422 calculation algorithm Methods 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 230000006872 improvement Effects 0.000 description 2
- 230000005012 migration Effects 0.000 description 2
- 238000013508 migration Methods 0.000 description 2
- 241000271566 Aves Species 0.000 description 1
- 241000283690 Bos taurus Species 0.000 description 1
- 241001494479 Pecora Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000009975 flexible effect Effects 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 244000144980 herd Species 0.000 description 1
- 238000011423 initialization method Methods 0.000 description 1
- 244000144972 livestock Species 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 238000009304 pastoral farming Methods 0.000 description 1
- 230000001846 repelling effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000002904 solvent Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Evolutionary Biology (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a visual target tracking self-adaptive template fusion method, and belongs to the technical field of target tracking. The method is used for judging whether to update the template or not by calculating the ratio of the extreme value to the mean value of the response graph, and avoids using frames with poor quality for updating the template due to filtering out frames with weak target response degree, thereby improving the quality of the template and obtaining better tracking effect. According to the method, the fusion coefficient of the template is calculated in a self-adaptive manner, and the template which has stronger response with the current frame can obtain larger updating weight, so that the target state is updated, the pollution of target blurring and shielding (the response with the current frame is weaker at the moment and the weight during updating is small) to the template is reduced, the video data is utilized more fully, the problems of target deformation and background pollution to the template in the target tracking process are effectively inhibited, and the template quality in the tracking process is improved.
Description
Technical Field
The invention belongs to the technical field of target tracking, and particularly relates to a visual target tracking self-adaptive template fusion method.
Background
Target tracking is widely used in the production and life fields. Object tracking is an important component, both in military and civilian applications. The visual target tracking technology has important significance in the fields of ecological environment protection, flight safety, animal husbandry automation and the like. For example, in the aspect of bird repelling in airports, flying birds in airports can also cause hidden dangers to airlines, so that huge economic loss is caused, and serious threats are brought to the safety of passengers. Therefore, flying birds in an airport need to be tracked as a basis for driving away. In addition, the unmanned aerial vehicle as a new intelligent aircraft has the characteristics of flexible action, low requirements on take-off and flight, no restriction of places, high lift-off speed, long hold-up time, easiness in acquisition, long control distance and the like. Tracking aiming at the unmanned aerial vehicle cluster is also an important means for applying the unmanned aerial vehicle in a large quantity and avoiding the influence of the unmanned aerial vehicle on important facilities such as airports and the like. Meanwhile, in the field of animal husbandry and animal research, many animals such as cattle, sheep, birds, etc. also move in a herd. Whether the migration mode of animals is researched or livestock is prevented from being dislocated in the grazing process, the migration and movement information of the animals needs to be effectively acquired. Researchers often need to track these target groups and obtain the motion trail of each individual.
In the multi-target tracking task, if the space distances between the targets are very close, the situations of shielding between the targets or crossing of motion tracks and the like can occur, and confusion is easy to generate. The animal target may deform to different degrees during the movement process, for example, the deformation of birds when swinging wings, and the shape and the size of the target change make it difficult to ensure the adaptability of the target template to the change of the target state. If a static template is used, the tracking performance may be poor, and a general dynamic template is easily polluted by the background, for example, relevant filtering, which still updates the template according to the target areas when the target is disturbed by motion blur, partial occlusion, and the like, so that template pollution is easily generated.
Disclosure of Invention
Aiming at the defects and the improvement requirements of the prior art, the invention provides a visual target tracking adaptive template fusion method, which aims to better resist the interference condition in the tracking process by judging whether the current template is updated or not and adaptively calculating the template weighting coefficient on the other hand.
To achieve the above object, according to a first aspect of the present invention, there is provided a visual target tracking adaptive template fusion method, comprising the steps of:
s1, performing convolution calculation on a target feature map of a previous frame and a feature map of a current frame to obtain a response map of the previous frame of the target;
s2, calculating the ratio of the maximum value to the mean value of the response image of the previous frame of the target;
s3, when the ratio is smaller than or equal to a set threshold, directly using the target template of the previous frame as the target template of the current frame, and when the ratio is larger than the set threshold, weighting and fusing the target template of the previous frame and the target feature map of the previous frame by the target template of the current frame;
and S4, performing weighted fusion on the target template of the current frame and the target feature map of the first frame to obtain a fusion template of the current frame, and updating the target template of the current frame.
Preferably, the calculation formula of the response map of the last frame of the target is as follows:
wherein z represents a target image and s represents a current frame image;representing a feature extraction function; b represents a two-dimensional bias matrix and m represents a current frame number.
Preferably, the feature extraction function is obtained by adopting artificial design or deep learning network training.
Preferably, the target template calculation formula of the current frame is as follows:
wherein z represents the target image, m represents the frame number of the current frame, β represents the template updating coefficient, and ratio represents the maximum of the response graph of the previous frame of the targetA ratio of the value to the mean value, σ denotes a set threshold value for determining whether the target template should be updated,the target feature map of the previous frame is represented,target template representing the previous frame, target template of the first frameTarget feature map initialized to first frame
Preferably, β is 0.5 and σ is 1.
Preferably, the fusion template calculation formula of the current frame is as follows:
wherein,a fusion template representing the current frame is shown,a target template representing the current frame is shown,and representing the target characteristic diagram of the first frame, z represents a target image, m represents the frame number of the current frame, and lambda represents a template weighting coefficient.
Preferably, the template weighting factor calculation formula is as follows:
wherein f is1(z, s) representsCalculated response map, f2(z, s) representsAnd (4) calculating a response graph.
To achieve the above object, according to a second aspect of the present invention, there is provided a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the visual target tracking adaptive template fusion method according to the first aspect.
Generally, by the above technical solution conceived by the present invention, the following beneficial effects can be obtained:
(1) the method is used for judging whether to update the template or not by calculating the ratio of the extreme value to the mean value of the response graph, and avoids using frames with poor quality for updating the template due to filtering out frames with weak target response degree, thereby improving the quality of the template and obtaining better tracking effect.
(2) According to the method, the fusion coefficient of the template is calculated in a self-adaptive manner, and the template which has stronger response with the current frame can obtain larger updating weight, so that the target state is updated, the pollution of target blurring and shielding (the response with the current frame is weaker at the moment and the weight during updating is small) to the template is reduced, the video data is utilized more fully, the problems of target deformation and background pollution to the template in the target tracking process are effectively inhibited, and the template quality in the tracking process is improved.
Drawings
FIG. 1 is a flow chart of a visual target tracking adaptive template fusion method provided by the present invention;
FIG. 2 is a graph of statistical result versus accuracy provided by the present invention;
fig. 3 is a graph of the comparison success rate of the statistical results provided by the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. In addition, the technical features involved in the respective embodiments of the present invention described below may be combined with each other as long as they do not conflict with each other.
As shown in fig. 1, the present invention provides a visual target tracking adaptive template fusion method, which includes the following steps:
and S1, carrying out convolution calculation on the target feature map of the previous frame and the feature map of the current frame to obtain a response map of the previous frame of the target.
Preferably, the calculation formula of the response map of the last frame of the target is as follows:
wherein z represents a target image and s represents a current frame image;representing a feature extraction function; b represents a two-dimensional bias matrix and m represents a current frame number.
Preferably, the feature extraction function is obtained by adopting artificial design or deep learning network training. The present embodiment adopts a twin network as a target template feature extraction network, but the present invention is not limited to this feature extraction method in the use process.
Training process of full convolution twin network
The hardware environment comprises: a CPU with model number of Intel (R) core (TM) i7-6850K, 6 core, and CPU of 12 threads, and main frequency of 3.60 GHz; two GPUs with model number of Nvidia GTX 1080 Ti; and 64GB in memory. The software environment for the experiment included: ubuntu 16.04 operating system, deep learning framework Tensorflow. The network's training data uses the ILSVRC-VID dataset. All 4417 video data in this data set were used to train and adjust network parameters. The iterative training method adopts a random gradient descent method. The parameter initialization method is an Xavier method. The training iteration turns were 50, 50000 image pairs were trained per turn. The batch size was set to 32. The initial learning rate was set to 0.01, the learning rate exponentially declined, and the lowest learning rate was 0.00001. The maximum interval frame between the input picture pairs is 100. When not otherwise specified, β is empirically 0.5.
And S2, calculating the ratio of the maximum value to the mean value of the response image of the previous frame of the target.
And taking the ratio of the maximum value to the mean value of the response image of the previous frame of the target as the measure of the difference degree of the targets in the current frame and the previous frame, wherein the smaller the ratio is, the larger the difference degree is.
And S3, when the ratio is smaller than or equal to a set threshold, directly using the target template of the previous frame as the target template of the current frame, and when the ratio is larger than the set threshold, weighting and fusing the target template of the previous frame and the target feature map of the previous frame by the target template of the current frame.
When the ratio is less than or equal to the set threshold, the difference between the two is large, which is caused by target blurring or shielding, therefore, in order to not introduce background pollution into the target template, the target template edge of the current frame is used as the target template of the previous frame; when the ratio is larger than a set threshold, the difference degree is small, and the target template of the current frame is a target template of the previous frame and a target feature map weighted fusion map of the previous frame.
Preferably, the target template calculation formula of the current frame is as follows:
wherein z represents a target image, m represents a frame number of a current frame, β represents a template updating coefficient, ratio represents a ratio of a maximum value to a mean value of a response map of a frame above a target, σ represents a set threshold for judging whether the target template should be updated,the target feature map of the previous frame is represented,target template representing the previous frame, target template of the first frameTarget feature map initialized to first frame
Preferably, β is 0.5 and σ is 1.
And S4, carrying out weighted fusion on the target template of the current frame and the target characteristic image of the first frame to obtain a fusion template of the current frame, and updating the target template of the current frame.
Preferably, the fusion template calculation formula of the current frame is as follows:
wherein,a fusion template representing the current frame is shown,a target template representing the current frame is shown,representing the target characteristic diagram of the first frame, z representing the target image, m representing the frame number of the current frameAnd λ denotes a template weighting coefficient.
Preferably, the template weighting factor is calculated as follows:
According to the method, the target template of the current frame and the target feature map of the first frame are subjected to weighted fusion to obtain the fusion template of the current frame, so that the influence of target blurring or shielding on the template in the tracking process is avoided. In particular, the amount of the solvent to be used,
the smaller the target pollution is, the target template of the current frameAnd feature map of the current frameThe higher the similarity is, the response result f1The larger (z, s), the max (f)1(z, s)) the larger the ratio of the current frame template to the fusion template is; if the current frame is interfered, the feature change of the target is large, and the target template of the current frameAnd feature map of the current frameThe lower the similarity is, the response result f1The smaller (z, s), the max (f)1(z, s)) is smaller, the proportion of the current frame template to the fusion template is smaller, the proportion of the first frame feature map to the fusion template is larger, and the first frame feature map is used as the most accurate priori knowledge, so that the introduction of interference information in the template updating process is avoided.
The denominator of the template weighting factor is for normalization, so that the template weighting factor λ is in the [0,1] interval.
After obtaining the fusion template of the current frame, when further used for target tracking, the method also comprises the following steps:
calculating response maps and tracking results using convolutionGet the response chart f againSEAnd (z, s) as the tracking result of the current target at the maximum value. If the tracking of all the targets of the current frame is not finished, continuing to perform the next target i ═ i +1, returning to the step S1, otherwise, judging whether the image sequence or the video data is finished, if not, then m ═ m +1, returning to the step S1, if so, indicating that the tracking is finished, and outputting the positions of all the targets in the other frames except the first frame in the video.
Experiments were performed according to the parameters of table 1.
TABLE 1
Fig. 2 and fig. 3 are a comparison accuracy chart and a success rate chart of the performance statistics of the fusion template-based visual target tracking algorithm, respectively. The abscissa of the accuracy map is the center position error distance threshold in pixels. The ordinate is the distance accuracy corresponding to the threshold. The abscissa of the success rate graph is the intersection ratio threshold. The ordinate is the overlay accuracy corresponding to the threshold. Table 2 shows the statistical results of the experiment. Compared with experiments Test1 and Test2, the visual target tracking algorithm based on the fusion template has obvious advantages. For the experiment Fusion of the Fusion template, the accuracy curve area is improved by 0.1133, and the success rate curve area is improved by 0.0766, so that the effect of the Fusion template is very obvious. This experimental result demonstrates the effectiveness of the algorithm of the present invention. It can be seen that the curves of the inventive algorithm perform significantly better.
TABLE 2
Serial number | Name of experiment | Area of accuracy curve | Area of success rate curve |
1 | Test1 | 0.4859 | 0.4108 |
2 | Test2 | 0.4895 | 0.4094 |
3 | Fusion | 0.6266 | 0.5092 |
Furthermore, the present invention also provides a computer readable storage medium having stored thereon a computer program which, when being executed by a processor, implements the visual target tracking adaptive template fusion method according to the first aspect.
It will be understood by those skilled in the art that the foregoing is only a preferred embodiment of the present invention, and is not intended to limit the invention, and that any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the scope of the present invention.
Claims (8)
1. A visual target tracking adaptive template fusion method is characterized by comprising the following steps:
s1, performing convolution calculation on a target feature map of a previous frame and a feature map of a current frame to obtain a response map of the previous frame of the target;
s2, calculating the ratio of the maximum value to the mean value of the response image of the previous frame of the target;
s3, when the ratio is smaller than or equal to a set threshold, directly using the target template of the previous frame as the target template of the current frame, and when the ratio is larger than the set threshold, weighting and fusing the target template of the previous frame and the target feature map of the previous frame by the target template of the current frame;
and S4, carrying out weighted fusion on the target template of the current frame and the target characteristic image of the first frame to obtain a fusion template of the current frame, and updating the target template of the current frame.
2. The method of claim 1, wherein the response map of the previous frame of the target is calculated as follows:
3. The method of claim 2, wherein the feature extraction function is obtained by artificial design or deep learning network training.
4. A method as claimed in any one of claims 1 to 3, wherein the target template for the current frame is calculated as follows:
wherein z represents a target image, m represents a frame number of a current frame, β represents a template updating coefficient, ratio represents a ratio of a maximum value to a mean value of a response map of a frame above the target, λ represents a set threshold for judging whether the target template should be updated,the target feature map of the previous frame is represented,target template representing the previous frame, target template of the first frameTarget feature map initialized to first frame
5. The method of claim 4, wherein β is 0.5 and σ is 1.
6. The method of any one of claims 1 to 3, wherein the fusion template for the current frame is calculated as follows:
wherein,a fusion template representing the current frame is shown,a target template representing the current frame is shown,and representing the target characteristic diagram of the first frame, z represents a target image, m represents the frame number of the current frame, and lambda represents a template weighting coefficient.
8. A computer-readable storage medium, characterized in that a computer program is stored thereon, which computer program, when being executed by a processor, implements the visual target tracking adaptive template fusion method according to any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010810873.4A CN111985375B (en) | 2020-08-12 | 2020-08-12 | Visual target tracking self-adaptive template fusion method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010810873.4A CN111985375B (en) | 2020-08-12 | 2020-08-12 | Visual target tracking self-adaptive template fusion method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111985375A CN111985375A (en) | 2020-11-24 |
CN111985375B true CN111985375B (en) | 2022-06-14 |
Family
ID=73434184
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010810873.4A Active CN111985375B (en) | 2020-08-12 | 2020-08-12 | Visual target tracking self-adaptive template fusion method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111985375B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113129335B (en) * | 2021-03-25 | 2023-03-14 | 西安电子科技大学 | Visual tracking algorithm and multi-template updating strategy based on twin network |
CN114372997B (en) * | 2021-12-06 | 2024-09-10 | 燕山大学 | Target tracking method based on quality and similarity evaluation online template updating |
CN115731516A (en) * | 2022-11-21 | 2023-03-03 | 国能九江发电有限公司 | Behavior recognition method and device based on target tracking and storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107767405A (en) * | 2017-09-29 | 2018-03-06 | 华中科技大学 | A kind of nuclear phase for merging convolutional neural networks closes filtered target tracking |
CN110084836A (en) * | 2019-04-26 | 2019-08-02 | 西安电子科技大学 | Method for tracking target based on the response fusion of depth convolution Dividing Characteristics |
CN111161324A (en) * | 2019-11-20 | 2020-05-15 | 山东工商学院 | Target tracking method based on adaptive multi-mode updating strategy |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180129934A1 (en) * | 2016-11-07 | 2018-05-10 | Qualcomm Incorporated | Enhanced siamese trackers |
-
2020
- 2020-08-12 CN CN202010810873.4A patent/CN111985375B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107767405A (en) * | 2017-09-29 | 2018-03-06 | 华中科技大学 | A kind of nuclear phase for merging convolutional neural networks closes filtered target tracking |
CN110084836A (en) * | 2019-04-26 | 2019-08-02 | 西安电子科技大学 | Method for tracking target based on the response fusion of depth convolution Dividing Characteristics |
CN111161324A (en) * | 2019-11-20 | 2020-05-15 | 山东工商学院 | Target tracking method based on adaptive multi-mode updating strategy |
Non-Patent Citations (2)
Title |
---|
一种改进的SAD目标跟踪算法;赵柏山等;《微处理机》;20181231(第002期);全文 * |
特征融合自适应目标跟踪;钟国崇等;《图学学报》;20181015(第05期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN111985375A (en) | 2020-11-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111985375B (en) | Visual target tracking self-adaptive template fusion method | |
Baum et al. | Extended object and group tracking with elliptic random hypersurface models | |
CN109102511B (en) | Cerebrovascular segmentation method, system and electronic equipment | |
CN109859209B (en) | Remote sensing image segmentation method and device, storage medium and server | |
CN110442129A (en) | A kind of control method and system that multiple agent is formed into columns | |
Zhang et al. | A bionic dynamic path planning algorithm of the micro UAV based on the fusion of deep neural network optimization/filtering and hawk-eye vision | |
Li et al. | Sequential dynamic leadership inference using Bayesian Monte Carlo methods | |
CN111488552A (en) | Close-proximity multi-target tracking method based on Gaussian mixture probability hypothesis density | |
CN109190693B (en) | Variant target high-resolution range profile recognition method based on block sparse Bayesian learning | |
CN115533897A (en) | Mechanical arm obstacle avoidance planning method based on dynamic system and Gaussian cluster ellipsoid | |
Ollinger et al. | Maximum likelihood reconstruction in fully 3D PET via the SAGE algorithm | |
CN114565861A (en) | Airborne downward-looking target image positioning method based on probability statistic differential homoembryo set matching | |
CN103584888B (en) | Ultrasonic target motion method for tracing | |
CN110298868A (en) | A kind of multiscale target tracking of high real-time | |
CN112215869A (en) | Group target tracking method and system based on graph similarity constraint | |
CN117116096A (en) | Airport delay prediction method and system based on multichannel traffic image and depth CNN | |
CN113379666B (en) | Image fusion method based on interval type fuzzy set and CHMM model | |
CN116560401A (en) | Method for determining control instruction of plane in unmanned plane formation and terminal equipment | |
CN111582299B (en) | Self-adaptive regularization optimization processing method for image deep learning model identification | |
Godsill | Particle filters for continuous-time jump models in tracking applications | |
Shaikh et al. | Otsu-Based Differential Evolution Method for Image Segmentation | |
CN105184811B (en) | Multi-scale target tracking method based on structure support vector machine | |
Tongyan et al. | Drogue detection for autonomous aerial refueling via hybrid pigeon-inspired optimized color opponent and saliency aggregation | |
CN111461160B (en) | Infrared imaging seeker target tracking method for preventing cloud and fog interference | |
Yang et al. | V-formation as optimal control |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |