Abstract
In visual tracking, correlation Filters (CFs) have attracted increasing research attention and achieved superior performance. However, owing to the larger search area, more background information is introduced to the shifted samples, meaning that tracking errors are prone to appear in the detection stage. Accordingly, in this work, firstly, hand-crafted features and deep features extracted from pre-trained convolutional networks are combined to improve the representation ability of object appearance. For deep features, we use two different VGG networks for extraction. Secondly, in an attempt to solve the problem of the object background of the traditional CF model not being modeled over time, and owing to the lack of spatial-temporal information of the image, we propose a new background-aware and spatial-temporal regularized correlation filters model (BSTCF) that introduces the background constraint and spatial-temporal regularization. The proposed BSTCF can effectively model not only the background but also variations in the background over time. Finally, we transform the objective function of BSTCF into an unconstrained Augmented Lagrange multiplier formular to promote convergence to the global optimum solution. Moreover, we adopt the alternating direction multiplier method (ADMM) to produce three sub-problems with closed-form solution, then propose a corresponding algorithm. Based on the above, we construct an intelligent tracking system and carry out extensive experiments to test its performance on OTB-2013, OTB-2015, TC128, UAV123, and VOT2016 public datasets. The experimental results demonstrate that the tracking algorithm achieves superior performance.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Zhang J, Yuan T, He Y, Wang J (2022) A background-aware correlation filter with adaptive saliency-aware regularization for visual tracking. Neural Comput Appl 34(8):6359–5376. https://doi.org/10.1007/s00521-021-06771-4
Li C, Sun X, Wang X, Zhang L, Tang J (2017) Grayscale-thermal object tracking via multitask laplacian sparse representation. IEEE Trans Syst Man Cybern: Syst 47(4):673–681
Zhang B, Li Z, Cao X, Ye Q, Chen C, Shen L, Perina A, Jill R (2017) Output constraint transfer for kernelized correlation filter in tracking. IEEE Trans Syst Man Cybern: Syst 47(4):693–703. https://doi.org/10.1109/TSMC.2016.2629509
Zhou Z, Zhang W, Zhao J (2019) Robust visual tracking using discriminative sparse collaborative map. Int J Mach Learn Cyber 10(11):3201–3212
Bruni V, Vitulano D (2014) An improvement of kernel-based object tracking based on human perception. IEEE Trans Syst Man Cybern: Syst 44(11):1474–1485
Zhang J, Sun J, Wang J, Li Z, Chen X (2022) An object tracking framework with recapture based on correlation filters and siamese networks. Comput Electr Eng 98:107730. https://doi.org/10.1016/j.compeleceng.2022.107730
Danelljan M, Hager G, Shahbaz KF, Felsberg M (2015) Learning spatially regularized correlation filters for visual tracking. In: Proceedings of the IEEE international conference on computer vision, pp 4310–4318
Cheng X, Zhang Y, Cui J, Zhou L (2016) Object tracking via temporal consistency dictionary learning. IEEE Trans Systems Man Cybern: Syst 47(4):628–638
Li F, Tian C, Zuo W, Zhang L, Yang M (2018) Learning spatial-temporal regularized correlation filters for visual tracking. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4904–4913
Ma C, Huang J, Yang X, Yang M (2015) Hierarchical convolutional features for visual tracking. In: Proceedings of the IEEE international conference on computer vision, pp 3074–3082
Zhang J, Sun J, Wang J, Yue X (2021) Visual object tracking based on residual network and cascaded correlation filters. J Ambient Intell Humanized Comput 12(8):8427–8440
Zhang J, Feng W, Yuan T, Wang J, Sangaiah A (2022) Scstcf: spatial-channel selection and temporal regularized correlation filters for visual tracking. Appl Soft Comput 118:108485. https://doi.org/10.1016/j.asoc.2022.108485
Danelljan M, Robinson A, Khan FS, Felsberg M (2016) Beyond correlation filters: learning continuous convolution operators for visual tracking. In: European conference on computer vision. Springer, pp 472–488
Henriques JF, Caseiro R, Martins P, Batista J (2014) High-speed tracking with kernelized correlation filters. IEEE Trans Pattern Anal Mach Intell 37(3):583–596
Galoogahi HK, Fagg A, Lucey S (2017) Learning background-aware correlation filters for visual tracking. In: 2017 IEEE international conference on computer vision (ICCV), pp 1144–1152. https://doi.org/10.1109/ICCV.2017.129
Wu Y, Lim J, Yang M (2013) Online object tracking: a benchmark. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2411–2418
Wu Y, Lim J, Yang M (2015) Object tracking benchmark. IEEE Trans Pattern Anal Mach Intell 37:1–1. https://doi.org/10.1109/TPAMI.2014.2388226
Liang P, Blasch E, Ling H (2015) Encoding color information for visual tracking: algorithms and benchmark. IEEE Trans Image Process 24(12):5630–5644
Mueller M, Smith N, Ghanem B (2016) A benchmark and simulator for uav tracking. In: European conference on computer vision. Springer, pp 445–461
Kristan M et al (2016) The visual object tracking vot2016 challenge results. In: Hua G, Jégou H (eds) Computer vision – ECCV 2016 workshops. Springer, pp 777–823
Zhang J, Jin X, Sun J, Wang J, Sangaiah AK (2020) Spatial and semantic convolutional features for robust visual object tracking. Multimed Tools Appl 79(21):15095–15115
Danelljan M, Häger G, Khan FS, Felsberg M (2015) Convolutional features for correlation filter based visual tracking. In: 2015 IEEE International conference on computer vision workshop (ICCVW), pp 621–629. https://doi.org/10.1109/ICCVW.2015.84https://doi.org/10.1109/ICCVW.2015.84
Danelljan M, Bhat G, Khan FS, Felsberg M (2017) Eco: efficient convolution operators for tracking. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 6931–6939. https://doi.org/10.1109/CVPR.2017.733
Nam H, Han B (2016) Learning multi-domain convolutional neural networks for visual tracking. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4293–4302
Bertinetto L, Valmadre J, Henriques JF, Vedaldi A, Torr PHS (2016) Fully-convolutional siamese networks for object tracking. In: European conference on computer vision. Springer, pp 850–865
Guo Q, Feng W, Zhou C, Huang R, Wan L, Wang S (2017) Learning dynamic siamese network for visual object tracking. In: 2017 IEEE international conference on computer vision (ICCV), pp 1781–1789. https://doi.org/10.1109/ICCV.2017.196
Li B, Yan J, Wu W, Zhu Z, Hu X (2018) High performance visual tracking with siamese region proposal network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8971–8980
Zhang J, Huang B, Ye Z, Kuang L, Ning X (2021) Siamese anchor-free object tracking with multiscale spatial attentions. Scientific Reports 11(1):22908. https://doi.org/10.1038/s41598-021-02095-4
Valmadre J, Bertinetto L, Henriques J, Vedaldi A, Torr PHS (2017) End-to-end representation learning for correlation filter based tracking. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 5000–5008. https://doi.org/10.1109/CVPR.2017.531
Bolme DS, Beveridge JR, Draper BA, Lui YM (2010) Visual object tracking using adaptive correlation filters. In: 2010 IEEE computer society conference on computer vision and pattern recognition. IEEE, pp 2544–2550
Henriques JF, Caseiro R, Martins P, Batista J (2012) Exploiting the circulant structure of tracking-by-detection with kernels. In: European conference on computer vision. Springer, pp 702–715
Bertinetto L, Valmadre J, Golodetz S, Miksik O, Torr PHS (2016) Staple: complementary learners for real-time tracking. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1401–1409
Danelljan M, Häger G, Khan F, Felsberg M (2014) Accurate scale estimation for robust visual tracking. In: British machine vision conference, Nottingham. Bmva Press, 1-5 September, 2014
Li Y, Zhu J (2014) A scale adaptive kernel correlation filter tracker with feature integration. In: European conference on computer vision. Springer, pp 254–265
Bhat G, Johnander J, Danelljan M, Khan FS, Felsberg M (2018) Unveiling the power of deep tracking. In: Computer vision – ECCV 2018, pp 493–509
Boyd S, Parikh N, Chu E (2011) Distributed optimization and statistical learning via the alternating direction method of multipliers. Now Publishers Inc
Wang N, Zhou W, Song Y, Ma J, Liu W, Liu H (2021) Unsupervised deep representation learning for real-time tracking. Int J Comput Vis 129(2):400–418. https://doi.org/10.1007/s11263-020-01357-4
Li P, Chen B, Ouyang W, Wang D, Yang X, Lu H (2019) GradNet: gradient-guided network for visual object tracking. In: 2019 IEEE/CVF international conference on computer vision (ICCV), pp 6161–6170. https://doi.org/10.1109/ICCV.2019.00626
Zhang T, Xu C, Yang M (2017) Multi-task correlation particle filter for robust object tracking. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 4819–4827. https://doi.org/10.1109/CVPR.2017.512
Feng S, Hu K, Fan E, Zhao L, Wu C (2021) Kalman filter for spatial-temporal regularized correlation filters. IEEE Trans Image Process 30:3263–3278. https://doi.org/10.1109/TIP.2021.3060164
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 61972056, the Basic Research Fund of Zhongye Changtian International Engineering Co., Ltd. under Grant 2020JCYJ07, the Postgraduate Training Innovation Base Construction Project of Hunan Province under Grant 2019-248-51, the “Double First-class” International Cooperation and Development Scientific Research Project of CSUST under Grant 2019IC34, the Enterprise-University Joint Postgraduate Scientific Research Innovation Fund of Hunan Province under Grant QL20210205, and the Postgraduate Scientific Research Innovation Fund of CSUST under Grant CX2021SS70.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declare no conflict of interest.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhang, J., He, Y., Feng, W. et al. Learning background-aware and spatial-temporal regularized correlation filters for visual tracking. Appl Intell 53, 7697–7712 (2023). https://doi.org/10.1007/s10489-022-03868-8
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03868-8