引入高斯掩膜的相关滤波目标跟踪算法
A Gaussian mask-based correlation filter tracking algorithm
- 2023年28卷第10期 页码:3092-3106
纸质出版日期: 2023-10-16
DOI: 10.11834/jig.220856
移动端阅览
浏览全部资源
扫码关注微信
纸质出版日期: 2023-10-16 ,
移动端阅览
王法胜, 尹双双, 贺冰, 孙福明. 2023. 引入高斯掩膜的相关滤波目标跟踪算法. 中国图象图形学报, 28(10):3092-3106
Wang Fasheng, Yin Shuangshuang, He Bing, Sun Fuming. 2023. A Gaussian mask-based correlation filter tracking algorithm. Journal of Image and Graphics, 28(10):3092-3106
目的
2
由于相关滤波(correlation filter,CF)跟踪算法使用循环移位操作增加训练样本,不可避免会引起边界效应问题。研究者大多采用余弦窗来抑制边界效应,但余弦窗的引入会导致样本污染,降低了算法的性能。为解决该问题,提出了嵌入高斯形状掩膜的相关滤波跟踪算法。
方法
2
在空间正则化相关滤波跟踪框架中嵌入高斯形状掩膜,对靠近目标中心和搜索区域边缘的样本重新分配权重,提高中心样本的重要性,同时降低远离中心的边缘样本的重要性,增加中心样本的响应,从而降低样本污染的影响,增强滤波器的判别能力。建立了高斯掩膜相关滤波跟踪算法的目标公式,然后使用交替方向乘子法(alternating direction of multiplier method,ADMM)求解滤波器及空间权重的闭合解。
结果
2
为评估所提算法的跟踪性能,在OTB2013(online tracking benchmark)、TC128(temple color)、UAV123(unmmaned aerial vehicle)及Got-10k(general object tracking)等多个基准数据集上进行了大量实验,并与多个先进的相关滤波跟踪算法对比。结果表明,在OTB2013数据集上精度和成功率分别为90.2%和65.2%,其中精度比基准算法提高0.5%;在TC128数据集上精度和成功率分别为77.9%和57.7%,其中成功率提高0.4%; UAV123数据集上的两个指标数据分别为74.1%和50.8%,精度提高0.3%;在Got-10k数据集上成功率提高了0.2%。
结论
2
与其他相关滤波跟踪算法相比,本文算法在各数据集上的精度和成功率表现出较强的竞争力,显著提高了基准算法的跟踪性能。
Objective
2
Object tracking has been concerned more as such a domain in relevance to computer vision. In the past decade, correlation filter has been recognized for object tracking due to its high accuracy and speed. For traditional correlation filter tracking algorithms, training sample set is generated using circulant shift operation on the underlying image patches, and the circulant matrix is used to compute in the Fourier domain as well, which can improve the learning efficiency further. However, due to Fourier transform is used periodically, the boundary of the image is not involved in computing. Consequently, the boundary of the object-tracking sample is discontinuous, resulting in the annoying boundary effect. The error and inquality of samples can be generated, which affects the filter ability severely. Two sorts of popular solutions are employed to resolve the boundary effects currently. The first one is oriented to draw real negative samples from the sampling area. But, all of the negative and positive samples are assigned equal weights, which cannot well suppress the boundary effect. The other one is focused on an integrated spatial regularization term relevant to the correlation filter framework. A defined spatial weight map is used to deal with boundary area-near spatial weights of the samples. However, spatial regularization weight map is defined in the first frame, and such related invariant can be kept during tracking, which is not melted into the variation of different object. Both of the two sorts of solutions are required to suppress boundary effects using cosine window. The introduction of cosine window is linked to the problem of sample contamination simultaneously. The generated samples are distorted by redundant pixels, for which the tracking performance is eroded. To resolve this problem, we develop a Gaussian mask-based correlation filter tracking algorithm.
Method
2
Gaussian shape mask is embedded into the spatially regularized correlation filter framework. It redefines the weight of the samples close to the object center and those next to the boundary of the sampling region, which strengthens the weight of the samples the object center and weakens the weight of the samples far away from the object center. Thus, the response of the samples close to the object center could be enhanced while those boundary area-near samples are suppressed. To improve the tracking performance, its related strategy can be implemented to suppress the boundary effect and shrink cosine window-caused sample contamination problem. The objective function of the proposed Gaussian mask correlation filter is configured out, and to obtain closed-form solutions of the filter and spatial weights, it is then solved using alternating direction of multiplier method. To deal with scale estimation problem, two sort of correlation filters are trained in terms of location and scale filters. The location filter is trained using histogram of gradient (HOG) feature and Visual Geometry Group (VGG) network-derived deep features, while the scale filter is trained using five-scale HOG features. The former one is mainly used for estimating the object location, while the latter one is used to determine the best scale of the target object.
Result
2
To evaluate the proposed Gaussian mask correlation filter, extensive experiments are carried out on several tracking benchmark datasets relevant to OTB2013, TC128, UAV123 and Got-10k, and comparative analysis is linked to such popular correlation filter based tracking methods like LADCF, ASRCF, RHCF and ECO. We use the general one-pass evaluation standard to evaluate all the tracking methods. For the evaluation metrics, the precision score and success rate are used and all the tracking methods are ranked according to the precision score in the precision plots and area under curve score in the success plots. The experimental results demonstrate that each of the OTB2013 dataset-relevant precision and success score is reached to 90.2% and 65.2%, in which the precision score is optimized by 0.5% over the baseline tracking method. For TC128 dataset, the values of the two metrics are 77.9% and 57.7%, where the success score is improved by 0.4 over the baseline tracking method. For UAV123 dataset, each of the two sorts of scores can be reached to 74.1% and 50.8%, where the precision score is improved by 0.3% compared to the baseline tracking method. For Got-10k dataset, the success rate is improved by 0.2% compared to the baseline tracking method. The qualitative results illustrate that our proposed method can be used to deal with such challenging issues like deformation, occlusion, out of view and fast motion. However, the limitation shows that computational cost of the proposed method is inevitable due to the introduction of the Gaussian mask. In addition, VGG network based deep features can increase the computational cost as well. For Got-10k dataset, the frame rate is about 5.2 frames per second, while the baseline tracking method is 5.43 frames per second, which realizes 0.23 frames per second higher.
Conclusion
2
The Gaussian mask is melted into the correlation filter framework, which can suppress the boundary effect and resolve the sample contamination problem. The sample weights are redistributed to highlight the samples near the object center, while the weight of the samples far away from the object center are suppressed. Extensive experimental results demonstrate that the proposed strategy is effective to some extent. Compared to other related correlation filter based trackers, our proposed method has its potentials on four public benchmark datasets in terms of tracking precision and success rate, which can improves the tracking performance of the baseline tracking method further. Guassian mask can also be extended to other related correlation filter based tracking framework potentially. The introduction of the Gaussian mask can decrease the tracking speed. Our future work will focus on this problem.
目标跟踪相关滤波(CF)边界效应样本污染高斯形状掩膜
object trackingcorrelation filter (CF)boundary effectsample contaminationGaussian shape mask
Abbass M Y, Kwon K C, Kim N, Abdelwahab S A, El-Samie F E A and Khalaf A A M. 2021. A survey on online learning for visual tracking. The Visual Computer, 37(5): 993-1014 [DOI: 10.1007/s00371-020-01848-yhttp://dx.doi.org/10.1007/s00371-020-01848-y]
Bertinetto L, Valmadre J, Golodetz S, Miksik O and Torr P H S. 2016. Staple: complementary learners for real-time tracking//Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, USA: IEEE: 1401-1409 [DOI: 10.1109/CVPR.2016.156http://dx.doi.org/10.1109/CVPR.2016.156]
Bolme D S, Beveridge J R, Draper B A and Lui Y M. 2010. Visual object tracking using adaptive correlation filters//Proceedings of 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. San Francisco, USA: IEEE: 2544-2550 [DOI: 10.1109/CVPR.2010.5539960http://dx.doi.org/10.1109/CVPR.2010.5539960]
Dai K N, Wang D, Lu H C, Sun C and Li J H. 2019. Visual tracking via adaptive spatially-regularized correlation filters//Proceedings of 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Long Beach, USA: IEEE: 4665-4674 [DOI: 10.1109/CVPR.2019.00480http://dx.doi.org/10.1109/CVPR.2019.00480]
Danelljan M, Bhat G, Khan F S and Felsberg M. 2017a. ECO: efficient convolution operators for tracking//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, USA: IEEE: 6931-6939 [DOI: 10.1109/CVPR.2017.733http://dx.doi.org/10.1109/CVPR.2017.733]
Danelljan M, Häger G, Khan F S and Felsberg M. 2015. Learning spatially regularized correlation filters for visual tracking//Proceedings of 2015 IEEE International Conference on Computer Vision (ICCV). Santiago, Chile: IEEE: 4310-4318 [DOI: 10.1109/ICCV.2015.490http://dx.doi.org/10.1109/ICCV.2015.490]
Danelljan M, Häger G, Khan F S and Felsberg M. 2017b. Discriminative scale space tracking. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(8): 1561-1575 [DOI: 10.1109/TPAMI.2016.2609928http://dx.doi.org/10.1109/TPAMI.2016.2609928]
Elayaperumal D and Joo Y H. 2021. Aberrance suppressed spatio-temporal correlation filters for visual object tracking. Pattern Recognition, 115: #107922 [DOI: 10.1016/j.patcog.2021.107922http://dx.doi.org/10.1016/j.patcog.2021.107922]
Feng W, Han R Z, Guo Q, Zhu J K and Wang S. 2019. Dynamic saliency-aware regularization for correlation filter-based object tracking. IEEE Transactions on Image Processing, 28(7): 3232-3245 [DOI: 10.1109/TIP.2019.2895411http://dx.doi.org/10.1109/TIP.2019.2895411]
Fu C H, Xu J T, Lin F L, Guo F Y, Liu T C and Zhang Z J. 2020. Object saliency-aware dual regularized correlation filter for real-time aerial tracking. IEEE Transactions on Geoscience and Remote Sensing, 58(12): 8940-8951 [DOI: 10.1109/TGRS.2020.2992301http://dx.doi.org/10.1109/TGRS.2020.2992301]
Galoogahi H K, Fagg A and Lucey S. 2017. Learning background-aware correlation filters for visual tracking//Proceedings of 2017 IEEE International Conference on Computer Vision (ICCV). Venice, Italy: IEEE: 1144-1152 [DOI: 10.1109/ICCV.2017.129http://dx.doi.org/10.1109/ICCV.2017.129]
Han R Z, Guo Q and Feng W. 2018. Content-related spatial regularization for visual object tracking//Proceedings of 2018 IEEE International Conference on Multimedia and Expo (ICME). San Diego, USA: IEEE: 1-6 [DOI: 10.1109/ICME.2018.8486487http://dx.doi.org/10.1109/ICME.2018.8486487]
Huang L H, Zhao X and Huang K Q. 2021. GOT-10k: a large high-diversity benchmark for generic object tracking in the wild. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(5): 1562-1577 [DOI: 10.1109/TPAMI.2019.2957464http://dx.doi.org/10.1109/TPAMI.2019.2957464]
Huang Z Y, Fu C H, Li Y M, Lin F L and Lu P. 2019. Learning aberrance repressed correlation filters for real-time UAV tracking//Proceedings of 2019 IEEE/CVF International Conference on Computer Vision (ICCV). Seoul, Korea (South): IEEE: 2891-2900 [DOI: 10.1109/ICCV.2019.00298http://dx.doi.org/10.1109/ICCV.2019.00298]
Jiang W T, Tu C and Liu W J. 2021. Background and direction-aware correlation filter tracking. Journal of Image and Graphics, 26(3): 527-541
姜文涛, 涂潮, 刘万军. 2021. 背景与方向感知的相关滤波跟踪. 中国图象图形学报, 26(3): 527-541 [DOI: 10.11834/jig.200139http://dx.doi.org/10.11834/jig.200139]
Li B W, Fu C H, Ding F Q, Ye J J and Lin F L. 2023. All-day object tracking for unmanned aerial vehicle. IEEE Transactions on Mobile Computing, 22(8): 4515-4529 [DOI: 10.1109/TMC.2022.3162892http://dx.doi.org/10.1109/TMC.2022.3162892]
Li F, Tian C, Zuo W M, Zhang L and Yang M H. 2018. Learning spatial-temporal regularized correlation filters for visual tracking//Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, USA: IEEE: 4904-4913 [DOI: 10.1109/CVPR.2018.00515http://dx.doi.org/10.1109/CVPR.2018.00515]
Li F, Wu X H, Zuo W M, Zhang D and Zhang L. 2020a. Remove cosine window from correlation filter-based visual trackers: when and how. IEEE Transactions on Image Processing, 29: 7045-7060 [DOI: 10.1109/TIP.2020.2997521http://dx.doi.org/10.1109/TIP.2020.2997521]
Li Y M, Fu C H, Ding F Q, Huang Z Y and Lu G. 2020b. AutoTrack: towards high-performance visual tracking for UAV with automatic spatio-temporal regularization//Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, USA: IEEE: 11920-11929 [DOI: 10.1109/CVPR42600.2020.01194http://dx.doi.org/10.1109/CVPR42600.2020.01194]
Liang P P, Blasch E and Ling H B. 2015. Encoding color information for visual tracking: algorithms and benchmark. IEEE Transactions on Image Processing, 24(12): 5630-5644 [DOI: 10.1109/TIP.2015.2482905http://dx.doi.org/10.1109/TIP.2015.2482905]
Liu S, Liu D Y, Srivastava G, Połap D and Woźniak M. 2021. Overview and methods of correlation filter algorithms in object tracking. Complex and Intelligent Systems, 7(4): 1895-1917 [DOI: 10.1007/s40747-020-00161-4http://dx.doi.org/10.1007/s40747-020-00161-4]
Mueller M, Smith N and Ghanem B. 2016. A benchmark and simulator for UAV tracking//Proceedings of the 14th European Conference on Computer Vision. Amsterdam, the Netherland: Springer International Publishing: 445-461 [DOI: 10.1007/978-3-319-46448-0_27http://dx.doi.org/10.1007/978-3-319-46448-0_27]
Pu L, Feng X X, Hou Z Q, Zha Y F and Yu W S. 2020. Correlation filter tracking based on deep spatial regularization. Acta Electronica Sinica, 48(10): 2025-2032
蒲磊, 冯新喜, 侯志强, 查宇飞, 余旺盛. 2020. 基于深度空间正则化的相关滤波跟踪算法. 电子学报, 48(10): 2025-2032 [DOI: 10.3969/j.issn.0372-2112.2020.10.021http://dx.doi.org/10.3969/j.issn.0372-2112.2020.10.021]
Qiang Z and Shi F H. 2020. Dynamic update correlation filter tracking based on appearance representation analysis. Journal of Image and Graphics, 25(6): 1209-1220
强壮, 石繁槐. 2020. 外观表征分析下动态更新相关滤波跟踪. 中国图象图形学报, 25(6): 1209-1220 [DOI: 10.11834/jig.190454http://dx.doi.org/10.11834/jig.190454]
Qiu Z L, Zha Y F, Wu M and Wang Q. 2020. Learning attentional regularized correlation filter for visual tracking. Acta Electronica Sinica, 48(9): 1762-1768
仇祝令, 查宇飞, 吴敏, 王青. 2020. 基于注意力学习的正则化相关滤波跟踪算法. 电子学报, 48(9): 1762-1768 [DOI: 10.3969/j.issn.0372-2112.2020.09.014http://dx.doi.org/10.3969/j.issn.0372-2112.2020.09.014]
Wu Y, Lim J and Yang M H. 2013. Online object tracking: a benchmark//Proceedings of 2013 IEEE Conference on Computer Vision and Pattern Recognition. Portland, USA: IEEE: 2411-2418 [DOI: 10.1109/CVPR.2013.312http://dx.doi.org/10.1109/CVPR.2013.312]
Xu T Y, Feng Z H, Wu X J and Kittler J. 2019. Learning adaptive discriminative correlation filters via temporal consistency preserving spatial feature selection for robust visual object tracking. IEEE Transactions on Image Processing, 28(11): 5596-5609 [DOI: 10.1109/TIP.2019.2919201http://dx.doi.org/10.1109/TIP.2019.2919201]
Xu T Y, Feng Z H, Wu X J and Kittler J. 2021. Adaptive channel selection for robust visual object tracking with discriminative correlation filters. International Journal of Computer Vision, 129(5): 1359-1375 [DOI: 10.1007/s11263-021-01435-1http://dx.doi.org/10.1007/s11263-021-01435-1]
Ye J J, Fu C H, Lin F L, Ding F Q, An S and Lu G. 2022. Multi-regularized correlation filter for UAV tracking and self-localization. IEEE Transactions on Industrial Electronics, 69(6): 6004-6014 [DOI: 10.1109/TIE.2021.3088366http://dx.doi.org/10.1109/TIE.2021.3088366]
Yuan D, Shu X and He Z Y. 2020. TRBACF: learning temporal regularized correlation filters for high performance online visual object tracking. Journal of Visual Communication and Image Representation, 72: #102882 [DOI: 10.1016/j.jvcir.2020.102882http://dx.doi.org/10.1016/j.jvcir.2020.102882]
Zeng M Y, Shang Z H, Liu H and Li J P. 2020. Target tracking algorithm based on adaptive updating of multilayer convolution features. Laser and Optoelectronics Progress, 57(2): #021008
曾梦媛, 尚振宏, 刘辉, 李健鹏. 2020. 融合多层卷积特征自适应更新的目标跟踪算法. 激光与光电子学进展, 57(2): #021008 [DOI: 10.3788/LOP.021008http://dx.doi.org/10.3788/LOP.021008]
Zhang Y H, Yang Y J, Zhou W N, Shi L F and Li D M. 2018. Motion-aware correlation filters for online visual tracking. Sensors, 18(11): #3937 [DOI: 10.3390/s18113937http://dx.doi.org/10.3390/s18113937]
Zhao J W, Wei F Y, Chen N N and Zhou Z H. 2022. Spatial and long-short temporal attention correlation filters for visual tracking. IET Image Processing, 16(11): 3011-3024 [DOI: 10.1049/ipr2.12535http://dx.doi.org/10.1049/ipr2.12535]
Zheng G Z, Fu C H, Ye J J, Lin F L and Ding F Q. 2021. Mutation sensitive correlation filter for real-time UAV tracking with adaptive hybrid label//Proceedings of 2021 IEEE International Conference on Robotics and Automation (ICRA). Xi’an, China: IEEE: 503-509 [DOI: 10.1109/ICRA48506.2021.9561931http://dx.doi.org/10.1109/ICRA48506.2021.9561931]
相关作者
相关机构