变分调整约束下的反向低秩稀疏学习目标跟踪
Object tracking via reverse low-rank sparse learning with variation regularization
- 2020年25卷第6期 页码:1142-1149
收稿:2019-06-21,
修回:2019-10-30,
录用:2019-11-6,
纸质出版:2020-06-16
DOI: 10.11834/jig.190271
移动端阅览

浏览全部资源
扫码关注微信
收稿:2019-06-21,
修回:2019-10-30,
录用:2019-11-6,
纸质出版:2020-06-16
移动端阅览
目的
2
低秩稀疏学习目标跟踪算法在目标快速运动和严重遮挡等情况下容易出现跟踪漂移现象,为此提出一种变分调整约束下的反向低秩稀疏学习目标跟踪算法。
方法
2
采用核范数凸近似低秩约束描述候选粒子间的时域相关性,去除不相关粒子,适应目标外观变化。通过反向稀疏表示描述目标表观,用候选粒子稀疏表示目标模板,减少在线跟踪中L
1
优化问题的数目,提高跟踪效率。在有界变差空间利用变分调整对稀疏系数差分建模,约束目标表观在相邻帧间具有较小变化,但允许连续帧间差异存在跳跃不连续性,以适应目标快速运动。
结果
2
实验利用OTB(object tracking benchmark)数据集中的4组涵盖了严重遮挡、快速运动、光照和尺度变化等挑战因素的标准视频序列进行测试,定性和定量对比了本文算法与5种热点算法的跟踪效果。定性分析基于视频序列的主要挑战因素进行比较,定量分析通过中心点位置误差(central pixel error,CPE)比较跟踪算法的精度。与CNT(convolutional networks training)、SCM(sparse collaborative model)、IST(inverse sparse tracker)、DDL(discriminative dictionary learning)和LLR(locally low-rank representation)算法相比,平均CPE值分别提高了2.80、4.16、13.37、35.94和41.59。实验结果表明,本文算法达到了较高的跟踪精度,对上述挑战因素更具鲁棒性。
结论
2
本文提出的跟踪算法,综合了低秩稀疏学习和变分优化调整的优势,在复杂场景下具有较高的跟踪精度,特别是对严重遮挡和快速运动情况的有效跟踪更具鲁棒性。
Objective
2
Visual object tracking is an important research subject in computer vision. It has extensive applications that include surveillance
human-computer interaction
and medical imaging. The goal of tracking is to estimate the states of a moving target in a video sequence. Considerable effort has been devoted to this field
but many challenges remain due to appearance variations caused by heavy occlusion
illumination variation
and fast motion. Low-rank approximation can acquire the underlying structure of a target because some candidate particles have extremely similar appearance. This approximation can prune irrelevant particles and is robust to global appearance changes
such as pose change and illumination variation. Sparse representation formulates candidate particles by using a linear combination of a few dictionary templates. This representation is robust against local appearance changes
e.g.
partial occlusions. Therefore
combining low-rank approximation with sparsity representation can improve the efficiency and effectiveness of object tracking. However
object tracking via low-rank sparse learning easily results in tracking drift when facing objects with fast motion and severe occlusions. Therefore
reverse low-rank sparse learning with a variation regularization-based tracking algorithm is proposed.
Method
2
First
a low-rank constraint is used to restrain the temporal correlation of the objective appearance
and thus
remove the uncorrelated particles and adapt the object appearance change. The rank minimization problem is known to be computationally intractable. Hence
we resort to minimizing its convex envelope via a nuclear norm. Second
the traditional sparse representation method requires solving numerous L
1
optimization problems. Computational cost increases linearly with the number of candidate particles. We build an inverse sparse representation formulation for object appearance using candidate particles to represent the target template inversely
reducing the number of L
1
optimization problems for online tracking from candidate particles to one. Third
variation regularization is introduced to model the sparse coefficient difference. The variation method can model the variable selection problem in bounded variation space
which can restrict object appearance with only a slight difference between consecutive frames but allow the existing difference between individual frames to jump discontinuously
and thus
adapt to fast object motion. Lastly
an online updating scheme based on alternating iteration is proposed for tracking computation. Each iteration updates one variable at a time. Meanwhile
the other variables are fixed to their latest values. To accommodate target appearance change
we also use a local updating scheme to update the local parts individually. This scheme captures changes in target appearance even when heavy occlusion occurs. In such case
the unoccluded local parts are still updated in the target template and the occluded ones are discarded. Consequently
we can obtain a representation coefficient for the observation model and realize online tracking.
Result
2
To evaluate our proposed tracker
qualitative and quantitative analyses are performed using MATLAB on benchmark tracking sequences (occlusion 1
David
boy
deer) obtained from OTB (object tracking benchmark) datasets. The selected videos include many challenging factors in visual tracking
such as occlusion
fast motion
illumination
and scale variation. The experimental results show that when faced with these challenging situations in the benchmark tracking dataset
the proposed algorithm can perform tracking effectively in complicated scenes. Comparative studies with five state-of-the-art visual trackers
namely
DDL (discriminative dictionary learning)
SCM (sparse collaborative model)
LLR (locally low-rank representation)
IST (inverse sparse tracker)
and CNT (convolutional networks training) are conducted. To achieve fair comparison
we use publicly available source codes or results provided by the authors. Among these trackers
DDL
SCM
LLR
and IST are the most relevant. Compared with the CNT tracker
we mostly consider that deep networks have attracted considerable attention in complicated visual tracking tasks. For qualitative comparison
the representative tracking results are discussed on the basis of the major challenging factors in each video. For quantitative comparison
the central pixel error (CPE)
which records the Euclidean distance between the central location of the tracked target and the manually labeled ground truth
is used. When the metric value is smaller
the tracking results are more accurate. From the evolution process of CPE versus frame number
our tracker achieves the best results in these challenging sequences. In particular
our tracker outperforms the SCM
IST
and CNT trackers in terms of occlusion
illumination
and scale variation sequences. It outperforms the LLR
IST
and DDL trackers in terms of fast motion sequences. These results demonstrate the effectiveness and robustness of our tracker to occlusion
illumination
scale variation
and fast motion.
Conclusion
2
Qualitative and quantitative evaluations demonstrate that the proposed algorithm achieves higher precision compared with many state-of-the-art algorithms. In particular
it exhibits better adaptability for objects with fast motion. In the future
we will extend our tracker by applying deep learning to enhance its discriminatory ability.
Elnakeeb A and Mitra U. 2018. Line constrained estimation with applications to target tracking:exploiting sparsity and low-rank. IEEE Transactions on Signal Processing, 66(24):6488-6502[DOI:10.1109/TSP.2018.2876305]
Fan B J, Li X M, Cong Y and Tang Y D. 2018. Structured and weighted multi-task low rank tracker. Pattern Recognition, 81:528-544[DOI:10.1016/j.patcog.2018.04.002]
He Y J, Li M, Zhang J L and Yao J P. 2016. Infrared target tracking based on robust low-rank sparse learning. IEEE Geoscience and Remote Sensing Letters, 13(2):232-236[DOI:10.1109/LGRS.2015.2506758]
He Z Y, Yi S Y, Cheung Y M, You X and Tang Y Y. 2017. Robust object tracking via key patch sparse representation. IEEE Transactions on Cybernetics, 47(2):354-364[DOI:10.1109/TCYB.2016.2514714]
Larsson V and Olsson C. 2016. Convex low rank approximation. International Journal of Computer Vision, 120(2):194-214[DOI:10.1007/s11263-016-0904-7]
Qi Y K, Qin L, Zhang J, Zhang S P, Huang Q M and Yang M H. 2018. Structure-aware local sparse coding for visual tracking. IEEE Transactions on Image Processing, 27(8):3857-3869[DOI:10.1109/TIP.2018.2797482]
Sui Y, Wang G H, Zhang L and Yang M H. 2018. Exploiting spatial-temporal locality oftracking via structured dictionary learning. IEEE Transactions on Image Processing, 27(3):1282-1296[DOI:10.1109/TIP.2017.2779275]
Sui Y and Zhang L. 2016. Robust tracking via locally structured representation. International Journal of Computer Vision, 119(2):110-144[DOI:10.1007/s11263-016-0881-x]
Tian D, Zhang G S and Sun S S. 2019. Object tracking based on fused Lasso with inverse low-rank and sparse constraint. Journal of Tianjin University (Science and Technology), 52(12):1278-1284
田丹, 张国山, 孙申申. 2019.反向低秩稀疏约束下的融合Lasso目标跟踪算法.天津大学学报(自然科学与工程技术版), 52(12):1278-1284[DOI:10.11784/tdxbz201811035]
Wang D, Lu H C, Xiao Z Y and Yang M H. 2015. Inverse sparse tracker with a locally weighted distance metric. IEEE Transactions on Image Processing, 24(9):2646-2657[DOI:10.1109/TIP.2015.2427518]
Yang Y H, Hu W R, Xie Y, Zhang W S and Zhang T Z. 2017. Temporal restricted visual tracking via reverse-low-rank sparse learning. IEEE Transactions on Cybernetics, 47(2):485-498[DOI:10.1109/TCYB.2016.2519532]
Zhang K H, Liu Q S, Wu Y and Yang M H. 2016. Robust visual tracking via convolutional networks without training. IEEE Transactions on Image Processing, 25(4):1779-1792[DOI:10.1109/TIP.2016.2531283]
Zhang T Z, Liu S, Ahuja N, Yang M H and Ghanem B. 2015. Robust visual tracking via consistent low-rank sparse learning. International Journal of Computer Vision, 111(2):171-190[DOI:10.1007/s11263-014-0738-0]
Zhao Z Q, Feng P, Wang T J, Liu F, Yang C H, Guo J J, Zhao Z J and Cui Z M. 2017. Dual-scale structural local sparse appearance model for robust object tracking. Neurocomputing, 237:101-113[DOI:10.1016/j.neucom.2016.09.031]
Zhong W, Lu H C and Yang M H. 2014. Robust object tracking via sparse collaborative appearance model. IEEE Transactions on Image Processing, 23(5):2356-2368[DOI:10.1109/TIP.2014.2313227]
Zhou T, Liu F H, Bhaskar H and Yang J. 2018. Robust visual tracking via online discriminative and low-rank dictionary learning. IEEE Transactions on Cybernetics, 48(9):2643-2655[DOI:10.1109/TCYB.2017.2747998]
相关作者
相关机构
京公网安备11010802024621