Nowadays, object tracking is one of the hot topics in computer vision and has been widely used in many engineering applications, such as satellites[
Chinese Optics Letters, Volume. 17, Issue 3, 031001(2019)
An improved long-term correlation tracking method with occlusion handling
By improving the long-term correlation tracking (LCT) algorithm, an effective object tracking method, improved LCT (ILCT), is proposed to address the issue of occlusion. If the object is judged being occluded by the designed criterion, which is based on the characteristic of response value curve, an added re-detector will perform re-detection, and the tracker is ordered to stop. Besides, a filtering and adoption strategy of re-detection results is given to choose the most reliable one for the re-initialization of the tracker. Extensive experiments are carried out under the conditions of occlusion, and the results demonstrate that ILCT outperforms some state-of-the-art methods in terms of accuracy and robustness.
Nowadays, object tracking is one of the hot topics in computer vision and has been widely used in many engineering applications, such as satellites[
In recent years, many robust trackers based on correlation filters were proposed, in which the minimum output sum of squared error (MOSSE) filter[
However, these trackers do not handle occlusion (OCC) well or only aim at partial OCC (50% coverage or less) and temporal full OCC. A robust tracking algorithm requires a detection module to recover the target from potential tracking failures caused by heavy OCC[
Sign up for Chinese Optics Letters TOC Get the latest issue of Advanced Photonics delivered right to you!Sign up now
ILCT uses the motion correlation filter
LCT decomposes the tracking task into translation estimating (to get a new position) and scale estimating (to get a new scale)[
This filter,
After mapping and fast Fourier transform (FFT)
When new frame comes, the filter will perform correlation on the new patch
The filter
Each scale has its size of
Therefore, the accepted tracking result of LCT must have two highest response values, i.e., the response values of
Figure 1.Response maps. (a) Object is intact; (b) object is occluded.
The motion correlation filter
Adopting the tracking-by-detection framework is also the critical factor showing that LCT is robust for SV, illumination variation (IV), background clutters (BCs), fast motion (FM), etc. An online support vector machine (SVM) classifier is used for recovering targets, and the color channels are quantized as features for detector learning[
In addition, the cosine window is used in translating estimation to remove the boundary discontinuities of the response map[
The tracking result is adopted according to the values of response maps. If the object is intact and undisturbed, the response map is clear, and the white point is obvious. On the contrary, the map is dim, and the point is obscure, for example, when OCC occurs, as shown in Fig.
When the OCC begins, the tracker may still locate the object successfully based on previous training. However, as time goes on, the coverage increases, which aggravates the correlation filter so that the tracker will fail to re-track the object after its quitting from OCC.
To design an OCC criterion, several sequences[
Figure 2.Curves of response values with different attributes. (a) Full OCC; (b) DEF; (c) partial OCC; (d) BCs; (e) FM; (f) IV; (g) SV. (Best view in PDF.)
In Fig.
When the OCC criterion triggers, we set five free frames so that no operation is carried out on these frames to (1) let the object be fully occluded, (2) avoid the filters being polluted, and (3) improve the real time.
For convenience, we name the frame and the time at which it reaches the OCC criterion as
In an image with a complex background, a huge number of proposal bounding boxes may be obtained, which takes more computation time, and most of them are not the object bounding boxes we want. So, a threshold is set that
While among these
Figure 3.Detection for proposal boxes. (a) Bounding box of object; (b)
In our algorithm,
The whole flowchart of our method is shown in Fig.
Figure 4.Flowchart of ILCT. (Best view in PDF.)
To demonstrate the performance of the improved tracker, experiments are performed on eight sequences with the attributes of OCC, etc. Eight state-of-the-art trackers are compared with ILCT. They are KCF[
The annotated attributes of the eight sequences include OCC, FM, moving camera (MC), SV, BCs, IV, DEF, out-of-plane rotation (OPR), and motion blur (MB). Their information is listed in Table
|
The parameters of the LCT part are set to the default values:
In the OCC criterion,
In the re-detector, we use the default parameters of edge boxes[
The tracking results of 12 trackers are shown in Fig.
Figure 5.Tracking results of 12 trackers. (a) Carchase1; (b) road; (c) carchase2; (d) group; (e) motorcycle; (f) triumphal arch; (g) jogging; (h) wandering. (Best view in PDF.)
Through Fig.
Center location error (CLE) is used for quantitative evaluation, which is defined as the percentage of frames whose Euclidean distance
The results of CLE are listed in Table
|
From Table
In conclusion, an effective tracking method that can handle OCC is proposed. Based on the motion and appearance correlation filters of LCT, ILCT employs a designed OCC criterion and a re-detector to judge the OCC and perform re-detection, respectively. Once the object is identified as occluded, the tracker stops, and the re-detector is activated. Then, the detection result with high confidence will re-initialize the tracker. Extensive experiments have been performed, and the results of qualitative and quantitative evaluation indicate that ILCT outperforms some state-of-the-art trackers in terms of accuracy and robustness. In future work, the efficiency and real-time performance have to be addressed to make the tracker perfect.
[4] C. Ma, X. Yang, C. Zhang, M. H. Yang. Proceedings of Computer Vision and Pattern Recognition, 5388(2015).
[5] T. Zhang, B. Ghanem, S. Liu, N. Ahuja. Proceedings of Computer Vision and Pattern Recognition, 2042(2012).
[6] D. S. Bolme, J. R. Beveridge, B. A. Draper, Y. M. Lui. Proceedings of Computer Vision and Pattern Recognition, 2544(2010).
[8] W. H. Press, S. A. Teukolsky, W. T. Vetterling, B. P. Flannery. Numerical Recipes in C(1988).
[9] J. F. Henriques, R. Caseiro, P. Martins, J. Batista. Proceedings of European Conference on Computer Vision, 702(2012).
[11] M. Danelljan, G. Häger, F. Khan, M. Felsberg. Proceedings of British Machine Vision Conference, 1(2014).
[12] J. Choi, H. J. Chang, J. Jeong, Y. Demiris, J. Y. Choi. Proceedings of Computer Vision and Pattern Recognition, 4321(2016).
[15] R. Zabih, J. Woodfill. Proceedings of European Conference on Computer Vision, 151(1994).
[16] C. Ma, X. Yang, C. Zhang, M. H. Yang. Long-term correlation tracking(2016).
[17] Y. Wu, J. Lim, M. H. Yang. Proceedings of Computer Vision and Pattern Recognition, 2411(2013).
[18] C. L. Zitnick, P. Dollár. Proceedings of European Conference on Computer Vision, 391(2014).
[21] S. Hare, A. Saffari, P. H. S. Torr. Proceedings of International Conference on Computer Vision, 263(2011).
[22] C. Bao, Y. Wu, H. Ling, H. Ji. Proceedings of Computer Vision and Pattern Recognition, 1830(2012).
[24] K. Zhang, L. Zhang, M. H. Yang. Proceedings of European Conference on Computer Vision, 864(2012).
[26] M. Mueller, N. Smith, B. Ghanem. Proceedings of European Conference on Computer Vision, 445(2016).
[27] O. Quaker. Anaheim California Police Chase 06/18/2017—Reckless Driver(2017).
Get Citation
Copy Citation Text
Junhao Zhao, Gang Xiao, Xingchen Zhang, D. P. Bavirisetti, "An improved long-term correlation tracking method with occlusion handling," Chin. Opt. Lett. 17, 031001 (2019)
Category: Image processing
Received: Jul. 28, 2018
Accepted: Dec. 20, 2018
Posted: Dec. 25, 2018
Published Online: Mar. 8, 2019
The Author Email: Gang Xiao (xiaogang@sjtu.edu.cn)