Abstract
Correlation Filter-based trackers have achieved excellent performance and run at high frame rates. Recently, Staple, which utilizing a simple combination of a Correlation Filter (using HOG features) and a global color histogram, has achieved excellent performance. It shows strong robustness in challenging situations including motion blur, illumination changes and deformation changes. However, Staple is only a linear combination of two methods. It is not reliable to determine the confidence level only by the peak. In this paper, we propose Gaussian-Staple that utilize a more sensible way of fusion without destroying the response distribution after fusion. Gaussian prior is added to the response of the output, which is used to determine whether to fine tune by local search. Extensive experiments on a commonly used tracking benchmark show that the proposed method significantly improves Staple, and achieves a better performance than other state-of-the-art trackers.
B. Luo—This work was supported in part by National Natural Science Foundation of China under Grant 61472002, 61572030 and 61671018, and Collegiate Natural Science Fund of Anhui Province under Grant KJ2017A014.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bertinetto, L., Valmadre, J., Golodetz, S., Miksik, O., Torr, P.H.S.: Staple: complementary learners for real-time tracking. 38(2), 1401–1409 (2015)
Bolme, D.S., Beveridge, J.R., Draper, B.A., Lui, Y.M.: Visual object tracking using adaptive correlation filters. In: CVPR, pp. 2544–2550 (2010)
Danelljan, M., Hager, G., Khan, F.S., Felsberg, M.: Accurate scale estimation for robust visual tracking. In: British Machine Vision Conference, pp. 65.1–65.11 (2014)
Dinh, T.B., Vo, N., Medioni, G.: Context tracker: exploring supporters and distracters in unconstrained environments. In: CVPR, pp. 1177–1184 (2011)
Han, J., Zhang, D., et al.: Object detection in optical remote sensing images based on weakly supervised learning and high-level feature learning. IEEE TGRS 53(6), 3325–3337 (2015)
Hare, S., Saffari, A., Torr, P.H.S.: Struck: structured output tracking with kernels. In: IEEE ICCV, pp. 263–270 (2012)
Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: High-speed tracking with kernelized correlation filters. IEEE T-PAMI 37(3), 583 (2015)
Henriques, J.F., Caseiro, R., Martins, P., Batista, J.: Exploiting the circulant structure of tracking-by-detection with kernels. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7575, pp. 702–715. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33765-9_50
Kalal, Z., Matas, J., Mikolajczyk, K.: P-N learning: bootstrapping binary classifiers by structural constraints. In: CVPR, pp. 49–56 (2010)
Kristan, M., Pflugfelder, R., et al.: The visual object tracking VOT2013 challenge results. In: IEEE ICCV Workshops, pp. 98–111 (2013)
Liu, Q.: Decontaminate feature for tracking: adaptive tracking via evolutionary feature subset. J. Electron. Imaging 26(6), 1 (2017)
Possegger, H., Mauthner, T., Bischof, H.: In defense of color-based model-free tracking. In: CVPR, pp. 2113–2120 (2015)
Pérez, P., Hue, C., Vermaak, J., Gangnet, M.: Color-based probabilistic tracking. In: Heyden, A., Sparr, G., Nielsen, M., Johansen, P. (eds.) ECCV 2002. LNCS, vol. 2350, pp. 661–675. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-47969-4_44
Ren, J., Orwell, J., Jones, G.A., Xu, M.: Tracking the soccer ball using multiple fixed cameras. Comput. Vis. Image Underst. 113(5), 633–642 (2009)
Ren, J., Xu, M., Orwell, J., Jones, G.A.: Real-time modeling of 3-D soccer ball trajectories from multiple fixed cameras. IEEE T-CSVT 18(3), 350–362 (2008)
Ren, J., Xu, M., et al.: Multi-camera video surveillance for real-time analysis and reconstruction of soccer games. Mach. Vis. Appl. 21(6), 855–863 (2010)
Sevilla-Lara, L., Learned-Miller, E.: Distribution fields for tracking. In: IEEE Conference on CVPR, pp. 1910–1917 (2012)
Wang, Z., Ren, J., Zhang, D., Sun, M., Jiang, J.: A deep-learning based feature hybrid framework for spatiotemporal saliency detection inside videos. Neurocomputing 287, 68–83 (2018)
Wu, Y., Lim, J., Yang, M.H.: Online object tracking: a benchmark. In: IEEE CVPR, pp. 2411–2418 (2013)
Wu, Y., Lim, J., Yang, M.H.: Object tracking benchmark. IEEE T-PAMI 37(9), 1834–1848 (2015)
Yan, Y., Ren, J., et al.: Cognitive fusion of thermal and visible imagery for effective detection and tracking of pedestrians in videos. Cogn. Comput. 10(1), 94–104 (2018)
Zhang, K., Zhang, L., Yang, M.H.: Real-time compressive tracking. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7574, pp. 864–877. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33712-3_62
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Chen, SB., Ding, CY., Luo, B. (2018). Gaussian-Staple for Robust Visual Object Real-Time Tracking. In: Ren, J., et al. Advances in Brain Inspired Cognitive Systems. BICS 2018. Lecture Notes in Computer Science(), vol 10989. Springer, Cham. https://doi.org/10.1007/978-3-030-00563-4_36
Download citation
DOI: https://doi.org/10.1007/978-3-030-00563-4_36
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-00562-7
Online ISBN: 978-3-030-00563-4
eBook Packages: Computer ScienceComputer Science (R0)