Skip to main content

Visual Tracking by Local Superpixel Matching with Markov Random Field

  • Conference paper
  • First Online:
Book cover Advances in Multimedia Information Processing - PCM 2016 (PCM 2016)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 9916))

Included in the following conference series:

Abstract

In this paper, we propose a novel method to track non-rigid and/or articulated objects using superpixel matching and markov random field (MRF). Our algorithm consists of three stages. First, a superpixel dataset is constructed by segmenting training frames into superpixels, and each superpixel is represented by multiple features. The appearance information of target is encoded in the superpixel database. Second, each new frame is segmented into superpixels and then its object-background confidence map is derived by comparing its superpixels with k-nearest neighbors in superpixel dataset. Taking context information into account, we utilize MRF to further improve the accuracy of confidence map. In addition, the local context information is incorporated through a feedback to refine superpixel matching. In the last stage, visual tracking is achieved via finding the best candidate by maximum a posterior estimate based on the confidence map. Experiments show that our method outperforms several state-of-the-art trackers.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The surrounding region is a square area centered at the location of target \(X_{t}^{c}\), and its side length is equal to \(\lambda _{s}[X_{s}^{t}]^{\frac{1}{2}}\), where \(X_{t}^{c}\) represents the center location of target region \(X_{t}\) and \(X_{t}^{s}\) denotes its size. The parameter \(\lambda _{s}\) is a constant variable, which determines the size of this surrounding region.

References

  1. Adam, A., Rivlin, E., Shimshoni, I.: Robust fragments based tracking using the integral histogram. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 798–805 (2006)

    Google Scholar 

  2. Avidan, S.: Ensemble tracking. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 29(2), 261–271 (2007)

    Article  Google Scholar 

  3. Babenko, B., Yang, M.-H., Belongie, S.: Robust object tracking with online multiple instance learning. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 33(8), 1619–1632 (2011)

    Article  Google Scholar 

  4. Boykov, Y., Veksler, O., Zabih, R.: Fast approximate energy minimization via graph cuts. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 23(11), 1222–1239 (2001)

    Article  Google Scholar 

  5. Fan, H., Xiang, J.: Robust visual tracking with multitask joint dictionary learning. IEEE Trans. Circ. Syst. Video Technol. PP, 1 (2016)

    Google Scholar 

  6. Fan, H., Xiang, J., Zhao, L.: Robust visual tracking via bag of superpixels. Multimedia Tools Appl. 75, 8781 (2015)

    Article  Google Scholar 

  7. Fan, H., Xiang, J., Liao, H., Du, X.: Robust tracking based on local structural cell graph. J. Vis. Commun. Image Represent. 31, 54–63 (2015)

    Article  Google Scholar 

  8. Fan, H., Xiang, J., Ni, F.: Multilayer feature combination for visual tracking. In: Asian Conference on Pattern Recognition (ACPR), pp. 589–593 (2015)

    Google Scholar 

  9. Fan, H., Xiang, J.: Patch-based visual tracking with two-stage multiple Kernel learning. In: Zhang, Y.-J. (ed.) ICIG 2015. LNCS, vol. 9219, pp. 20–33. Springer, Heidelberg (2015). doi:10.1007/978-3-319-21969-1_3

    Chapter  Google Scholar 

  10. Fan, H., Xiang, J., Xu, J., Liao, H.: Part-based visual tracking via online weighted P-N learning. Sci. World J. (2014)

    Google Scholar 

  11. Felzenszwalb, P.F., Huttenlocher, D.P.: Efficient graph based image segmentation. Int. J. Comput. Vis. (IJCV) 59(2), 167–181 (2004)

    Article  Google Scholar 

  12. Jia, X., Lu, H., Yang, M.-H.: Visual tracking via adaptive structural local sparse appearance model. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1822–1829 (2012)

    Google Scholar 

  13. Kalal, Z., Mikolajczyk, K., Matas, J.: Tracking-learning-detection. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 34(7), 1409–1422 (2012)

    Article  Google Scholar 

  14. Liang, P., Blasch, E., Ling, H.: Encoding color information for visual tracking: algorithms and benchmark. IEEE Trans. Image Process. (TIP) 24(12), 5630–5644 (2015)

    Article  MathSciNet  Google Scholar 

  15. Mei, X., Ling, H.: Robust visual tracking and vehicle classification via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 33(11), 2259–2272 (2011)

    Article  Google Scholar 

  16. Ross, D.A., Lim, J., Lin, R.S., Yang, M.-H.: Incremental learning for robust visual tracking. Int. J. Comput. Vis. (IJCV) 77(1), 125–141 (2008)

    Article  Google Scholar 

  17. Wang, D., Lu, H.: Visual tracking via probability continuous outlier model. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3478–3485 (2014)

    Google Scholar 

  18. Yang, F., Lu, H., Yang, M.-H.: Robust superpixel tracking. IEEE Trans. Image Process. (TIP) 23(4), 1639–1651 (2014)

    Article  MathSciNet  Google Scholar 

  19. Yang, J., Price, B., Cohen, S., Yang, M.-H.: Context driven scene parsing with attention to rare classes. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3294–3301 (2014)

    Google Scholar 

  20. Zhang, K., Zhang, L., Yang, M.-H.: Fast compressive tracking. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 36(10), 2002–2015 (2014)

    Article  Google Scholar 

  21. Zhang, K., Zhang, L., Liu, Q., Zhang, D., Yang, M.-H.: Fast visual tracking via dense spatio-temporal context learning. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 127–141. Springer, Heidelberg (2014). doi:10.1007/978-3-319-10602-1_9

    Google Scholar 

  22. Zhang, T., Ghanem, B., Liu, S., Ahuja, N.: Robust visual tracking via structured multi-task sparse learning. Int. J. Comput. Vis. (IJCV) 101, 367–383 (2013)

    Article  MathSciNet  Google Scholar 

  23. Zhong, W., Lu, H., Yang, M.-H.: Robust object tracking via sparsity-based collaborative model. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1838–1845 (2012)

    Google Scholar 

Download references

Acknowledgement

This work was primarily supported by Foundation Research Funds for the Central Universities (Program No. 2662016PY008 and Program No. 2662014PY052).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jinhai Xiang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing AG

About this paper

Cite this paper

Fan, H., Xiang, J., Chen, Z. (2016). Visual Tracking by Local Superpixel Matching with Markov Random Field. In: Chen, E., Gong, Y., Tie, Y. (eds) Advances in Multimedia Information Processing - PCM 2016. PCM 2016. Lecture Notes in Computer Science(), vol 9916. Springer, Cham. https://doi.org/10.1007/978-3-319-48890-5_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-48890-5_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-48889-9

  • Online ISBN: 978-3-319-48890-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics