Enhancing Network Flow for Multi-target Tracking with Detection Group Analysis
Multi-target tracking (MTT) has been a research hotspot in the field of computer vision. The objective is forming the trajectory of multiple targets in a given video. However, the useful detection and tracklet relationship during the tracking process are not fully explored in most current algorithms and it leads to the accumulation of errors. We introduce a novel Detection Group, which includes the detections within a temporal and spatial threshold and then model the relationship between Detection Group(DG) and close tracklets. Although the minimum-cost network flow algorithm has been proven to be a successful strategy for multi-target tracking, but it still has one main drawback: due to the fact that useful corresponding detection and tracklet relationships are not well modeled, the network flow based tracker can only model low-level detection relationship without high-level detection set information. To cope with this problem, we extend the classical minimum-cost network flow algorithm within the tracking-by-detection paradigm by incorporating additional constraints. In our experiment, we achieved encouraging result on the MOT17 benchmark and our result is comparable to the current state of the art trackers.
KeywordsNetwork flow Detection group Multi-target tracking
This study is partially supported by the National Key R&D Program of China (No. 2017YFC0803700), the National Natural Science Foundation of China (No. 61 472019), the Macao Science and Technology Development Fund (No. 138/2016/A3), the Program of Introducing Talents of Discipline to Universities and the Open Fund of the State Key Laboratory of Software Development Environment under grant SKLSDE-2017ZX-09, the Project of Experimental Verification of the Basic Commonness and Key Technical Standards of the Industrial Internet network architecture. Thank you for the support from HAWKEYE Group.
- 1.Gurobi optimization. http://www.gurobi.com/
- 2.Bae, S.H., Yoon, K.J.: Robust online multi-object tracking based on tracklet confidence and online discriminative appearance learning. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1218–1225 (2014)Google Scholar
- 3.Bernardin, K., Stiefelhagen, R.: Evaluating multiple object tracking performance: the clear mot metrics. EURASIP J. Image Video Process. 2008(1), 246309 (2008)Google Scholar
- 4.Chari, V., Lacostejulien, S., Laptev, I., Sivic, J.: On pairwise costs for network flow multi-object tracking. In: Computer Vision and Pattern Recognition, pp. 5537–5545 (2015)Google Scholar
- 5.Chen, J., Sheng, H., Zhang, Y., Xiong, Z.: Enhancing detection model for multiple hypothesis tracking. In: Computer Vision and Pattern Recognition Workshops, pp. 2143–2152 (2017)Google Scholar
- 7.Kim, C., Li, F., Ciptadi, A., Rehg, J.M.: Multiple hypothesis tracking revisited. In: ICCV, pp. 4696–4704 (2015)Google Scholar
- 8.Mclaughlin, N., Martinez, Del Rincon, J., Miller, P.: Enhancing linear programming with motion modeling for multi-target tracking. In: Applications of Computer Vision, pp. 71–77 (2016)Google Scholar
- 9.Milan, A., Leal-Taixé, L., Schindler, K., Reid, I.: Joint tracking and segmentation of multiple targets. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 5397–5406 (2015)Google Scholar
- 10.Milan, A., Schindler, K., Roth, S.: Detection- and trajectory-level exclusion in multiple object tracking. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 3682–3689 (2013)Google Scholar
- 11.Sanchez-Matilla, R., Poiesi, F., Cavallaro, A.: Online multi-target tracking with strong and weak detections. In: European Conference on Computer Vision, pp. 84–99 (2016)Google Scholar
- 12.Shi, X., Ling, H., Xing, J., Hu, W.: Multi-target tracking by rank-1 tensor approximation. In: Computer Vision and Pattern Recognition, pp. 2387–2394 (2013)Google Scholar
- 13.Zhang, L., Li, Y., Nevatia, R.: Global data association for multi-object tracking using network flows. In: Computer Vision and Pattern Recognition IEEE Conference on 2008 CVPR 2008, pp. 1–8 (2008)Google Scholar