Abstract
Video processing and more particularly motion tracking algorithms present a necessary tool for various domains related to computer vision such as motion recognition, depth estimation and event detection. However, the use of high definitions videos (HD, Full HD, 4K, etc.) cause that current implementations, even running on modern hardware, no longer respect the requirements of real-time treatment. In this context, several solutions have been proposed to overcome this constraint, by exploiting graphic processing units (GPUs). Although, they benefit from the high power of GPU, none of them is able to provide efficient dense and sparse motion tracking within high definition videos efficiently. In this work, we propose a GPU and Multi-GPU based method for both sparse and dense optical flow motion tracking using the Lucas-Kanade algorithm. Our method presents an efficient exploitation and management of single or/and multiple GPU memories, according to the type of applied implementation: sparse or dense. The sparse implementation allows tracking meaningful pixels, which are detected with the Harris corner detector. The dense implementation requires more computation since it is applied on each pixel of the video. Within our approach, high definition videos are processed on GPUs while low resolution videos are treated on CPUs. As result, our method allows real-time sparse and dense optical flow computation on videos in Full HD or even 4K format. The exploitation of multiple GPUs presents performance that scale up very well. In addition to these performances, the parallel implementations offered lower power consumption as result of the fast treatment.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
OpenMP. The OpenMP API specification for parallel programming. www.openmp.org.
- 3.
NVIDIA Quadro SDI Capture: http://www.nvidia.com/object/product_quadro_sdi_capture_us.html.
References
Baker, S., Roth, S., Scharstein, D., Black, M., Lewis, J.P., Szeliski, R.: A database and evaluation methodology for optical flow. Int. J. Comput. Vision 92(1), 1–8 (2011)
Mahmoudi, S.A., et al.: Towards a smart selection of resources in the cloud for low-energy multimedia processing. Concurr. Comput. Pract. Exp. 30(12), 1–13 (2018)
Bouguet, J.Y.: Pyramidal Implementation of the Lucas Kanade Feature Tracker. Intel Corporation Microprocessor Research Labs (2000)
Ferhat, O., Vilarino, F.: A cheap portable eye-tracker solution for common setups. In: 17th European Conference on Eye Movements (2013)
Gibson, J.: The Perception of the Visual World. Houghton Mifflin, Boston (1950)
Gwosdek, P., Zimmer, H., Grewenig, S., Bruhn, A., Weickert, J.: A highly efficient GPU implementation for variational optic flow based on the Euler-Lagrange framework, vol. 6554, pp. 372–383 (2012)
Harris, C., Stephens, M.: A combined corner and edge detector. In: The 4th Alvey Vision Conference, vol. 15, pp. 147–151 (1988)
Horn, B.K.P., Schunk, B.G.: Determining optical flow. Artif. Intell. 2, 185–203 (1981)
Huang, J., Ponce, S., Park, S., Cao, Y., Quek, F.: GPU-accelerated computation for robust motion tracking using CUDA framework. In: Proceedings of the IET International Conference on Visual Information Engineering (2008)
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. (IJCV) 60(2), 91–110 (2004)
Lucas, B., Kanade, T.: An iterative image registration technique with an application to stereo vision. In: Proceedings of Imaging Understanding Workshop, pp. 121–130 (1981)
Mahmoudi, S.A., Kierzynka, M., Manneback, P., Kurowski, K.: Real-time motion tracking using optical flow on multiple GPUs. Bull. Pol. Acad. Sci. Tech. Sci. 62, 139–150 (2014)
Marzat, J., Dumortier, Y., Ducrot, A.: Real-time dense and accurate parallel optical flow using CUDA. In: Proceedings of WSCG, pp. 105–111 (2009)
Mizukami, Y., Tadamura, K.: Optical flow computation on compute unified device architecture. In: Proceedings of the 14th International Conference on Image Analysis and Processing, pp. 179–184 (2007)
Mahmoudi, S.A., Manneback, P.: Multi-GPU based event detection and localization using high definition videos. In: International Conference on Multimedia Computing and Systems (ICMCS), pp. 81–86 (2014)
Ready, J.M., Taylor, C.N.: GPU acceleration of real-time feature based algorithms, p. 8 (2007)
Mahmoudi, S.A., Manneback, P.: Multi-CPU/multi-GPU based framework for multimedia processing. In: Computer Science and Its Applications, vol. 456, pp. 54–65 (2015)
Sinha, S.N., Fram, J.-M., Pollefeys, M., Genc, Y.: GPU-based video feature tracking and matching. In: EDGE, Workshop on Edge Computing Using New Commodity Architectures (2006)
Sundaram, N., Brox, T., Keutzer, K.: Dense point trajectories by GPU-accelerated large displacement optical flow (2010). http://www.eecs.berkeley.edu/Pubs/TechRpts/2010/EECS-2010-104.html
Tomasi, C., Kanade, T.: Detection and tracking of point features. Technical Report CMU-CS-91-132, CMU, pp. 1–4 (1991)
Wang, T., Snoussi, H.: Histograms of optical flow orientation for visual abnormal events detection, pp. 13–18, September 2012
Mahmoudi, S.A., et al.: Real-time GPU-based motion detection and tracking using full HD videos. In: International Conference on Intelligent Technologies for Interactive Entertainment, Belgium, pp. 12–21 (2013)
PZEM: AC Digital Display Multifunction Meter. https://www.circuitspecialists.com/content/189799/ac004.pdf. Accessed 01 Mar 2018
Possa, P.R., et al.: A new self-adapting architecture for feature detection. In: 22nd International Conference on Field Programmable Logic and Applications (FPL), pp. 643–646 (2012)
Mahmoudi, S.A., Manneback, P.: Efficient exploitation of heterogeneous platforms for images features extraction. In: 3rd International Conference on Image Processing Theory, Tools and Applications, pp. 91–96 (2012)
Acknowledgements
If you want to include acknowledgments of assistance and the like at the end of an individual chapter please use the acknowledgement environment – it will automatically render Springer’s preferred layout.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Mahmoudi, S.A., Belarbi, M.A., Manneback, P. (2019). Towards a Smart Exploitation of GPUs for Low Energy Motion Estimation Using Full HD and 4K Videos. In: Zbakh, M., Essaaidi, M., Manneback, P., Rong, C. (eds) Cloud Computing and Big Data: Technologies, Applications and Security. CloudTech 2017. Lecture Notes in Networks and Systems, vol 49. Springer, Cham. https://doi.org/10.1007/978-3-319-97719-5_18
Download citation
DOI: https://doi.org/10.1007/978-3-319-97719-5_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-97718-8
Online ISBN: 978-3-319-97719-5
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)