Abstract
Smart mobile devices can be easily transformed into driving assistance tools or traffic monitoring systems. These devices are placed behind the windshield such that the camera is facing forward to observe the traffic. For the visual information to be useful, the camera must be calibrated, and a proper calibration is laborious and difficult to perform for the average user. In this paper, we propose a calibration technique that requires no input from the user and is able to estimate the extrinsic parameters of the camera: yaw, pitch and roll angles and the height of the camera above the road. The calibration algorithm is based on detecting vehicles using CNN based classifiers, and using statistics about their size and position in the image to estimate the extrinsic parameters via Extended Kalman filters.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Danescu, R., Itu, R., Petrovai, A.: Generic dynamic environment perception using smart mobile devices. Sensors 16, 1–21 (2016). Article no. 1721
Caprile, B., Torre, V.: Using vanishing points for camera calibration. Int. J. Comput. Vis. 4, 127–139 (1990)
Magee, M., Aggarwal, J.: Determining vanishing points from perspective images. Comput. Vis. Graph. Image Process. 26, 256–267 (1984)
Bazin, J., Pollefeys, M.: 3-line RANSAC for orthogonal vanishing point detection. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4282–4287 (2012)
Itu, R., Borza, D., Danescu, R.: Automatic extrinsic camera parameters calibration using Convolutional Neural Networks. In: 2017 IEEE 13th International Conference on Intelligent Computer Communication and Processing (ICCP 2017), pp. 273–278 (2017)
Bileschi, S.: Fully automatic calibration of LiDAR and video streams from a vehicle. In: IEEE International Conference on Computer Vision Workshops (ICCV), pp. 1457–1464 (2009)
Levinson, J., Thrun, S.: Automatic online calibration of cameras and lasers. In: Robotics Science Systems Conference, pp. 1–8 (2013)
Abadi, M., et al.: TensorFlow: large-scale machine learning on heterogeneous systems. arXiv:1603.04467 (2016) (preprint)
Howard, A., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861 (2017) (preprint)
Lin, T., et al.: Microsoft COCO: common objects in context. In: European Conference on Computer Vision (ECCV), pp. 740–755 (2014)
Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? In: Computer Vision and Pattern Recognition Conference (CVPR), pp. 3354–3361 (2012)
Udacity Vehicle Dataset: https://github.com/udacity/self-driving-car/tree/master/annotations
Rosten, E., Porter, R., Drummond, T.: Faster and better: a machine learning approach to corner detection. IEEE Trans. Pattern Anal. Mach. Intell. 32, 105–119 (2010)
Fischler, M., Bolles, R.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24, 381–395 (1981)
Acknowledgment
This work was supported by a grant of Ministry of Research and Innovation, CNCS - UEFISCDI, project number PN-III-P1-1.1-TE-2016-0440, within PNCDI III.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Itu, R., Danescu, R. (2019). Machine Learning Based Automatic Extrinsic Calibration of an Onboard Monocular Camera for Driving Assistance Applications on Smart Mobile Devices. In: Dubbert, J., Müller, B., Meyer, G. (eds) Advanced Microsystems for Automotive Applications 2018. AMAA 2018. Lecture Notes in Mobility. Springer, Cham. https://doi.org/10.1007/978-3-319-99762-9_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-99762-9_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-99761-2
Online ISBN: 978-3-319-99762-9
eBook Packages: EngineeringEngineering (R0)