Abstract
In this paper we propose a novel method for the exploitation of High Density Localization (HDL) maps obtained by Mobile Laser Scanning in order to increase the performance of state-of-the-art real time dynamic object detection (RTDOD) methods utilizing Rotating Multi-Beam (RMB) Lidar measurements. First, we align the onboard measurements to the 3D HDL map with a multimodal point cloud registration algorithm operating in the Hough space. Next we apply a grid based probabilistic step to filter out the object regions on the RMB Lidar data which were falsely predicted as dynamic objects by RTDOD, although they are part of the static background scene. On the other hand, to find objects erroneously missed by the RTDOD predictions, we implement a Markov Random Field based point level change detection approach between the map and the current onboard measurement frame. Finally, to analyse the changed but previously unclassified segments of the RMB Lidar clouds, we apply a geometric blob separation and a Support Vector Machine based classification to distinguish the different object types. Comparative tests are provided in high traffic road sections of Budapest, Hungary, and we show an improvement of \(5,96\%\) in precision, \(9,21\%\) in recall and \(7,93\%\) in F-score metrics against the state-of-the-art RTDOD algorithm.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Börcs, A., Nagy, B., Benedek, C.: Instant object detection in Lidar point clouds. IEEE Geosci. Remote Sens. Lett. 14(7), 992–996 (2017)
Gálai, B., Benedek, C.: Change detection in urban streets by a real time Lidar scanner and MLS reference data. In: Karray, F., Campilho, A., Cheriet, F. (eds.) ICIAR 2017. LNCS, vol. 10317, pp. 210–220. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-59876-5_24
Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2012)
Kidono, K., Miyasaka, T., Watanabe, A., Naito, T., Miura, J.: Pedestrian recognition using high-definition LIDAR. J. Robot. Soc. Japan 29, 405–410 (2011). https://doi.org/10.1109/IVS.2011.5940433
Lang, A., Vora, S., Caesar, H., Zhou, L., Yang, J., Beijbom, O.: PointPillars: fast encoders for object detection from point clouds. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)
Matthaei, R., Bagschik, G., Maurer, M.: Map-relative localization in lane-level maps for ADAS and autonomous driving. In: IEEE Intelligent Vehicles Symposium Proceedings, Dearborn, MI, USA, pp. 49–55, June 2014. https://doi.org/10.1109/IVS.2014.6856428
Nagy, B., Benedek, C.: 3D CNN-based semantic labeling approach for mobile laser scanning data. IEEE Sens. J. 19(21), 10034–10045 (2019)
Nagy, B., Benedek, C.: Real-time point cloud alignment for vehicle localization in a high resolution 3D map. In: Leal-Taixé, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11129, pp. 226–239. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-11009-3_13
Seif, H.G., Hu, X.: Autonomous driving in the iCity-HD maps as a key challenge of the automotive industry. Engineering 2(2), 159–162 (2016). https://doi.org/10.1016/J.ENG.2016.02.010
Shin, K., Kwon, Y., Tomizuka, M.: . RoarNet: a robust 3D object detection based on region approximation refinement. In: IEEE Intelligent Vehicles Symposium, pp. 2510–2515 (2018)
Simon, M., Milz, S., Amende, K., Groß, H.M.: Complex-YOLO: real-time 3D object detection on point clouds. arXiv abs/1803.06199 (2018)
Yan, Y., Mao, Y., Li, B.: SECOND: sparsely embedded convolutional detection. Sensors 18, 3337 (2018). https://doi.org/10.3390/s18103337
Yang, B., Liang, M., Urtasun, R.: HDNET: exploiting HD maps for 3D object detection. In: Conference on Robot Learning. Proceedings of Machine Learning Research, 29–31 Oct 2018, vol. 87, pp. 146–155. http://proceedings.mlr.press/v87/yang18b.html
Yang, B., Luo, W., Urtasun, R.: PIXOR: real-time 3D object detection from point clouds. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7652–7660 (2018). https://doi.org/10.1109/CVPR.2018.00798
Zhou, Y., Tuzel, O.: VoxelNet: end-to-end learning for point cloud based 3D object detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4490–4499 (2018). https://doi.org/10.1109/CVPR.2018.00472
Acknowledgements
This work was supported by the National Research, Development and Innovation Fund under grant number K-120233, by the European Union and the Hungarian Government from the projects Thematic Fundamental Research Collaborations Grounding Innovation in Informatics and Infocommunications under grant number EFOP-3.6.2-16-2017-00013 (Örkény Zováthi and Balázs Nagy) and Intensification of the activities of HU-MATHS-IN - Hungarian Service Network of Mathematics for Industry and Innovation under grant number EFOP-3.6.2-16-2017-00015, and by the Michelberger Master Award of the Hungarian Academy of Engineering (Csaba Benedek).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Zováthi, Ö., Nagy, B., Benedek, C. (2020). Exploitation of Dense MLS City Maps for 3D Object Detection. In: Campilho, A., Karray, F., Wang, Z. (eds) Image Analysis and Recognition. ICIAR 2020. Lecture Notes in Computer Science(), vol 12131. Springer, Cham. https://doi.org/10.1007/978-3-030-50347-5_34
Download citation
DOI: https://doi.org/10.1007/978-3-030-50347-5_34
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50346-8
Online ISBN: 978-3-030-50347-5
eBook Packages: Computer ScienceComputer Science (R0)