Skip to main content

Robots Perception Through 3D Point Cloud Sensors

  • Chapter
  • First Online:
Robot Operating System (ROS)

Abstract

This chapter brings a tutorial about use of Point Cloud data for the environment perception of mobile robots. Point Cloud is a powerful tool that gives robots the ability to perceive the world around them through a dense measurement. One advantage of this kind of sensor is the large measuring space, with faint or no external light. Although there are several works about Point Clouds, only a few of them speak about how this kind of information can be obtained and what can be extracted. This chapter aims to fill this gap and clarify how Point Clouds can be acquired, processed, transformed between coordinate systems and which these information can be easily extracted using ROS and Matlab. The codes used in this chapter are available in GitHub and can be found at https://github.com/air-lasca/ros_book_point_cloud. The videos developed with the experiments can be seen on YouTube, in Robot LASCA channel that can be accessed at https://www.youtube.com/channel/UCtgnBqaodQAGtbh0HW9nJEA.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 249.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Oliver, A., S. Kang, B.C. Wünsche, and B. MacDonald. 2012. Using the kinect as a navigation sensor for mobile robotics. In Proceedings of the 27th conference on image and vision computing New Zealand, CM, 509–514.

    Google Scholar 

  2. Endres, F., J. Hess, N. Engelhard, J. Sturm, D. Cremers, and W. Burgard. 2012. An evaluation of the RGB-d slam system. In 2012 IEEE international conference on robotics and automation (ICRA), 1691–1696. New York: IEEE.

    Google Scholar 

  3. Whelan, T., M. Kaess, M. Fallon, H. Johannsson, J. Leonard, and J. McDonald. 2012. Kintinuous: Spatially extended kinectfusion.

    Google Scholar 

  4. Whelan, T., M. Kaess, J.J. Leonard, and J. McDonald. 2013. Deformation-based loop closure for large scale dense RGB-D slam. In 2013 IEEE/RSJ international conference on intelligent robots and systems (IROS), 548–555. New York: IEEE.

    Google Scholar 

  5. Wallace, L., A. Lucieer, Z. Malenovskỳ, D. Turner, and P. Vopěnka. 2016. Assessment of forest structure using two UAV techniques: A comparison of airborne laser scanning and structure from motion (SFM) point clouds. Forests 7 (3): 62.

    Article  Google Scholar 

  6. Wang, Q., L. Wu, Z. Xu, H. Tang, R. Wang, and F. Li. 2014. A progressive morphological filter for point cloud extracted from UAV images. In IEEE international geoscience and remote sensing symposium (IGARSS), 2014, 2023–2026. New York: IEEE.

    Google Scholar 

  7. Nagai, M., T. Chen, R. Shibasaki, H. Kumagai, and A. Ahmed. 2009. UAV-borne 3-d mapping system by multisensor integration. IEEE Transactions on Geoscience and Remote Sensing 47 (3): 701–708.

    Article  Google Scholar 

  8. Tao, W., Y. Lei, and P. Mooney. 2011. Dense point cloud extraction from UAV captured images in forest area. In 2011 IEEE international conference on spatial data mining and geographical knowledge services (ICSDM), 389–392. New York: IEEE.

    Google Scholar 

  9. Matlab, “Matlab”. 2016. http://www.mathworks.com/products/matlab/.

  10. GitHub. 2016. Installation from GitHub on debian. https://github.com/Singular/Sources/wiki/Installation-from-GitHub-on-Debian.

  11. PCL. 2016. What is PCL? Abr. http://pointclouds.org/about/.

  12. Li, L. 2014. Time-of-flight camera–an introduction, Technical White Paper.

    Google Scholar 

  13. PCL. 2016. Module io, Abr. http://docs.pointclouds.org/trunk/group__io.html.

  14. ROS. 2016. PCL overview, Abr. http://wiki.ros.org/pcl/Overview.

  15. Kreylos. 2016. Kinect hacking. http://idav.ucdavis.edu/~okreylos/ResDev/Kinect/.

  16. HEPTAGON. 2016. Our history. http://hptg.com/about-us/#history.

  17. ROS. 2016. About ROS, Abr. http://www.ros.org/about-ros/.

  18. ROS. 2016. ROS jade installation instructions, Abr. http://wiki.ros.org/ROS/Installation.

  19. ubuntu. 2016. About ubuntu, Abr. http://www.ubuntu.com/about/about-ubuntu.

  20. Ubuntu. 2016. Ubuntu 14.04.4 LTS (trusty Tahr), Abr. http://releases.ubuntu.com/14.04/.

  21. Ubuntu. 2016. Install Ubuntu 16.04 LTS, Abr. http://www.ubuntu.com/download/desktop/install-ubuntu-desktop.

  22. ROS. 2016. Ubuntu install of ROS indigo, Abr. http://wiki.ros.org/indigo/Installation/Ubuntu.

  23. ROS. 2016. freenect_camera. http://wiki.ros.org/freenect_camera.

  24. ROS. 2016. freenect_launch. http://wiki.ros.org/freenect_launch.

  25. UBUNTU. 2016. Package: libfreenect-dev (1:0.0.1+20101211+2-3). http://packages.ubuntu.com/precise/libfreenect-dev.

  26. ROS. 2016. openni_launch. http://wiki.ros.org/openni_launch.

  27. Robot, B.R. 2016. Kinect basics. http://sdk.rethinkrobotics.com/wiki/Kinect_basics.

  28. ROS. 2016. Kinect: Using microsoft kinect on the evarobot. http://wiki.ros.org/Robots/evarobot/Tutorials/indigo/Kinect

  29. HEPTAGON. 2016. Swissranger. http://hptg.com/industrial/.

  30. ROS. 2016. swissranger_camera. http://wiki.ros.org/swissranger_camera.

  31. ROS. 2016. cob_camera_sensors. http://wiki.ros.org/cob_camera_sensors.

  32. ROS. 2016. Care-o-bot: Configuring and using the swissranger 3000 or 4000 depth sensor. http://wiki.ros.org/cob_camera_sensors/Mesa_Swissranger.

  33. M. imaging. 2016. Sr4000/sr4500 user manual. http://www.realtechsupport.org/UB/SR/range_finding/SR4000_SR4500_Manual.pdf.

  34. ROS. 2016. visualization_msgs/marker message. http://docs.ros.org/api/visualization_msgs/html/msg/Marker.html.

  35. ROS. 2016. visualization_msgs/markerarray message. http://docs.ros.org/api/visualization_msgs/html/msg/MarkerArray.html.

  36. ROS. 2016. The marker message. http://wiki.ros.org/rviz/DisplayTypes/Marker.

  37. OctoMap. 2016. An efficient probabilistic 3d mapping framework based on octrees. http://octomap.github.io/.

  38. MathWorks. 2016. find. http://www.mathworks.com/help/matlab/ref/find.html.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Marco Antonio Simões Teixeira .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this chapter

Cite this chapter

Teixeira, M.A.S., Santos, H.B., de Oliveira, A.S., Arruda, L.V., Neves, F. (2017). Robots Perception Through 3D Point Cloud Sensors. In: Koubaa, A. (eds) Robot Operating System (ROS). Studies in Computational Intelligence, vol 707. Springer, Cham. https://doi.org/10.1007/978-3-319-54927-9_16

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-54927-9_16

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-54926-2

  • Online ISBN: 978-3-319-54927-9

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics