Advertisement

Quantifying the Effects of Environmental Conditions on Autonomy Algorithms for Unmanned Ground Vehicles

  • Phillip J. DurstEmail author
  • Justin Carrillo
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11472)

Abstract

Autonomy for commercial applications is developing at a rapid pace; however, autonomous navigation of unmanned ground vehicles (UGVs) for military applications has been deployed to a limited extent. Delaying the use of autonomy for military applications is the environment in which military UGVs must operate. Military operations take place in unstructured environments under adverse environmental conditions. Military UGVs are infrequently tested harsh conditions; therefore, there exists a lack of understanding in how autonomy reacts to challenging environmental conditions. Using high-fidelity modeling and simulation (M&S), autonomy algorithms can be exercised quickly and inexpensively in realistic operational conditions. The presented research introduces the M&S tools available for simulating adverse environmental conditions. Simulated camera images generated using these M&S tools are run through two typical autonomy algorithms, road lane detection and object classification, to assess the impact environmental conditions have on autonomous operations. Furthermore, the presented research proposes a methodology for quantifying these environmental effects.

Keywords

Autonomy Autonomous ground vehicles Environment effects Perception 

Notes

Acknowledgments

Permission to publish was granted by Director, Geotechnical and Structures Laboratory.

References

  1. 1.
    Wright, A.: Automotive autonomy. Commun. ACM 54(7), 16–18 (2011)CrossRefGoogle Scholar
  2. 2.
    Luettel, T., et al.: Autonomous ground vehicles-concepts and a path to the future. Proc. IEEE 100(Centennial Issue), 1831–1839 (2012)Google Scholar
  3. 3.
    Jones, R., et al.: Virtual autonomous navigation environment (VANE). In: Earth & Space 2008: Engineering, Science, Construction, and Operations in Challenging Environments, pp. 1–8 (2008)Google Scholar
  4. 4.
    Cummins, C.: Virtual autonomous navigation environment. DTIC Document (2008)Google Scholar
  5. 5.
    Goodin, C., et al.: Sensor modeling for the virtual autonomous navigation environment. In: IEEE Sensors, pp. 1588–1592 (2009)Google Scholar
  6. 6.
    Goodin, C., Durst, P.J., Gates, B., Cummins, C., Priddy, J.: High fidelity sensor simulations for the virtual autonomous navigation environment. In: Ando, N., Balakirsky, S., Hemker, T., Reggiani, M., von Stryk, O. (eds.) SIMPAR 2010. LNCS (LNAI), vol. 6472, pp. 75–86. Springer, Heidelberg (2010).  https://doi.org/10.1007/978-3-642-17319-6_10CrossRefGoogle Scholar
  7. 7.
    He, X.D., et al.: A comprehensive physical model for light reflection. In: ACM SIGGRAPH Comput. Graphics, vol. 25. no. 4 (1991)Google Scholar
  8. 8.
    Hosek, L., Wilkie, A.: An analytic model for full spectral sky-dome radiance. ACM Trans. Graph. (TOG) 31, 95 (2012)CrossRefGoogle Scholar
  9. 9.
    Bird, R.E., Riordan, C.: Simple solar spectral model for direct and diffuse irradiance on horizontal and tilted planes at the earth’s surface for cloudless atmospheres. J. Clim. Appl. Meteorol. 25(1), 87–97 (1986)CrossRefGoogle Scholar
  10. 10.
    Narasimhan, S.G., Nayar, S.K.: Contrast restoration of weather degraded images. IEEE Trans. Pattern Anal. Mach. Intell. 25(6), 713–724 (2003)CrossRefGoogle Scholar
  11. 11.
    Kang, L.-W., Lin, C.-W., Fu, Y.-H.: Automatic single-image-based rain streaks removal via image decomposition. IEEE Trans. Image Process. 21(4), 1742 (2012)MathSciNetCrossRefGoogle Scholar
  12. 12.
    Bertozzi, M., Broggi, A., Fascioli, A.: Vision-based intelligent vehicles: state of the art and perspectives. Robot. Auton. Syst. 32(1), 1–16 (2000)CrossRefGoogle Scholar
  13. 13.
    Abadi, M., et al.: Tensorflow: a system for large-scale machine learning. In: OSDI, vol. 16 (2016)Google Scholar
  14. 14.
    Szegedy, C., Liu, W., Jia, Y., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015)Google Scholar
  15. 15.
    Hillel, A.B., et al.: Recent progress in road and lane detection: a survey. Mach. Vis. Appl. 25(3), 727–745 (2014)CrossRefGoogle Scholar
  16. 16.
    Fritsch, J., et al.: A new performance measure and evaluation benchmark for road detection algorithms. In: 16th International IEEE Conference on Intelligent Transportation Systems. IEEE (2013)Google Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  1. 1.U.S. Army Engineer Research and Development CenterVicksburgUSA

Personalised recommendations