Abstract
This paper presents an approach to endow a humanoid robot with the capability of learning new objects and recognizing them in an unstructured environment. New objects are learnt, whenever an unrecognized one is found within a certain (small) distance from the robot head. Recognized objects are mapped to an ego-centric frame of reference, which together with a simple short-term memory mechanism, makes this mapping persistent. This allows the robot to be aware of their presence even if temporarily out of the field of view, thus providing a primary spatial model of the environment (as far as known objects are concerned). SIFT features are used, not only for recognizing previously learnt objects, but also to allow the robot to estimate their distance (depth perception). The humanoid platform used for the experiments was the iCub humanoid robot. This capability functions together with iCub’s low-level attention system: recognized objects enact salience thus attracting the robot attention, by gazing at them, each one in turn. We claim that the presented approach is a contribution towards linking a bottom-up attention system with top-down cognitive information.
This work was supported by the European Commission, Project IST-004370 RobotCub, and by the Portuguese Government — FCT (ISR/IST plurianual funding), and through the project BIO-LOOK, PTDC / EEA-ACR / 71032 / 2006.
Chapter PDF
References
Hamilton, C.M., Hedberg, S.: Modern masters of an ancient game. AI magazine 18(4), 137–144 (1997)
Nilsson, N.J.: Human-level artificial intelligence? be serious! AI magazine 26(4), 68–75 (2005)
Vernon, D., Metta, G., Sandini, G.: A survey of artificial cognitive systems: Implications for the autonomous development of mental capabilities in computational agents. IEEE Transactions on Evolutionary Computation 11(2), 151–180 (2007)
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60(2), 91–110 (2004)
Ruesch, J., Lopes, M., Bernardino, A., Hornstein, J., Santos-Victor, J., Pfeifer, R.: Multimodal saliency-based bottom-up attention, a framework for the humanoid robot icub. In: IEEE International Conference on Robotics and Automation Pasadena, CA, USA (May 2008)
Dankers, A., Barnes, N., Zelinsky, A.: A reactive vision system: Active-dynamic saliency. In: Proc. of the 5th International Conference on Computer Vision Systems, ICVS, Bielefeld, Germany (March 2007)
Metta, G., Sandini, G., Vernon, D., Caldwell, D., Tsagarakis, N., Beira, R., Santos-Victor, J., Ijspeert, A., Righetti, L., Cappiello, G., Stellin, G., Becchi, F.: The RobotCub project: an open framework for research in embodied cognition. In: Proceedings of the 2005 IEEE-RAS International Conference on Humanoid Robots (2005)
Coradeschi, S., Saffiotti, A.: Perceptual anchoring of symbols for action. In: Proceedings of the 17th International Conference on Artificial Intelligence (IJCAI 2001), Seattle, WA, pp. 407–412 (2001)
Chella, A., Coradeschi, S., Frixione, M., Safotti, A.: Perceptual anchoring via conceptual spaces. In: Proceedings of the AAAI 2004 Workshop on Anchoring Symbols to Sensor Data. AAAI Press, Menlo Park (2004)
Stoytchev, A.: Toward learning the binding affordances of objects: A behavior-grounded approach. In: Proceedings of AAAI Symposium on Developmental Robotics, pp. 17–22. Stanford University, Menlo Park (2005)
Montesano, L., Lopes, M., Bernardino, A., Santos-Victor, J.: Learning object affordances: From sensory motor maps to imitation. IEEE Transactions on Robotics 24(1), 15–26 (2008)
Itti, L., Koch, C.: A saliency-based search mechanism for overt and covert shifts of visual attention. Vision Research 40(10-12), 1489–1506 (2000)
Schiele, B., Crowley, J.L.: Object recognition using multidimensional receptive field histograms. In: European Conference on Computer Vision, vol. I, pp. 610–619 (1996)
Bay, H., Tuytelaars, T., Van Gool, L.: SURF: Speeded up robust features. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3951, pp. 404–417. Springer, Heidelberg (2006)
Bauer, J., Sunderhauf, N., Protzel, P.: Comparing several implementations of two recently published feature detectors. In: International Conference on Intelligent and Autonomous Systems (IAV), Toulouse, France (2007)
Bernardino, A., Santos-Victor, J.: A binocular stereo algorithm for log-polar foveated systems. In: Biological Motivated Computer Vision, Tuebingen, Germany (November 2002)
Prazdny, K.: Detection of binocular disparities. Biological Cybernetics 52(2), 93–99 (1985)
Figueira, D., Lopes, M., Ventura, R., Ruesch, J.: From pixels to objects: Enabling a spatial model for humanoid social robots. In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA 2009), Kobe, Japan (May 2009)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Figueira, D., Lopes, M., Ventura, R., Ruesch, J. (2009). Towards a Spatial Model for Humanoid Social Robots. In: Lopes, L.S., Lau, N., Mariano, P., Rocha, L.M. (eds) Progress in Artificial Intelligence. EPIA 2009. Lecture Notes in Computer Science(), vol 5816. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-04686-5_24
Download citation
DOI: https://doi.org/10.1007/978-3-642-04686-5_24
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-04685-8
Online ISBN: 978-3-642-04686-5
eBook Packages: Computer ScienceComputer Science (R0)