Abstract
Object segmentation in infant’s egocentric videos is a fundamental step in studying how children perceive objects in early stages of development. From the computer vision perspective, object segmentation in such videos poses quite a few challenges because the child’s view is unfocused, often with large head movements, effecting in sudden changes in the child’s point of view which leads to frequent change in object properties such as size, shape and illumination. In this paper, we develop a semi-automated, domain specific method, to address these concerns and facilitate the object annotation process for cognitive scientists, allowing them to select and monitor the object under segmentation. The method starts with an annotation of the desired object by user and employs graph cut segmentation and optical flow computation to predict the object mask for subsequent video frames automatically. To maintain accurate segmentation of objects, we use domain specific heuristic rules to re-initialize the program with new user input whenever object properties change dramatically. The evaluations demonstrate the high speed and accuracy of the presented method for object segmentation in voluminous egocentric videos. We apply the proposed method to investigate potential patterns in object distribution in child’s view at progressive ages.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Pereira, A.F., Smith, L.B., Yu, C.: A bottom-up view of toddler word learning. Psychonomic bulletin & review 21(1), 178–185 (2014)
Pereira, A.F., Yu, C., Smith, L.B., Shen, H.: A first-person perspective on a parent-child social interaction during object play. In: Proceedings of the 31st Annual Meeting of the Cognitive Science Society (2009)
Smith, L.B., Yu, C., Pereira, A.F.: Not your mothers view: The dynamics of toddler visual experience. Developmental science 14(1), 9–17 (2011)
Bambach, S., Crandall, D.J., Yu, C.: Understanding embodied visual attention in child-parent interaction. In: Development and Learning and Epigenetic Robotics (ICDL), 2013 IEEE Third Joint International Conference on. pp. 1–6. IEEE (2013)
Burling, J.M., Yoshida, H., Nagai, Y.: The significance of social input, early motion experiences, and attentional selection. In: Development and Learning and Epigenetic Robotics (ICDL), 2013 IEEE Third Joint International Conference on. pp. 1–2. IEEE (2013)
Xu, T., Chen, Y., Smith, L.: It’s the child’s body: The role of toddler and parent in selecting toddler’s visual experience. In: Development and Learning (ICDL), 2011 IEEE International Conference on. vol. 2, pp. 1–6. IEEE (2011)
Yoshida, H., Smith, L.B.: What’s in view for toddlers? Using a head camera to study visual experience. Infancy 13(3), 229–248 (2008)
Smith, L., Yu, C., Yoshida, H., Fausey, C.M.: Contributions of Head-Mounted Cameras to Studying the Visual Environments of Infants and Young Children. Journal of Cognition and Development (just-accepted) (2014)
Bambach, S.: A Survey on Recent Advances of Computer Vision Algorithms for Egocentric Video. arXiv preprint arXiv:1501.02825 (2015)
Ren, X., Gu, C.: Figure-ground segmentation improves handled object recognition in egocentric video. In: Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on. pp. 3137–3144. IEEE (2010)
Ren, X., Philipose, M.: Egocentric recognition of handled objects: Benchmark and analysis. In: Computer Vision and Pattern Recognition Workshops, 2009. CVPR Workshops 2009. IEEE Computer Society Conference on. pp. 1–8. IEEE (2009)
Boykov, Y., Kolmogorov, V.: An experimental comparison of min-cut/max-flow algorithms for energy minimization in vision. Pattern Analysis and Machine Intelligence, IEEE Transactions on 26(9), 1124–1137 (2004)
Boykov, Y., Veksler, O., Zabih, R.: Fast approximate energy minimization via graph cuts. Pattern Analysis and Machine Intelligence, IEEE Transactions on 23(11), 1222–1239 (2001)
Horn, B.K., Schunck, B.G.: Determining optical flow. In: 1981 Technical Symposium East. pp. 319–331. International Society for Optics and Photonics (1981)
Yoshida, H., Burling, J.M.: Dynamic shift in isolating referents: From social to self-generated input. In: Development and Learning and Epigenetic Robotics (ICDL), 2013 IEEE Third Joint International Conference on. pp. 1–2. IEEE (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer Science+Business Media Singapore
About this paper
Cite this paper
Mirsharif, Q., Sadani, S., Shah, S., Yoshida, H., Burling, J. (2017). A Semi-automated Method for Object Segmentation in Infant’s Egocentric Videos to Study Object Perception. In: Raman, B., Kumar, S., Roy, P., Sen, D. (eds) Proceedings of International Conference on Computer Vision and Image Processing. Advances in Intelligent Systems and Computing, vol 460. Springer, Singapore. https://doi.org/10.1007/978-981-10-2107-7_6
Download citation
DOI: https://doi.org/10.1007/978-981-10-2107-7_6
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-2106-0
Online ISBN: 978-981-10-2107-7
eBook Packages: EngineeringEngineering (R0)