Active Image Capturing and Dynamic Scene Visualization by Cooperative Distributed Vision

  • Takashi Matsuyama
  • Toshikazu Wada
  • Shogo Tokai
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 1554)


This paper addresses active image capturing and dynamic scene visualization by Cooperative Distributed Vision (CDV, in short). The concept of CDV was proposed by our five years project starting from 1996. From a practical point of view, the goal of CDV is summarized as follows: Embed in the real world a group of network-connected Observation Stations (real time video image processor with active camera(s)) and mobile robots with vision. And realize 1) wide-area dynamic scene understanding and 2) versatile scene visualization. Applications of CDV include real time wide-area surveillance, remote conference and lecturing systems, interactive 3D TV and intelligent TV studio, navigation of (non-intelligent) mobile robots and disabled people, cooperative mobile robots, and so on. In this paper, we first define the framework of CDV and give a brief retrospective view of the computer vision research to show the background of CDV. Then we present technical research results so far obtained: 1) fixed viewpoint pan-tilt-zoom camera for wide-area active imaging, 2) moving object detection and tracking for reactive image acquisition, 3) multi-viewpoints object imaging by cooperative observation stations, and 4) scenario-based cooperative camera-work planning for dynamic scene visualization. Prototype systems demonstrate the effectiveness and practical utilities of the proposed methods.


Mobile Robot Dynamic Scene Camera Control Move Object Detection Background Scene 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Matsuyama, T.: Cooperative Distributed Vision-Dynamic Integration of Visual Perception, Action, and Communication-, Proc. of Image Understanding Workshop, Monterey CA, 1998.11Google Scholar
  2. 2.
    Aloimonos, Y. (ed.): Special Issue on Purposive, Qualitative, Active Vision, CVGIP: Image Understanding, Vol.56 No.1, 1992.Google Scholar
  3. 3.
    Aloimonos, Y. (ed.): Active Perception, Lawrence Erlbaum Associates Publisher, 1993Google Scholar
  4. 4.
    Yagi Y. and Yachida M.: Real-Time Generation of Environmental Map and Obstacle Avoidance Using Omnidirectional Image Sensor with Conic Mirror, Prof. of CVPR, 160–165, 1991.Google Scholar
  5. 5.
    Yamazawa K., Yagi Y. and Yachida M.: Obstacle Detection with Omnidirectional Image Sensor HyperOmni Vision, Proc. of ICRA, pp.1062–1067, 1995.Google Scholar
  6. 6.
    Peri V. N. and Nayar S. K.: Generation of Perspective and Panoramic Video from Omnidirectional Video, Proc. of IUW, pp. 243–245, 1997.Google Scholar
  7. 7.
    Murray, D. and Basu, A.: Motion Tracking with an Active Camera, IEEE Trans. of PAMI, Vol. 16, No. 5, pp. 449–459, 1994.Google Scholar
  8. 8.
    Wada T. and Matsuyama T.: Appearance Sphere: Background Model for Pan-Tilt-Zoom Camera, Proc. of ICPR, Vol. A, pp. 718–722, 1996.Google Scholar
  9. 9.
    Lavest, J.M., Delherm, C., Peuchot, B, and Daucher, N.: Implicit Reconstruction by Zooming, Computer Vision and Image Understanding, Vol.66, No.3, pp.301–315, 1997.CrossRefGoogle Scholar
  10. 10.
    Hall R.: Hybrid Techniques for Rapid Image Synthesis, in Image Rendering Tricks (Whitted T. and Cook R. eds.), Course Notes 16 for SIGGRAPH’86, 1986.Google Scholar
  11. 11.
    Greene N.: Environment Mapping and Other Applications of World Projections, CGA, 6(11), pp. 21–29, 1986.Google Scholar
  12. 12.
    Chen S.E.: QuickTime VR-An Image-Based Approach to Virtual Environment Navigation, Proc. of SIGGRAPH’95, pp. 29–38, 1995.Google Scholar
  13. 13.
    Nakai, H.: Robust Object Detection Using A-Posteriori Probability, Tech. Rep. of IPSJ, SIG-CV90-1, 1994 (in Japanese).Google Scholar
  14. 14.
    Grimson, E.: A Forest of Sensors, Proc. of VSAM Workshop, 1997.Google Scholar
  15. 15.
    Davis, L.: Visual Surveillance and Monitoring, Proc. of VSAM Workshop, 1997.Google Scholar
  16. 16.
    Habe, H., Ohya, T., and Matsuyama, T.: A Robust Background Subtraction Method for Non-Stationary Scenes, Proc. of MIRU’98, Vol.1, pp.467–472, 1998 (in Japanese).Google Scholar
  17. 17.
    Yamaashi, K., Cooperstock, J.R., Narine, T., and Buxton, W.: Beating the Limitations of Camera-Monitor Mediated Telepresence with Extra Eyes, Proc. of CHI, pp.50–57, 1996.Google Scholar
  18. 18.
    Hiura, S. and Matsuyama, T.: Depth Measurement by the Multi-Focus Camera, Proc. of CVPR, pp.953–959, 1998Google Scholar
  19. 19.
    Mikoshi, Y.: 3D Image Measurement Based on Planes, Master Thesis, Kyoto University, 1998 (in Japanese)Google Scholar
  20. 20.
    Arijon, D.: Grammar of the Film Language, Focal Press Ltd., London, 1976Google Scholar
  21. 21.
    He, L., Cohen, M.F., and Salesin, D.H.: The Virtual Cinematographer: A Paradigm for Automatic Real-Time Camera Control and Directing, SIGGRAPH’96, pp.217–224, 1996.Google Scholar
  22. 22.
    Christianson,D.B., Anderson, S.E., He, L., Weld, D.S., Cohen, M.F., and Salesin, D.H.: Declarative Camera Control for Automatic Cinematography, Proceedings of AAAI’ 96, pp.148–155, 1996.Google Scholar
  23. 23.
    Mase, K., Pinhanez, C.S., and Bobick, A.F.: Scripting Method Based on Temporal Intervals for Designing Interactive Systems, Trans. of IPSJ, Vol.39, No.5, pp.1403–1413, 1998 (in Japanese).Google Scholar
  24. 24.
    Allen, J.F.: Towards a General Theory of Action and Time, Artificial Intelligence, Vol.23, pp.123–154, 1984.zbMATHCrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 1999

Authors and Affiliations

  • Takashi Matsuyama
    • 1
  • Toshikazu Wada
    • 1
  • Shogo Tokai
    • 1
  1. 1.Department of Intelligence Science and TechnologyKyoto UniversityKyotoJapan

Personalised recommendations