Dynamic Three-Dimensional Human Model

  • Y. Iwadate


Three-dimensional (3-D) content is widely used in movies and video games, thanks to advances in computer graphics, motion capture, 3-D scanners, and so on. However, producing high quality 3-D content is expensive and involves many different technologies. The dynamic 3-D model discussed in this chapter contains shape information about a moving subject. Thus, it can express the motions of humans in detail, including movements of clothes. It is useful as a component in video synthesis, because of its flexibility in viewpoint and view direction. In addition, it can be transformed into stereoscopic images such as binocular, lenticular, integral photographic, etc.

To generate a dynamic 3-D model, 10 or more cameras are set up in a dome surrounding a subject to synchronously take video images. The model is calculated frame by frame from these video images. There are many choices from which to choose a shape restoration algorithm. Of these, the popular volume-intersection method...


Texture Image Camera Image Depth Image Approximate Model Texture Mapping 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.



The support of the Ministry of Education, Culture and Sports, Science and Technology Japan is gratefully acknowledged.


  1. [1]
    A. Laurentini: “The Visual Hull Concept for Silhouette-based Image Understanding," IEEE Trans. on PAMI, 16, 2, 150–162 (1994)CrossRefGoogle Scholar
  2. [2]
    P. Rander: “A Multi-Camera Method for 3D Digitization of Dynamic Real-World Events,” CMU-RI-TR-9812 (1998)Google Scholar
  3. [3]
    Y. Iwadate, M. Katayama, K. Tomiyama, and H. Imaizumi: “VRML Animation from Multi-Viewpoint Images,” Proc. of ICME2002, 1, 881–884 (2002)Google Scholar
  4. [4]
    S.M. Seitz and C.R. Dyer: “Photorealistic Scene Reconstruction by Voxel Coloring," Int. J. Computer Vis., 35, 2, 1–23 (1999)CrossRefGoogle Scholar
  5. [5]
    W. Matusik, C. Buehler, R. Rasker, S.J. Gortler, and L. McMillan: “Image-Based Visual Hulls," In Proc. of SIGGRAPH2000,11–20 (2000)Google Scholar
  6. [6]
    G. Slabaugh, R.W. Schafer, and M.C. Hans: “Image-Based Photo Hulls," 1st International Symposium on 3D Processing Visualization and transmission, 704–708 (2002)Google Scholar
  7. [7]
    S. Nobuhara and T. Matsuyama: “Deformable Mesh Model for Complex Multi-Object 3D Motion Estimation from Multi-Viewpoint Video," 3DPVT2006 (2006)Google Scholar
  8. [8]
    P.E. Debvec, C.J. Taylor, and J. Malik: “Modeling and Rendering Architecture from Photographs: A Hybrid Geometry- and Image-Based Approach," Proc. SIGGRAPH1996, 11–20 (1996)Google Scholar
  9. [9]
    K. Pulli, M. Cohen, T. Duchamp, H. Hoppe, L. Shapiro, and W. Stuetzle: “View-based Rendering: Visualizing Real Objects from Scanned Range and Color Data," Proc. of 8th Eurographics Workshop on Rendering, 23–34 (1997)Google Scholar
  10. [10]
    P.E. Debevec, G. Borshukov, and Y. Yu: “Efficient View-Dependent Image-Based Rendering with Projective Texture-Mapping," Proc. 9th Eurographics Workshop on Rendering, 105–116 (1998)Google Scholar
  11. [11]
    K. Tomiyama, M. Katayama, Y. Orihara, and Y. Iwadate: “Arbitrary Viewpoint Images for Performances of Japanese Traditional Art," The 2nd European Conference on Visual Production, 68–75 (2005)Google Scholar
  12. [12]
    M. Katayama, K. Tomiyama, Y. Orihara, and Y. Iwadate: “Archiving Performances of Japanese Traditional Dramatic Arts with a Dynamic 3D Model," Proc. EVA 2006 Vienna, 51–58 (2006)Google Scholar
  13. [13]
    R. Y. Tsai: “A versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using Off-the-Shelf TV Cameras and Lenses," IEEE Journal of Robotics and Automation, Vol. RA-3, No.4, 323–344 (1987)Google Scholar
  14. [14]
    Y. Kenmochi, K. Kotani, and A. Imiya: “Marching Cubes Method with Connectivity ," ICIP-99, 361–365 (1999)Google Scholar
  15. [15]
    The Japan Arts Council: “An Introduction to the World of Noh & kyogen,"

Copyright information

© Springer Science+Business Media, LLC 2009

Authors and Affiliations

  1. 1.NHK (Japan Broadcasting Corporation) Science & Technical Research Laboratories1-10-11, KinutaSetagaya-kuJapan

Personalised recommendations