“Unstructured lumigraph rendering” by Buehler, Bosse, McMillan, Gortler and Cohen

  • ©Chris Buehler, Michael Bosse, Leonard McMillan, Steven J. Gortler, and Michael Cohen




    Unstructured lumigraph rendering



    We describe an image based rendering approach that generalizes many current image based rendering algorithms, including light field rendering and view-dependent texture mapping. In particular, it allows for lumigraph-style rendering from a set of input cameras in arbitrary configurations (i.e., not restricted to a plane or to any specific manifold). In the case of regular and planar input camera positions, our algorithm reduces to a typical lumigraph approach. When presented with fewer cameras and good approximate geometry, our algorithm behaves like view-dependent texture mapping. The algorithm achieves this flexibility because it is designed to meet a set of specific goals that we describe. We demonstrate this flexibility with a variety of examples.


    1. Jin-Xiang Chai, Xin Tong, Shing-Chow Chan, and Heung-Yeung Shum. Plenoptic sampling. SIGGRAPH 00, pages 307-318.
    2. Lucia Darsa, Bruno Costa Silva, and Amitabh Varshney. Navigating static environments using image-space simplification and morphing. 1997 Symposium on Interactive 3D Graphics, pages 25-34.
    3. P. Debevec, C. Taylor, and J. Malik. Modeling and rendering architecture from photographs. SIGGRAPH 96, pages 11-20.
    4. Paul E. Debevec, Yizhou Yu, and George D. Borshukov. Efficient viewdependent image-based rendering with projective texture-mapping. Eurographics Rendering Workshop 1998.
    5. Steven J. Gortler, Radek Grzeszczuk, Richard Szeliski, and Michael F. Cohen. The lumigraph. SIGGRAPH 96, pages 43-54.
    6. P. Heckbert and H. Moreton. Interpolation for polygon texture mapping and shading. State of the Art in Computer Graphics: Visualization and Modeling, 1991.
    7. B. Heigl, R. Koch, M. Pollefeys, J. Denzler, and L. Van Gool. Plenoptic modeling and rendering from image sequences taken by hand-held camera. Proc. DAGM 99, pages 94-101.
    8. Y. Horry, K. Anjyo, and K. Arai. Tour into the picture: Using a spidery mesh interface to make animation from a single image. SIGGRAPH 97, pages 225- 232.
    9. A Isaksen, L. McMillan, and S. Gortler. Dynamically reparameterized light fields. SIGGRAPH ’00, pages 297-306.
    10. M. Levoy and P. Hanrahan. Light field rendering. SIGGRAPH 96, pages 31-42.
    11. Marc Levoy, Kari Pulli, Brian Curless, Szymon Rusinkiewicz, David Koller, Lucas Pereira, Matt Ginzton, Sean Anderson, James Davis, Jeremy Ginsberg, Jonathan Shade, and Duane Fulk. The digital michelangelo project: 3d scanning of large statues. SIGGRAPH 2000, pages 131-144.
    12. F. Pighin, J. Hecker, D. Lischinski, R. Szeliski, and D. H. Salesin. Synthesizing realistic facial expressions from photographs. SIGGRAPH 98, pages 75-84.
    13. Kari Pulli, Michael Cohen, Tom Duchamp, Hugues Hoppe, Linda Shapiro, and Werner Stuetzle. View-based rendering: Visualizing real objects from scanned range and color data. Eurographics Rendering Workshop 1997, pages 23-34.
    14. Ramesh Raskar, Michael S. Brown, Ruigang Yang, Wei-Chao Chen, Greg Welch, Herman Towles, Brent Seales, and Henry Fuchs. Multi-projector displays using camera-based registration. IEEE Visualization ’99, pages 161-168.
    15. Pedro V. Sander, Xianfeng Gu, Steven J. Gortler, Hugues Hoppe, and John Snyder. Silhouette clipping. SIGGRAPH 2000, pages 327-334.
    16. Jonathan Richard Shewchuk. Triangle: Engineering a 2d quality mesh generator and delaunay triangulator. First Workshop on Applied Computational Geometry, pages 124-133, 1996.
    17. Heung-Yeung Shum and Li-Wei He. Rendering with concentric mosaics. SIG- GRAPH 99, pages 299-306.
    18. Daniel N. Wood, Daniel I. Azuma, Ken Aldinger, Brian Curless, Tom Duchamp, David H. Salesin, and Werner Stuetzle. Surface light fields for 3d photography. SIGGRAPH 2000, pages 287-296.

ACM Digital Library Publication: