“Unstructured lumigraph rendering” by Buehler, Bosse, McMillan, Gortler and Cohen
Conference:
Type(s):
Title:
- Unstructured lumigraph rendering
Presenter(s)/Author(s):
Abstract:
We describe an image based rendering approach that generalizes many current image based rendering algorithms, including light field rendering and view-dependent texture mapping. In particular, it allows for lumigraph-style rendering from a set of input cameras in arbitrary configurations (i.e., not restricted to a plane or to any specific manifold). In the case of regular and planar input camera positions, our algorithm reduces to a typical lumigraph approach. When presented with fewer cameras and good approximate geometry, our algorithm behaves like view-dependent texture mapping. The algorithm achieves this flexibility because it is designed to meet a set of specific goals that we describe. We demonstrate this flexibility with a variety of examples.
References:
1. Jin-Xiang Chai, Xin Tong, Shing-Chow Chan, and Heung-Yeung Shum. Plenoptic sampling. SIGGRAPH 00, pages 307-318.
2. Lucia Darsa, Bruno Costa Silva, and Amitabh Varshney. Navigating static environments using image-space simplification and morphing. 1997 Symposium on Interactive 3D Graphics, pages 25-34.
3. P. Debevec, C. Taylor, and J. Malik. Modeling and rendering architecture from photographs. SIGGRAPH 96, pages 11-20.
4. Paul E. Debevec, Yizhou Yu, and George D. Borshukov. Efficient viewdependent image-based rendering with projective texture-mapping. Eurographics Rendering Workshop 1998.
5. Steven J. Gortler, Radek Grzeszczuk, Richard Szeliski, and Michael F. Cohen. The lumigraph. SIGGRAPH 96, pages 43-54.
6. P. Heckbert and H. Moreton. Interpolation for polygon texture mapping and shading. State of the Art in Computer Graphics: Visualization and Modeling, 1991.
7. B. Heigl, R. Koch, M. Pollefeys, J. Denzler, and L. Van Gool. Plenoptic modeling and rendering from image sequences taken by hand-held camera. Proc. DAGM 99, pages 94-101.
8. Y. Horry, K. Anjyo, and K. Arai. Tour into the picture: Using a spidery mesh interface to make animation from a single image. SIGGRAPH 97, pages 225- 232.
9. A Isaksen, L. McMillan, and S. Gortler. Dynamically reparameterized light fields. SIGGRAPH ’00, pages 297-306.
10. M. Levoy and P. Hanrahan. Light field rendering. SIGGRAPH 96, pages 31-42.
11. Marc Levoy, Kari Pulli, Brian Curless, Szymon Rusinkiewicz, David Koller, Lucas Pereira, Matt Ginzton, Sean Anderson, James Davis, Jeremy Ginsberg, Jonathan Shade, and Duane Fulk. The digital michelangelo project: 3d scanning of large statues. SIGGRAPH 2000, pages 131-144.
12. F. Pighin, J. Hecker, D. Lischinski, R. Szeliski, and D. H. Salesin. Synthesizing realistic facial expressions from photographs. SIGGRAPH 98, pages 75-84.
13. Kari Pulli, Michael Cohen, Tom Duchamp, Hugues Hoppe, Linda Shapiro, and Werner Stuetzle. View-based rendering: Visualizing real objects from scanned range and color data. Eurographics Rendering Workshop 1997, pages 23-34.
14. Ramesh Raskar, Michael S. Brown, Ruigang Yang, Wei-Chao Chen, Greg Welch, Herman Towles, Brent Seales, and Henry Fuchs. Multi-projector displays using camera-based registration. IEEE Visualization ’99, pages 161-168.
15. Pedro V. Sander, Xianfeng Gu, Steven J. Gortler, Hugues Hoppe, and John Snyder. Silhouette clipping. SIGGRAPH 2000, pages 327-334.
16. Jonathan Richard Shewchuk. Triangle: Engineering a 2d quality mesh generator and delaunay triangulator. First Workshop on Applied Computational Geometry, pages 124-133, 1996.
17. Heung-Yeung Shum and Li-Wei He. Rendering with concentric mosaics. SIG- GRAPH 99, pages 299-306.
18. Daniel N. Wood, Daniel I. Azuma, Ken Aldinger, Brian Curless, Tom Duchamp, David H. Salesin, and Werner Stuetzle. Surface light fields for 3d photography. SIGGRAPH 2000, pages 287-296.