Results 1 - 10
of
1,025
Recovering High Dynamic Range Radiance Maps from Photographs
"... We present a method of recovering high dynamic range radiance maps from photographs taken with conventional imaging equipment. In our method, multiple photographs of the scene are taken with different amounts of exposure. Our algorithm uses these differently exposed photographs to recover the respon ..."
Abstract
-
Cited by 859 (15 self)
- Add to MetaCart
(Show Context)
We present a method of recovering high dynamic range radiance maps from photographs taken with conventional imaging equipment. In our method, multiple photographs of the scene are taken with different amounts of exposure. Our algorithm uses these differently exposed photographs to recover the response function of the imaging process, up to factor of scale, using the assumption of reciprocity. With the known response function, the algorithm can fuse the multiple photographs into a single, high dynamic range radiance map whose pixel values are proportional to the true radiance values in the scene. We demonstrate our method on images acquired with both photochemical and digital imaging processes. We discuss how this work is applicable in many areas of computer graphics involving digitized photographs, including image-based modeling, image compositing, and image processing. Lastly, we demonstrate a few applications of having high dynamic range radiance maps, such as synthesizing realistic motion blur and simulating the response of the human visual system.
Photo tourism: Exploring photo collections in 3D
- IN PROC. ACM SIGGRAPH
, 2006
"... We present a system for interactively browsing and exploring large unstructured collections of photographs of a scene using a novel 3D interface. Our system consists of an image-based modeling front end that automatically computes the viewpoint of each photograph as well as a sparse 3D model of th ..."
Abstract
-
Cited by 677 (37 self)
- Add to MetaCart
We present a system for interactively browsing and exploring large unstructured collections of photographs of a scene using a novel 3D interface. Our system consists of an image-based modeling front end that automatically computes the viewpoint of each photograph as well as a sparse 3D model of the scene and image to model correspondences. Our photo explorer uses image-based rendering techniques to smoothly transition between photographs, while also enabling full 3D navigation and exploration of the set of images and world geometry, along with auxiliary information such as overhead maps. Our system also makes it easy to construct photo tours of scenic or historic locations, and to annotate image details, which are automatically transferred to other relevant images. We demonstrate our system on several large personal photo collections as well as images gathered from Internet photo sharing sites.
Layered depth images
, 1997
"... In this paper we present an efficient image based rendering system that renders multiple frames per second on a PC. Our method performs warping from an intermediate representation called a layered depth image (LDI). An LDI is a view of the scene from a single input camera view, but with multiple pix ..."
Abstract
-
Cited by 456 (29 self)
- Add to MetaCart
In this paper we present an efficient image based rendering system that renders multiple frames per second on a PC. Our method performs warping from an intermediate representation called a layered depth image (LDI). An LDI is a view of the scene from a single input camera view, but with multiple pixels along each line of sight. When n input images are preprocessed to create a single LDI, the size of the representation grows linearly only with the observed depth complexity in the n images, instead of linearly with n. Moreover, because the LDI data are represented in a single image coordinate system, McMillan's warp ordering algorithm can be successfully adapted. As a result, pixels are drawn in the output image in back to front order. No z-buffer is required, so alpha-compositing can be done efficiently without depth sorting. This makes splatting an efficient solution to the resampling problem.
Rendering Synthetic Objects into Real Scenes: Bridging Traditional and Image-based Graphics with Global Illumination and High Dynamic Range Photography
, 1998
"... We present a method that uses measured scene radiance and global illumination in order to add new objects to light-based models with correct lighting. The methodusesahighdynamicrangeimagebasedmodelofthescene, ratherthansyntheticlightsources,toilluminatethe new objects. Tocomputetheillumination,thesc ..."
Abstract
-
Cited by 438 (17 self)
- Add to MetaCart
We present a method that uses measured scene radiance and global illumination in order to add new objects to light-based models with correct lighting. The methodusesahighdynamicrangeimagebasedmodelofthescene, ratherthansyntheticlightsources,toilluminatethe new objects. Tocomputetheillumination,thesceneis consideredasthreecomponents:thedistantscene, the localscene, andthesyntheticobjects. The
distant scene is assumed
tobephotometricallyunaffectedbytheobjects, obviatingtheneedforreflectancemodelinformation. Thelocalsceneisendowedwithestimatedreflectancemodel
informationsothatitcancatchshadows andreceivereflectedlightfromthenewobjects. Renderings are createdwithastandardglobalilluminationmethodby simulating theinteractionoflightamongstthethreecomponents.
A differentialrenderingtechniqueallowsforgoodresults
to be obtained when only an estimate ofthelocalscenereflectancepropertiesisknown. Weapplythegeneralmethodtotheproblemofrendering
syntheticobjectsintorealscenes.
The light-based model is constructed from an approximategeometricmodelofthesceneandbyusinga lightprobetomeasuretheincidentilluminationatthe locationof thesyntheticobjects.
Theglobalilluminationsolutionisthen
compositedintoaphotographofthesceneusing
thedifferentialrenderingtechnique.
Weconcludebydiscussingtherelevance of the technique to recovering surface reflectance properties in uncontrolled lighting situations. Applications of the method include visual effects, interior design, and architectural visualization.
Creating Full View Panoramic Image Mosaics and Environment Maps
, 1997
"... This paper presents a novel approach to creating full view panoramic mosaics from image sequences. Unlike current panoramic stitching methods, which usually require pure horizontal camera panning, our system does not require any controlled motions or constraints on how the images are taken (as long ..."
Abstract
-
Cited by 340 (29 self)
- Add to MetaCart
This paper presents a novel approach to creating full view panoramic mosaics from image sequences. Unlike current panoramic stitching methods, which usually require pure horizontal camera panning, our system does not require any controlled motions or constraints on how the images are taken (as long as there is no strong motion parallax). For example, images taken from a hand-held digital camera can be stitched seamlessly into panoramic mosaics. Because we represent our image mosaics using a set of transforms, there are no singularity problems such as those existing at the top and bottom of cylindrical or spherical maps. Our algorithm is fast and robust because it directly recovers 3D rotations instead of general 8 parameter planar perspective transforms. Methods to recover camera focal length are also presented. We also present an algorithm for efficiently extracting environment maps from our image mosaics. By mapping the mosaic onto an artibrary texture-mapped polyhedron surrounding t...
Unstructured lumigraph rendering
- In Computer Graphics, SIGGRAPH 2001 Proceedings
, 2001
"... We describe an image based rendering approach that generalizes many image based rendering algorithms currently in use including light field rendering and view-dependent texture mapping. In particular it allows for lumigraph style rendering from a set of input cameras that are not restricted to a pla ..."
Abstract
-
Cited by 291 (11 self)
- Add to MetaCart
(Show Context)
We describe an image based rendering approach that generalizes many image based rendering algorithms currently in use including light field rendering and view-dependent texture mapping. In particular it allows for lumigraph style rendering from a set of input cameras that are not restricted to a plane or to any specific manifold. In the case of regular and planar input camera positions, our algorithm reduces to a typical lumigraph approach. In the case of fewer cameras and good approximate geometry, our algorithm behaves like view-dependent texture mapping. Our algorithm achieves this flexibility because it is designed to meet a set of desirable goals that we describe. We demonstrate this flexibility with a variety of examples. Keyword Image-Based Rendering 1
Efficient View-Dependent Image-Based Rendering with Projective Texture-Mapping
"... . This paper presents how the image-based rendering technique of viewdependent texture-mapping (VDTM) can be efficiently implemented using projective texture mapping, a feature commonly available in polygon graphics hardware. VDTM is a technique for generating novel views of a scene with approximate ..."
Abstract
-
Cited by 277 (10 self)
- Add to MetaCart
(Show Context)
. This paper presents how the image-based rendering technique of viewdependent texture-mapping (VDTM) can be efficiently implemented using projective texture mapping, a feature commonly available in polygon graphics hardware. VDTM is a technique for generating novel views of a scene with approximately known geometry making maximal use of a sparse set of original views. The original presentation of VDTM in by Debevec, Taylor, and Malik required significant per-pixel computation and did not scale well with the number of original images. In our technique, we precompute for each polygon the set of original images in which it is visibile and create a "view map" data structure that encodes the best texture map to use for a regularly sampled set of possible viewing directions. To generate a novel view, the view map for each polygon is queried to determine a set of no more than three original images to blended together in order to render the polygon with projective texture-mapping. Invisible t...
Modeling the World from Internet Photo Collections
- INT J COMPUT VIS
, 2007
"... There are billions of photographs on the Internet, comprising the largest and most diverse photo collection ever assembled. How can computer vision researchers exploit this imagery? This paper explores this question from the standpoint of 3D scene modeling and visualization. We present structure-fro ..."
Abstract
-
Cited by 267 (6 self)
- Add to MetaCart
There are billions of photographs on the Internet, comprising the largest and most diverse photo collection ever assembled. How can computer vision researchers exploit this imagery? This paper explores this question from the standpoint of 3D scene modeling and visualization. We present structure-from-motion and image-based rendering algorithms that operate on hundreds of images downloaded as a result of keyword-based image search queries like “Notre Dame ” or “Trevi Fountain.” This approach, which we call Photo Tourism, has enabled reconstructions of numerous well-known world sites. This paper presents these algorithms and results as a first step towards 3D modeling of the world’s well-photographed sites, cities, and landscapes from Internet imagery, and discusses key open problems and challenges for the research community.