liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Next Generation Image Based Lighting using HDR Video
Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, The Institute of Technology. (Computer Graphics and Image Processing)ORCID iD: 0000-0002-7765-1747
Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, The Institute of Technology. (Computer Graphics and Image Processing)ORCID iD: 0000-0003-2559-6122
Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, The Institute of Technology. (Computer Graphics and Image Processing)
Linköping University, Department of Science and Technology, Media and Information Technology. Linköping University, The Institute of Technology. (Computer Graphics and Image Processing)
Show others and affiliations
2011 (English)In: Proceeding SIGGRAPH '11 ACM SIGGRAPH 2011 Talks, ACM Special Interest Group on Computer Science Education, 2011, article no 60- p.Conference paper, Published paper (Refereed)
Abstract [en]

We present an overview of our recently developed systems pipeline for capture, reconstruction, modeling and rendering of real world scenes based on state-of-the-art high dynamic range video (HDRV). The reconstructed scene representation allows for photo-realistic Image Based Lighting (IBL) in complex environments with strong spatial variations in the illumination. The pipeline comprises the following essential steps:

1.) Capture - The scene capture is based on a 4MPixel global shutter HDRV camera with a dynamic range of more than 24 f-stops at 30 fps. The HDR output stream is stored as individual un-compressed frames for maximum flexibility. A scene is usually captured using a combination of panoramic light probe sequences [1], and sequences with a smaller field of view to maximize the resolution at regions of special interest in the scene. The panoramic sequences ensure full angular coverage at each position and guarantee that the information required for IBL is captured. The position and orientation of the camera is tracked during capture.

2.) Scene recovery - Taking one or more HDRV sequences as input, a geometric proxy model of the scene is built using a semi-automatic approach. First, traditional computer vision algorithms such as structure from motion [2] and Manhattan world stereo [3] are used. If necessary, the recovered model is then modified using an interaction scheme based on visualizations of a volumetric representation of the scene radiance computed from the input HDRV sequence. The HDR nature of this volume also enables robust extraction of direct light sources and other high intensity regions in the scene.

3.) Radiance processing - When the scene proxy geometry has been recovered, the radiance data captured in the HDRV sequences are re-projected onto the surfaces and the recovered light sources. Since most surface points have been imaged from a large number of directions, it is possible to reconstruct view dependent texture maps at the proxy geometries. These 4D data sets describe a combination of detailed geometry that has not been recovered and the radiance reflected from the underlying real surfaces. The view dependent textures are then processed and compactly stored in an adaptive data structure.

4.) Rendering - Once the geometric and radiometric scene information has been recovered, it is possible to place virtual objects into the real scene and create photo-realistic renderings as illustrated above. The extracted light sources enable efficient sampling and rendering times that are fully comparable to that of traditional virtual computer graphics light sources. No previously described method is capable of capturing and reproducing the angular and spatial variation in the scene illumination in comparable detail.

We believe that the rapid development of high quality HDRV systems will soon have a large impact on both computer vision and graphics. Following this trend, we are developing theory and algorithms for efficient processing HDRV sequences and using the abundance of radiance data that is going to be available.

Place, publisher, year, edition, pages
ACM Special Interest Group on Computer Science Education, 2011. article no 60- p.
Keyword [en]
High Dynamic Range Video, Image Based Lighting
National Category
Media Engineering
Identifiers
URN: urn:nbn:se:liu:diva-75043DOI: 10.1145/2037826.2037906ISBN: 978-1-4503-0974-5 (print)OAI: oai:DiVA.org:liu-75043DiVA: diva2:502963
Conference
SIGGRAPH 2011, Vancouver, British Columbia, Canada, August 7–11, 2011
Available from: 2012-02-14 Created: 2012-02-14 Last updated: 2015-09-22Bibliographically approved

Open Access in DiVA

No full text

Other links

Publisher's full text

Authority records BETA

Unger, JonasGustavson, StefanKronander, JoelLarsson, Per

Search in DiVA

By author/editor
Unger, JonasGustavson, StefanKronander, JoelLarsson, Per
By organisation
Media and Information TechnologyThe Institute of Technology
Media Engineering

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 137 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • harvard1
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf