Depth-of-field rendering with multiview synthesis

  • Authors:
  • Sungkil Lee;Elmar Eisemann;Hans-Peter Seidel

  • Affiliations:
  • Max-Planck-Institut für Informatik;Saarland University / MPI Informatik;Max-Planck-Institut für Informatik

  • Venue:
  • ACM SIGGRAPH Asia 2009 papers
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a GPU-based real-time rendering method that simulates high-quality depth-of-field effects, similar in quality to multiview accumulation methods. Most real-time approaches have difficulties to obtain good approximations of visibility and view-dependent shading due to the use of a single view image. Our method also avoids the multiple rendering of a scene, but can approximate different views by relying on a layered image-based scene representation. We present several performance and quality improvements, such as early culling, approximate cone tracing, and jittered sampling. Our method achieves artifact-free results for complex scenes and reasonable depth-of-field blur in real time.