An approximate image-space approach for real-time rendering of deformable translucent objects

  • Authors:
  • Yi Gong;Wei Chen;Long Zhang;Yun Zeng;Qunsheng Peng

  • Affiliations:
  • State Key Lab of CAD&CG, Zhejiang University, Hangzhou, China;State Key Lab of CAD&CG, Zhejiang University, Hangzhou, China;State Key Lab of CAD&CG, Zhejiang University, Hangzhou, China;State Key Lab of CAD&CG, Zhejiang University, Hangzhou, China;State Key Lab of CAD&CG, Zhejiang University, Hangzhou, China

  • Venue:
  • CGI'06 Proceedings of the 24th international conference on Advances in Computer Graphics
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Although lots of works have been engaged in interactive and realistic rendering of translucent materials, efficient processing for deformable models remains a challenging problem. In this paper we introduce an approximate image-space approach for real-time rendering of deformable translucent models by taking account of diffuse multiple sub-surface scattering. We decompose the process into two stages, called the Gathering and Scattering corresponding to the computations for incident and exiting irradiance respectively. We derive a simplified all-frequency illumination model for the gathering of the incident irradiance, which is amenable for deformable models using two auxiliary textures. We introduce two modes for efficient accomplishment of the view-dependent scattering. We implement our approach by fully exploiting the capabilities of graphics processing units (GPUs). Our implementation achieves visually plausible results and real-time frame rates for deformable models on commodity desktop PCs.