Refocusing images captured from a stereoscopic camera

  • Authors:
  • Chia-Lun Ku;Yu-Shuen Wang;Chia-Sheng Chang;Hung-Kuo Chu;Chih-Yuan Yao

  • Affiliations:
  • National Chiao Tung University;National Chiao Tung University;National Tsing Hua University;National Tsing Hua University;National Taiwan University of Science and Technology

  • Venue:
  • ACM SIGGRAPH 2013 Posters
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Traditional photography projects a 3D scene to a 2D image without recording the depth of each local region, which prevents users from changing the focus plane of a photograph once it has been taken. To tackle this problem, Ng et al. [2005] presented light-field cameras that record all focus planes of a scene and synthesized the refocused image using ray tracing. Nevertheless, the captured photographs are of low resolution because the image sensor is divided into sub-cells. Levin et al. [2007] embedded a coded aperture on the camera lens and recover depth information from blur patterns in a single image. However, the coded aperture blocks around 50% of light. Their system requires longer exposition time when taking pictures. Liang et al. [2008] also embedded a coded aperture on the camera lens to capture the scene but with multiple exposures. It produces high quality depth maps yet is not suitable to hand-held devices. Recently, Microsoft Kinect directly estimates the depth information using infrared light, which works only in a indoor environment.