Performance capture of interacting characters with handheld kinects

  • Authors:
  • Genzhi Ye;Yebin Liu;Nils Hasler;Xiangyang Ji;Qionghai Dai;Christian Theobalt

  • Affiliations:
  • Department of Automation, Tsinghua Univercity, Beijing, China;Department of Automation, Tsinghua Univercity, Beijing, China;Max-Planck Institute for Informatics, Saarbrücken, Germany;Department of Automation, Tsinghua Univercity, Beijing, China;Department of Automation, Tsinghua Univercity, Beijing, China;Max-Planck Institute for Informatics, Saarbrücken, Germany

  • Venue:
  • ECCV'12 Proceedings of the 12th European conference on Computer Vision - Volume Part II
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present an algorithm for marker-less performance capture of interacting humans using only three hand-held Kinect cameras. Our method reconstructs human skeletal poses, deforming surface geometry and camera poses for every time step of the depth video. Skeletal configurations and camera poses are found by solving a joint energy minimization problem which optimizes the alignment of RGBZ data from all cameras, as well as the alignment of human shape templates to the Kinect data. The energy function is based on a combination of geometric correspondence finding, implicit scene segmentation, and correspondence finding using image features. Only the combination of geometric and photometric correspondences and the integration of human pose and camera pose estimation enables reliable performance capture with only three sensors. As opposed to previous performance capture methods, our algorithm succeeds on general uncontrolled indoor scenes with potentially dynamic background, and it succeeds even if the cameras are moving.