Markerless Human Motion Transfer

  • Authors:
  • German K. M. Cheung;Simon Baker;Jessica Hodgins;Takeo Kanade

  • Affiliations:
  • Carnegie Mellon University, Pittsburgh PA;Carnegie Mellon University, Pittsburgh PA;Carnegie Mellon University, Pittsburgh PA;Carnegie Mellon University, Pittsburgh PA

  • Venue:
  • 3DPVT '04 Proceedings of the 3D Data Processing, Visualization, and Transmission, 2nd International Symposium
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we develop a computer vision-based system to transfer human motion from one subject to another. Our system uses a network of eight calibrated and synchronized cameras. We first build detailed kinematic models of the subjects based on our algorithms for extracting shape from silhouette across time [A 3D reconstruction algorithm combining shape-frame-shilhouette]. These models are then used to capture the motion (joint angles) of the subjects in new video sequences. Finally we describe an image-based rendering algorithm to render the captured motion applied to the articulated model of another person. Our rendering algorithm uses an ensemble of spatially and temporally distributed images to generate photo-realistic video of the transferred motion. We demonstrate the performance of the system by rendering throwing and kungfu motions on subjects who did not perform them.