Social interactive human video synthesis

  • Authors:
  • Dumebi Okwechime;Eng-Jon Ong;Andrew Gilbert;Richard Bowden

  • Affiliations:
  • CVSSP, University of Surrey, Guildford, Surrey, UK;CVSSP, University of Surrey, Guildford, Surrey, UK;CVSSP, University of Surrey, Guildford, Surrey, UK;CVSSP, University of Surrey, Guildford, Surrey, UK

  • Venue:
  • ACCV'10 Proceedings of the 10th Asian conference on Computer vision - Volume Part I
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we propose a computational model for social interaction between three people in a conversation, and demonstrate results using human video motion synthesis. We utilised semi-supervised computer vision techniques to label social signals between the people, like laughing, head nod and gaze direction. Data mining is used to deduce frequently occurring patterns of social signals between a speaker and a listener in both interested and not interested social scenarios, and the mined confidence values are used as conditional probabilities to animate social responses. The human video motion synthesis is done using an appearance model to learn a multivariate probability distribution, combined with a transition matrix to derive the likelihood of motion given a pose configuration. Our system uses social labels to more accurately define motion transitions and build a texture motion graph. Traditional motion synthesis algorithms are best suited to large human movements like walking and running, where motion variations are large and prominent. Our method focuses on generating more subtle human movement like head nods. The user can then control who speaks and the interest level of the individual listeners resulting in social interactive conversational agents.