Automatic annotation of tennis action for content-based retrieval by integrated audio and visual information

  • Authors:
  • Hisashi Miyamori

  • Affiliations:
  • Keihanna Human Info-Communication Research Center, Communications Research Laboratory, Kyoto, Japan

  • Venue:
  • CIVR'03 Proceedings of the 2nd international conference on Image and video retrieval
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper proposes a method of automatically annotating tennis action through the integrated use of audio and video information. The proposed method extracts ball-hitting times called "impact times" using audio information, and evaluates the position relations between the player and the ball at the impact time to identify the player's basic actions, such as forehand swing, overhead swing, etc. Simulation results show that the detection rate for impact time influences the recognition rate of the player's basic actions. They also reveal that using audio information avoids some event recognition failures that cannot be averted when using only video information, demonstrating the performance and the validity of our approach.