Extracting spatio-temporal local features considering consecutiveness of motions

  • Authors:
  • Akitsugu Noguchi;Keiji Yanai

  • Affiliations:
  • Department of Computer Science, The University of Electro-Communications, Tokyo, Japan;Department of Computer Science, The University of Electro-Communications, Tokyo, Japan

  • Venue:
  • ACCV'09 Proceedings of the 9th Asian conference on Computer Vision - Volume Part II
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recently spatio-temporal local features have been proposed as image features to recognize events or human actions in videos. In this paper, we propose yet another local spatio-temporal feature based on the SURF detector, which is a lightweight local feature. Our method consists of two parts: extracting visual features and extracting motion features. First, we select candidate points based on the SURF detector. Next, we calculate motion features at each point with local temporal units divided in order to consider consecutiveness of motions. Since our proposed feature is intended to be robust to rotation, we rotate optical flow vectors to the main direction of extracted SURF features. In the experiments, we evaluate the proposed spatio-temporal local feature with the common dataset containing six kinds of simple human actions. As the result, the accuracy achieves 86%, which is almost equivalent to state-of-the-art. In addition, we make experiments to classify large amounts of Web video clips downloaded from Youtube.