Fast action recognition using negative space features

  • Authors:
  • Shah Atiqur Rahman;Insu Song;M. K. H. Leung;Ickjai Lee;Kyungmi Lee

  • Affiliations:
  • School of Business and IT, James Cook University Australia, Singapore 574421, Singapore;School of Business and IT, James Cook University Australia, Singapore 574421, Singapore;FICT, Universiti Tunku Abdul Rahman (Kampar), Malaysia;School of Business (IT), James Cook University Cairns, QLD 4870, Australia;School of Business (IT), James Cook University Cairns, QLD 4870, Australia

  • Venue:
  • Expert Systems with Applications: An International Journal
  • Year:
  • 2014

Quantified Score

Hi-index 12.05

Visualization

Abstract

Due to the number of potential applications and their inherent complexity, automatic capture and analysis of actions have become an active research area. In this paper, an implicit method for recognizing actions in a video is proposed. Existing implicit methods work on the regions of subjects, but our proposed system works on the surrounding regions, called negative spaces, of the subjects. Extracting features from negative spaces facilitates the system to extract simple, yet effective features for describing actions. These negative-space based features are robust to deformed actions, such as complex boundary variations, partial occlusions, non-rigid deformations and small shadows. Unlike other implicit methods, our method does not require dimensionality reduction, thereby significantly improving the processing time. Further, we propose a new method to detect cycles of different actions automatically. In the proposed system, first, the input image sequence is background segmented and shadows are eliminated from the segmented images. Next, motion based features are computed for the sequence. Then, the negative space based description of each pose is obtained and the action descriptor is formed by combining the pose descriptors. Nearest Neighbor classifier is applied to recognize the action of the input sequence. The proposed system was evaluated on both publically available action datasets and a new fish action dataset for comparison, and showed improvement in both its accuracy and processing time. Moreover, the proposed system showed very good accuracy for corrupted image sequences, particularly in the case of noisy segmentation, and lower frame rate. Further, it has achieved highest accuracy with lowest processing time compared with the state-of-art methods.