Comparing evaluation protocols on the KTH dataset

  • Authors:
  • Zan Gao;Ming-Yu Chen;Alexander G. Hauptmann;Anni Cai

  • Affiliations:
  • School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing, P.R. China;School of Computer Science, Carnegie Mellon University, PA;School of Computer Science, Carnegie Mellon University, PA;School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing, P.R. China

  • Venue:
  • HBU'10 Proceedings of the First international conference on Human behavior understanding
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Human action recognition has become a hot research topic, and a lot of algorithms have been proposed. Most of researchers evaluated their performances on the KTH dataset, but there is no unified standard how to evaluate algorithms on this dataset. Different researchers have employed different test setups, so the comparison is not accurate, fair or complete. In order to know how much difference there is when different experimental setups are used, we take our own spatio-temporal MoSIFT feature as an example to assess its performance on the KTH dataset using different test scenarios and different partitioning of the data. In all experiments, support vector machine (SVM) with a chi-square kernel is adopted. First, we evaluate performance changes resulting from differing vocabulary sizes of the codebook, and then decide on a suitable vocabulary size of codebook. Then, we train the models using different training dataset partitions, and test the performances one the corresponding held-out test sets. Experiments show that the best performance of MoSIFT can reach 96.33% on the KTH dataset. When different n-fold cross-validation methods are used, there can be up to 10.67% difference in the result. And when different dataset segmentations are used (such as KTH1 and KTH2), the difference in results can be up to 5.8% absolute. In addition, the performance changes dramatically when different scenarios are used in the training and test dataset. When training on KTH1 S1+S2+S3+S4 and testing on KTH1 S1 and S3 scenarios, the performance can reach 97.33% and 89.33% respectively. This paper shows how different test configurations can skew results, even on standard data set. The recommendation is to use a simple leave-one-out as the most easily replicable clear-cut partitioning.