The influence of cross-validation on video classification performance

  • Authors:
  • Jan C. van Gemert;Cees G. M. Snoek;Cor J. Veenman;Arnold W. M. Smeulders

  • Affiliations:
  • Informatics Institute University of Amsterdam, Amsterdam, The Netherlands;Informatics Institute University of Amsterdam, Amsterdam, The Netherlands;Informatics Institute University of Amsterdam, Amsterdam, The Netherlands;Informatics Institute University of Amsterdam, Amsterdam, The Netherlands

  • Venue:
  • MULTIMEDIA '06 Proceedings of the 14th annual ACM international conference on Multimedia
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Digital video is sequential in nature. When video data is used in a semantic concept classification task, the episodes are usually summarized with shots. The shots are annotated as containing, or not containing, a certain concept resulting in a labeled dataset. These labeled shots can subsequently be used by supervised learning methods (classifiers) where they are trained to predict the absence or presence of the concept in unseen shots and episodes. The performance of such automatic classification systems is usually estimated with cross-validation. By taking random samples from the dataset for training and testing as such, part of the shots from an episode are in the training set and another part from the same episode is in the test set. Accordingly, data dependence between training and test set is introduced, resulting in too optimistic performance estimates. In this paper, we experimentally show this bias, and propose how this bias can be prevented using episode-constrained crossvalidation. Moreover, we show that a 17% higher classifier performance can be achieved by using episode constrained cross-validation for classifier parameter tuning.