Information theoretical Kernels for generative embeddings based on hidden Markov models

  • Authors:
  • André F. T. Martins;Manuele Bicego;Vittorio Murino;Pedro M. Q. Aguiar;Mário A. T. Figueiredo

  • Affiliations:
  • Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal;Computer Science Department, University of Verona, Verona, Italy and Istituto Italiano di Tecnologia, Genova, Italy;Computer Science Department, University of Verona, Verona, Italy and Istituto Italiano di Tecnologia, Genova, Italy;Instituto de Sistemas e Robótica, Instituto Superior Técnico, Lisboa, Portugal;Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal

  • Venue:
  • SSPR&SPR'10 Proceedings of the 2010 joint IAPR international conference on Structural, syntactic, and statistical pattern recognition
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Many approaches to learning classifiers for structured objects (e.g., shapes) use generative models in a Bayesian framework. However, state-of-the-art classifiers for vectorial data (e.g., support vector machines) are learned discriminatively. A generative embedding is a mapping from the object space into a fixed dimensional feature space, induced by a generative model which is usually learned from data. The fixed dimensionality of these feature spaces permits the use of state of the art discriminative machines based on vectorial representations, thus bringing together the best of the discriminative and generative paradigms. Using a generative embedding involves two steps: (i) defining and learning the generative model used to build the embedding; (ii) discriminatively learning a (maybe kernel) classifier on the adopted feature space. The literature on generative embeddings is essentially focused on step (i), usually adopting some standard off-the-shelf tool (e.g., an SVM with a linear or RBF kernel) for step (ii). In this paper, we follow a different route, by combining several HiddenMarkov Models-based generative embeddings (including the classical Fisher score) with the recently proposed non-extensive information theoretic kernels. We test this methodology on a 2D shape recognition task, showing that the proposed method is competitive with the state-of-art.