Riemannian geometry and statistical machine learning

  • Authors:
  • Guy Lebanon;John Lafferty

  • Affiliations:
  • Carnegie Mellon University;Carnegie Mellon University

  • Venue:
  • Riemannian geometry and statistical machine learning
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Statistical machine learning algorithms deal with the problem of selecting an appropriate statistical model from a model space Θ based on a training set xiN i=1 ⊂ X or xi,y iN i=1 ⊂ X × Y. In doing so they either implicitly or explicitly make assumptions on the geometries of the model space Θ and the data space X. Such assumptions are crucial to the success of the algorithms as different geometries are appropriate for different models and data spaces. By studying these assumptions we are able to develop new theoretical results that enhance our understanding of several popular learning algorithms. Furthermore, using geometrical reasoning we are able to adapt existing algorithms such as radial basis kernels and linear margin classifiers to non-Euclidean geometries. Such adaptation is shown to be useful when the data space does not exhibit Euclidean geometry. In particular, we focus in our experiments on the space of text documents that is naturally associated with the Fisher information metric on corresponding multinomial models.