Data-Dependent Margin-Based Generalization Bounds for Classification

  • Authors:
  • Balázs Kégl;Tamás Linder;Gábor Lugosi

  • Affiliations:
  • -;-;-

  • Venue:
  • COLT '01/EuroCOLT '01 Proceedings of the 14th Annual Conference on Computational Learning Theory and and 5th European Conference on Computational Learning Theory
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

We derive new margin-based inequalities for the probability of error of classifiers.The main feature of these bounds is that they can be calculated using the training data and therefore may be effectively used for model selection purposes.In particular, the bounds involve quantities such as the empirical fat-shattering dimension and covering number measured on the training data, as opposed to their worst-case counterparts traditionally used in such analyses, and appear to be sharper and more general than recent results involving empirical complexity measures.In addition, we also develop an alternative data-based bound for the generalization error of classes of convex combinations of classifiers involving an empirical complexity measure that is more easily computable than the empirical covering number or fat-shattering dimension.W e also show an example of efficient computation of the new bounds.