Active learning with irrelevant examples

  • Authors:
  • Dominic Mazzoni;Kiri L. Wagstaff;Michael C. Burl

  • Affiliations:
  • Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA;Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA;Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA

  • Venue:
  • ECML'06 Proceedings of the 17th European conference on Machine Learning
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Active learning algorithms attempt to accelerate the learning process by requesting labels for the most informative items first. In real-world problems, however, there may exist unlabeled items that are irrelevant to the user's classification goals. Queries about these points slow down learning because they provide no information about the problem of interest. We have observed that when irrelevant items are present, active learning can perform worse than random selection, requiring more time (queries) to achieve the same level of accuracy. Therefore, we propose a novel approach, Relevance Bias, in which the active learner combines its default selection heuristic with the output of a simultaneously trained relevance classifier to favor items that are likely to be both informative and relevant. In our experiments on a real-world problem and two benchmark datasets, the Relevance Bias approach significantly improves the learning rate of three different active learning approaches.