Bias/variance analysis for relational domains

  • Authors:
  • Jennifer Neville;David Jensen

  • Affiliations:
  • Departments of Computer Science and Statistics, Purdue University;Department of Computer Science, University of Massachusetts Amherst

  • Venue:
  • ILP'07 Proceedings of the 17th international conference on Inductive logic programming
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Bias/variance analysis [1] is a useful tool for investigating the performance of machine learning algorithms. Conventional analysis decomposes loss into errors due to aspects of the learning process with an underlying assumption that there is no variation in model predictions due to the inference process used for prediction. This assumption is often violated when collective inference models are used for classification of relational data. In relational data, when there are dependencies among the class labels of related instances, the inferences about one object can be used to improve the inferences about other related objects. Collective inference techniques exploit these dependencies by jointly inferring the class labels in a test set. This approach can produce more accurate predictions than conditional inference for each instance independently, but it also introduces an additional source of error, both through the use of approximate inference algorithms and through variation in the availability of test set information. To date, the impact of inference error on relational model performance has not been investigated.