Random artificial incorporation of noise in a learning classifier system environment

  • Authors:
  • Ryan Urbanowicz;Nicholas Sinnott-Armstrong;Jason Moore

  • Affiliations:
  • Dartmouth College, Hanover, NH, USA;Dartmouth College, Hanover, NH, USA;Dartmouth College, Hanover, NH, USA

  • Venue:
  • Proceedings of the 13th annual conference companion on Genetic and evolutionary computation
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Effective rule generalization in learning classifier systems (LCSs) has long since been an important consideration. In noisy problem domains, where attributes do not precisely determine class, overemphasis on accuracy without sufficient generalization leads to over-fitting of the training data, and a large discrepancy between training and testing accuracies. This issue is of particular concern within noisy bioinformatic problems such as complex disease, gene association studies. In an effort to promote effective generalization we introduce and explore a simple strategy which seeks to discourage over-fitting via the probabilistic incorporation of random noise within training instances. We evaluate a variety of noise models and magnitudes which either specify an equal probability of noise per attribute, or target higher noise probability to the attributes which tend to be more frequently generalized. Our results suggest that targeted noise incorporation can reduce training accuracy without eroding testing accuracy. In addition, we observe a slight improvement in our power estimates (i.e. ability to detect the true underlying model(s)).