A markov chain framework for the simple genetic algorithm

  • Authors:
  • Thomas E. Davis;Jose C. Principe

  • Affiliations:
  • WL/MNGS Eglin AFB, FL 32542;Electrical Engineering Department University of Florida, Gainesville, FL 32611 principe@brain.ee.ufl.edu

  • Venue:
  • Evolutionary Computation
  • Year:
  • 1993

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper develops a theoretical framework for the simple genetic algorithm (combinations of the reproduction, mutation, and crossover operators) based on the asymptotic state behavior of a nonstationary Markov chain algorithm model. The methodology borrows heavily from that of simulated annealing. We prove the existence of a unique asymptotic probability distribution (stationary distribution) for the Markov chain when the mutation probability is used with any constant nonzero probability value. We develop a Cramer's Rule representation of the stationary distribution components for all nonzero mutation probability values and then extend the representation to show that the stationary distribution possesses a zero mutation probability limit. Finally, we present a strong ergodicity bound on the mutation probability sequence that ensures that the nonstationary algorithm (which results from varying mutation probability during algorithm execution) achieves the limit distribution asymptotically. Although the focus of this work is on a nonstationary algorithm in which mutation probability is reduced asymptotically to zero via a schedule (in a fashion analogous to simulated annealing), the stationary distribution results (existence, Cramer's Rule representation, and zero mutation probability limit) are directly applicable to conventional, simple genetic algorithm implementations as well.