Information perspective of optimization

  • Authors:
  • Yossi Borenstein;Riccardo Poli

  • Affiliations:
  • University of Essex, Colchester, UK;University of Essex, Colchester, UK

  • Venue:
  • PPSN'06 Proceedings of the 9th international conference on Parallel Problem Solving from Nature
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we relate information theory and Kolmogorov Complexity (KC) to optimization in the black box scenario. We define the set of all possible decisions an algorithm might make during a run, we associate a function with a probability distribution over this set and define accordingly its entropy. We show that the expected KC of the set (rather than the function) is a better measure of problem difficulty. We analyze the effect of the entropy on the expected KC. Finally, we show, for a restricted scenario, that any permutation closure of a single function, the finest level of granularity for which a No Free Lunch Theorem can hold [7], can be associated with a particular value of entropy. This implies bounds on the expected performance of an algorithm on members of that closure.