Compressed-objective genetic algorithm

  • Authors:
  • Kuntinee Maneeratana;Kittipong Boonlong;Nachol Chaiyaratana

  • Affiliations:
  • Department of Mechanical Engineering, Chulalongkorn University, Bangkok, Thailand;Department of Mechanical Engineering, Chulalongkorn University, Bangkok, Thailand;Research and Development Center for Intelligent Systems, King Mongkut's Institute of Technology North Bangkok, Bangkok, Thailand

  • Venue:
  • PPSN'06 Proceedings of the 9th international conference on Parallel Problem Solving from Nature
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

A strategy for solving an optimisation problem with a large number of objectives by transforming the original objective vector into a two-objective vector during survival selection is presented. The transformed objectives, referred to as preference objectives, consist of a winning score and a vicinity index. The winning score, a maximisation criterion, describes the difference of the number of superior and inferior objectives between two solutions. The minimisation vicinity index describes the level of solution clustering around a search location, particularly the best value of each individual objective, is used to encourage the results to spread throughout the Pareto front. With this strategy, a new multi-objective algorithm, the compressed-objective genetic algorithm (COGA), is introduced. COGA is subsequently benchmarked against a non-dominated sorting genetic algorithm II (NSGA-II) and an improved strength Pareto genetic algorithm (SPEA-II) in six scalable DTLZ benchmark problems with three to six objectives. The results reveal that the proposed strategy plays a crucial role in the generation of a superior solution set compared to the other two techniques in terms of the solution set coverage and the closeness to the true Pareto front. Furthermore, the spacing of COGA solutions is very similar to that of SPEA-II solutions. Overall, the functionality of the multi-objective evolutionary algorithm (MOEA) with preference objectives is effectively demonstrated.