Two entropy-based methods for learning unsupervised gaussian mixture models

  • Authors:
  • Antonio Peñalver;Francisco Escolano;Juan M. Sáez

  • Affiliations:
  • Robot Vision Group, Alicante University, Spain;Robot Vision Group, Alicante University, Spain;Robot Vision Group, Alicante University, Spain

  • Venue:
  • SSPR'06/SPR'06 Proceedings of the 2006 joint IAPR international conference on Structural, Syntactic, and Statistical Pattern Recognition
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we address the problem of estimating the parameters of a Gaussian mixture model. Although the EM (Expectation-Maximization) algorithm yields the maximum-likelihood solution it requires a careful initialization of the parameters and the optimal number of kernels in the mixture may be unknown beforehand. We propose a criterion based on the entropy of the pdf (probability density function) associated to each kernel to measure the quality of a given mixture model. Two different methods for estimating Shannon entropy are proposed and a modification of the classical EM algorithm to find the optimal number of kernels in the mixture is presented. We test our algorithm in probability density estimation, pattern recognition and color image segmentation.