Minimum Classification Error Training Incorporating Automatic Loss Smoothness Determination

  • Authors:
  • Hideyuki Watanabe;Jun'ichi Tokuno;Tsukasa Ohashi;Shigeru Katagiri;Miho Ohsaki;Shigeki Matsuda;Hideki Kashioka

  • Affiliations:
  • National Institute of Information and Communications Technology, Kyoto, Japan 619-0289;Graduate School of Engineering, Doshisha University, Kyoto, Japan 610-0394;Graduate School of Engineering, Doshisha University, Kyoto, Japan 610-0394;Graduate School of Engineering, Doshisha University, Kyoto, Japan 610-0394;Graduate School of Engineering, Doshisha University, Kyoto, Japan 610-0394;National Institute of Information and Communications Technology, Kyoto, Japan 619-0289;National Institute of Information and Communications Technology, Kyoto, Japan 619-0289

  • Venue:
  • Journal of Signal Processing Systems
  • Year:
  • 2014

Quantified Score

Hi-index 0.00

Visualization

Abstract

Minimum Classification Error (MCE) training, which has been widely used as one of the recent standards of discriminative training for classifiers, is characterized by a smooth sigmoidal-form classification error count loss. The smoothness of this loss function effectively increases training robustness to unseen samples, well approximates the ultimate, minimum classification error probability status, and leads to accurate classification over unseen samples. However, few rational methods have been developed for controlling the smoothness, which is often determined through many repetitions of the experimental setting; this empirical approach has been a disincentive to the increased popularization of MCE training. To alleviate this long-standing problem, we propose a new MCE training method that automatically determines loss smoothness. The proposed method is based on Parzen-estimation-based MCE re-formalization, and the loss smoothness degree is determined so that Parzen distribution can be an accurate approximation to the unknown true distribution, whose positive-domain integration corresponds to classification error probability, in one-dimensional misclassification measure space. Through systematic experiments, we show that the proposed method efficiently yields a classification accuracy that nearly matches the best accuracy obtained by the conventional, trial-and-error-mode repetitions of smoothness setting.