Automatic allocation of training data for rapid prototyping of speech understanding based on multiple model combination

  • Authors:
  • Kazunori Komatani;Masaki Katsumaru;Mikio Nakano;Kotaro Funakoshi;Tetsuya Ogata;Hiroshi G. Okuno

  • Affiliations:
  • Kyoto University;Kyoto University;Honda Research Institute Japan Co., Ltd.;Honda Research Institute Japan Co., Ltd.;Kyoto University;Kyoto University

  • Venue:
  • COLING '10 Proceedings of the 23rd International Conference on Computational Linguistics: Posters
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

The optimal choice of speech understanding method depends on the amount of training data available in rapid prototyping. A statistical method is ultimately chosen, but it is not clear at which point in the increase in training data a statistical method become effective. Our framework combines multiple automatic speech recognition (ASR) and language understanding (LU) modules to provide a set of speech understanding results and selects the best result among them. The issue is how to allocate training data to statistical modules and the selection module in order to avoid overfitting in training and obtain better performance. This paper presents an automatic training data allocation method that is based on the change in the coefficients of the logistic regression functions used in the selection module. Experimental evaluation showed that our allocation method outperformed baseline methods that use a single ASR module and a single LU module at every point while training data increase.