Detection of language (model) errors

  • Authors:
  • K. Y. Hung;R. W. P. Luk;D. Yeung;K. F. L. Chung;W. Shu

  • Affiliations:
  • Hong Kong Polytechnic University, Hong Kong;Hong Kong Polytechnic University, Hong Kong;Hong Kong Polytechnic University, Hong Kong;Hong Kong Polytechnic University, Hong Kong;Hong Kong Polytechnic University, Hong Kong

  • Venue:
  • EMNLP '00 Proceedings of the 2000 Joint SIGDAT conference on Empirical methods in natural language processing and very large corpora: held in conjunction with the 38th Annual Meeting of the Association for Computational Linguistics - Volume 13
  • Year:
  • 2000

Quantified Score

Hi-index 0.00

Visualization

Abstract

The bigram language models are popular, in much language processing applications, in both Indo-European and Asian languages. However, when the language model for Chinese is applied in a novel domain, the accuracy is reduced significantly, from 96% to 78% in our evaluation. We apply pattern recognition techniques (i.e. Bayesian, decision tree and neural network classifiers) to discover language model errors. We have examined 2 general types of features: model-based and language-specific features. In our evaluation, Bayesian classifiers produce the best recall performance of 80% but the precision is low (60%). Neural network produced good recall (75%) and precision (80%) but both Bayesian and Neural network have low skip ratio (65%). The decision tree classifier produced the best precision (81%) and skip ratio (76%) but its recall is the lowest (73%).