Question classification based on co-training style semi-supervised learning

  • Authors:
  • Zhengtao Yu;Lei Su;Lina Li;Quan Zhao;Cunli Mao;Jianyi Guo

  • Affiliations:
  • The School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650051, China and The Institute of Intelligent Information Processing, Computer Technolo ...;Department of Software, Yunnan University, Kunming 650091, China;The School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650051, China;The School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650051, China;The School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650051, China;The School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650051, China and The Institute of Intelligent Information Processing, Computer Technolo ...

  • Venue:
  • Pattern Recognition Letters
  • Year:
  • 2010

Quantified Score

Hi-index 0.10

Visualization

Abstract

In statistical question classification, semi-supervised learning that can exploit the abundant unlabeled samples has received substantial attention in recent years. In this paper, a novel question classification approach with the co-training style semi-supervised learning is proposed. In particular, the method extracts high-frequency keywords as classification features, and uses the word semantic similarity to adjust the feature weights. The classifiers are initially trained from labeled data and then the learned models are refined using unlabeled data which can get labeled if the classifiers agree on the labeling. Experiments on the Chinese question answering system in tourism domain were conducted by employing different feature selections, different supervised and semi-supervised algorithms, different feature dimensions and different unlabeled rates. The experimental results show the proposed method can effectively improve the classification accuracy. Specifically, under the 40% unlabeled rate of training set, the average accuracy rates reach 88.9% on coarse types and 78.2% on fine types, respectively, which get an improvement of around 2-4% points.