Training conditional random fields with unlabeled data and limited number of labeled examples

  • Authors:
  • Tak-Lam Wong;Wai Lam

  • Affiliations:
  • Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Shatin, Hong Kong;Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Shatin, Hong Kong

  • Venue:
  • ICMLC'05 Proceedings of the 4th international conference on Advances in Machine Learning and Cybernetics
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Conditional random fields is a probabilistic approach which has been applied to sequence labeling task achieving good performance. We attempt to extend the model so that human effort in preparing labeled training examples can be reduced by considering unlabeled data. Instead of maximizing the conditional likelihood, we aim at maximizing the likelihood of the observation of the sequences from both of the labeled and unlabeled data. We have conducted extensive experiments in two different data sets to evaluate the performance. The experimental results show that our model learned from both labeled and unlabeled data has a better performance over the model learned by only considering labeled training examples.