Multinomial event model based abstraction for sequence and text classification

  • Authors:
  • Dae-Ki Kang;Jun Zhang;Adrian Silvescu;Vasant Honavar

  • Affiliations:
  • Artificial Intelligence Research Laboratory, Department of Computer Science, Iowa State University, Ames, IA;Artificial Intelligence Research Laboratory, Department of Computer Science, Iowa State University, Ames, IA;Artificial Intelligence Research Laboratory, Department of Computer Science, Iowa State University, Ames, IA;Artificial Intelligence Research Laboratory, Department of Computer Science, Iowa State University, Ames, IA

  • Venue:
  • SARA'05 Proceedings of the 6th international conference on Abstraction, Reformulation and Approximation
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

In many machine learning applications that deal with sequences, there is a need for learning algorithms that can effectively utilize the hierarchical grouping of words. We introduce Word Taxonomy guided Naive Bayes Learner for the Multinomial Event Model (WTNBL-MN) that exploits word taxonomy to generate compact classifiers, and Word Taxonomy Learner (WTL) for automated construction of word taxonomy from sequence data. WTNBL-MN is a generalization of the Naive Bayes learner for the Multinomial Event Model for learning classifiers from data using word taxonomy. WTL uses hierarchical agglomerative clustering to cluster words based on the distribution of class labels that co-occur with the words. Our experimental results on protein localization sequences and Reuters text show that the proposed algorithms can generate Naive Bayes classifiers that are more compact and often more accurate than those produced by standard Naive Bayes learner for the Multinomial Model.