Multinomial naive bayes for text categorization revisited

  • Authors:
  • Ashraf M. Kibriya;Eibe Frank;Bernhard Pfahringer;Geoffrey Holmes

  • Affiliations:
  • Department of Computer Science, University of Waikato, Hamilton, New Zealand;Department of Computer Science, University of Waikato, Hamilton, New Zealand;Department of Computer Science, University of Waikato, Hamilton, New Zealand;Department of Computer Science, University of Waikato, Hamilton, New Zealand

  • Venue:
  • AI'04 Proceedings of the 17th Australian joint conference on Advances in Artificial Intelligence
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents empirical results for several versions of the multinomial naive Bayes classifier on four text categorization problems, and a way of improving it using locally weighted learning More specifically, it compares standard multinomial naive Bayes to the recently proposed transformed weight-normalized complement naive Bayes classifier (TWCNB) [1], and shows that some of the modifications included in TWCNB may not be necessary to achieve optimum performance on some datasets However, it does show that TFIDF conversion and document length normalization are important It also shows that support vector machines can, in fact, sometimes very significantly outperform both methods Finally, it shows how the performance of multinomial naive Bayes can be improved using locally weighted learning However, the overall conclusion of our paper is that support vector machines are still the method of choice if the aim is to maximize accuracy.