On the Optimality of the Simple Bayesian Classifier under Zero-One Loss
Machine Learning - Special issue on learning with probabilistic representations
Data mining: practical machine learning tools and techniques with Java implementations
Data mining: practical machine learning tools and techniques with Java implementations
Why Discretization Works for Naive Bayesian Classifiers
ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning
Estimating continuous distributions in Bayesian classifiers
UAI'95 Proceedings of the Eleventh conference on Uncertainty in artificial intelligence
Machine learning: a review of classification and combining techniques
Artificial Intelligence Review
Ranking the Uniformity of Interval Pairs
ECML PKDD '08 Proceedings of the 2008 European Conference on Machine Learning and Knowledge Discovery in Databases - Part I
Bayesian classifiers based on kernel density estimation: Flexible classifiers
International Journal of Approximate Reasoning
Supervised Machine Learning: A Review of Classification Techniques
Proceedings of the 2007 conference on Emerging Artificial Intelligence Applications in Computer Engineering: Real Word AI Systems with Applications in eHealth, HCI, Information Retrieval and Pervasive Technologies
Comparing data mining methods with logistic regression in childhood obesity prediction
Information Systems Frontiers
Naïve Bayes ensembles with a random oracle
MCS'07 Proceedings of the 7th international conference on Multiple classifier systems
MC-tree: Improving Bayesian anytime classification
SSDBM'10 Proceedings of the 22nd international conference on Scientific and statistical database management
A 'non-parametric' version of the naive Bayes classifier
Knowledge-Based Systems
Bulk loading hierarchical mixture models for efficient stream classification
PAKDD'10 Proceedings of the 14th Pacific-Asia conference on Advances in Knowledge Discovery and Data Mining - Volume Part II
BT*: an advanced algorithm for anytime classification
SSDBM'12 Proceedings of the 24th international conference on Scientific and Statistical Database Management
Hi-index | 0.00 |
There are three main methods for handling continuous variables in naive Bayes classifiers, namely, the normal method (parametric approach), the kernel method (non parametric approach) and discretization In this article, we perform a methodologically sound comparison of the three methods, which shows large mutual differences of each of the methods and no single method being universally better This suggests that a method for selecting one of the three approaches to continuous variables could improve overall performance of the naive Bayes classifier We present three methods that can be implemented efficiently v-fold cross validation for the normal, kernel and discretization method Empirical evidence suggests that selection using 10 fold cross validation (especially when repeated 10 times) can largely and significantly improve over all performance of naive Bayes classifiers and consistently outperform any of the three popular methods for dealing with continuous variables on their own This is remarkable, since selection among more classifiers does not consistently result in better accuracy.