Semi-naive Bayesian classifier
EWSL-91 Proceedings of the European working session on learning on Machine learning
C4.5: programs for machine learning
C4.5: programs for machine learning
Artificial Intelligence Review - Special issue on lazy learning
Wrappers for feature subset selection
Artificial Intelligence - Special issue on relevance
On the Optimality of the Simple Bayesian Classifier under Zero-One Loss
Machine Learning - Special issue on learning with probabilistic representations
Machine Learning - Special issue on learning with probabilistic representations
A probabilistic framework for memory-based reasoning
Artificial Intelligence
Using Model Trees for Classification
Machine Learning
On Bias, Variance, 0/1—Loss, and the Curse-of-Dimensionality
Data Mining and Knowledge Discovery
Machine Learning
Induction of Recursive Bayesian Classifiers
ECML '93 Proceedings of the European Conference on Machine Learning
Estimating continuous distributions in Bayesian classifiers
UAI'95 Proceedings of the Eleventh conference on Uncertainty in artificial intelligence
Geometric Properties of Naive Bayes in Nominal Domains
EMCL '01 Proceedings of the 12th European Conference on Machine Learning
Conditionally Independent Component Extraction for Naive Bayes Inference
ICANN '01 Proceedings of the International Conference on Artificial Neural Networks
Toward Bayesian Classifiers with Accurate Probabilities
PAKDD '02 Proceedings of the 6th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining
Augmenting naive Bayes for ranking
ICML '05 Proceedings of the 22nd international conference on Machine learning
Software Defect Association Mining and Defect Correction Effort Prediction
IEEE Transactions on Software Engineering
Enhancing software quality estimation using ensemble-classifier based noise filtering
Intelligent Data Analysis
Evaluating noise elimination techniques for software quality estimation
Intelligent Data Analysis
Extension of Bayesian Network Classifiers to Regression Problems
IBERAMIA '08 Proceedings of the 11th Ibero-American conference on AI: Advances in Artificial Intelligence
Naive Bayes for optimal ranking
Journal of Experimental & Theoretical Artificial Intelligence
Software quality analysis by combining multiple projects and learners
Software Quality Control
An effort prediction framework for software defect correction
Information and Software Technology
Conditional Density Estimation with Class Probability Estimators
ACML '09 Proceedings of the 1st Asian Conference on Machine Learning: Advances in Machine Learning
Using locally weighted learning to improve SMOreg for regression
PRICAI'06 Proceedings of the 9th Pacific Rim international conference on Artificial intelligence
Model tree learning for query term weighting in question answering
ECIR'07 Proceedings of the 29th European conference on IR research
IDA'07 Proceedings of the 7th international conference on Intelligent data analysis
Applying REC analysis to ensembles of sigma-point kalman filters
ICANN'06 Proceedings of the 16th international conference on Artificial Neural Networks - Volume Part II
Machine learning for query formulation in question answering
Natural Language Engineering
Learning naive bayes for probability estimation by feature selection
AI'06 Proceedings of the 19th international conference on Advances in Artificial Intelligence: Canadian Society for Computational Studies of Intelligence
Hi-index | 0.00 |
Despite its simplicity, the naive Bayes learning scheme performs well on most classification tasks, and is often significantly more accurate than more sophisticated methods. Although the probability estimates that it produces can be inaccurate, it often assigns maximum probability to the correct class. This suggests that its good performance might be restricted to situations where the output is categorical. It is therefore interesting to see how it performs in domains where the predicted value is numeric, because in this case, predictions are more sensitive to inaccurate probability estimates.This paper shows how to apply the naive Bayes methodology to numeric prediction (i.e., regression) tasks by modeling the probability distribution of the target value with kernel density estimators, and compares it to linear regression, locally weighted linear regression, and a method that produces “model trees”—decision trees with linear regression functions at the leaves. Although we exhibit an artificial dataset for which naive Bayes is the method of choice, on real-world datasets it is almost uniformly worse than locally weighted linear regression and model trees. The comparison with linear regression depends on the error measure: for one measure naive Bayes performs similarly, while for another it is worse. We also show that standard naive Bayes applied to regression problems by discretizing the target value performs similarly badly. We then present empirical evidence that isolates naive Bayes' independence assumption as the culprit for its poor performance in the regression setting. These results indicate that the simplistic statistical assumption that naive Bayes makes is indeed more restrictive for regression than for classification.