Quantifying inductive bias: AI learning algorithms and Valiant's learning framework
Artificial Intelligence
Mistake bounds and logarithmic linear-threshold learning algorithms
Mistake bounds and logarithmic linear-threshold learning algorithms
Original Contribution: Stacked generalization
Neural Networks
C4.5: programs for machine learning
C4.5: programs for machine learning
Circuits of the mind
The nature of statistical learning theory
The nature of statistical learning theory
Artificial Intelligence - Special issue on relevance
On PAC learning using Winnow, Perceptron, and a Perceptron-like algorithm
COLT '99 Proceedings of the twelfth annual conference on Computational learning theory
A Winnow-Based Approach to Context-Sensitive Spelling Correction
Machine Learning - Special issue on natural language learning
Identification of partial disjunction, parity, and threshold functions
Theoretical Computer Science
Data mining: practical machine learning tools and techniques with Java implementations
Data mining: practical machine learning tools and techniques with Java implementations
Computational sample complexity and attribute-efficient learning
Journal of Computer and System Sciences
Machine Learning
A neuroidal architecture for cognitive computation
Journal of the ACM (JACM)
Neural Networks: A Comprehensive Foundation
Neural Networks: A Comprehensive Foundation
Learning to recognize three-dimensional objects
Neural Computation
Relational Learning for NLP using Linear Threshold Elements
IJCAI '99 Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence
Pattern Classification (2nd Edition)
Pattern Classification (2nd Edition)
Stacked generalization: when does it work?
IJCAI'97 Proceedings of the Fifteenth international joint conference on Artifical intelligence - Volume 2
Hi-index | 0.00 |
Excessive information is known to degrade the classification performance of many machine learning algorithms. Attribute-efficient learning algorithms can tolerate irrelevant attributes without their performance being affected too much. Valiant's projection learning is a way to combine such algorithms so that this desired property is maintained. The archetype attribute-efficient learning algorithm Winnow and, especially, combinations of Winnow have turned out empirically successful in domains containing many attributes. However, projection learning as proposed by Valiant has not yet been evaluated empirically. We study how projection learning relates to using Winnow as such and with an extended set of attributes. We also compare projection learning with decision tree learning and Naive Bayes on UCI data sets.Projection learning systematically enhances the classification accuracy of Winnow, but the cost in time and space consumption can be high. Balanced Winnow seems to be a better alternative than the basic algorithm for learning the projection hypotheses. However, Balanced Winnow is not well suited for learning the second level (projective disjunction) hypothesis. The on-line approach projection learning does not fall far behind in classification accuracy from batch algorithms such as decision tree learning and Naive Bayes on the UCI data sets that we used.