The Strength of Weak Learnability
Machine Learning
Knowledge acquisition from databases
Knowledge acquisition from databases
Discovering informative patterns and data cleaning
Advances in knowledge discovery and data mining
Two Variations on Fisher's Linear Discriminant for Pattern Recognition
IEEE Transactions on Pattern Analysis and Machine Intelligence
Machine Learning
Machine Learning
Estimating a Kernel Fisher Discriminant in the Presence of Label Noise
ICML '01 Proceedings of the Eighteenth International Conference on Machine Learning
ICML '97 Proceedings of the Fourteenth International Conference on Machine Learning
Experiments with Noise Filtering in a Medical Domain
ICML '99 Proceedings of the Sixteenth International Conference on Machine Learning
Support Vector Data Description
Machine Learning
AAAI'96 Proceedings of the thirteenth national conference on Artificial intelligence - Volume 1
Identifying and eliminating mislabeled training instances
AAAI'96 Proceedings of the thirteenth national conference on Artificial intelligence - Volume 1
The reduced nearest neighbor rule (Corresp.)
IEEE Transactions on Information Theory
IEEE Transactions on Pattern Analysis and Machine Intelligence
Journal of Multivariate Analysis
Label noise-tolerant hidden Markov models for segmentation: application to ECGs
ECML PKDD'11 Proceedings of the 2011 European conference on Machine learning and knowledge discovery in databases - Volume Part I
Estimating mutual information for feature selection in the presence of label noise
Computational Statistics & Data Analysis
Hi-index | 0.01 |
In machine learning, class noise occurs frequently and deteriorates the classifier derived from the noisy data set. This paper presents two promising classifiers for this problem based on a probabilistic model proposed by Lawrence and Scholkopf (2001). The proposed algorithms are able to tolerate class noise, and extend the earlier work of Lawrence and Scholkopf in two ways. First, we present a novel incorporation of their probabilistic noise model in the Kernel Fisher discriminant; second, the distribution assumption previously made is relaxed in our work. The methods were investigated on simulated noisy data sets and a real world comparative genomic hybridization (CGH) data set. The results show that the proposed approaches substantially improve standard classifiers in noisy data sets, and achieve larger performance gain in non-Gaussian data sets and small size data sets.