Artificial Intelligence Review - Special issue on lazy learning
Feature Subset Selection within a Simulated Annealing DataMining Algorithm
Journal of Intelligent Information Systems
Wrappers for feature subset selection
Artificial Intelligence - Special issue on relevance
Data selection for support vector machine classifiers
Proceedings of the sixth ACM SIGKDD international conference on Knowledge discovery and data mining
A Tutorial on Support Vector Machines for Pattern Recognition
Data Mining and Knowledge Discovery
Feature Selection via Concave Minimization and Support Vector Machines
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
Signal Processing - Special issue: Genomic signal processing
Artificial Intelligence Review
Pattern Classification (2nd Edition)
Pattern Classification (2nd Edition)
Large margin nearest local mean classifier
Signal Processing
To obtain orthogonal feature extraction using training data selection
Proceedings of the 18th ACM conference on Information and knowledge management
Generalized re-weighting local sampling mean discriminant analysis
Pattern Recognition
A sparse nearest mean classifier for high dimensional multi-class problems
Pattern Recognition Letters
Robustness analysis of eleven linear classifiers in extremely high–dimensional feature spaces
ANNPR'10 Proceedings of the 4th IAPR TC3 conference on Artificial Neural Networks in Pattern Recognition
Metric learning for large scale image classification: generalizing to new classes at near-zero cost
ECCV'12 Proceedings of the 12th European conference on Computer Vision - Volume Part II
Hi-index | 0.14 |
In this paper, we specifically focus on high-dimensional data sets for which the number of dimensions is an order of magnitude higher than the number of objects. From a classifier design standpoint, such small sample size problems have some interesting challenges. The first challenge is to find, from all hyperplanes that separate the classes, a separating hyperplane which generalizes well for future data. A second important task is to determine which features are required to distinguish the classes. To attack these problems, we propose the LESS (Lowest Error in a Sparse Subspace) classifier that efficiently finds linear discriminants in a sparse subspace. In contrast with most classifiers for high-dimensional data sets, the LESS classifier incorporates a (simple) data model. Further, by means of a regularization parameter, the classifier establishes a suitable trade-off between subspace sparseness and classification accuracy. In the experiments, we show how LESS performs on several high-dimensional data sets and compare its performance to related state-of-the-art classifiers like, among others, linear ridge regression with the LASSO and the Support Vector Machine. It turns out that LESS performs competitively while using fewer dimensions.