Nonlinear total variation based noise removal algorithms
Proceedings of the eleventh annual international conference of the Center for Nonlinear Studies on Experimental mathematics : computational issues in nonlinear science: computational issues in nonlinear science
Support vector regression with ANOVA decomposition kernels
Advances in kernel methods
Least Squares Support Vector Machine Classifiers
Neural Processing Letters
Atomic Decomposition by Basis Pursuit
SIAM Review
Choosing Multiple Parameters for Support Vector Machines
Machine Learning
Structural Modelling with Sparse Kernels
Machine Learning
Training v-support vector regression: theory and algorithms
Neural Computation
Ridge Regression Learning Algorithm in Dual Variables
ICML '98 Proceedings of the Fifteenth International Conference on Machine Learning
Classes of kernels for machine learning: a statistics perspective
The Journal of Machine Learning Research
Convex Optimization
A Compression Approach to Support Vector Model Selection
The Journal of Machine Learning Research
The evidence framework applied to classification networks
Neural Computation
Nomograms for visualizing support vector machines
Proceedings of the eleventh ACM SIGKDD international conference on Knowledge discovery in data mining
Handling missing values in support vector machine classifiers
Neural Networks - 2005 Special issue: IJCNN 2005
Componentwise support vector machines for structure detection
ICANN'05 Proceedings of the 15th international conference on Artificial neural networks: formal models and their applications - Volume Part II
Hi-index | 0.01 |
This paper proposes a new method to obtain sparseness and structure detection for a class of kernel machines related to least-squares support vector machines (LS-SVMs). The key method is to adopt an hierarchical modeling strategy. Here, the first level consists of an LS-SVM substrate which is based upon an LS-SVM formulation with additive regularization trade-off. This regularization trade-off is determined at higher levels such that sparse representations and/or structure detection are obtained. Using the necessary and sufficient conditions for optimality given by the Karush-Kuhn-Tucker conditions, one can guide the interaction between different levels via a well-defined set of hyper-parameters. From a computational point of view, all levels can be fused into a single convex optimization problem. Furthermore, the principle is applied in order to optimize the validation performance of the resulting kernel machine. Sparse representations as well as structure detection are obtained, respectively, by using an L"1 regularization scheme and a measure of maximal variation at the second level. A number of case studies indicate the usefulness of these approaches both with respect to interpretability of the final model as well as generalization performance.