Introduction to statistical pattern recognition (2nd ed.)
Introduction to statistical pattern recognition (2nd ed.)
Adjustment Learning and Relevant Component Analysis
ECCV '02 Proceedings of the 7th European Conference on Computer Vision-Part IV
Laplacian Eigenmaps for dimensionality reduction and data representation
Neural Computation
Think globally, fit locally: unsupervised learning of low dimensional manifolds
The Journal of Machine Learning Research
Kernel Methods for Pattern Analysis
Kernel Methods for Pattern Analysis
Convex Optimization
Learning the Kernel Matrix with Semidefinite Programming
The Journal of Machine Learning Research
A fast iterative algorithm for fisher discriminant using heterogeneous kernels
ICML '04 Proceedings of the twenty-first international conference on Machine learning
Multiple kernel learning, conic duality, and the SMO algorithm
ICML '04 Proceedings of the twenty-first international conference on Machine learning
Multi-task feature and kernel selection for SVMs
ICML '04 Proceedings of the twenty-first international conference on Machine learning
Matching Theory (North-Holland mathematics studies)
Matching Theory (North-Holland mathematics studies)
A DC-programming algorithm for kernel selection
ICML '06 Proceedings of the 23rd international conference on Machine learning
Discriminative cluster analysis
ICML '06 Proceedings of the 23rd international conference on Machine learning
Optimal kernel selection in Kernel Fisher discriminant analysis
ICML '06 Proceedings of the 23rd international conference on Machine learning
Nonstationary kernel combination
ICML '06 Proceedings of the 23rd international conference on Machine learning
Large Scale Multiple Kernel Learning
The Journal of Machine Learning Research
Adaptive dimension reduction using discriminant analysis and K-means clustering
Proceedings of the 24th international conference on Machine learning
Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining
An efficient algorithm for local distance metric learning
AAAI'06 Proceedings of the 21st national conference on Artificial intelligence - Volume 1
An adaptive kernel method for semi-supervised clustering
ECML'06 Proceedings of the 17th European conference on Machine Learning
Learning a Mahalanobis distance metric for data clustering and classification
Pattern Recognition
Semi-supervised clustering with metric learning: An adaptive kernel method
Pattern Recognition
Transfer metric learning by learning task relationships
Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining
Image analysis with nonlinear adaptive dimension reduction
Proceedings of the Third International Conference on Internet Multimedia Computing and Service
Using Clustering and Metric Learning to Improve Science Return of Remote Sensed Imagery
ACM Transactions on Intelligent Systems and Technology (TIST)
Transfer Metric Learning with Semi-Supervised Extension
ACM Transactions on Intelligent Systems and Technology (TIST)
Regularized soft K-means for discriminant analysis
Neurocomputing
Hi-index | 0.00 |
A good distance metric is crucial for many data mining tasks. To learn a metric in the unsupervised setting, most metric learning algorithms project observed data to a low-dimensional manifold, where geometric relationships such as pairwise distances are preserved. It can be extended to the nonlinear case by applying the kernel trick, which embeds the data into a feature space by specifying the kernel function that computes the dot products between data points in the feature space. In this paper, we propose a novel unsupervised Nonlinear Adaptive Metric Learning algorithm, called NAML, which performs clustering and distance metric learning simultaneously. NAML firstmaps the data to a high-dimensional space through a kernel function; then applies a linear projection to find a low-dimensional manifold where the separability of the data is maximized; and finally performs clustering in the low-dimensional space. The performance of NAML depends on the selection of the kernel function and the projection. We show that the joint kernel learning, dimensionality reduction, and clustering can be formulated as a trace maximization problem, which can be solved via an iterative procedure in the EM framework. Experimental results demonstrated the efficacy of the proposed algorithm.