Sciweavers

Share
ICML
2003
IEEE

Learning Metrics via Discriminant Kernels and Multidimensional Scaling: Toward Expected Euclidean Representation

11 years 11 days ago
Learning Metrics via Discriminant Kernels and Multidimensional Scaling: Toward Expected Euclidean Representation
Distance-based methods in machine learning and pattern recognition have to rely on a metric distance between points in the input space. Instead of specifying a metric a priori, we seek to learn the metric from data via kernel methods and multidimensional scaling (MDS) techniques. Under the classification setting, we define discriminant kernels on the joint space of input and output spaces and present a specific family of discriminant kernels. This family of discriminant kernels is attractive because the induced metrics are Euclidean and Fisher separable, and MDS techniques can be used to find the lowdimensional Euclidean representations (also called feature vectors) of the induced metrics. Since the feature vectors incorporate information from both input points and their corresponding labels and they enjoy Fisher separability, they are appropriate to be used in distance-based classifiers.
Zhihua Zhang
Added 17 Nov 2009
Updated 17 Nov 2009
Type Conference
Year 2003
Where ICML
Authors Zhihua Zhang
Comments (0)
books