By Topic

Convex Optimizations for Distance Metric Learning and Pattern Classification [Applications Corner]

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
Weinberger, K.Q. ; Dept. of Comput. Sci. & Eng., Washington Univ., St. Louis, MO, USA ; Fei Sha ; Saul, L.K.

The goal of machine learning is to build automated systems that can classify and recognize complex patterns in data. The representation of the data plays an important role in determining what types of patterns can be automatically discovered. Many algorithms for machine learning assume that the data are represented as elements in a metric space. The performance of these algorithms can depend sensitively on the manner in which distances are measured. When data are represented as points in a multidimensional vector space, simple Euclidean distances are often used to measure the dissimilarity between different examples. However, such distances often do not yield reliable judgments; in addition, they cannot highlight the distinctive features that play a role in certain types of classification, but not others. Naturally, for different types of clustering, different ways of measuring dissimilarity were needed. In particular, different metrics for computing distances between feature vectors. This paper describes two algorithms for learning such distance metrics based on recent developments in convex optimization.

Published in:

Signal Processing Magazine, IEEE  (Volume:27 ,  Issue: 3 )