TY - JOUR
T1 - Large margin nearest neighbor classifiers
AU - Domeniconi, Carlotta
AU - Gunopulos, Dimitrios
AU - Peng, Jing
PY - 2005/7
Y1 - 2005/7
N2 - The nearest neighbor technique is a simple and appealing approach to addressing classification problems. It relies on the assumption of locally constant class conditional probabilities. This assumption becomes invalid in high dimensions with a finite number of examples due to the curse of dimensionality. Severe bias can be introduced under these conditions when using the nearest neighbor rule. The employment of a locally adaptive metric becomes crucial in order to keep class conditional probabilities close to uniform, thereby minimizing the bias of estimates. We propose a technique that computes a locally flexible metric by means of support vector machines (SVMs). The decision function constructed by SVMs is used to determine the most discriminant direction in a neighborhood around the query. Such a direction provides a local feature weighting scheme. We formally show that our method increases the margin in the weighted space where classification takes place. Moreover, our method has the important advantage of online computational efficiency over competing locally adaptive techniques for nearest neighbor classification. We demonstrate the efficacy of our method using both real and simulated data.
AB - The nearest neighbor technique is a simple and appealing approach to addressing classification problems. It relies on the assumption of locally constant class conditional probabilities. This assumption becomes invalid in high dimensions with a finite number of examples due to the curse of dimensionality. Severe bias can be introduced under these conditions when using the nearest neighbor rule. The employment of a locally adaptive metric becomes crucial in order to keep class conditional probabilities close to uniform, thereby minimizing the bias of estimates. We propose a technique that computes a locally flexible metric by means of support vector machines (SVMs). The decision function constructed by SVMs is used to determine the most discriminant direction in a neighborhood around the query. Such a direction provides a local feature weighting scheme. We formally show that our method increases the margin in the weighted space where classification takes place. Moreover, our method has the important advantage of online computational efficiency over competing locally adaptive techniques for nearest neighbor classification. We demonstrate the efficacy of our method using both real and simulated data.
KW - Feature relevance
KW - Margin
KW - Nearest neighbor classification
KW - Support vector machines (SVMs)
UR - http://www.scopus.com/inward/record.url?scp=23044490834&partnerID=8YFLogxK
U2 - 10.1109/TNN.2005.849821
DO - 10.1109/TNN.2005.849821
M3 - Article
C2 - 16121731
AN - SCOPUS:23044490834
SN - 1045-9227
VL - 16
SP - 899
EP - 909
JO - IEEE Transactions on Neural Networks
JF - IEEE Transactions on Neural Networks
IS - 4
ER -