Adaptive kernel metric nearest neighbor classification

Jing Peng, Douglas R. Heisterkamp, H. K. Dai

Research output: Contribution to journalArticlepeer-review

34 Scopus citations

Abstract

Nearest neighbor classification assumes locally constant class conditional probabilities. This assumption becomes invalid in high dimensions due to the curse-of-dimensionality. Severe bias can be introduced under these conditions when using the nearest neighbor rule. We propose an adaptive nearest neighbor classification method to try to minimize bias. We use quasiconformal transformed kernels to compute neighborhoods over which the class probabilities tend to be more homogeneous. As a result, better classification performance can be expected. The efficacy of our method is validated and compared against other competing techniques using a variety of data sets.

Original languageEnglish
Pages (from-to)33-36
Number of pages4
JournalProceedings - International Conference on Pattern Recognition
Volume16
Issue number3
StatePublished - 2002

Fingerprint

Dive into the research topics of 'Adaptive kernel metric nearest neighbor classification'. Together they form a unique fingerprint.

Cite this