Nearest neighbor voting in high dimensional data: Learning from past occurrences
Author(s) -
Nenad Tomašev,
Dunja Mladenić
Publication year - 2012
Publication title -
computer science and information systems
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.244
H-Index - 24
eISSN - 2406-1018
pISSN - 1820-0214
DOI - 10.2298/csis111211014t
Subject(s) - computer science , k nearest neighbors algorithm , curse of dimensionality , best bin first , voting , large margin nearest neighbor , nearest neighbor chain algorithm , nearest neighbor search , nearest neighbor graph , data mining , pattern recognition (psychology) , artificial intelligence , algorithm , cluster analysis , canopy clustering algorithm , correlation clustering , politics , political science , law
Hubness is a recently described aspect of the curse of dimensionality inherent to nearest-neighbor methods. This paper describes a new approach for exploiting the hubness phenomenon in k-nearest neighbor classification. We argue that some of the neighbor occurrences carry more information than others, by the virtue of being less frequent events. This observation is related to the hubness phenomenon and we explore how it affects high-dimensional k-nearest neighbor classification. We propose a new algorithm, Hubness Information k-Nearest Neighbor (HIKNN), which introduces the k-occurrence informativeness into the hubness-aware k-nearest neighbor voting framework. The algorithm successfully overcomes some of the issues with the previous hubness-aware approaches, which is shown by performing an extensive evaluation on several types of high-dimensional data.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom