Distance‐preserving dimensionality reduction
Author(s) -
Yang Li
Publication year - 2011
Publication title -
wiley interdisciplinary reviews: data mining and knowledge discovery
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.506
H-Index - 47
eISSN - 1942-4795
pISSN - 1942-4787
DOI - 10.1002/widm.39
Subject(s) - dimensionality reduction , computer science , preprocessor , strengths and weaknesses , data pre processing , computation , reduction (mathematics) , data mining , curse of dimensionality , data reduction , artificial intelligence , machine learning , algorithm , mathematics , philosophy , geometry , epistemology
Abstract This paper presents an overview of basic concepts and principles that deal with the problem of mapping high‐dimensional data to low‐dimensional space such that distances between all or some pairs of data points are preserved. It introduces related techniques and systematizes today's methods into linear methods, methods using iterative optimization, methods preserving exact distances, methods using geodesic distances, and methods using alignments of local models. It discusses these methods by focusing on their basic ideas, by summarizing their common features and differences, and by comparing their strengths and weaknesses. This paper assumes no familiarity with dimensionality reduction. The main text should be readable by people with little technical background. Technical information of important algorithms is briefly presented in sidebars, the reading of which assumes basics in statistics and matrix computation. © 2011 John Wiley & Sons, Inc. WIREs Data Mining Knowl Discov 2011 1 369–380 DOI: 10.1002/widm.39 This article is categorized under: Technologies > Data Preprocessing Technologies > Machine Learning