Premium
A randomized generalized low rank approximations of matrices algorithm for high dimensionality reduction and image compression
Author(s) -
Li Ke,
Wu Gang
Publication year - 2021
Publication title -
numerical linear algebra with applications
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.02
H-Index - 53
eISSN - 1099-1506
pISSN - 1070-5325
DOI - 10.1002/nla.2338
Subject(s) - singular value decomposition , mathematics , dimensionality reduction , algorithm , singular value , rank (graph theory) , curse of dimensionality , dimension (graph theory) , randomized algorithm , reduction (mathematics) , image compression , image (mathematics) , computer science , artificial intelligence , image processing , eigenvalues and eigenvectors , physics , statistics , geometry , quantum mechanics , combinatorics , pure mathematics
Summary High‐dimensionality reduction techniques are very important tools in machine learning and data mining. The method of generalized low rank approximations of matrices (GLRAM) is a popular technique for dimensionality reduction and image compression. However, it suffers from heavily computational overhead in practice, especially for data with high dimension. In order to reduce the cost of this algorithm, we propose a randomized GLRAM algorithm based on randomized singular value decomposition (RSVD). The theoretical contribution of our work is threefold. First, we discuss the decaying property of singular values of the matrices during iterations of the GLRAM algorithm, and provide a target rank required in the RSVD process from a theoretical point of view. Second, we establish the relationship between the reconstruction errors generated by the standard GLRAM algorithm and the randomized GLRAM algorithm. It is shown that the reconstruction errors generated by the former and the latter are comparable, even if the solutions are computed inaccurately during iterations. Third, the convergence of the randomized GLRAM algorithm is investigated. Numerical experiments on some real‐world data sets illustrate the superiority of our proposed algorithm over its original counterpart and some state‐of‐the‐art GLRAM‐type algorithms.