z-logo
open-access-imgOpen Access
How isotropic kernels perform on simple invariants
Author(s) -
Jonas Paccolat,
Stefano Spigler,
Matthieu Wyart
Publication year - 2021
Publication title -
machine learning: science and technology
Language(s) - English
Resource type - Journals
ISSN - 2632-2153
DOI - 10.1088/2632-2153/abd485
Subject(s) - algorithm , artificial intelligence , support vector machine , computer science , machine learning
We investigate how the training curve of isotropic kernel methods depends on the symmetry of the task to be learned, in several settings. (i) We consider a regression task, where the target function is a Gaussian random field that depends only on d ∥ variables, fewer than the input dimension d . We compute the expected test error ϵ that follows ϵ ∼ p − β where p is the size of the training set. We find that β  ∼ 1/ d independently of d ∥ , supporting previous findings that the presence of invariants does not resolve the curse of dimensionality for kernel regression. (ii) Next we consider support-vector binary classification and introduce the stripe model , where the data label depends on a single coordinate y ( x _ ) = y ( x 1 ) , corresponding to parallel decision boundaries separating labels of different signs, and consider that there is no margin at these interfaces. We argue and confirm numerically that, for large bandwidth, β = d − 1 + ξ 3 d − 3 + ξ , where ξ  ∈ (0, 2) is the exponent characterizing the singularity of the kernel at the origin. This estimation improves classical bounds obtainable from Rademacher complexity. In this setting there is no curse of dimensionality since β → 1 / 3 as d → ∞ . (iii) We confirm these findings for the spherical model , for which y ( x _ ) = y ( | | x _ | | ) . (iv) In the stripe model, we show that, if the data are compressed along their invariants by some factor λ (an operation believed to take place in deep networks), the test error is reduced by a factor λ − 2 ( d − 1 ) 3 d − 3 + ξ .

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here