Research Library

open-access-imgOpen AccessOnline Laplace Model Selection Revisited
Author(s)
Jihao Andreas Lin,
Javier Antorán,
José Miguel Hernández-Lobato
Publication year2024
The Laplace approximation provides a closed-form model selection objectivefor neural networks (NN). Online variants, which optimise NN parameters jointlywith hyperparameters, like weight decay strength, have seen renewed interest inthe Bayesian deep learning community. However, these methods violate Laplace'smethod's critical assumption that the approximation is performed around a modeof the loss, calling into question their soundness. This work re-derives onlineLaplace methods, showing them to target a variational bound on a mode-correctedvariant of the Laplace evidence which does not make stationarity assumptions.Online Laplace and its mode-corrected counterpart share stationary points where1. the NN parameters are a maximum a posteriori, satisfying the Laplacemethod's assumption, and 2. the hyperparameters maximise the Laplace evidence,motivating online methods. We demonstrate that these optima are roughlyattained in practise by online algorithms using full-batch gradient descent onUCI regression datasets. The optimised hyperparameters prevent overfitting andoutperform validation-based early stopping.
Language(s)English

Seeing content that should not be on Zendy? Contact us.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here