Quantifying the mini-batching error in Bayesian inference for Adaptive Langevin dynamics - Inria - Institut national de recherche en sciences et technologies du numérique Access content directly
Journal Articles Journal of Machine Learning Research Year : 2023

Quantifying the mini-batching error in Bayesian inference for Adaptive Langevin dynamics

Abstract

The computational cost of usual Monte Carlo methods for sampling a posteriori laws in Bayesian inference scales linearly with the number of data points. One option to reduce it to a fraction of this cost is to resort to mini-batching in conjunction with unadjusted discretizations of Langevin dynamics, in which case only a random fraction of the data is used to estimate the gradient. However, this leads to an additional noise in the dynamics and hence a bias on the invariant measure which is sampled by the Markov chain. We advocate using the so-called Adaptive Langevin dynamics, which is a modification of standard inertial Langevin dynamics with a dynamical friction which automatically corrects for the increased noise arising from mini-batching. We investigate the practical relevance of the assumptions underpinning Adaptive Langevin (constant covariance for the estimation of the gradient), which are not satisfied in typical models of Bayesian inference, and quantify the bias induced by minibatching in this case. We also show how to extend AdL in order to systematically reduce the bias on the posterior distribution by considering a dynamical friction depending on the current value of the parameter to sample.

Dates and versions

hal-03386488 , version 1 (19-10-2021)

Identifiers

Cite

Inass Sekkat, Gabriel Stoltz. Quantifying the mini-batching error in Bayesian inference for Adaptive Langevin dynamics. Journal of Machine Learning Research, 2023, 24 (329), pp.1-58. ⟨hal-03386488⟩
45 View
0 Download

Altmetric

Share

Gmail Facebook X LinkedIn More