On the geometry of Stein variational gradient descent

Andrew Duncan, Lukasz Szpruch, Nikolas Nusken

Research output: Contribution to journalArticlepeer-review

16 Downloads (Pure)

Abstract

Bayesian inference problems require sampling or approximating high-dimensional probability dis- tributions. The focus of this paper is on the recently introduced Stein variational gradient descent methodology, a class of algorithms that rely on iterated steepest descent steps with respect to a reproducing kernel Hilbert space norm. This construction leads to interacting particle systems, the mean-field limit of which is a gradient flow on the space of probability distributions equipped with a certain geometrical structure. We leverage this viewpoint to shed some light on the convergence properties of the algorithm, in particular addressing the problem of choosing a suitable positive definite kernel function. Our analysis leads us to considering certain nondifferentiable kernels with adjusted tails. We demonstrate significant performance gains of these in various numerical experiments.
Original languageEnglish
Pages (from-to)1-39
Number of pages39
JournalJOURNAL OF MACHINE LEARNING RESEARCH
Volume24
Issue number56
Publication statusPublished - 1 Jan 2023

Fingerprint

Dive into the research topics of 'On the geometry of Stein variational gradient descent'. Together they form a unique fingerprint.

Cite this