A Language and Smoothed Semantics for Convergent Stochastic Gradient DescentRemote
We study the optimisation of expectations, a problem which arises for examples in variational inference using the reparametrisation trick. As our starting point we take a variant of the simply-typed lambda calculus with reals and conditionals to express the function the expectation of which is taken. Unfortunately, it is well-known that the reparametrisation gradient estimator is biased in this setting due to fact that the model is not differentiable in general. As a consequence, stochastic gradient descent, which is commonly used to solve such problems, may converge to incorrect results. We have devised a type system restricting the language, and designed a stochastic gradient descent-like procedure for solving the optimisation problem which provably converges to stationary points for typable programs. Our approach is based on a smoothed denotational semantics in the cartesian closed category of Frölicher spaces.
Slides (Lafi2022_Dominik_Wagner_dsgd.pdf) | 492KiB |
Sun 16 JanDisplayed time zone: Eastern Time (US & Canada) change
15:05 - 16:20 | |||
15:05 18mTalk | Towards Denotational Semantics of AD for Higher-Order, Recursive, Probabilistic LanguagesRemote LAFI Alexander K. Lew Massachusetts Institute of Technology, USA, Mathieu Huot Oxford University, Vikash K. Mansinghka MIT File Attached | ||
15:23 18mTalk | A Language and Smoothed Semantics for Convergent Stochastic Gradient DescentRemote LAFI File Attached | ||
15:42 18mTalk | Nonparametric Involutive Markov Chain Monte CarloRemote LAFI Carol Mak University of Oxford, Fabian Zaiser University of Oxford, C.-H. Luke Ong University of Oxford File Attached | ||
16:01 18mTalk | Rigorous Approximation of Posterior Inference for Probabilistic ProgramsRemote LAFI Fabian Zaiser University of Oxford, Raven Beutner CISPA Helmholtz Center for Information Security, Germany, C.-H. Luke Ong University of Oxford File Attached |