Skip to yearly menu bar Skip to main content


Poster

Variational Learning of Fractional Posteriors

Kian Ming Chai · Edwin V. Bonilla

[ ]
Wed 16 Jul 11 a.m. PDT — 1:30 p.m. PDT

Abstract:

We introduce a novel one-parameter variational objective that lower bounds the data evidence and enables the estimation of approximate fractional posteriors. We extend this framework to hierarchical and Bayes posteriors, offering a versatile tool for probabilistic modelling. We demonstrate two cases where gradients can be obtained analytically and a simulation study on mixture models showing that our fractional posteriors can be used to achieve better calibration compared to posteriors from the conventional variational bound. When applied to variational autoencoders (VAEs), our approach attains tighter evidence bounds and enables learning of high-performing approximate Bayes posteriors jointly with fractional posteriors. We show that VAEs trained with fractional posteriors produce decoders that are better aligned for generation from the prior.

Live content is unavailable. Log in and register to view live content