Proximity variational inference

Jaan Altosaar, Rajesh Ranganath, David M. Blei

Research output: Contribution to conferencePaperpeer-review


Variational inference is a powerful approach for approximate posterior inference. However, it is sensitive to initialization and can be subject to poor local optima. In this paper, we develop proximity variational inference (pvi). pvi is a new method for optimizing the variational objective that constrains subsequent iterates of the variational parameters to robustify the optimization path. Consequently, pvi is less sensitive to initialization and optimization quirks and finds better local optima. We demonstrate our method on four proximity statistics. We study pvi on a Bernoulli factor model and sigmoid belief network fit to real and synthetic data and compare to deterministic annealing (Katahira et al., 2008). We highlight the flexibility of pvi by designing a proximity statistic for Bayesian deep learning models such as the variational autoencoder (Kingma and Welling, 2014; Rezende et al., 2014) and show that it gives better performance by reducing overpruning. pvi also yields improved predictions in a deep generative model of text. Empirically, we show that pvi consistently finds better local optima and gives better predictive performance.

Original languageEnglish (US)
Number of pages9
StatePublished - 2018
Event21st International Conference on Artificial Intelligence and Statistics, AISTATS 2018 - Playa Blanca, Lanzarote, Canary Islands, Spain
Duration: Apr 9 2018Apr 11 2018


Conference21st International Conference on Artificial Intelligence and Statistics, AISTATS 2018
CityPlaya Blanca, Lanzarote, Canary Islands

ASJC Scopus subject areas

  • Statistics and Probability
  • Artificial Intelligence


Dive into the research topics of 'Proximity variational inference'. Together they form a unique fingerprint.

Cite this