Abstract
Variational inference is a powerful tool for approximate inference, and it has been recently applied for representation learning with deep generative models. We develop the variational Gaussian process (VGP), a Bayesian nonparametric variational family, which adapts its shape to match complex posterior distributions. The VGP generates approximate posterior samples by generating latent inputs and warping them through random non-linear mappings; the distribution over random mappings is learned during inference, enabling the transformed outputs to adapt to varying complexity. We prove a universal approximation theorem for the VGP, demonstrating its representative power for learning any model. For inference we present a variational objective inspired by auto-encoders and perform black box inference over a wide class of models. The VGP achieves new state-of-the-art results for unsupervised learning, inferring models such as the deep latent Gaussian model and the recently proposed DRAW.
Original language | English (US) |
---|---|
State | Published - 2016 |
Event | 4th International Conference on Learning Representations, ICLR 2016 - San Juan, Puerto Rico Duration: May 2 2016 → May 4 2016 |
Conference
Conference | 4th International Conference on Learning Representations, ICLR 2016 |
---|---|
Country/Territory | Puerto Rico |
City | San Juan |
Period | 5/2/16 → 5/4/16 |
ASJC Scopus subject areas
- Education
- Computer Science Applications
- Linguistics and Language
- Language and Linguistics