TU Darmstadt / ULB / TUbiblio

Efficient Gradient-Free Variational Inference using Policy Search

Arenz, Oleg ; Neumann, Gerhard ; Zhong, Mingjun (2022):
Efficient Gradient-Free Variational Inference using Policy Search. (Publisher's Version)
80, In: Proceedings of Machine Learning Research, pp. 234-243,
Darmstadt, PMLR, 35th International Conference on Machine Learning (ICML 2018), Stockholm, Sweden, 10.-15.07.2018, e-ISSN 2640-3498,
DOI: 10.26083/tuprints-00022925,
[Conference or Workshop Item]

Abstract

Inference from complex distributions is a common problem in machine learning needed for many Bayesian methods. We propose an efficient, gradient-free method for learning general GMM approximations of multimodal distributions based on recent insights from stochastic search methods. Our method establishes information-geometric trust regions to ensure efficient exploration of the sampling space and stability of the GMM updates, allowing for efficient estimation of multi-variate Gaussian variational distributions. For GMMs, we apply a variational lower bound to decompose the learning objective into sub-problems given by learning the individual mixture components and the coefficients. The number of mixture components is adapted online in order to allow for arbitrary exact approximations. We demonstrate on several domains that we can learn significantly better approximations than competing variational inference methods and that the quality of samples drawn from our approximations is on par with samples created by state-of-the-art MCMC samplers that require significantly more computational resources.

Item Type: Conference or Workshop Item
Erschienen: 2022
Creators: Arenz, Oleg ; Neumann, Gerhard ; Zhong, Mingjun
Origin: Secondary publication service
Status: Publisher's Version
Title: Efficient Gradient-Free Variational Inference using Policy Search
Language: English
Abstract:

Inference from complex distributions is a common problem in machine learning needed for many Bayesian methods. We propose an efficient, gradient-free method for learning general GMM approximations of multimodal distributions based on recent insights from stochastic search methods. Our method establishes information-geometric trust regions to ensure efficient exploration of the sampling space and stability of the GMM updates, allowing for efficient estimation of multi-variate Gaussian variational distributions. For GMMs, we apply a variational lower bound to decompose the learning objective into sub-problems given by learning the individual mixture components and the coefficients. The number of mixture components is adapted online in order to allow for arbitrary exact approximations. We demonstrate on several domains that we can learn significantly better approximations than competing variational inference methods and that the quality of samples drawn from our approximations is on par with samples created by state-of-the-art MCMC samplers that require significantly more computational resources.

Book Title: Proceedings of Machine Learning Research
Series Volume: 80
Place of Publication: Darmstadt
Publisher: PMLR
Collation: 10 ungezählte Seiten
Uncontrolled Keywords: Machine Learning, ICML, Variational Inference, Sampling, Policy Search, MCMC, Markov Chain Monte Carlo
Divisions: 20 Department of Computer Science
20 Department of Computer Science > Intelligent Autonomous Systems
Event Title: 35th International Conference on Machine Learning (ICML 2018)
Event Location: Stockholm, Sweden
Event Dates: 10.-15.07.2018
Date Deposited: 02 Dec 2022 12:46
DOI: 10.26083/tuprints-00022925
URL / URN: https://tuprints.ulb.tu-darmstadt.de/22925
URN: urn:nbn:de:tuda-tuprints-229250
Additional Information:

Presentation video: https://vimeo.com/294656117

PPN:
Corresponding Links:
Export:
Suche nach Titel in: TUfind oder in Google
Send an inquiry Send an inquiry

Options (only for editors)
Show editorial Details Show editorial Details