liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Markovian Score Climbing: Variational Inference with KL(p||q)
Columbia University.
Linköping University, Department of Computer and Information Science, The Division of Statistics and Machine Learning. Linköping University, Faculty of Science & Engineering.
Columbia University.
2020 (English)In: Advances in Neural Information Processing Systems 33 (NeurIPS 2020), 2020, Vol. 33, p. 15499-15510Conference paper, Published paper (Refereed)
Abstract [en]

Modern variational inference (VI) uses stochastic gradients to avoid intractable expectations, enabling large-scale probabilistic inference in complex models. VI posits a family of approximating distributions q and then finds the member of that family that is closest to the exact posterior p. Traditionally, VI algorithms minimize the “exclusive Kullback-Leibler (KL)” KL(q||p), often for computational convenience. Recent research, however, has also focused on the “inclusive KL” KL(p||q), which has good statistical properties that makes it more appropriate for certain inference problems. This paper develops a simple algorithm for reliably minimizing the inclusive KL using stochastic gradients with vanishing bias. This method, which we call Markovian score climbing (MSC), converges to a local optimum of the inclusive KL. It does not suffer from the systematic errors inherent in existing methods, such as Reweighted Wake-Sleep and Neural Adaptive Sequential Monte Carlo, which lead to bias in their final estimates. We illustrate convergence on a toy model and demonstrate the utility of MSC on Bayesian probit regression for classification as well as a stochastic volatility model for financial data.

Place, publisher, year, edition, pages
2020. Vol. 33, p. 15499-15510
National Category
Probability Theory and Statistics Computer Sciences
Identifiers
URN: urn:nbn:se:liu:diva-184163ISI: 001207696402032OAI: oai:DiVA.org:liu-184163DiVA, id: diva2:1650024
Conference
Neural Information Processing Systems
Funder
Swedish Research Council, 2016-04278Swedish Foundation for Strategic Research, ICA16- 0015Wallenberg AI, Autonomous Systems and Software Program (WASP)Available from: 2022-04-05 Created: 2022-04-05 Last updated: 2024-11-18

Open Access in DiVA

No full text in DiVA

Other links

https://proceedings.neurips.cc/paper/2020/hash/b20706935de35bbe643733f856d9e5d6-Abstract.html

Search in DiVA

By author/editor
Lindsten, Fredrik
By organisation
The Division of Statistics and Machine LearningFaculty of Science & Engineering
Probability Theory and StatisticsComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 47 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf