Large Language Bayes

2
citations
#1951
in NEURIPS 2025
of 5858 papers
1
Top Authors
7
Data Points

Top Authors

Abstract

Many domain experts do not have the time or expertise to write formal Bayesian models. This paper takes an informal problem description as input, and combines a large language model and a probabilistic programming language to define a joint distribution over formal models, latent variables, and data. A posterior over latent variables follows by conditioning on observed data and integrating over formal models. This presents a challenging inference problem. We suggest an inference recipe that amounts to generating many formal models from the large language model, performing approximate inference on each, and then doing a weighted average. This is justified and analyzed as a combination of self-normalized importance sampling, MCMC, and importance-weighted variational inference. Experimentally, this produces sensible predictions from only data and an informal problem description, without the need to specify a formal model.

Citation History

Jan 26, 2026
0
Jan 26, 2026
2+2
Jan 27, 2026
2
Feb 3, 2026
2
Feb 13, 2026
2
Feb 13, 2026
2
Feb 13, 2026
2