Skip to yearly menu bar Skip to main content


Poster

posteriordb: Testing, Benchmarking and Developing Bayesian Inference Algorithms

Måns Magnusson · Jakob Torgander · Paul Bürkner · Lu Zhang · Bob Carpenter · Aki Vehtari

Hall A-E 9
[ ]
Mon 5 May 1 a.m. PDT — 4 a.m. PDT
 
Oral presentation: Oral Session 5: Probabilistic Inference and Optimzation
Sun 4 May 8:30 p.m. PDT — 9:30 p.m. PDT

Abstract:

The general applicability and robustness of posterior inference algorithms is critical to widely used probabilistic programming languages such as Stan, PyMC, Pyro, and Turing.jl. When designing a new inference algorithm, whether it involves Monte Carlo sampling or variational approximation, the fundamental problem is evaluating its accuracy and efficiency across a range of representative target posteriors. To solve this problem, we propose posteriordb, a database of models and data sets defining target densities along with reference Monte Carlo draws. We further provide a guide to the best practices in using posteriordb for algorithm evaluation and comparison. To provide a wide range of realistic posteriors, posteriordb currently comprises 120 representative models with data, and has been instrumental in developing several inference algorithms.

Live content is unavailable. Log in and register to view live content