Edward provides a testbed for rapid experimentation and research with probabilistic models. Here we show how to apply this process for complicated learning tasks.

Bayesian linear regression
A fundamental model for supervised learning.

Gaussian process classification
Learning a distribution over functions for supervised classification.

Mixture models
Unsupervised learning by clustering data points.

Latent space models
Analyzing connectivity patterns in neural data.

Mixture density networks
A neural density estimator for solving inverse problems.

Generative adversarial networks
Building a deep generative model of MNIST digits.

Probabilistic decoder
A model of latent codes in information theory.

Inference networks
How to amortize computation for training and testing models.

Bayesian neural network
Bayesian analysis with neural networks.

If you’re interested in contributing a tutorial, checking out the contributing page. For more background and notation, see the pages below.