Giter Club home page Giter Club logo

rxinfer.jl's Introduction

Official page Stable Dev Examples Q&A Roadmap Build Status Coverage DOI Zenodo

Overview

RxInfer.jl is a Julia package for automatic Bayesian inference on a factor graph with reactive message passing.

Given a probabilistic model, RxInfer allows for an efficient message-passing based Bayesian inference. It uses the model structure to generate an algorithm that consists of a sequence of local computations on a Forney-style factor graph (FFG) representation of the model.

Performance and scalability

RxInfer.jl has been designed with a focus on efficiency, scalability and maximum performance for running Bayesian inference with message passing. Below is a comparison between RxInfer.jl and Turing.jl on latent state estimation in a linear multi-variate Gaussian state-space model. Turing.jl is a state-of-the-art Julia-based general-purpose probabilistic programming package and is capable of running inference in a broader class of models. Still, RxInfer.jl executes the inference task in various models faster and more accurately. RxInfer.jl accomplishes this by taking advantage of any conjugate likelihood-prior pairings in the model, which have analytical posteriors that are known by RxInfer.jl. As a result, in models with conjugate pairings, RxInfer.jl often beats general-purpose probabilistic programming packages in terms of computational load, speed, memory and accuracy. Note, however, that RxInfer.jl also supports non-conjugate inference and is continually improving in order to support a larger class of models.

Turing comparison Scalability performance

Faster inference with better results

RxInfer.jl not only beats generic-purpose Bayesian inference methods in conjugate models, executes faster, and scales better, but also provides more accurate results. Check out the documentation for more examples!

Inference with RxInfer Inference with HMC

The benchmark and accuracy experiment, which generated these plots, is available in the benchmarks/ folder. Note, that the execution speed and accuracy of the HMC estimator heavily depends on the choice of hyper-parameters. In this example, RxInfer executes exact inference consistently and does not depend on any hyper-parameters.

References

Installation

Install RxInfer through the Julia package manager:

] add RxInfer

Optionally, use ] test RxInfer to validate the installation by running the test suite.

Getting Started

There are examples available to get you started in the examples/ folder. Alternatively, preview the same examples in the documentation.

Coin flip simulation

Here we show a simple example of how to use RxInfer.jl for Bayesian inference problems. In this example we want to estimate a bias of a coin in a form of a probability distribution in a coin flip simulation.

Let's start by creating some dataset. For simplicity in this example we will use static pre-generated dataset. Each sample can be thought of as the outcome of single flip which is either heads or tails (1 or 0). We will assume that our virtual coin is biased, and lands heads up on 75% of the trials (on average).

First let's setup our environment by importing all needed packages:

using RxInfer, Random

Next, let's define our dataset:

n = 500  # Number of coin flips
p = 0.75 # Bias of a coin

distribution = Bernoulli(p) 
dataset      = float.(rand(Bernoulli(p), n))

Model specification

In a Bayesian setting, the next step is to specify our probabilistic model. This amounts to specifying the joint probability of the random variables of the system.

Likelihood

We will assume that the outcome of each coin flip is governed by the Bernoulli distribution, i.e.

$$y_i \sim \mathrm{Bernoulli}(\theta)$$

where $y_i = 1$ represents "heads", $y_i = 0$ represents "tails". The underlying probability of the coin landing heads up for a single coin flip is $\theta \in [0,1]$.

Prior

We will choose the conjugate prior of the Bernoulli likelihood function defined above, namely the beta distribution, i.e.

$$\theta \sim Beta(a, b)$$

where $a$ and $b$ are the hyperparameters that encode our prior beliefs about the possible values of $\theta$. We will assign values to the hyperparameters in a later step.

Joint probability

The joint probability is given by the multiplication of the likelihood and the prior, i.e.

$$P(y_{1:N}, \theta) = P(\theta) \prod_{i=1}^N P(y_i | \theta).$$

Now let's see how to specify this model using GraphPPL's package syntax.

# GraphPPL.jl export `@model` macro for model specification
# It accepts a regular Julia function and builds an FFG under the hood
@model function coin_model(n)

    # `datavar` creates data 'inputs' in our model
    # We will pass data later on to these inputs
    # In this example we create a sequence of inputs that accepts Float64
    y = datavar(Float64, n)
    
    # We endow θ parameter of our model with some prior
    θ ~ Beta(2.0, 7.0)
    
    # We assume that outcome of each coin flip 
    # is governed by the Bernoulli distribution
    for i in 1:n
        y[i] ~ Bernoulli(θ)
    end
    
end

As you can see, RxInfer offers a model specification syntax that resembles closely to the mathematical equations defined above. We use datavar function to create "clamped" variables that take specific values at a later date. $\theta \sim \mathrm{Beta}(2.0, 7.0)$ expression creates random variable $θ$ and assigns it as an output of $\mathrm{Beta}$ node in the corresponding FFG.

Inference specification

Once we have defined our model, the next step is to use RxInfer API to infer quantities of interests. To do this we can use a generic infer function from RxInfer.jl that supports static datasets.

result = infer(
    model = coin_model(length(dataset)),
    data  = (y = dataset, )
)

Coin Flip

Where to go next?

There are a set of examples available in RxInfer repository that demonstrate the more advanced features of the package. Alternatively, you can head to the documentation that provides more detailed information of how to use RxInfer to specify more complex probabilistic models.

Additionally, checkout our video from JuliaCon 2023 for a high-level overview of the package

Roadmap

Our high-level project roadmap outlines the key milestones and focus areas for the upcoming years:

Q1/Q2 2024 Q3/Q4 2024 2025 2026
🧩 Nested models with GraphPPL.jl 🛡️ Robustness (NaN, Inf free) 🌐 Stochastic Processes 🔄 Automated inference with non-exponential family
🔄 Automated inference with ExponentialFamily.jl 🧠 Memory-efficiency 🚀 Resource-adaptive inference 📊 Inference over graph structure

For a more granular view of our progress and ongoing tasks, check out our project board.

Contributing

We welcome contributions from the community. If you are interested in contributing to the development of RxInfer.jl, please check out our contributing guide, the contributing guidelines, or look at the issues linked with the good first issue label to get started.

Ecosystem

The RxInfer framework consists of three core packages developed by reactivebayes:

  • ReactiveMP.jl - the underlying message passing-based inference engine
  • GraphPPL.jl - model and constraints specification package
  • Rocket.jl - reactive extensions package for Julia

License

MIT License Copyright (c) 2021-2024 BIASlab, 2024-present ReactiveBayes

rxinfer.jl's People

Contributors

bvdmitri avatar albertpod avatar bartvanerp avatar hoangmhnguyen avatar chengfeng-jia avatar mhidalgoaraya avatar wmkouw avatar nimrais avatar thijsvdlaar avatar wouterwln avatar sepideh-adamiat avatar raphael-tresor avatar magnuskoudahl avatar alstat avatar dhuige avatar kronosthelate avatar pitmonticone avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.