Bayesian Experimental Design (BED)
A model-based approach to choose some design which maximizes the information gained about some model parameters from the outcome after performing experiment . The design is something the experimenter can control, meanwhile is the latent variable of interest - whose posterior is
Given the likelihood and prior
, the objective is to choose a design that
yields the greatest reduction in our uncertainty i.e. the difference
between the prior entropy and the posterior entropy. This value,
called the Information Gain (IG) (/Bayesian surprise) is large if the posterior reduces our uncertainty by a large amount. As a function of it is
Or more generally in an iterative setting, at time , where and is the entropy operator. This quantity
requires knowing posterior distributions which is in general an
intractable task in itself. It also depends on which we
have not yet seen. Integrating out (i.e. taking an expectation
w.r.t gives us the Expected
Information Gain (EIG)
#TODO Add to above.
The above formula is derived after:
- Dropping the first term
since it
does not depend on .
- The double expectation
is the
same as .
- drops out since
the term is integrated over just the function 1, and the DAG
model leaves and independent.
This amounts to using the model to both define the utility function we optimize and to approximate the true distribution over outcomes .
The Bayes optimal design is then defined as:
In general, the objective need not be EIG, any other utility function
could be used, as long as we integrate out .
Since the EIG is itself an expectation over the IG, the actual objective we optimize for here is doubly intractable. A Monte Carlo approximation involves sampling for the outer expectation and for each such sample, approximate the marginal likelihood .
The EIG can also be interpreted as:
- Mutual information between the parameters and the outcomes.
- Expected utility with the KL divergence utility function.
- Expected reduction in predictive uncertainty (BALD score)
((, , Equation 2)).
To perform adaptive experiments, we condition on in both the prior and posterior entropies in the IG and posterior predictive distributions. This gives rise to the greedy adpative BOED loop:
- Choose new design by optimizing conditional EIG.
- Update posterior .
- Sample new outcome .
Emacs 29.4 (Org mode 9.6.15)