Sequential bayesian experimental design with variable cost structure
Sequential Bayesian Experimental Design With Variable Cost Structure
Sue Zheng, David Hayden, Jason Pacheco, John W Fisher III
Advances in Neural Information Processing Systems 33, 2020.
[pdf] [poster] [slides]

Summary: Mutual information (MI) is a commonly adopted utility function in Bayesian optimal experimental design [see Blackwell 1950, Lindley 1956 & Bernardo 1979]. While theoretically appealing, MI evaluation poses a significant computational burden for most real world applications. As a result, many algorithms utilize MI bounds as proxies that lack regret-style guarantees. Here, we utilize two-sided bounds to provide such guarantees. Bounds are successively refined/tightened through additional computation until a desired guarantee is achieved. We consider the problem of adaptively allocating computational resources in BOED. Our approach achieves the same guarantee as existing methods, but with fewer evaluations of the costly MI reward. We adapt knapsack optimization of best arm identification problems, with important differences that impact overall algorithm design and performance. First, observations of MI rewards are biased. Second, evaluating experiments incurs shared costs amongst all experiments (posterior sampling) in addition to per experiment costs that may vary with increasing evaluation. We propose and demonstrate an algorithm that accounts for these variable costs in the refinement decision. Specifically, we

  • incorporate both lower and upper bounds on the intractable MI utility into BOED to provide performance guarantees relative to the optimal design and show how they can be maintained with minimal additional computation,
  • introduce a cost-sensitive sequential optimization that iteratively refines bounds to target a desired performance level,
  • formulate a greedy knapsack optimization that elegantly trades off performance for computation, and
  • evaluate our BOED method against adaptive allocation techniques in Gaussian models as well as the challenging problem of multi-target tracking.

Related SLI papers: Williams formulates information gathering in distributed networks in Williams et al 2007 as a Markov Decision Process subject to resource constraints and subsequently develops theoretical guarantees on MI based sensor selection Williams at al 2007 (AIStats). Zheng develops MCMC-based information planning Zheng et al algorithms along with analysis of estimator deviation bounds. Pacheco develops a variational formulation of information planning including bounds on the variational approximation of the information reward in Pacheco et al.


Video Overview

Sequential BOED with Variable Cost Structure Overview presented at Neurips 2020

The approach makes use of two Monte Carlo estimators of MI: $$\mathit{l} = \frac{1}{N} \sum_{n=1}^N \log \frac{p(y_n | x_n)}{\frac{1}{N} \sum_{m=1}^N p(y_n | x_m)}, \quad \mathit{u} = \frac{1}{N} \sum_{n=1}^N \log \frac{p(y_n | x_n)}{\frac{1}{N-1} \sum_{m \neq n} p(y_n | x_m)}$$ and the fact that they are (in expectation) lower and upper bounds Foster 2019 and 2020 $$\mathbb{E} \left[\mathit{l} \right] \leq I\left(X;Y\right) \leq \mathbb{E}\left[\mathit{u}\right]$$

Algorithms

TSP

The standard BOED algorithm is shown above where a sample-based estimator is used as a proxy for the MI reward. In contrast, the variable cost structure algorithm below incorporates sample-based bounds on MI combined with an iterative refinement scheme that yields reduced computation and performance guarantees. Equation numbers correspond to the published paper.

TSP

Experimental Result

TSP
Realized Performance Across Motifs in BOED. Top-row: Three distributions over MI rewards characterize different problem structures or ``motifs''. Bottom-row: Median and quartiles performance (information gain relative to optimal) under proxy-based BOED framework for two proxies under various budgets: lower bound (bed-lb) and upper bound (bed-ub). Motifs are: scarce mostly uninformative (left), broad similarly informative (middle), and abundant mostly informative (right). Performance varies greatly depending on the proxy used and the problem structure, leading to wasted compute cycles when the specified budget is too large (scarce with bed-lb) or poor performance when not large enough (abundant with bed-ub).

References

[1] Jason Pacheco, John W. Fisher III; "Variational information planning for sequential decision making," in International Conference on Artificial Intelligence and Statistics (Aistats), 2019. [pdf]
[2] Sue Zheng, Jason Pacheco, John W Fisher III; "A robust approach to sequential information theoretic planning," in International Conference on Machine Learning (ICML), 2018. [pdf]
[3] Jason L. Williams, John W. Fisher III, Alan S. Willsky; "Approximate dynamic programming for communication-constrained sensor network management," in IEEE Transactions on Signal Processing, 2007. [pdf]
[4] Jason L. Williams, John W. Fisher III, Alan S. Willsky; "Performance guarantees for information theoretic active inference," in Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics, 2007. [pdf]
[5] Adam Foster, Martin Jankowiak, Matthew O’Meara, Yee Whye Teh, Tom Rainforth; "A unified stochastic gradient approach to designing bayesian-optimal experiments," in AISTATS, 2020.
[6] Adam Foster, Martin Jankowiak, Elias Bingham, Paul Horsfall, Yee Whye Teh, Thomas Rainforth, Noah Goodman; "Variational bayesian optimal experimental design," in NeurIPS, 2019.
[7] J. M. Bernardo; "Expected information as expected utility," in Ann. Stat., 1979.
[8] D. V. Lindley; "On a measure of the information provided by an experiment," in The Annals of Mathematical Statistics, 1956.
[9] D. Blackwell; "Comparison of experiments," in 2nd Bsmsp, 1950.