Browse by author
Lookup NU author(s): Professor James WasonORCiD
Full text for this publication is not currently held within this repository. Alternative links are provided below where available.
© Institute of Mathematical Statistics, 2015. Multi-armed bandit problems (MABPs) are a special type of optimal control problem well suited to model resource allocation under uncertainty in a wide variety of contexts. Since the first publication of the optimal solution of the classic MABP by a dynamic index rule, the bandit literature quickly diversified and emerged as an active research topic. Across this literature, the use of bandit models to optimally design clinical trials became a typical motivating application, yet little of the resulting theory has ever been used in the actual design and analysis of clinical trials. To this end, we review two MABP decision-theoretic approaches to the optimal allocation of treatments in a clinical trial: the infinite-horizon Bayesian Bernoulli MABP and the finite-horizon variant. These models possess distinct theoretical properties and lead to separate allocation rules in a clinical trial design context. We evaluate their performance compared to other allocation rules, including fixed randomization. Our results indicate that bandit approaches offer significant advantages, in terms of assigning more patients to better treatments, and severe limitations, in terms of their resulting statistical power. We propose a novel bandit-based patient allocation rule that overcomes the issue of low power, thus removing a potential barrier for their use in practice.
Author(s): Villar SS, Bowden J, Wason J
Publication type: Article
Publication status: Published
Journal: Statistical Science
Print publication date: 01/05/2015
Acceptance date: 01/01/1900
ISSN (print): 0883-4237
ISSN (electronic): 2168-8745
Publisher: Institute of Mathematical Statistics
Altmetrics provided by Altmetric