Abstract

In a rare life-threatening disease setting the number of patients in the trial is a high proportion of all patients with the condition (if not all of them). Further, this number is usually not enough to guarantee the required statistical power to detect a treatment effect of a meaningful size. In such a context, the idea of prioritizing patient benefit over hypothesis testing as the goal of the trial can lead to a trial design that produces useful information to guide treatment, even if it does not do so with the standard levels of statistical confidence. The idealised model to consider such an optimal design of a clinical trial is known as a classic multi-armed bandit problem with a finite patient horizon and a patient benefit objective function. Such a design maximises patient benefit by balancing the learning and earning goals as data accumulates and given the patient horizon. On the other hand, optimally solving such a model has a very high computational cost (many times prohibitive) and more importantly, a cumbersome implementation, even for populations as small as a hundred patients. Several computationally feasible heuristic rules to address this problem have been proposed over the last 40 years in the literature. In this article we study a novel heuristic approach to solve it based on the reformulation of the problem as a Restless bandit problem and the derivation of its corresponding Whittle index rule. Such rule was recently proposed in the context of a clinical trial in Villar et al (2015). We perform extensive computational studies to compare through both exact value calculations and simulated values the performance of this rule, other index rules and simpler heuristics previously proposed in the literature. Our results suggest that for the two and three-armed case and a patient horizon less or equal than a hundred patients, all index rules are a priori practically identical in terms of the expected proportion of success attained when all arms start with a uniform prior. However, we find that a posteriori, for specific values of the parameters of interest, the index policies outperform the simpler rules in every instance and specially so in the case of many arms and a larger, though still relatively small, total number of patients with the diseases. The very good performance of bandit rules in terms of patient benefit (i.e. expected number of successes and mean number of patients allocated to the best arm, if it exists) makes them very appealing in context of the challenge posed by drug development for rare life threatening diseases.

Highlights

  • Developing specific statistical learning methods for drug development for rare diseases is one of the most pressing modern clinical needs

  • We have extended the exact numerical results included in Berry [2] in Table 5 by including the results for the other rules considered in this paper: Myopic Index (MI), Gittins Index (GI)(d = 0.9) and Whittle Index (WI)

  • The results indicate that the MI rule and Feldman’s index (FI) are a priori practically equivalent in their performance ( MI appears to slightly outperform FI for N > 4)

Read more

Summary

INTRODUCTION

Developing specific statistical learning methods for drug development for rare diseases is one of the most pressing modern clinical needs. The resulting paradigm provides an alternative and feasible method to evaluate new therapies for rare and fatal diseases and to balance the need for experimentation with the desire to guide treatment selection toward the best treatment of a population Implementing such a dual learning–earning goal into a trial can be done in several ways. We explain how to derive near optimal heuristics for the finite-horizon Bernoulli Multiarmed Bandit problem based on a Restless bandit reformulation of the problem and on the Whittle and Gittins indices We illustrate how this approach manages to reduce the suboptimality gap (when compared with that of Feldman’s approach in Berry [2]), being computationally feasible and relatively simple to interpret and implement. We compare it with other heuristics and we perform various exact and simulated calculations in different contexts to evaluate when their application is more appropriate

Background
The WI
Other Index Strategies
Two-armed Trials
Multi-armed Trials
Understanding the WI Rule
DISCUSSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call