Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Sep 12;17(9):e0274272.
doi: 10.1371/journal.pone.0274272. eCollection 2022.

Some performance considerations when using multi-armed bandit algorithms in the presence of missing data

Affiliations

Some performance considerations when using multi-armed bandit algorithms in the presence of missing data

Xijin Chen et al. PLoS One. .

Abstract

When comparing the performance of multi-armed bandit algorithms, the potential impact of missing data is often overlooked. In practice, it also affects their implementation where the simplest approach to overcome this is to continue to sample according to the original bandit algorithm, ignoring missing outcomes. We investigate the impact on performance of this approach to deal with missing data for several bandit algorithms through an extensive simulation study assuming the rewards are missing at random. We focus on two-armed bandit algorithms with binary outcomes in the context of patient allocation for clinical trials with relatively small sample sizes. However, our results apply to other applications of bandit algorithms where missing data is expected to occur. We assess the resulting operating characteristics, including the expected reward. Different probabilities of missingness in both arms are considered. The key finding of our work is that when using the simplest strategy of ignoring missing data, the impact on the expected performance of multi-armed bandit strategies varies according to the way these strategies balance the exploration-exploitation trade-off. Algorithms that are geared towards exploration continue to assign samples to the arm with more missing responses (which being perceived as the arm with less observed information is deemed more appealing by the algorithm than it would otherwise be). In contrast, algorithms that are geared towards exploitation would rapidly assign a high value to samples from the arms with a current high mean irrespective of the level observations per arm. Furthermore, for algorithms focusing more on exploration, we illustrate that the problem of missing responses can be alleviated using a simple mean imputation approach.

PubMed Disclaimer

Conflict of interest statement

The authors have declared that no competing interests exist.

Figures

Fig 1
Fig 1. Allocation procedure of bandit algorithms.
Performance of different bandit algorithms over a single simulation (with p0 = 0.7, p1 = 0.7, and n = 200) under the null. The two colors reflect different arms that could be regarded as the same under the null.
Fig 2
Fig 2. Simulation results of E[p*] for TTS, CB and UCB under the null.
Expectations of 104 replications are taken for CB and UCB and 103 replications for TTS under different combinations of missingness probabilities, with p0 = p1 = 0.9 and n = 200.
Fig 3
Fig 3. Simulation results under the null.
Simulation results of E[p*] under the null for different missing data combinations. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm.
Fig 4
Fig 4. Simulation results under the alternative.
Simulation results of E[p*] under the alternative for different missing data combinations. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm.
Fig 5
Fig 5. Imputation results under the null.
Imputation results of E[p*] under the null for different missing data combinations with initial value p^k,0=0.5. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.
Fig 6
Fig 6. Imputation results under the alternative.
Imputation results of E[p*] under the alternative for different missing data combinations initial value p^k,0=0.5. Grey lines correspond to the case of equal missingness probability in both arms; Blue lines correspond to missingness in the control arm; Red lines correspond to missingness in the experimental arm. Solid lines correspond to the results without mean imputation, while the dashed lines correspond to the results with mean imputation.

References

    1. Chen IY, Joshi S, Ghassemi M, Ranganath R. Probabilistic machine learning for healthcare. Annual Review of Biomedical Data Science. 2020;4. - PubMed
    1. Bastani H, Bayati M. Online decision making with high-dimensional covariates. Operations Research. 2020;68(1):276–294. doi: 10.1287/opre.2019.1902 - DOI
    1. Scott I, Carter S, Coiera E. Clinician checklist for assessing suitability of machine learning applications in healthcare. BMJ Health & Care Informatics. 2021;28(1). doi: 10.1136/bmjhci-2020-100251 - DOI - PMC - PubMed
    1. Thompson WR. On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika. 1933;25(3/4):285–294. doi: 10.1093/biomet/25.3-4.285 - DOI
    1. Villar SS, Bowden J, Wason J. Multi-armed bandit models for the optimal design of clinical trials: benefits and challenges. Statistical science: a review journal of the Institute of Mathematical Statistics. 2015;30(2):199. doi: 10.1214/14-STS504 - DOI - PMC - PubMed

Publication types