---
title: "Bayesian Response-Adaptive Design Analysis with BRADA"
author: "Riko Kelter"
output:
rmarkdown::html_vignette:
#css: style.css
toc: FALSE
vignette: >
%\VignetteIndexEntry{Brada4}
%\VignetteEngine{knitr::rmarkdown}
%\VignetteEncoding{UTF-8}
---
```{r, include = FALSE}
old <- options()
options(rmarkdown.html_vignette.check_title = FALSE)
# chunk option dev="png" is the default raster graphics format for HTML output
knitr::opts_chunk$set(dev="png")
options(mc.cores=2)
```
# Introduction
The getting started vignette illustrated the basic features of the `brada` package. In this vignette, we illustrate how to monitor a running trial with the `brada` package.
Note that there are more vignettes which illustrate
- how to apply and calibrate the predictive evidence value design with the `brada` package. This vignette is hosted at the
Open Science Foundation.
- how to monitor a running clinical trial with a binary endpoint by means of the `brada` package
# Monitoring a trial
To apply the package, first, load the \code{brada} package:
```{r, warning=FALSE, message=FALSE, class.source="chunks"}
library(brada)
```
Monitoring a trial with the `brada` package is straightforward through the `monitor` function. Suppose we have analyzed and calibrated a design according to our requirements, and end up with the following design:
```{r, class.source="chunks"}
design = brada(Nmax = 30, batchsize = 5, nInit = 10,
p_true = 0.4, p0 = 0.4, p1 = 0.4,
nsim = 100,
theta_T = 0.90, theta_L = 0.1, theta_U = 1,
method = "PP",
cores = 2)
```
Now, suppose the trial is performed and the first ten patients show the response pattern $(0,1,0,0,0,0,0,1,0,0)$, where $1$ encodes a response and $0$ no response. Thus, there are $2$ responses out of `nInit=10` observations. To check whether the trial can be stopped for futility or efficacy based on `theta_L=0.1` and `theta_U=1`, we run the `monitor` function as follows:
```{r, class.source="chunks", class.output="output"}
monitor(design, obs = c(0,1,0,0,0,0,0,1,0,0))
```
Thus, the results indicate that we should stop for efficacy. This is intuitively in agreement with the notion that $2$ responses out of $10$ observations are quite unlikely if $H_1:p>0.4$ would hold.
Note that it is not important which value the `p_true` or `nsim` arguments had in the `brada` call which returned the object `design`. We could also have simulated data under `p_true=0.2` and `nsim=3000` or some other values, the monitor function only takes the `brada` object and applies the design specified in the `method` argument of the object, in this case, the predictive probability design. All necessary arguments are identified by the `monitor` function automatically. The predictive evidence value design can be monitored analogue, for details on the design and its calibration see the
Open Science Foundation.
***
```{r, include = FALSE}
options(old)
```
## References
Berry, S. M. (2011). Bayesian Adaptive Methods for Clinical Trials. CRC Press.
Kelter, R. (2022). The Evidence Interval and the Bayesian Evidence Value - On a unified theory for Bayesian hypothesis testing and interval estimation. British Journal of Mathematical and Statistical Psychology (2022). https://doi.org/10.1111/bmsp.12267
Kelter, R. (2021b). Bayesian Hodges-Lehmann tests for statistical equivalence in the two-sample setting: Power analysis, type I error rates and equivalence boundary selection in biomedical research. BMC Medical Research Methodology, 21(171). https://doi.org/10.1186/s12874-021-01341-7
Kelter, R. (2021a). fbst: An R package for the Full Bayesian Significance Test for testing a sharp null hypothesis against its alternative via the e-value. Behav Res (2021). https://doi.org/10.3758/s13428-021-01613-6
Kelter, R. (2020). Analysis of Bayesian posterior significance and effect size indices for the two-sample t-test to support reproducible medical research. BMC Medical Research Methodology, 20(88). https://doi.org/https://doi.org/10.1186/s12874-020-00968-2
Morris, T. P., White, I. R., & Crowther, M. J. (2019). Using simulation studies to evaluate statistical methods. Statistics in Medicine, 38(11), 2074–2102. https://doi.org/10.1002/SIM.8086
Pereira, C. A. d. B., & Stern, J. M. (2020). The e-value: a fully Bayesian significance measure for precise statistical hypotheses and its research program. São Paulo Journal of Mathematical Sciences, 1–19. https://doi.org/10.1007/s40863-020-00171-7
Rosner, G. L. (2021). Bayesian Thinking in Biostatistics. Chapman and Hall/CRC.
Rosner, G. L. (2020). Bayesian Adaptive Designs in Drug Development. In E. Lesaffre, G. Baio, & B. Boulanger (Eds.), Bayesian Methods in Pharmaceutical Research (pp. 161–184). CRC Press.
Rouder, Jeffrey N., Paul L. Speckman, Dongchu Sun, Richard D. Morey, and Geoffrey Iverson. 2009. “Bayesian t tests for accepting and rejecting the null hypothesis.” Psychonomic Bulletin and Review 16 (2): 225–37. https://doi.org/10.3758/PBR.16.2.225