Five preventative HIV vaccine efficacy trials have been conducted over the last 12 years, all of which evaluated vaccine efficacy (VE) to prevent HIV infection for a single vaccine regimen versus placebo. Now that one of these trials has supported partial VE of a prime-boost vaccine regimen, there is interest in conducting efficacy trials that simultaneously evaluate multiple prime-boost vaccine regimens against a shared placebo group in the same geographic region, for accelerating the pace of vaccine development. This article proposes such a design, which has main objectives (1) to evaluate VE of each regimen versus placebo against HIV exposures occurring near the time of the immunizations; (2) to evaluate durability of VE for each vaccine regimen showing reliable evidence for positive VE; (3) to expeditiously evaluate the immune correlates of protection if any vaccine regimen shows reliable evidence for positive VE; and (4) to compare VE among the vaccine regimens. The design uses sequential monitoring for the events of vaccine harm, non-efficacy, and high efficacy, selected to weed out poor vaccines as rapidly as possible while guarding against prematurely weeding out a vaccine that does not confer efficacy until most of the immunizations are received. The evaluation of the design shows that testing multiple vaccine regimens is important for providing a well-powered assessment of the correlation of vaccine-induced immune responses with HIV infection, and is critically important for providing a reasonably powered assessment of the value of identified correlates as surrogate endpoints for HIV infection.
A flexible sample size computation is desired for a binomial outcome consisting of repeated binary measures with autocorrelation over time. This type of outcome is common in viral shedding studies, in which each individual's outcome is a proportion: the number of samples on which virus is detected out of number of samples assessed. Autocorrelation between proximal samples occurs in some conditions such as herpes infection, in which reactivation is episodic. We determine a sample size computation that accounts for: (1) participant-level differences in outcome frequency, (2) autocorrelation in time between samples, and (3) varying number of samples per participant. In addition, we develop a computation appropriate for crossover designs that accounts for the dependence of the investigational treatment effect on the pretreatment detection frequency. The computations are validated through comparison with real and simulated data, and sensitivity to misspecification of parameter values is examined graphically.