Accounting for treatment by center interaction in sample size determinations and the use of surrogate outcomes in the pessary for the prevention of preterm birth trial: a simulation study
 Andrew R. Willan^{1}Email author
Received: 30 June 2015
Accepted: 8 June 2016
Published: 5 July 2016
Abstract
Background
The Pessary for the Prevention of Preterm Birth Study (PS3) is an international, multicenter, randomized clinical trial designed to examine the effectiveness of the Arabin pessary in preventing preterm birth in pregnant women with a short cervix. During the design of the study two methodological issues regarding power and sample size were raised. Since treatment in the Standard Arm will vary between centers, it is anticipated that so too will the probability of preterm birth in that arm. This will likely result in a treatment by center interaction, and the issue of how this will affect the sample size requirements was raised. The sample size requirements to examine the effect of the pessary on the baby’s clinical outcome was prohibitively high, so the second issue is how best to examine the effect on clinical outcome. The approaches taken to address these issues are presented.
Results
Simulation and sensitivity analysis were used to address the sample size issue. The probability of preterm birth in the Standard Arm was assumed to vary between centers following a Beta distribution with a mean of 0.3 and a coefficient of variation of 0.3. To address the second issue a Bayesian decision model is proposed that combines the information regarding the betweentreatment difference in the probability of preterm birth from PS3 with the data from the Multiple Courses of Antenatal Corticosteroids for Preterm Birth Study that relate preterm birth and perinatal mortality/morbidity. The approach provides a betweentreatment comparison with respect to the probability of a bad clinical outcome. The performance of the approach was assessed using simulation and sensitivity analysis.
Accounting for a possible treatment by center interaction increased the sample size from 540 to 700 patients per arm for the base case. The sample size requirements increase with the coefficient of variation and decrease with the number of centers. Under the same assumptions used for determining the sample size requirements, the simulated mean probability that pessary reduces the risk of perinatal mortality/morbidity is 0.98. The simulated mean decreased with coefficient of variation and increased with the number of clinical sites.
Conclusion
Employing simulation and sensitivity analysis is a useful approach for determining sample size requirements while accounting for the additional uncertainty due to a treatment by center interaction. Using a surrogate outcome in conjunction with a Bayesian decision model is an efficient way to compare important clinical outcomes in a randomized clinical trial in situations where the direct approach requires a prohibitively high sample size.
Keywords
Background
The Pessary for the Prevention of Preterm Birth Study (P3S) is an international, multicenter, randomized clinical trial (RCT) to be conducted in 80 clinical centers by the Centre for Mother, Infant, and Child Research at the Sunnybrook Research Institute in Toronto, and is designed to answer the following primary question: Does the introduction of the Arabin pessary reduce the probability of preterm birth in pregnant women with a short cervix? The Arabin pessary is a simple, noninvasive therapeutic option for preventing preterm birth. Closing the cervix with a silicone ring (that is removed at term gestation) is a simple, relatively painless procedure that can be performed in an office setting and does not require surgery, hospitalization or anesthesia. To be eligible for randomization, women must present with a singleton pregnancy between 14 and 24 weeks gestation, and be identified by ultrasound to have a cervical length of between 10 and 24 mm. Randomization will be by randomsized blocks, stratified by center and gestational age. The primary outcome is preterm birth (i.e., birth at or prior to 33^{+6/7} weeks’ gestation). Since there is no agreedupon standard of care, women will be randomized between (1) the current management of short cervix as determined by the center’s policy or at the attending physician’s discretion (Standard) and (2) Arabin pessary together with any additional care at the attending physician’s discretion (Treatment).
Two issues regarding sample size and power arose in the design and proposed analysis of P3S. The investigators assume that since treatment in the Standard Arm will vary between clinical centers, so might the probability of preterm birth in that arm. However, since all patients in the Treatment Arm are allocated to pessary, the investigators assume that due to a floor effect, if treatment does reduce the probability of preterm birth, the additional effect of any other interventions will be minimal and the probability of preterm birth in the Treatment Arm will not vary substantially between centers. As a result the treatment effect will vary between centers. This will increase the uncertainty with which the overall treatment effect is estimated and require a larger sample size to maintain type I and II error probabilities. To account for this in the analysis a randomeffects model that allows the treatment effect to vary by clinical center (i.e, treatment by center interaction) should be used. The above assumption for the Treatment Arm is not required for the analysis using a randomeffects model to be valid, although relaxing it could increase the uncertainty further. A bigger issue, and the one addressed in this paper, is what effect will a treatment by center interaction have on the sample size requirements.
Many authors have discussed the issue of center effect in RCTs [1–13]. Investigators who suspect that the treatment effect varies by clinical center (i.e., the existence of treatment by center interaction) have three strategies to choose from. The first is to ignore the issue. The second is to adopt a model with a fixed effect for center. The third is to adopt a model with random effect for center. The disadvantage of the “ignore it” strategy is that if a treatment by center interaction does exist then the uncertainty regarding the estimated treatment effect will be underestimated, leading to an unknown inflation of the type I error probability. The disadvantage of a fixedeffects model is that the inference is restricted to the centers in the trial. As argued by Feaster et al. [13] and others, inference from a fixedeffects model is more appropriate for an early phase III trial with a small number of centers, the positive results of which should lead to a more pragmatic trial with more centers and employing a randomeffects model. Furthermore, for a fixedeffects model the inference for each center is based solely on their own data since there is no parameter in the model that represents an overall treatment effect. On the other hand, the randomeffects model does contain a parameter for overall treatment effect and adjusts the uncertainty to account for the betweencenter variance in treatment effect. In addition, shrinkagetype estimation, which uses the data from all centers, can be used to provide centerspecific inference [14]. One must assume that the centers in the trial are a random sample of centers to which inference is to be made. Although this may seem a strong assumption, it is reasonable to expect that most RCTs are done to make inference to a population broader than just the centers in the trial, and without this assumption, valid inference beyond the centers in the trial is problematic. One disadvantage of the randomeffects model is that it is generally accepted that a reasonably large number of centers are required to provide a robust estimate of the betweencenter variance [12]. In addition, since the betweencenter variance is usually unknown in advance, sample size determinations that account for it can be challenging. Raudenbush and Liu [8] provide power and sample size solutions adjusting for treatment by center interaction when the outcome variable is continuous. Less well developed are methods for adjusting sample size requirements for a binary outcome variable.
The second issue is how best to answer the question: Does the introduction of the pessary reduce the probability of a bad clinical outcome for the babies? The sample size to answer this question directly by observing the clinical outcome of the infants in P3S is prohibitively high. As an alternative the investigators propose a Bayesian approach, combining the information regarding the betweentreatment difference in the probability of preterm birth from P3S with the data from the Multiple Courses of Antenatal Corticosteroids for Preterm Birth Study (MACS) [15] that relate preterm birth and perinatal mortality/morbidity. This analysis will provide an estimate of the reduction in the probability of the clinical outcome from the introduction of the pessary. The question of concern to the investigators is: Will the proposed analysis answer the question regarding clinical outcome with sufficient certainty?
Many authors [16–30] have discussed the issues related to using surrogate outcomes in RCTs. Ellenberg and Hamilton [16] discuss the use of a tumor marker as a surrogate for tumor response and emphasize the value of identifying valid surrogates for clinical outcomes that have insufficient sensitivity and specificity. Although Wittes, Lakatos and Probstfield [17] conclude that the use of valid surrogate outcomes can reduce sample size and shorten trial durations, they point out several problems that can arise, especially regarding informative censoring. A criterion proposed by Prentice [18], and supported by others [27, 28], for a valid surrogate is stated as follows: a variable for which a valid test of the null hypothesis of no treatment effect is also a valid test of the corresponding hypothesis based on the clinical outcome. The author discusses the implementation of the criterion and applies it to examples given in [16, 17, 21]. Freedman, Graubard and Schatzkin [19] take issue with the implementation of the Prentice criterion, arguing that unless the observed treatment effect exceeds its standard error by a factor of four, the procedure will usually only lead to a weak validation of the surrogate outcome. Fleming et al. [20] discuss the implementation of the Prentice criterion in the context of cancer and AIDS trials. Day and Duffy [22] demonstrate how, in the British Breast Screening Frequency Trial, the use of surrogate outcomes can dramatically reduce the variance of the estimate of treatment effect and shorten the duration of the trial. Fleming and DeMets [23] argue that the Prentice criterion for a valid surrogate is often not meet and use examples to support their argument. They point out that numerous pathways might affect the clinical outcome, not all which will be mediated through the surrogate. The authors also raise the issue that the interventions might affect the clinical outcome in ways other than through the surrogate. Daniels and Hughes [24] proposed a Bayesian metaanalysis approach for examining the association between the betweentreatment difference in the surrogate and the betweentreatment difference in the clinical outcome. Buyse and Molenberghs [25] take issue with the Prentice procedure for validating surrogate outcomes. They propose an alternative method for validation and provide an illustration for the situation where the surrogate and clinical outcome are either both normal or both binary. These methods are extended by Molenberghs, Geys and Buyse [29] for the situation where the surrogate is binary and the clinical outcome is continuous, and vice versa. Begg and Leung [26] argue that conceptual strategy for the Prentice procedure is flawed and propose an alternative structure for validating surrogate outcomes. Baker, Izmirlian and Kipnis [30] address the issues raised by Day and Duffy [22] and Begg and Leung [26].
The issue of sample size inflation is addressed in the next section where simulations are used to determine the required sample size in the presence of treatment by center interaction. In the section following that, a Bayesian model is proposed for comparing treatment arms with respect to perinatal mortality/morbidity by combining the information regarding the betweentreatment arm difference in the probability of preterm birth from P3S with the information relating preterm birth and perinatal mortality/morbidity from MACS. The performance of the proposed analysis is assessed using simulations. Although the two issues are seemingly separate, they are related since both are associated with the same trial, and both deal with power and sample size questions. The first issue deals with loss of power due to betweencenter variation in treatment effect, while the second deals with lack of power because the clinical outcome is so rare.
Methods
The P3S is currently under review for funding from the Canadian Institutes for Health Research. If funding is received then the trial will be registered and the appropriate ethics approval will be sought by the coordinating center and all the clinical centers. If P3S is funded then informed consent will be sought for each participant enrolled.
Between center treatmenteffect variation and sample size
Simulations were used to determine sample size under the assumption that the treatment effect, if it exists, will vary between centers. To model the assumption for the Standard Arm, the preterm probability was assumed to vary between centers following a Beta distribution, with parameters a and b. That is, π_{ Si } ~ Beta(a, b), where π_{ Si } is the preterm probability on the Standard Arm in the i ^{th} clinical center. The Beta distribution was chosen because it yields values between 0 and 1, and as illustrated below, the values for a and b can be chosen to provide the desired mean and betweencenter variance. Furthermore, the Beta distribution is used in Bayesian analysis for proportions since it is conjugate to binomial sampling. That is, if the prior distribution for a proportion is Beta, then with binomial sampling, the posterior distribution will also be Beta. The parameters a and b were chosen to provide (1) a mean 0.3 that reflects the current preterm probability as determined by available evidence [15, 31] and (2) as the base case, a coefficient of variation (standard deviation/mean) of 0.3 to reflect a reasonable amount of betweencenter variation. In the results section solutions for values of the coefficient of variation ranging from 0.1 to 0.4 are given for comparison.
Since π_{ Si } ~ Beta(a, b),
\( E\left({\pi}_{Si}\right)=\frac{a}{a+b} \) and \( V\left({\pi}_{Si}\right)=\frac{ab}{{\left(a+b\right)}^2\left(a+b+1\right)}, \)
where E and V are the expectation and variance functions, respectively. Therefore, for P3S:
\( E\left({\pi}_{Si}\right)=\frac{a}{a+b}=0.3 \) and \( CV\left({\pi}_{Si}\right)=\frac{\sqrt{V\left({\pi}_{Si}\right)}}{E\left({\pi}_{Si}\right)}=\sqrt{\frac{ab}{{\left(a+b\right)}^2\left(a+b+1\right)}}/\frac{a}{a+b}=0.3, \)
Thus, a = 7.478 and b = 17.45, and for these parameter values there is a 95 % probability that a center’s preterm probability lies between 0.16 and 0.46, representing a reasonably wide range. (This is the variation in the true probabilities, not the variation in the observed proportions, which also include sampling variation.) The preterm probability in the Treatment Arm, denoted π _{ T }, for which the investigators required sufficient power was set at 0.225 and fixed across clinical centers, yielding under the alternative hypothesis, a mean relative risk reduction of 0.25 and a mean number needed to treat of about 13. By assuming a CV(π _{ Si }) = 0.3, 79 % of the centers in the Standard Arm have a probability of preterm birth greater than 0.225.

A chosen sample size of patients was allocated randomly between the 80 centers with equal probability among centers with the constraint that there had to be at least six patients in each center

Because randomization is to be stratifiedblocked by center, within each center, half the patients were assigned to the Standard Arm and half to the Treatment Arm

Those assigned to the Treatment Arm had their outcome determined by a Bernoulli distribution with probability 0.225

Those assigned to Standard in i ^{th} center had their outcome determined by a Bernoulli distribution with a probability that was sampled from Beta(7.478, 17.45)

The resulting data were analyzed using a randomeffects model, allowing treatment effect to vary between centers, using risk difference as the measure of treatment effect

The simulation was replicated 10,000 times, and the proportion of replicates in which statistical significance (twosided, 5 % level) was achieved was recorded

If the proportion of replicates which achieved statistical significance was less (greater) than 80 %, the sample size was increased (decreased) and the procedure was repeated
Assessing the treatment effect on clinical outcome
Required sample size per arm as a function of the coefficient of variation and the number of clinical centers
Coefficient of variation  Probability that π_{ Si } >0.225  Number of clinical centers  

20  40  60  80  100  
0  1  610  580  560  540  520 
0.1  0.99  660  600  580  560  530 
0.2  0.90  880  700  650  600  560 
0.3  0.79  1840  1000  800  700  640 
0.4  0.71  50,000  1700  1200  900  800 
In this analysis the parameter of interest is Δ = p _{ T } − p _{ S }, where p _{ T } and p _{ S } are the probabilities of perinatal mortality/morbidity in babies whose mothers are randomized to the Treatment and Standard Arms, respectively. Thus, Δ is the betweentreatment arm difference in the probability of perinatal mortality/morbidity.

π _{ ptbT } is the probability of preterm birth for mothers randomized to the Treatment Arm

τ _{ pmmptb } is the probability of perinatal mortality/morbidity in a baby who is born preterm

π _{ tbT } = 1 − π _{ ptbT } is the probability of term birth for mothers randomized to the Treatment Arm

τ _{ pmmtb } is the probability of perinatal mortality/morbidity in a baby who is born at term

π _{ ptbS } is the probability of preterm birth for mothers randomized to the Standard Arm

π _{ tbS } = 1 − π _{ ptbS } is the probability of term birth for mothers randomized to the Standard Arm

τ _{ pmmptb } ~ Beta(y _{ ptb } + 1, n _{ ptb } − y _{ ptb } + 1), where from MACS n _{ ptb } is number of preterm births of whom y _{ ptb } experienced the clinical outcome. Recalling that n _{ ptb } = 530 and y _{ ptb } = 166, then τ _{ pmmptb } ~ Beta(167,365)

τ _{ pmmtb } ~ Beta(y _{ tb } + 1, n _{ tb } − y _{ tb } + 1), where from MACS n _{ tb } is number of term births of whom y _{ tb } experienced the clinical outcome. Recalling that n _{ tb } = 933 and y _{ tb } = 5, then τ _{ pmmtb } ~ Beta(6,929)
The distribution for (π _{ ptbT } − π _{ ptbS }) will come from a randomeffects regression model for the outcome of preterm birth using the P3S data. The model will include a fixed effect for Treatment Arm and a random effect for center, allowing for the treatment effect to vary by center. By assuming an uninformative normal prior, the posterior distribution for (π _{ ptbT } − π _{ ptbS }) will be Normal(μ, ν), where μ is the estimate of the coefficient for treatment group and ν is the associated variance. Using the above distributions, a distribution for Δ will be sampled. From the sampled distribution, the probability that Δ is less than 0 (i.e., the probability that pessary reduces the risk of the perinatal mortality/morbidity) will be determined. In addition, 95 % credible intervals for Δ will be determined. By using uninformative priors the inference regarding Δ will be conditional on using the data from P3S and MACS only.

The trial of 700 per arm was simulated 10,000 times using the same assumptions for deriving the sample size given above

For each simulation:

o The mean and variance of the estimator of π _{ ptbT } − π _{ ptbS } were determined

o 10,000 samples were taken from a normal distribution with that mean and variance

o 10,000 samples were taken from the Beta distributions for τ _{ pmmptb } and τ _{ pmmtb }, as given above

o For each sample Δ was determined as (π _{ ptbT } − π _{ ptbS }) × (τ _{ pmmptb } − τ _{ pmmtb }) from the sampled values

o The proportion of samples for which Δ was negative was recorded

Results
Betweencenter treatmenteffect variation and sample size
For the basecase parameters listed in the next sentence the required sample size is 700 patients per arm. E(π _{ Si }) = 0.3; CV = 0.3; π _{ T } = 0.225; number of centers = 80; type I error probability = 0.05 (twosided); and type II error probability = 0.2. To examine robustness with respect to the CV and number of centers, the associated sample size requirements are given in Table 1. The sample size required increases with the coefficient of variation, since larger CVs result in greater uncertainty, and decreases with the number of centers, as noted by Feaster et al. [13] and Raudenbush and Liu [8]. The effect of the number of centers is explained in the Appendix. For 80 centers and a CV of 0, the required sample size was 540 per arm. Therefore, the increase in total required sample size to account for the betweencenters variance is 320.
For this example the assumption has been made that there is no betweencenter variation in the probability of preterm birth in the Treatment Arm. This assumption was made on the assessment that if the pessary does decrease the probability of preterm birth by 0.075 (i.e., the alternative hypothesis is true), a floor effect is expected, meaning that any additional interventions are unlikely to decrease the probability much further, and so even if the additional interventions vary between center, it is unlikely to result in betweencenter variation in the probability in preterm birth. To examine the robustness of the assumption in this example consider the following.
From the Appendix, under the assumption that each center randomizes n subjects, the variance of the estimated treatment effect is given by \( V=\frac{2\left({\sigma}_S^2+{\sigma}_T^2\right)}{N}+\frac{v_S+{v}_T}{K} \),
where K is the number of centers;
N is the total sample size, i.e., N = n × K;
σ _{ S } ^{2} = π _{ S }(1 − π _{ S }) is the betweenpatient variance in the Standard Arm;
σ _{ T } ^{2} = π _{ T }(1 − π _{ T }) is the betweenpatient variance in the Treatment Arm;
v _{ S } and v _{ T } are the betweencenter variances in the Standard and Treatment Arms, respectively, where v _{ i } = (p _{ i } × CV _{ i })^{2}, for i = S, T.
Under the alternative hypothesis, π _{ S } = 0.3 and π _{ T } = 0.225. In this example K = 80 and N = 1400. Thus, if there is no betweencenter variation in either arm (i.e., CV _{ S } = CV _{ T } = 0, and therefore v _{ S } = v _{ T } = 0), then V = 5.49 × 10^{− 4}. Based on our assumption that there is betweencenter variation in the Standard Arm but not the Treatment Arm (i.e., CV _{ S } = 0.3 and CV _{ T } = 0), V = 6.50 × 10^{− 4}, representing a 18.4 % increase. Now if we allow some residual betweencenter variation in the Treatment Arm, say CV _{ T } = 0.1, then V = 6.57 × 10^{− 4}, which is a mere 0.97 % increase. The small effect of the betweencenter variability in the Treatment Arm relates to the fact that the betweencenter variance, v _{ T }, is the square of the product of π _{ T } and CV _{ T }, both of which are smaller than the corresponding quantities in the Standard Arm. Even a CV _{ T } of 0.15 increases the variance by only 2.19 %. The conclusion is that in this example the assumption of no betweencenter variation is reasonably robust, especially considering that 0.3 was selected as an upper bound for CV _{ S }.
Assessing the treatment effect on clinical outcome
The average and (5th percentile) of the proportion of negative Δ’s as a function of the coefficient of variation and the number of clinical centers
Coefficient of variation  Probability that π_{ Si } > 0.225  Number of clinical centers  

20  40  60  80  100  
0  1  0.98 (0.88)  0.99 (0.94)  0.98 (0.93)  0.99 (0.93)  0.99 (0.94) 
0.1  0.99  0.98 (0.87)  0.98 (0.92)  0.99 (0.92)  0.98 (0.92)  0.99 (0.94) 
0.2  0.90  0.97 (0.88)  0.97 (0.85)  0.98 (0.90)  0.98 (0.90)  0.99 (0.92) 
0.3  0.79  0.97 (0.87)  0.96 (0.79)  0.97 (0.85)  0.98 (0.87)  0.98 (0.90) 
0.4  0.71  0.97 (0.86)  0.94 (0.73)  0.96 (0.79)  0.97 (0.85)  0.98 (0.89) 
The SAS programs used to generate Tables 1 and 2 can be found at www.andywillan.com/downloads.
Discussion
In a multicenter RCT a treatment by center interaction may exist for any number of reasons. It can exist because of differences in patient referral patterns, clinician skills, supportive care, implementation of the Treatment Arm, clinical evaluation or, as in P3S, with variations in the Standard Arm. The risk of a treatment by center interaction is often ignored [13]. This can lead to inflated type I error probabilities due to underestimating uncertainty, and inflated type II error probabilities due to inadequate sample sizes. It is well known that randomeffect models provide valid estimates of variances of the estimated treatment effects, and the models can be adopted after a trial is completed if a treatment by center interaction is observed. However, sample sizes cannot be adjusted if an interaction is observed once the trial has been completed. In which case the inflation of type II error probability (i.e., reduction in power) has to be accepted. Modeling the effect that a treatment by center interaction has on sample size is challenging since it depends on many factors, many of which might be unknown at the time of trial planning. These include the number of centers, the degree to which the treatment effect varies between centers and the distribution of patients between centers. Other factors include the nature of the outcome variable and the metric used for the treatment comparison. Employing simulation and sensitivity analysis, as illustrated on the P3S above, is a fruitful, albeit timeconsuming, approach. The simulation process depends on the specific causes of the betweencenter variability and other factors as discussed above. Consequently, many applications require a custom programming code.
The Bayesian decision model approach proposed in the previous section uses a surrogate outcome from an RCT in combination with data from other sources to make treatment comparison with respect to more clinical outcomes. To meet the Prentice criterion for a valid surrogate it must be on the causal pathway between the intervention (Treatment versus Standard) and the clinical outcome, and the entire effect of the intervention on the clinical outcome must be through its effect on the surrogate. Investigators are well advised to consider these issues carefully. Regarding P3S, it is firmly accepted that preterm birth is a risk factor for bad clinical outcome and that interventions that reduce the probability of preterm birth will reduce the risk such an outcome [33–35].
The use of a surrogate outcome in a Bayesian decision model is particularly useful when the clinical outcomes are rare or occur well in to the future. For binary outcomes the required sample size is dominated by the one over the square of the risk difference; the smaller the risk difference, the greater the required sample size. Assuming constant relative risks, the ratio of the risk difference for a surrogate to that for the clinical outcome equals the ratio of the corresponding risks of the outcomes in the Standard Arm. For example, in the P3S trial the risks in the Standard Arm for surrogate and clinical outcome are 0.3 and 0.01, respectively, and consequently one over the square of the risk difference for the clinical outcome is 900 times that for the surrogate. Even when the surrogate and clinical risks do not differ greatly, the decision model approach has distinct advantages for clinical outcomes that occur much later in time, such as survival. If longterm survival is the clinical outcome of interest, it might be the case that the research question will no longer be relevant after the required number of events have been observed since interest may have shifted to new interventions.
The requirement that the entire effect of the intervention (Treatment versus Standard) on the clinical outcome must be through its effect on the surrogate can be easily overlooked. For example, if relapse is used as a surrogate for survival, care must be taken that treatment does not negatively affect survival through some longterm adverse outcome. The other limiting factors are the relevance and strength of the evidence relating the surrogate to the clinical outcome. Critical appraisal of the evidence, including appropriateness of the sample and the methodology used is crucial for establishing the credibility of any application of a decision model.
The proposed Bayesian approach mirrors the decision analytic approaches used in health economics for assessing the costeffectiveness of health care interventions [36]. A number of issues come to the fore concerning the quality of such analyses [37], and investigators must be prepared to defend the assumptions made in building the model. For P3S the model itself was very simple. Nonetheless, issues regarding the appropriateness and exhaustiveness of the “other sources” need to be addressed. These issues aside, the approach provides a very efficient means of addressing important clinical questions. For example, the sample size for P3S to compare Treatment Arms with respect to the clinical outcome total 75,000, more than 50 times the planned recruitment of 1400.
Conclusion
Simulation techniques were used to address two issues relating to sample size and power for the pessary trial. Modeling the effect that a treatment by center interaction has on sample size is challenging because it depends on many factors, most of which might be unknown at the time of trial planning. Nonetheless, as illustrated above employing simulation and sensitivity analysis is a useful approach for determining sample size requirements while accounting for the additional uncertainty due to a treatment by center interaction.
It is often the case that an important clinical outcome is very rare or occurs many years after randomization. Using a surrogate outcome in conjunction with a Bayesian decision model is an efficient way to compare important clinical outcomes in a randomized clinical trial in situations where the direct approach requires a prohibitively high sample size or impossibly long followup. For particular examples simulations can be used to determine the power properties of employing such an approach.
Abbreviations
CV, coefficient of variation; MACS, The Multiple Courses of Antenatal Corticosteroids for Preterm Birth Study; P3S, The Pessary for the Prevention of Preterm Birth Study; RCT, randomized clinical trial
Declarations
Acknowledgements
The following members of the steering committee contributed to the design of P3S: Drs. Jon Barrett, Elizabeth Asztalos, Kelley Murphy, Tony Armson, KS Joseph and Amiram Gafni. The author is a coinvestigator on projects funded by the Canadian Institutes for Health Research.
Competing interests
The author declares that he has no competing interests.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
Authors’ Affiliations
References
 Fleiss JL. Analysis of data from multiclinic trials. Control Clin Trials. 1986;7(4):267–75.View ArticlePubMedGoogle Scholar
 Kallen A. Treatmentbycenter interaction: what is the issue? Drug Inf J. 1997;31(3):927–36.Google Scholar
 Senn S. Some controversies in planning and analysing multicentre trials. Stat Med. 1998;17(15–16):1753–65.View ArticlePubMedGoogle Scholar
 Jones B, Teather D, Wang J, Lewis J. A comparison of various estimators of a treatment difference for a multicentre clinical trial. Stat Med. 1998;17(15–16):1767–77.View ArticlePubMedGoogle Scholar
 Gould AL. Multicentre trial analysis revisited. Stat Med. 1998;17(15–16):1779–97.View ArticlePubMedGoogle Scholar
 Agresti A, Hartzel J. Strategies for comparing treatments on a binary response with multi‐centre data. Stat Med. 2000;19(8):1115–39.View ArticlePubMedGoogle Scholar
 Lin Z. The number of centers in a multicenter clinical study: effects on statistical power. Drug Inf J. 2000;34(2):379–86.Google Scholar
 Raudenbush SW, Liu X. Statistical power and optimal design for multisite randomized trials. Psychol Methods. 2000;5(2):199–213.View ArticlePubMedGoogle Scholar
 Localio AR, Berlin JA, Ten Have TR, Kimmel SE. Adjustments for center in multicenter studies: an overview. Ann Intern Med. 2001;135(2):112–23.View ArticlePubMedGoogle Scholar
 Moerbeek M, van Breukelen GJ, Berger MP. A comparison between traditional methods and multilevel regression for the analysis of multicenter intervention studies. J Clin Epidemiol. 2003;56(4):341–50.View ArticlePubMedGoogle Scholar
 Dragalin V, Fedorov V. Design of multicentre trials with binary response. Stat Med. 2006;25(16):2701–19.View ArticlePubMedGoogle Scholar
 Nixon RM, Thompson SG. Methods for incorporating covariate adjustment, subgroup analysis and betweencentre differences into costeffectiveness evaluations. Health Econ. 2005;14(12):1217–29.View ArticlePubMedGoogle Scholar
 Feaster DJ, MikulichGilbertson S, Brincks AM. Modeling site effects in the design and analysis of multisite trials. Am J Drug Alcohol Ab. 2011;37(5):383–91.View ArticleGoogle Scholar
 Morris C. Parametric empirical Bayes inference: theory and applications. J Am Stat Assoc. 1983;78(1):47–55.View ArticleGoogle Scholar
 Murphy KE, Hannah ME, Willan AR, et al. Multiple courses of antenatal corticosteroids for preterm birth study (MACS): a randomised controlled trial. Lancet. 2008;372(9656):2143–51.View ArticlePubMedGoogle Scholar
 Ellenberg SS, Hamilton JM. Surrogate endpoints in clinical trials: cancer. Stat Med. 1989;8(4):405–13.View ArticlePubMedGoogle Scholar
 Wittes J, Lakatos E, Probstfield J. Surrogate endpoints in clinical trials: cardiovascular diseases. Stat Med. 1989;8(4):415–25.View ArticlePubMedGoogle Scholar
 Prentice RL. Surrogate endpoints in clinical trials: definition and operational criteria. Stat Med. 1998;8(4):431–40.View ArticleGoogle Scholar
 Freedman LS, Graubard BI, Schatzkin A. Statistical validation of intermediate endpoints for chronic diseases. Stat Med. 1992;11(2):167–78.View ArticlePubMedGoogle Scholar
 Fleming TR, Prentice RL, Pepe MS, Glidden D. Surrogate and auxiliary endpoints in clinical trials, with potential applications in cancer and AIDS research. Stat Med. 1994;13(9):955–68.View ArticlePubMedGoogle Scholar
 Hillis A, Seigel D. Surrogate endpoints in clinical trials: ophthalmologic disorders. Stat Med. 1989;8(4):427–30.View ArticlePubMedGoogle Scholar
 Day NE, Duffy SW. Trial design based on surrogate end points—application to comparison of different breast screening frequencies. 1996. J R Stat Soc Ser A. 1996;159(1):49–60.View ArticleGoogle Scholar
 Fleming TR, DeMets DL. Surrogate end points in clinical trials: are we being misled? Ann Intern Med. 1996;125(7):605–13.View ArticlePubMedGoogle Scholar
 Daniels MJ, Hughes MD. Metaanalysis for the evaluation of potential surrogate markers. Stat Med. 1997;16(17):1965–82.View ArticlePubMedGoogle Scholar
 Buyse M, Molenberghs G. Criteria for the validation of surrogate endpoints in randomized experiments. Biometrics. 1998;54(3):1014–29.View ArticlePubMedGoogle Scholar
 Begg CB, Leung DHY. On the use of surrogate endpoints in randomized trials. J R Stat Soc Ser AG. 2000;163(1):15–28.View ArticleGoogle Scholar
 Atkinson AJ, Colburn WA, DeGruttola VG, DeMets DL, Downing GJ, Hoth DF, et al. Biomarkers and surrogate endpoints: preferred definitions and conceptual framework. Clin Pharmaco Ther. 2001;69(3):89–95.View ArticleGoogle Scholar
 De Gruttola VG, Clax P, DeMets DL, Downing GJ, Ellenburg SS, Friedman L, et al. Considerations in the evaluation of surrogate endpoints in clinical trials: summary of a National Institutes of Health workshop. Control Clin Trials. 2001;22(5):485–502.View ArticlePubMedGoogle Scholar
 Molenberghs G, Geys H, Buyse M. Evaluation of surrogate endpoints in randomized experiments with mixed discrete and continuous outcomes. Stat Med. 2001;20(20):3023–38.View ArticlePubMedGoogle Scholar
 Baker SG, Izmirlian G, Kipnis V. Resolving paradoxes involving surrogate endpoints. J R Stat Soc Ser AG. 2005;168(4):753–62.View ArticleGoogle Scholar
 Goya M, Pratcorona L, Merced C, et al. Cervical pessary in pregnant women with a short cervix (PECEP): an openlabel randomised controlled trial. Lancet. 2012;379(9828):1800–06.View ArticlePubMedGoogle Scholar
 Gellman A, Carlin JB, Stern HS, Dunson DB, Vehtari A, Rubin DB, Bayesian Data Analysis, 3rd edition. Boca Raton, FL, USA: Chapman Hall/CRC; 2014Google Scholar
 Kirkby S, Greenspan JS, Kornhauser M, Schneiderman R. Clinical outcomes and cost of the moderately preterm infant. Adv Neonatal Care. 2007;7(2):80–7.View ArticlePubMedGoogle Scholar
 Johnston KM, Gooch K, Korol E, Vo P, Eyawo O, Bradt P, et al. The economic burden of prematurity in Canada. BMC Pediatr. 2014;14:93.View ArticlePubMedPubMed CentralGoogle Scholar
 Liu L, Oza S, Hogan D, Perin J, Rudan I, Lawn JE, et al. Global, regional, and national causes of child mortality in 2000–13, with projections to inform post2015 priorities: an updated systematic analysis. Lancet. 2015;385(9966):430–40.View ArticlePubMedGoogle Scholar
 Petitti DB. Metaanalysis, decision analysis, and costeffectiveness analysis: methods for quantitative synthesis in medicine. Oxford UK: Oxford University Press; 1999.Google Scholar
 Sculpher M, Fenwick E, Claxton K. Assessing quality in decision analytic costeffectiveness models. Pharmacoeconomics. 2000;17(5):461–77.View ArticlePubMedGoogle Scholar
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.