Premium
An examination of methods for sample size recalculation during an experiment
Author(s) -
Betensky Rebecca A.,
Tierney Camlin
Publication year - 1997
Publication title -
statistics in medicine
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.996
H-Index - 183
eISSN - 1097-0258
pISSN - 0277-6715
DOI - 10.1002/(sici)1097-0258(19971130)16:22<2587::aid-sim687>3.0.co;2-5
Subject(s) - oversampling , sample size determination , variance (accounting) , computer science , sample (material) , sampling (signal processing) , statistics , early stopping , sequential estimation , algorithm , mathematics , artificial intelligence , bandwidth (computing) , computer network , chemistry , accounting , chromatography , filter (signal processing) , artificial neural network , business , computer vision
In designing experiments, investigators frequently can specify an important effect that they wish to detect with high power, without the ability to provide an equally certain assessment of the variance of the response. If the experiment is designed based on a guess of the variance, an under‐powered study may result. To remedy this problem, there have been several procedures proposed that obtain estimates of the variance from the data as they accrue and then recalculate the sample size accordingly. One class of procedures is fully sequential in that it assesses after each response whether the current sample size yields the desired power based on the current estimate of the variance. This approach is efficient, but it is not practical or advisable in many situations. Another class of procedures involves only two or three stages of sampling and recalculates the sample size based on the observed variance at designated times, perhaps coinciding with interim efficacy analyses. The two‐stage approach can result in substantial oversampling, but it is feasible in many situations, whereas the three‐stage approach corrects the problem of oversampling, but is less feasible. We propose a procedure that aims to combine the advantages of both the fully sequential and the two‐stage approaches. This quasi‐sequential procedure involves only two stages of sampling and it applies the stopping rule from the fully sequential procedure to data beyond the initial sample which we obtain via multiple imputation. We show through simulations that when the initial sample size is substantially less than the correct sample size, the mean squared error of the final sample size calculated from the quasi‐sequential procedure can be considerably less than that from the two‐stage procedure. We compare the distributions of these recalculated sample sizes and discuss our findings for alternative procedures, as well. © 1997 by John Wiley & Sons, Ltd.