Tugas Ibu Sumah No 2 PDF
Tugas Ibu Sumah No 2 PDF
Assistant Professor
Department of Clinical Epidemiology & Biostatistics
Faculty of Health Sciences
McMaster University
University
Hamilton ON
November 19, 2004
1
Contents
1 Intro duction 1
1.1 Learning Ob jjeectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Introductory Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
7 Repo
porrtin
ting the resu
sult
ltss of sa
sam
mple si
sizze calcu
lculatio
ation
n in th
the
e protoc
otoco
ol 13
9 In
Ina
approp
ropriat
riate
e word
ordin
ing
g or repo
reporrtin
ting of Sample si
siz
ze calcu
lculatio
ation
ns 17
10 Im
Impo
porrtan
tant Rema marrks Abo bou ut Achiev eviing th
the
e req
equ
uired
red sasamp
mplele si
size
ze 19
10.1 Common Recruitment Strategies . . . . . . . . . . . . . . . . . . . . . . . . . 19
10
10.2
.2 Re
Reas
ason
onss ffor
or fa
fail
ilur
uree ttoo aacchiev
hievee tthe
he requ
requir
ired
ed samp
sample
le size
size . . . . . . . . . . . . . 19
10.3 Possible or Common Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . 20
12
12.2
.2 Ru
Rule
less of Th
Thum
umb
b for Rel
elat
atio
ions
nshi
hips
ps/A
/Ass
ssoc
ocia
iattion
ion . . . . . . . . . . . . . . . . . 23
i
13 Tips on Elicitation of Effect Sizes and Variances for Sample Size Calcula-
tions 24
14 Sa
Sampl
mple
e S
Siz
ize
e C
Calc
alcula
ulatio
tions
ns for Clu
Cluste
sterr R
Ran
andom
domiz
ized
ed Con
Contro
trolle
lled
d S
Stud
tudie
iess 25
14
14.1
.1 Re
Reas
ason
onss fo
forr Us
Usin
ingg Cl
Clus
ustter
er-r
-ran
ando
domi
mize
zed
d Desig
esigns
ns . . . . . . . . . . . . . . . . . 25
14
14.2
.2 Sa
Samp
mplele Size
Size For
orm
mulae
ulae fo
forr C
Clu
lust
ster
er-r
-ran
ando
domi
mize
zed
dDDes
esig
igns
ns . . . . . . . . . . . . . 26
ii
1 In
Intr
trodu
oduc
cti
tion
on
1.1
1.1 Lear
Learni
ning
ng Obje
Objecti
ctiv
ves
Specific Learning Objectives
• Learn about the important elements of a sample size calculation in the design of a
clinical trial
• Gain some knowledge about the basic statistical rules of thumb in sample size calcu-
lations
Learn how to report the results of sample size calculation for a granting agency
agency,, research
• ethics board submission, etc.
1.2
1.2 Intr
Introdu
oduct
ctor
ory
y Rema
Remark
rkss
Sample size calculations form an integral part of a vast majority of quantitative studies.
There are three main parts to sample size calculation: (i) sample size estimation, which de-
pends on a host of items (see Section 3.1); (ii) sample size justific
justification,
ation, whic
which
h ofte
often
n inv
involv
olves
es
justification of the calculated number in the light of budgetary and other biological consid-
erations, and (iii) sample size adjustment, increasing the sample to account for potential
dropouts or effect of covariates.
• Sample size calculations may not be required for some pilot or exploratory studies. It
is important to note that
1
• As part of sample size discussions, it is crucial to know what the consequences of
’getting it wrong’ are: these may be ethical, economic or scientific
• Sample size problems can be approached in two ways:
– Patients I need
need approa
approach
ch:: bas
based
ed on cal
calcul
culati
ation
on of the sampl
samplee siz
sizee for a giv
given
en
power, level of significance and clinically meaningful difference
– Patients I can get approach:
approach: based on calculation of power or detectable difference
for a given sample size and level of significance
2 Wh
Why
y is sa
samp
mple
le si
size
ze ccal
alcu
cula
lati
tion
on iimpor
mporta
tan
nt?
Sample size is important for two main reasons:
– If a trial with negative results has insufficient power (insufficient sample size),
size), a
clinically important (but statistically nonsignificant) effect is usually ignored or,
worse, is taken to mean that the treatment under study made no difference
Overall sample size calculation is an important part of the study design to ensure validity,
accuracy, reliability and, scientific and ethical integrity of the study.
3 Ap
Appr
proac
oache
hess to sam
sampl
ple
e si
size
ze cal
calcu
cula
lati
tion
on
There are two major classical approaches to sample size calculations in the design of quan-
titative studies:
• However, a prevalence of 10% from a sample of size 400 would have a 95% CI of (7%,
13%), which may be considered more accurate or informative.
3.2 Hypothesis
Hypothesis Testing
Testing of e
effects/
ffects/relati
relationship
onships:
s: Power
Power Analysis
Analysis
In studie
studiess con
concer
cerned
ned wit
with
h det
detect
ecting
ing an effect
effect (e.
(e.g.
g. a diff
differ
erenc
encee bet
betw
ween tw
twoo tre
treatm
atmen
ents,
ts,
or relative risk of a diagnosis if a certain risk factor is present versus absent), sample size
calculations are important to ensure that if an effect deemed to be clinically meaningful exists,
3
then the
then there
re is a hig
high
h chanc
chancee of it being det
detect
ected,
ed, i.e. tha
thatt the analy
analysis
sis will be sta
statis
tistic
ticall
ally
y
significan
signifi cant.
t. If the sample is too small, then even if large differen
differences
ces are observ
observed,ed, it will b
bee
impossible to show that these are due to anything more than sampling variation. There are
differen
diffe rentt types of hypot
hypothesis
hesis testi
testing
ng problems dependin
dependingg on the goal of the resear
research.
ch. Let
µS = mean of standard treatment, µT = mean of new treatment, and and δ
δ = the minimum
clinically important difference.
2. Test for Non-inferiority : To demonstrate that the new drug is as less effective as the
standard treatment (ie the difference between the new treatment and the standard is
less than the smallest clinically meaningful difference)
• If δ =
= 0, it is called the test of statistical
statistical superiority
superiority
• Equiv
Equivale alence
nce is tak
taken
en to be the alt alter
ernat
nativ ivee hypoth
hypothesi
esis,
s, and the nu
null
ll hypoth
hypothesi
esiss is
nonequivalence
4 In
Info
form
rmat
atio
ion
n req
requi
uire
red
d to ca
calc
lcul
ulat
ate
e a samp
sample
le si
size
ze
It is highly recommended that you ask a professional statistician to conduct the sample size
calculation.
4
4.1 Fact
actors
ors that
that influenc
influence
e sample
sample size calcul
calculati
ation:
on: A check
checklis
listt
1. The objective(s) of the research: Is the researc
research
h dealing with an estimation, h
hypothesis
ypothesis
or equivalence testing problem?
3. The outcome
outcome(s)
(s) of the rese
researc
arch:
h:
7. What is the re
researc
search
h design
design?? Is it
12. Wha
Whatt type of summar
summary
y or tes
testt stati
statisti
sticc will be use
used
d for analys
analysis?
is? Wil
Willl it be a one
one-- or
two-tailed test?
• Does this reflect the degree of benefit from the intervention against the control
over the specified time frame?
• Is it stated as
6
al [10]).
14. Justi
Justificati
fication:
on: Most importan
importantly
tly,, is the justi
justificati
fication
on provided on how the variou
variouss prior
estimates used in the calculations were obtained and their usefulness in the context of
the study? This also deals with the clinical rele
relevvance of the estimates
estimates depending on
the source (ie published data, previous work, review of records, expert opinions, etc).
5 Ex
Expl
plan
anat
atio
ion
n of S
Sta
tati
tist
stic
ical
al T
Ter
erms
ms
Below are some brief descriptions of some of above statistical terms.
2. P-value of a test : The p-value is the probability of obtaining the effect as extreme or
more extreme than what is observed in the study if the null hypothesis of no effect is
actually true. It is usually expressed as a proportion (e.g. p=0.001).
3. Significance level of a test : Also called the Type error probability, the significance level
is a cut-off point for the p-value, below which the null hypothesis will be rejected and
it will be concl
concluded
uded that there is evide
evidence
nce of an effect. The conv
convent
entional
ional significance
significance
7
• that the null hypothesis will be correctly rejected i.e. rejected when there is indeed
a real difference or association
• the test will detect a difference or association of a particular magnitude when it
exists
• 1−β , where β
where β is
is the P(Type II error), the chance of missing a clinically meaningful
difference
The higher the power, the lower the chance of missing a real effect. Power is typically
set to be at least 80%.
Reject H 0
Reject Accept H a
Accept H
H 0 True
rue Type I er
erro
rorr Corr
orrec
ectt De
Deci
cisi
sion
on
(Probability=α
(Probability= α)
H 0 Fal
alse
se Co
Corr
rrec
ectt De
Deci
cisi
sion
on Typ
ypee II erro
errorr
(Probability=β
(Probability= β )
states there is no effect, and the alternative hypothesis is that a difference exists in
either direction. In a one-sided test the alternative hypothesis does specify a direction,
for example that an active treatment is better than a placebo, and the null hypothesis
then includes both no effect and placebo better than active treatment.
• Two-sided tests should be used unless there is a very good reason for doing oth-
erwise.
• One-sided tests may be appropriate in situations where is it completely inconceiv-
able that the results could go in either direction, or the only true concern with
outcomes in one tail of the distribution.
(a) Exampl
Examples
es includ
include:
e:
8
i. tox
toxicit
icity
y studies
ii. safety ev
evaluations
aluations
iii. analys
analysis
is of occurr
occurrence
encess of advers
adversee drug react
reactions
ions
iv. risk an
analysis
alysis
(b) Refe
Referenc
rences:
es:
i. Bland JM, Altman DG. One and two sided tests of significance. BMJ
1994; 309: 248.
ii. Dubey SD. Some Thoughts on the One-sided and Two-sided
Two-sided Tests. Jour-
Tests. Jour-
nal of Biopharmaceutical Statistics 1991;
1991; 1: 139-150.
iii. Chow S-C
S-C,, Shao J, W
Wang
ang H. Sample Size Calculations in Clinical Researc
H. Sample Research
h
Marcel Dekker: New York, NY 2003
The expectation that the difference will be in a particular direction is not adequate
9
6 For
orm
mul
ulae
ae fo
forr Sam
Sampl
ple
e Siz
Size
e Cal
Calcu
cula
lati
tions
ons
Table 1: Formula
ormulae
e for Sample Size Calculations for Comparisons Betw
Between
een
Means
z α +zβ σ2
2
One-sample Equality µ − µ = 0
0 µ − µ = 0
0 n = (µ−µ0 )2
2
(zα +zβ ) σ2
Superiority µ − µ ≤ δ
0 µ − µ > δ
0 n = (µ−µ0 −δ )2
2
(zα +zβ ) σ2
Equivalence |µ − µ | ≥ δ |µ − µ | < δ
0 0 n = |−δ )2
(|µ−µ0 |−δ
2 z α +zβ σ2
Two-sample Parallel Equality µ1 − µ = 0
2 µ1 − µ = 0
2 ni =
2
(µ1 −µ2 )2
2
2(zα +zβ ) σ 2
Non-inferiority µ1 − µ ≥ δ
2 µ1 − µ < δ
2 ni = (µ1 −µ2 −δ )2
2
2(zα +zβ ) σ 2
Superiority µ1 − µ ≤ δ
2 µ1 − µ > δ
2 ni = (µ1 −µ2 −δ )2
2
2(zα +zβ ) σ 2
Equivalence |µ − µ | ≥ δ |µ − µ | < δ
1 2 1 2 ni = |−δ )2
(|µ1 −µ2 |−δ
2
z α +zβ σ2
2
Two-sample Crossover Equality µ1 µ2 = 0 µ1 µ2 = 0 ni = 2(µ1 −µ2 )2
2(µ
− − 2
(zα +zβ ) σ2
Non-inferiority µ1 − µ ≥ δ
2 µ1 − µ < δ
2 ni = 2(µ1 −µ2 −δ )2
2(µ
2
(zα +zβ ) σ2
Superiority µ1 − µ ≤ δ
2 µ1 − µ > δ
2 ni = 2(µ1 −µ2 −δ )2
2(µ
2
(zα +zβ ) σ2
Equivalence |µ − µ | ≥ δ |µ − µ | < δ
1 2 1 2 ni = 2( |−δ)2
2(||µ1 −µ2 |−δ
10
Table 2: Formula
ormulae
e for Sample Size Calculations for Comparisons Betw
Between
een
Proportions
2
(zα +zβ ) π (1
(1−
−π )
Superiority π − π ≤ δ
0 n = (π −π0 −δ )2
2
(zα +zβ ) π (1
(1−
−π )
Equivalence |π − π | ≥ δ 0 n = |−δ )2
(|π −π0 |−δ
2
z α +zβ (π1 (1
(1−
−π2 )+π
)+π2 (1
(1−−π2 ))
2
Two-sample Parallel Equality π1 − π = 0
2 ni = (π1 −π2 )2
2
(zα +zβ ) (π1 (1
(1−−π2 )+π
)+π2 (1
(1−−π2 ))
Non-inferiority π1 − π ≥ δ
2 ni = (π1 −π2 −δ )2
2
(zα +zβ ) (π1 (1
(1−−π2 )+π
)+π2 (1
(1−−π2 ))
Superiority π1 − π ≤ δ
2 ni = (π1 −π2 −δ )2
2
(zα +zβ ) (π1 (1
(1−−π2 )+π
)+π2 (1
(1−−π2 ))
Equivalence |π − π | ≥ δ
1 2 ni = |−δ )2
(|π1 −π2 |−δ
2
z α +zβ σd2
2
Two-sample Crossover Equality π1 − π = 0
2 ni = 2(π1 − π2 ) 2
2(π
2
(zα +zβ ) σ2
Non-inferiority π1 − π ≥ δ
2 ni = 2(
2(ππ1 −π2 −δ )d2
2
(zα +zβ ) σd2
Superiority π1 − π ≤ δ
2 ni = 2(
2(ππ1 −π2 −δ )2
2
(zα +zβ/ ) 2 σd2
Equivalence |π − π | ≥ δ
1 2 ni = 2( |−δ )2
2(||π1 −π2 |−δ
11
6.1
6.1 Sampl
Sample
e S
Siz
ize
e A
Adju
djust
stme
men
nts
It is important to note that sample-size problems will vary from study to study depending
on the context. The sample size may need to be adjusted to account for the effects of other
variables, and the uncertainty of predictable practical and ethical factors.
• Which variables should be included in the sample size calculation?
– The sample size calculation should relate to the study’s primary outcome variable.
– Ideally, separate sample size calculations should be provided for each important
variable : the sample size should also be suffici
sufficient
ent for the analyse
analysess of all importan
importantt
variables.
– A simpler conservative approach is to estimate sample sizes for all important
outcomes and then use the maximum estimate.
– As a rule of thumb, when the correlation of a covariate with the response variable
is ρ,, then the sample size can be reduced by a factor of 1 ρ2 . That is,
is ρ −
nnew = n(1
n(1 − ρ ).
2
• Multiplicity and Sample Size Adjustment : Multiplicity adjustment using the Bonferroni
method to the level of significance should be made when at least one significant result
(eg one of several primary outcomes or several pairwise comparisons) is required to
draw a conclusion
– loss to follow up
– lack of compliance
– any other unforseen reasons for loss of subjects
For example to adjust the sample size for the anticipated loss to follow-up rare: Suppose
n is the total number of subjects in each group not accounting for loss to follow-up,
and L
and L is the loss to follow-up rate, then the adjusted sample size is given by
n
nnew =
−
1 L
It is important to state clearly what factors were taken into consideration in the sample
size adjustment and the justification should also be explicit.
– First calculate n
calculate n the per group sample size assuming equal number per group
– If we require n
require n 1 /n2 = k
= k,, then
1 1
n2 = n(1 + 1/k
1/k)) and n1 = n(1 + k
+ k))
2 2
– The questio
question
n of whether to use unequal sample sizes matters when multiple
multiple sizes
can be obtained in one group (see Lachin 2000, van Belle 2003)
7 Re
Repor
porti
ting
ng the re
resu
sult
ltss of sam
sampl
ple
e si
size
ze cal
calcu
cula
lati
tion
on in
the protocol
The protocol should provide sufficient details on how the sample size was determined. This
should cover
1. clear state
statemen
ments
ts of the (primar
(primary)
y) ob
object
jective
ivess of the study
2. the desire
desired
d level of signific
significance
ance
3. the desi
desired
red pow
power
er
4. type of summary or test stati
statistic
stic will be used for analysis
5. whet
whether
her the test will one- or two-taile
two-tailed
d
6. the smalles
smallestt differenc
differencee and a clear statemen
statementt of whet
whether
her it is
• the smallest clinically important difference
• the difference that investigators think is worth detecting
• the difference that investigators think is likely to be detected
7. justi
justificati
fication
on provide
provided
d on how the var
various
ious prior estimate
estimatess of the varianc
ariancee and the effect
used in the calculations were obtained and their usefulness in the context of the study
8. clear state
statemen
ments
ts about the assumpti
assumptions
ons made about the distr
distributio
ibution
n or varia
variabilit
bility
y of
the outcomes
9. clear state
statemen
mentt about the sche
scheduled
duled durat
duration
ion of the study
10. clear state
statemen
ments
ts about how the sample size calculati
calculation
on was adjusted for
• the expected response rate
• loss to follow up
• lack of compliance
any other unforseen reasons for loss of subjects
•
11. any other infor
information
mation that forme
formed
d the basis for the sample size calculat
calculation.
ion.
13
8 Spec
Specifi
ificc ex
exam
ampl
ples
es of ssamp
ample
less si
size
ze ccal
alcu
cula
lati
tions
ons
If yo
your
ur stu
study
dy req
requir
uires
es the estim
estimati
ation
on of a sin
single
gle pro
proport
portion
ion,, com
compar
pariso
ison
n of tw
twoo mea
means,
ns,
or com
compar
pariso
ison
n of twtwoo pro
proport
portion
ions,
s, the sam
sample
ple siz
sizee cal
calcul
culati
ations
ons for the
these
se situat
situation
ionss are
(gener
(ge nerall
ally)
y) rel
relati
ative
vely
ly str
straig
aight
htfor
forwa
ward,
rd, and are the
there
refor
foree pre
presen
sented
ted her
here.
e. HoHowe
weve
ver,
r, it is
still strongly recommended that you ask a statistician to conduct the sample size calcu-
lation.
latio n. The follo
following
wing exampl
exampleses (take
(takenn from St. Georg
George’s e’s Hospital Medi
Medical
cal School website:
website:
http://www.sghms.ac.uk/depts/phs/guide/size.htm ) are meant to illustrate how to calcu-
late, justify and report sample size calculations.
8.1 Exampl
Example
e 1
1:: Compar
Comparing
ing tw
two
o propo
proporti
rtions
ons
• Goal: The following calculation only applies when you intend to compare two groups
of the same size.
• Scenario:
Scenario: A place
placebo-con
bo-controll
trolled
ed randomize
randomized
d trial proposes to assess the effectiv
effectivenes
enesss of
colony stimulating factors (CESS) in reducing sepsis in premature babies. A previous
study has shown the underlying rate of sepsis to be about 50% in such infants around 2
weeks after birth, and a reduction of this rate to 34% would be of clinical importance.
• Required information :
– Primary outcome variable = presence/absence of sepsis at 14 days after treatment
(treatment is for a maximum of 72 hours after birth).
– Hence, a categorical variable summarized by proportions.
– Size of difference of clinical importance = 16%, or 0.16 (i.e. 50%-34%)
– Significance level = 5%
– Power = 80%
14
[1 0.84]2
[1..96 + 0. [(0..50
[(0 0.50) + (0.
(0.34 0.66)]
n = × [0 ×
[0..16]2 × = 146
8.2
8.2 Exam
Exampl
ple
e 2: Comp
Compar
arin
ing
g ttw
wo mea
means
ns
• Goal : The following calculation only applies when you intend to compare two groups
of the same size.
• Scenario:
Scenario: A ran
random
domize
ized
d con
contro
trolle
lled
d tri
trial
al has been pla
planne
nnedd to ev
evalu
aluate
ate a bri
brief
ef psy-
chological intervention in comparison to usual treatment in the reduction of suicidal
ideation
ideation amongs
amongstt patie
patients
nts presen
presenting
ting at hospit
hospital
al with deliberate self-pois
self-poisoning.
oning. Suici
Suici--
dal ideation will be measured on the Beck scale; the standard deviation of this scale
in a previous study was 7.7, and a difference of 5 points is considered to be of clini-
cal importanc
importance.
e. It is an
antic
ticipa
ipate
ted
d tha
thatt aro
around
und one thi
third
rd of pat
patien
ients
ts may drop out of
treatment (Guthrie et al. 2001)
• Required information :
15
The formula for the sample size for comparison of 2 means (2-sided) is as follows: -
[z α + z ββ ]2 × 2σ 2
n = 2
δ 2
where n = the sample size required in each group (double this for total sample).
σ = standard deviation, of the primary outcome variable = 7.7.
δ =
= size of difference of clinical importance = 5.0.
z α = 1.96.
2
z ββ = 0.84.
Inserting the required information into the formula gives: -
0.84]2 2 7.72
[1..96 + 0.
[1 × ×
n = = 38
5.02
This gives
sample sizethe numberthis,
is double requi.e.
required
ired76.in eac
each
h of the trial
trial’s
’s two groups. Ther
Therefor
eforee the total
To allow
allow for the pred
predict
icted
ed dro
dropout
pout rat
ratee of aro
around
und one thi
third,
rd, the sampl
samplee siz
sizee wa
wass
increased to 60 in each group, a total sample of 120.
16
9 In
Inap
appr
prop
opri
riat
ate
e wor
ordi
ding
ng or re
repor
porti
ting
ng of Sa
Samp
mple
le siz
size
e
calculations
1. Example 1: ”A previous study in this area recruited 150 subjects and found highly
significant results (p=0.014), and therefore a similar sample size should be sufficient
here.”
• Why is this a problem? : Previous studies may have been ’lucky’ to find significant
results, due to random sampling variation
• Solution: Cal
Calcul
culati
ations
ons of sam
sample
ple size spec
specific
ific to the prese
present
nt,, pro
propose
posed
d stu
study
dy
should be provided
17
2. Example 2: ”Sample sizes are not provided because there is no prior information on
which to base them.”
• If the study is a preliminary pilot aimed at assessing feasibility or gathering the
information required
size calculations to calculate
are not necessarysample sizes for a full-scale study, then sample
• Where
Where prior informat
information
ion on standa
standard
rd devia
deviations
tions is unav
unavailabl
ailable,
e, then standard
deviation can be estimated from the range as
Max-Min
Standard deviation = .
4
Then sample size calculation
calculationss can be giv
given
en in ve
very
ry general terms
terms,, i.e. by giving
the size of difference that may be detected in terms of a number of standard
deviations
3. Example
Example 3: “The throughput of the clinic is around 50 patients a year, of whom
10% may refuse to take part in the study. Therefore over the 2 years of the study, the
sample size will be 90 patients. ”
• Although most studies need to balance feasibility with study power, the sample
size should not be decided on the number of available patients alone.
• If the number of available patients is a known limiting factor, a apply the patients
the patients
I can get approach
approach to indicate either
(a) the power which the study will hav
havee to detect the desired difference of clinical
importance,
importan ce, or
(b) the differe
difference
nce whic
which
h will b
bee detecte
detectedd when the desir
desired
ed power is applied
Where the number of available patients is too small to provide sufficient power to detect
differences of clinical importance, you may wish to consider extending the length of
the study, or collaborating with a colleague to conduct a multi-centre study.
4. Other Examples
• “The results of a pilot study have been submitted for publication and the revie
reviewers
wers
indicate that our sample size is adequate.”
• “W
“Wee aim to recruit 100 parti
participan
cipants.
ts. This sample size was determin
determined
ed to dete
detect
ct
a small to moderate mean difference of xx points between the treatment groups
on at least one of the key outcomes with a 70% power.”
18
10 Im
Import
portan
antt Rem
Remark
arkss About A
Acchi
hievi
eving
ng th
the
e req
requi
uire
red
d
sample size
10.1
10.1 Common
Common Rec
Recrui
ruitmen
tmentt Str
Strateg
ategies
ies
Recruitment of suitable subjects for participation in clinical trials can also achieved through
use of effective recruitment strategies (taken from the 2000 Office of Inspector General (OIG)
of the US Department of Health Services Report):
1. throu
through
gh the use of financ
financial
ial and non-fi
non-financia
nanciall incenti
incentives
ves
2. by physic
physicians
ians flagging patient
patientss in their pract
practice
ice through chart revie
reviews
ws or when they
appear for an appointment
3. by furnishing tr
trial
ial information to other local clinicians or to
to disease advocacy and other
groups
4. throu
through
gh advert
advertising
ising and promotion suc
such
h as
• media ads
• press releases
• televised segments
• speakers at local health fairs
Recruitment through the Internet or Web in also increasingly becoming a popular option.
10.2
10.2 Reason
Reasonss for failur
failure
e to achiev
achieve
e the req
requir
uired
ed sample
sample size
size
The sample size required for a clinical trial may be very hard to recruit or recruitment has
been much
can lead to slower than ending
pre-mature anticipated.
of theThis
trial,iswhich
quite common ininconclusive
could lead clinical studies. Sometimes
findings this
because of
lack of power. In order to avoid or address this problem, it is important to understand why
it happen:
• adverse media publicity: sometimes adverse media publicity about medicine in general
and trials in particular may discourage potential subjects from taking part in a research
endeavor
• lack of genuine commitment to the project: sometimes this which might be caused by
honest doubts about the safety or efficacy of the new treatment
• poor recruitment may also be due to staffing problems: eg clinical unit is under pressure
from excessive patient numbers or understaffing
10.3
10.3 Possi
Possible
ble or Common
Common Sol
Solutio
utions
ns
Possible solutions include
• Pilot studies are very helpful in providing insights into some of these issues
• good planning
planning:: dev
devise
ise a cle
clear
ar plan of how to mon
monito
itorr recru
recruitm
itmen
ent.
t. Thi
Thiss may also
help the proposal for funding since funders may well be impressed by a proposal which
shows that this issue has been considered by applicants and some there are plans to
deal with it
• to have recruitment targets (milestones) to enable the research team to monitor how
well recruitment is going, so that problems can be detected as soon as possible
11 Re
Retr
trospe
ospect
ctiv
ive
e Sam
Sampl
ple
e Si
Size
ze Ca
Calc
lcul
ulat
atio
ions
ns
Sometimes, people try to estimate the sample size or perform power analysis after the study
is completed.
– It is ass
associa
ociated
ted with thi
thiss com
common
mon mis
mis-in
-inter
terpre
pretat
tation
ion or mis
miscon
conce
cepti
ption:
on: if the
test is nonsignificant (pvalue is large), but the observed power is high, then this is
interpreted to mean that there is strong evidence in support of the null hypothesis
– It causes great confusion because it is ofte
often
n used inappr
inappropriat
opriately
ely to add interp
interpre-
re-
tation to a non-significant test result
12 Im
Import
portan
antt Ru
Rule
less of Th
Thum
umb/C
b/Caut
autio
ions
ns
12.1
2.1 Gener
enera
al
1. Multiplicity and Sample Size Adjustment : Multiplicity adjustment using the Bonferroni
method to the level of significance should be made when at least one significant result
(eg one of several primary outcomes or several pairwise comparisons) is required to
draw a conclusion
2. Overlapping Confidence Intervals Do not imply non-significance
Basic Rule: “Confidence intervals associated with statistics can overlap as much
as 29% and the statistics can still be significantly different” (van Belle 2002)
3. Sample size calculations should be based on the statistics used in the analysis . Example,
if sample size calculations were based on assumption that the outcome is continuous,
then dichotomizing the outcome for the analysis would not be appropriate. Why?
• using a different statistic for analysis may alter the anticipated power
• the anticipated treatment effect may no longer be meaningful in the scale of the
new statistic
4. The basic rule of thumb for estimating the sample size for testing equality
equality of two means
is 2
= n 2 = 8σ2 ; wh
n1 = n −
eree δ = µ
wher = µ1 µ2
δ
5. The basic rule of thumb for estimating the sample size for testing equality of two
proportions
proportions is
n1 = n
= n 2 =
−
8π(1 π)
; wh
wher
ere
e π =
π1 + π
+ π2
(π1 π2 )2
− 2
6. Since sample calcu
calculation
lationss are estimat
estimates,
es,
• it is usually better to be conservative . For example, it is usually better to assume
a two-sided test than a one-sided test
• it is better to adopt a simple approach even
even for complex problems. For example, it
is simpler to use difference between proportions than logistic regression in sample
size calculation
21
7. Althou
Althoughgh larger sample sizes are genera
generally
lly desired
desired,, it is important to be aware of the
to be
statistical (over-power), ethical, and economic consequences of too large a sample
9. It is wor
worth
th notin
notingg that observational or non-randomized studies looking for differences
that observational
or associations will generally require a much larger sample in
sample in order to allow adjustment
for confounding factors within the analysis
10. It is the absolute sample size which is of most interest , not the sample size as a pro-
portion of the whole population
• The adequacy of a sample size should be assessed according to the purpose of the
study. Check whether the purpose of the study is
– to test for no difference (equality)
– to test for non-inferiority
– to test for superiority
– to test for equivalence
Note that the sample size required to demonstrate equivalence will be larger than
that required to demon
demonstrat
stratee a differ
difference
ence..
• Sample size calculations should relate to the study’s stated objectives, and be
based on the study’s primary outcome variable or all important outcome variables
• Sample size calculations should also be consistent with the proposed method of
analysis
• For moderate to large effect size (ie 0.50 ≤effect size≤0.80), 30 subjects per group
are required
• For comparisons between three or more groups, then to detect an effect size of 0.5
(a) with 80% power, will requi
require
re 14 subjects
subjects/group
/group
(b) with 50% power, will requi
require
re 7 subjects
subjects/group
/group
13. Sensitivity Analysis : It is best to create a sample size table for different values of the
(α), power or different effect sizes, and then ponder the table to
level of significance (α
select the optimal sample size
22
12.2
12.2 Rules
Rules of Th
Thum
umb
b for Relat
Relation
ionshi
ships/
ps/Ass
Associa
ociatio
tion
n
problems, power refers to the ability to find a specified regression coefficient
In regression problems, power
2
R statistically significant at a specified level of significance and specified sample
or level of R
size.
1. For multi
multiple
ple regre
regression:
ssion: Hair et al, 2000 state
n > 50
> 50 + 8m
8m
n > 104
> 104 + m
+ m
3. Harri
Harriss (1985) recomm
recommends
ends
n > 104
> 104 + m
+ m
4. Large sample
sampless are needed (see T
Tabac
abachnic
hnick
k and Fidell, 1996) if
23
(d) step
stepwise
wise regre
regression
ssion is used
5. Rules for chi-
chi-squar
squared
ed testing
(a) Sample size shou
should
ld be suc
such
h that no expect
expected
ed freque
frequency
ncy in a cell shoul
should
d drop b
belo
elow
w
5: small expected cell frequencies can limit power substantially and inflate Type
I error.
(b) Ov
Overall
erall samp
sample
le size shoul
should
d be at least 20
(c) The num
number
ber of cells (ie degrees of freedom of the chi-sq
chi-squared
uared test) is indirectl
indirectly
y
related with power (see Cohen (1988))
6. Rules for F
Factor
actor Analy
Analysis
sis
(a) At least 300 case
cases/subject
s/subjectss (T
(Tabac
abachnic
hnick
k and Fidell, 1996)
(b) At least 50 partic
participan
ipants/subjec
ts/subjects
ts p
per
er variab
variable
le (Pedhaz
(Pedhazur
ur and Sch
Schmelk
melkin,
in, 1991)
(c) Comre
Comrey
y and Lee (1992) guide
guide::
•• n=100:
n=50: very poor
poor
• n=200: fair
• n=300: good
• n=500: very good
(d) The higher the cases-per-var
cases-per-variable
iable ratio, the smaller the chance of “overfitting”
“overfitting” (ie
creating factors that are not generalizable beyond the specific sample)
13 Tip
Tipss on E
Elic
licita
itatio
tion
n of E
Effec
ffectt Si
Sizes
zes and V
Vari
ariance
ancess fo
forr
Sample Size Calculations
The following tips are taken from “Some Practical Guidelines for Effective Sample Size
Determination” by Lenth (Lenth, 2001)
1. Elici
Elicitation
tation of informat
information
ion on effec
effectt sizes to calcu
calculate
late sample size: Importa
Important
nt questi
questions
ons
• What results do you expect (hope to) see?
• Would an effect of half that magnitude [specify] be of any scientific interest?
• Would a increase/decrease/difference of this magnitude [specify] be of any prac-
tical importance?
• What is the range of clinical indifference?
• If you were a patient, would the benefits of reducing/increasing the primary out-
come [specify] by this magnitude [specify] outweigh the cost, inconvenience and
potential
potential side effec
effects
ts of this treatmen
treatment?
t?
24
2. Elici
Elicitation
tation of infor
information
mation on stand
standard
ard devia
deviations
tions/v
/varianc
ariances
es to calcu
calculate
late sample size:
Important questions
14 Sam
Sampl
ple
e Size Cal
Calcu
cula
lati
tions
ons for Clus
Cluste
terr Rand
Random
omiz
ized
ed
Controlled Studies
Cluster randomized designs are increasingly used in community healthcare interventions and
health services research. Cluster randomized designs are designs in which intact social units
or clusters of other units are allocated to treatment groups or interventions.
14.1
14.1 Reason
Reasonss for Usin
Using
g Clu
Cluster
ster-ra
-rando
ndomiz
mized
ed Design
Designss
Reasons for using cluster-randomized designs include (from Hutton 2001; Donner and Klar
2000):
• Scientific reasons
25
• Ethical Reasons
1. Randomizing part of the family : For trials that deal with vacci
accines
nes or food int
inter-
er-
ventions, it would appear unethical to randomize some members of a family or
community to one intervention instead of the whole family.
14.2
14.2 Sample
Sample Size Form
Formula
ulae
e for Clust
Cluster-
er-ran
random
domize
ized
d Designs
Designs
Let
k = num
umbeberr of cl
clus
uste
ters
rs
m = av
aver
erag
agee cl
clus
uste
terr size
size
ρ = intra
intra-cl
-clust
uster
er cor
correl
relati
ation
on coeffi
coefficie
cient
nt
I F = −
1 + (m 1) ρ = infl
1)ρ inflati
ation
on fac
factor
tor
26
z 2
[π (1 + π
α
α/2 + z β
/2 + z β 1 + π (1 − π )] × I F
+ π ) + π 1 2 2
k = 2
m (π1 −π ) 2
where
C V 2 (λ21 + λ
+ λ22 ) t
I F t = 1 +
(λ1 + λ
+ λ2 )
σ1 σ2
C V = =
λ1 λ2
and σi2 is the between-cluster variation in incidence rates for the ith group, t is the
and
person-years, and and C V
V is
is the coefficient of variation, which plays the same role as the
intra-cl
intra-classass corre
correlation
lation coefficien
coefficient.
t. Note that if CV=0
CV=0,, then I F t = 1
then I
15 Sam
Sampl
ple
e Siz
Size
e Cal
Calcu
cula
lati
tions
ons for
for Ot
Othe
herr Ty
Types
pes of St
Stud-
ud-
ies
15.1
15.1 Analys
Analysis
is o
off Chang
Change
e From
From Baselin
Baseline
e
Let
y = Res
Response
ponse vari
ariabl
ablee
yb = Corre
Corresponding
sponding baseli
baseline
ne measu
measureme
rement
nt
27
15.2
15.2 An
Anal
alys
ysis
is o
off Times
Times to F
Fai
ailu
lure
re
Let
Assuming the survival times are exponentially distributed, the required sample size is given
by 2
2 Z αα/2 + Z ββ
/2 + Z
n =
M t /M c ))2
(ln(M
(ln(
15.3
15.3 Compar
Compariso
isons
ns of Means
Means for two
two Poisso
Poisson
n Pop
Popula
ulatio
tions
ns
Let
θ1 = Mea
Mean
n of Po
Popul
pulati
ation
on 1
θ1 = Mea
Mean
n of Po
Popul
pulati
ation
on 2
Using a two- sample test of equality of means based on samples from two Poisson populations,
the required number of observations per sample is
4
n = √ θ − √ θ
1 2 2
28
15.4
15.4 Test
esting
ing for
for a single
single cor
correl
relati
ation
on coefficie
coefficient
nt
Let
H 0 : ρ = 0
H a : ρ = 0
1
+ ρ
1 + ρ
Z = ln
2 1 ρ −
and normal approximation, then the required sample is
2
2 Z αα/2 + Z ββ
/2 + Z
n = 3 + 2
1+
1+ρ
ρ
ln
1−ρ
where ρ is regarded
where regarded as the clinica
clinically
lly meaningf
meaningful
ul value of the corre
correlation
lation coefficie
coefficient.
nt. Note
that the same formula can be used to determine the sample size for testing that the slope of
a regression line is not equal to zero.
15.5 Comparing
Comparing Correlatio
Correlation
n Coefficien
Coefficients
ts for Two
Two In
Independen
dependentt S
Sam-
am-
ples
Let
H 0 : ρ1 = ρ
= ρ2
H a : ρ1 = ρ2
The required number of observations per sample is given by
2
2 Z αα/2 + Z ββ
/2 + Z
n = 3 + ln − ln 2
1+ρ
1+ρ1 1+
1+ρρ2
1−ρ1 1−ρ2
15.6
15.6 Es
Esti
tima
mati
tion
on Prob
Proble
lems
ms
If there are no comparisons being made but a parameter is being estimated, then confidence
interval approach is used in calculating the sample size. Here we require the prior estimate
of the varian
ariance
ce and the margin of erro
errorr or required accurac
accuracy
y. For estimatin
estimatingg the population
29
−
mean µ, using a (1 α)100% confidence interval and the desired margin of error of E , the
mean
sample size is given by
2
Z αα/2
/2 σ
n =
E
where σ is
where σ is the prior estimate of the standard deviation of the population. The corresponding
formula for estimating the population proportion is given by
2
Z α/
α/2 2 π (1 − π)
n =
E 2
π is the prior estimate.
where π
where
16 Samp
Sample
le Si
Size
ze C
Calc
alcula
ulatio
tion
n bas
based
ed on Trans
ransfor
format
mation
ionss
Most of the statistical testing and corresponding sample size calculation procedures are based
on the norma
normall distr
distributi
ibution
on or some specifi
specificc distr
distributi
ibution
on of the responses or data. How
Howev
ever,
er,
quite often the assumed distribution may not fit the data, changing the scale of the original
data (transformations) and assuming the distribution for the transformed data may provides
a solution. Thus, if the analysis of the data is to be done on transformed data, it is equally
important to base the sample calculations on the scale of the transformed data.
p (1 − p ) 1 2
OR =
OR = .
p (1 − p ) 2 1
test H : OR
In this case the required sample size to test : OR = versus H : O
= 1 versus H
0 : OR
R=
a 1 is
2
Z α + Z β
2
β 1 1
n = log (OR
OR)) p (1 − p ) + p (1 − p )
2
1
1 2 2
2. The distri
distribution
bution of certai
certain
n outcomes suc
such
h as durat
duration
ion of sympt
symptoms,
oms, cost, etc, is often
skewed, but the log-transformation may normalize the distribution leading to a log-
normal
normal distr
distribu
ibutio
tion.
n. The
There
refor
foree it wo
would
uld be impo
importa
rtant
nt to perf
perform
orm the sam
sample
ple size
calculations on the transformed scale which would used for inferences.
• Most of theunder
developed statistical proceduresofdiscussed
the assumption normalitysoor
far, including
some those under s, have been
other distribution
30
For latest developments in sample size calculations for nonparametric tests, see Chapter 11
of
18 Sof
Softtware ffor
or Sa
Samp
mple
le Si
Size
ze Ca
Calc
lcul
ulat
atio
ions
ns
The following article by Len Thomas and Charles J. Krebs provides an indepth review of
some software of sample size calculation software:
• Thomas L, Krebs CJ. A Review of Statistical power analysis software. Bulletin of the
Ecological Society of America 1997;
1997; 78(2): 126-139.
• Commercial Software Many more options are provided by the commercial computer
package that include
1. nQue
nQuery
ry advisor: http :
advisor: http : //www.statsol.ie/nquery/samplesize.htm
2. Po
Power
wer and Precision: http :
Precision: http : //www.power
//www.power − analysis.com/home.htm
2002:: http :
3. PASS 2002 http : //www.ncss.com/pass.html
//www.ncss.com/pass.html
http://www.stat.ucla.edu/~jbond/HTMLPOWER/index.html
http://www.health.ucalgary.ca/~rollin/stats/ssize/
http://www.stat.uiowa.edu/%7Erlenth/Power/index.html
http://www.dssresearch.com/SampleSize/
http://www.stat.ucla.edu/calculators/powercalc/
http://hedwig.mgh.harvard.edu/sample_size/size.html
http://www.bobwheeler.com/stat/SSize/ssize.html
http://www.math.yorku.ca/SCS/Online/power/
http://www.surveysystem.com/sscalc.htm
http://www.researchinfo.com/docs/calculators/samplesize.cfm
http://espse.ed.psu.edu/spsy/Watkins/Watkins3.ssi
http://www.mc.vanderbilt.edu/prevmed/ps/index.htm
32
References
[1] Chow S-C, Shao J, Wang H. Sample Size Calcu
Calculatio
lations
ns in Clin
Clinic
ical
al Re
Rese
sear
arch
ch Marcel
Dekker: New York, NY 2003
[12] Lac
Lachin
hin Contr
JM. Int
trials. Contr
trials. Introducti
roduction
Clin on 1981;
Trials to sample
1981; size determ
1: 13-28. determinatio
ination
n and power analysis for clinical
[13]] Fle
[13 Fleiss
iss JL. Statistical methods for rates and proportions, 2nd ed. John Wiley & Sons:
JL. Statistical
New York, NY 1981
[14] Armit
Armitage
age PP,, Berry G, Matt
Matthews
hews JNS. Statistical Methods in Medical Research , 4th ed.
JNS. Statistical
Blackwell, Oxford, 2002.
[15] Bland JM, Altman DG. One and tw
twoo sided tests of significance.. BMJ
significance BMJ 1994;
1994; 309: 248.
M. An Introduction to Medical Statistics , 3rd. ed. Oxford University Press, Ox-
[16] Bland M. An
ford, 2000.
[17] Elasho
Elashoff
ff JD. nQuery Advisor Version 4.0 User’s Guide . Los Angeles, CA, 2000.
JD. nQuery
33
34
[33] Ada
Adayy LA. Chapter 7: Decid
Deciding
ing how man
many
y will be in the sample
sample.. In Designing and Con-
In Designing
ducting Health
Health Surveys: A Comprehensi
Comprehensive
ve Guide , 2nd Edition. San Fr Francisc
ancisco:
o: Josse
Jossey-
y-
Bass Publi
Publishers
shers,, 1996.
[34]] Gor
[34 Gordis L. Epidemiology . Philad
dis L. Philadelphi
elphia:
a: W.B. Saunders Company
Company,, 1996.
[35]] Fle
[35 Fletc
tcher
her RH, Fle
Fletc
tcher
her SW, Wagn
agner EH. Clinical
er EH. Clinical Epide
Epidemiolo
miology,
gy, The Esse
Essential
ntials
s , 3rd
Edition. Philadelphia: Williams & Wilkins, 1996.
[36] Campbel
Campbelll M, Grims
Grimshaw
haw J, Steen N, for the Changing Profess
Professional
ional Practi
Practice
ce in Europe
trials. J Health Serv Res Policy ,
Group. Sample size calculations for cluster randomised trials. J
2000;5(1):12-16.
[37] Ray JG, Verme
ermeulen
ulen MJ. Sample size estimatio
estimation
n for the sorce
sorcerers
rers apprentice.. Can Fam
apprentice
Phys , July 1999;45:1999.
[38] Lwa
Lwanganga S, Lemsho
Lemshow
w S. Sample Size Determination in Health Studies: a Practical Man-
S. Sample
ual . Geneva, Switzerland: World Health Organization, 1991.
[39] Sch
Schulz
ulz KF, Grimes DA. Sample size slippag
slippages
es in random
randomised
ised trials
trials:: Exclu
Exclusions
sions and the
wayward. Lancet 2002;
lost and wayward. Lancet 2002; 358: 781-5.
[40] Boen JR, ZaZahn DA.. The Human Side of Statistical Consulting , Life
hn DA Lifetime
time Learning Pub-
lications, Belmont, CA 1982.
[41] Boren
Borenstein
stein M, Roths
Rothstein
tein H, Cohen J. Power and Precision , Biostat, Teaneck, NJ, Soft-
J. Power
ware for MS-DOS systems 1997.
[42]] Cas
[42 Castel
telloe
loe J. Sam
Sample
ple Siz
Sizee Com
Comput
putati
ations
ons and PoPowe
werr Ana
Analys
lysis
is wit
with
h the SAS Sys
System
tem,,
in Proceedingsof the Twenty-Fifth Annual SAS Users Group International Conference
in Proceedingsof
2000, Cary,
Cary, NC, SAS Instit
Institute,
ute, Inc.,Pape
Inc.,Paperr 265-25.
[43] Cohen J. Statistical Power Analysis for the Behavioral Sciences , 2nd Edn, Academic
Press, New York, 1988.
[44] Desu MM, Ragha
Raghav D. Sample Size Methodology , Academic Press, Boston 1990.
varao D. Sample
[45]] Ela
[45 Elasho
shoff J. nQuery Advisor Release 4.0 , Statistical Solutions, Cork, Ireland, Software
ff J.
for MS-DOS systems 2000.
[46]] Frei
[46 reiman
man JA ChaChalme
lmers
rs TC, Smith (Jr) H, Kue
Kueble
blerr RR. The Impo
Importa
rtance
nce of Bet
Beta,a, the
Type II Error, and Sample Size in the Design and Interpretation of the Randomized
Controll
Con trolled
ed Trial: Surv
Survey
ey of 71 “Negative
“Negative T
Trials” in Medical Uses of Statistics , eds. J.
rials”,, in Medical
C. Bailar III and F. Mosteller, chap. 14, pp. 289304, NEJM Books, Waltham, Mass.
1986.
[47] Hin
Hintze
tze J. PASS 2000 , Number Cruncher Statistical Systems, Kaysville, UT, Software
J. PASS
for MS-DOS systems 2000.
[48] Hoenig JM, Heisey DM. The Abuse of Power: The Perv Pervasive
asive F
Fallacy
allacy of Power Calcula-
Analysis, The American Statistician 2001;
tionsin Data Analysis, The 2001; 55: 1924.
35
[49] Kraem
Kraemerer HC, Thi
Thiemann
emann S. How Many Subjects? Statistical Power Analysis in Research
S. How
,Sage Publications,
Publications, Newbury Park, CA 1987.
[50]
[50] Le
Len
nth RV. (2000
2000)), Java appl
applet
etss for po
pow
wer and
and samp
sample
le size
size.. Av
Aval
alia
ialb
lble
le at
http://www.stat.uiowa.edu/ rlenth/Power/ 2000. (Last accessed on September 7, 2003)
[51] Lipse
LipseyyMMW. Design Sensitivity: Statistical
W. Design Statistical Power for Exper
Experimental
imental Rese
Research
arch , Sage Pub-
lications, Newbury Park, CA 1990.
[52] Mace AE
AE.. (1964), Sample-size determination , Reinhold, New York, 1964.
(1964), Sample-size
[53] Muller KE, Benignus V. A. Increasing scien
scientific
tific power with statistical
statistical power, Neurotoxi-
power, Neurotoxi-
cology and Teratology 1992;
1992; 14: 211219.
[54] OBrie
OBrienn RG. UnifyPow.sas Version 98.08.25 , Department of Biostatistics and Epidemi-
RG. UnifyPow.sas
ology,Cleveland Clinic Foundation, Cleveland, OH, 1998. Available for download from
http://www.bio.ri.ccf.org/power.html (Last accessed on September 7, 2003)
[55] Odeh RE, F
Fox
ox M. Sample Size Choice: Charts for Experiments with Linear Models , 2nd
M. Sample
edn. Marcel Dekker, New York 1991
[56] Sch
Schuirmann
uirmann D. A compromise test for equivalence of average bioavailability,, ASA Pro-
bioavailability
ceedings of the Biopharmaceutical Section 1987;
1987; 137142.
[57] Tayl
aylor
or DJ, Mulle
Mullerr KE. Computin
Computingg Confid
Confidence
ence Bounds for Pow
Power er and Sample Size of
Model, The American Statistician 1995;
theGeneral Linear Univariate Model, The 1995; 49: 4347.
[58] Thoma
Thomass L. Retr
Retrospect
ospective
ive Pow
Power
er Analy
Analysis, Conservation Biology 1997;
sis, Conservation 1997; 11: 276280.
[59] Thomas L. Statistical po pow
wer analysis software. 1998. Avalialble at
http://w
http://www.f
ww.forest
orestry
ry.ubc.c
.ubc.ca/
a/ conse
conserv
rvation/
ation/pow
power/
er/ (Last acce
accessed
ssed on Septe
September
mber 7
2003)
[60] Thorn
Thornley
ley B, Adams C. ConConten
tentt and qualit
quality
y of 2000 con
controll
trolled
ed trials in sch
schizoph
izophrenia
renia
50years, BMJ 1998;
over 50years, BMJ 1998; 317: 11811184.
[61] Wheeler RE. Port
Portable
able Pow
Power, Technometrics 1974; 16:193-201.
er, Technometrics
[62] Wrigh
rightt TA. simple algorit
algorithm
hm for tighte
tighterr exact upper confid
confidence
ence bounds with rare at-
universes, Statistics and Probability Letters 1997;
tributes infinite universes, Statistics 1997; 36: 5967
[63]] Whi
[63 Whitle
tley
y E, Bal
Balll J. Sta
Statis
tistic
ticss rev
review
iew 4: Sam
Sample
ple size calcu
calculat
lations. Critical Care
ions. Care 2002;
6:335-341
[64] Dupont WD, Plumm
Plummerer WD, Jr. Pow
Powerer and Sample Size Calcula
Calculations:
tions: A Review and
Program. Controlled Clinical Trials 11:116-128,
Computer Program. Controlled 11:116-128, 1990
[65] Dupont WD, Plummer WD, Jr. Pow Power
er and Sample Size Calculations for Studies Involv-
Involv-
ing Linear Regression. Controlled
Regression. Controlled Clinical Trials 19:589-601,
19:589-601, 1998
[66] Sch
Schoenfel
oenfeldd DA, Rich
Richter
ter JR. Nomogr
Nomogramsams for calculat
calculating
ing the num
number
ber of patient
patientss neede
needed
d
endpoint. Biometrics 38:163-170,
for a clinical trial with survival as an endpoint. Biometrics 38:163-170, 1982
36
[67] Pea
Pearson
rson ES, Hartl
Hartley
ey HO. Biometrika Tables for Statisticians Vol.
HO. Biometrika Vol. I 3rd Ed. Cambridge:
Cambridge University Press, 1970
[68] Sch
Schlesse
lesselman JJ. Case-Control Studies: Design, Conduct, Analysis . New York: Oxford
lman JJ.
University Press, 1982
[69] Casagrande JT, Pike MC, Smith PG. An improved approximate
approximate formula for calculating
sample sizes for comparing two binomial distributions. Biometrics 1978;
1978; 34:483-
34:483-486
486
[70] Dupont WD. Po
Power
wer calcula
calculations
tions for matc
matched
hed case-con
case-control
trol studies. Biometrics 44:1157-
studies. Biometrics 44:1157-
1168,
[71] Gore SM
SM.. Assess
Assessing
ing clin
clinical
ical tr
trials
ials trial si
size. BMJ 1981;
ze. BMJ 1981; 282: 1687-1689.
[72] Friedm
riedman
an L, Furberg C, DeMets D. Fundamentals of clinical trials . 3rd ed. New York:
D. Fundamentals
Springer-Verlag; 1998.
[73] Gebski V, Marschne
Marschnerr I, Keech A
AC.C. Specifying objectives and outcomes for clinical trials.
Med J Aust 2002;
2002; 176: 491-492.
[83] Hutto
Huttonn LJ. Are Disti
Distinctiv
nctivee Ethic
Ethical
al Princ
Principles
iples Requir
Required
ed for Clust
Cluster
er Randomize
Randomized
d Con-
Trials. Statistics in Medicine 2001;
troled Trials. Statistics 2001; 20: 473-488
37
[84] Hay
Hayes
es RJ, Benne
Bennett
tt S. Simpl
Simplee Sample Size Calculati
Calculations
ons for Cluster-
Cluster-rando
randomized
mized T
Trials
rials..
Int J Epidermiology 1999;
1999; 28:319-
28:319-326
326
[85] Recrui
Recruiting
ting Huma
Human
n Subj
Subjeects: Sampl
Samplee Guid
Guideline
eliness for Pr
Practi
actice
ce . OEI-0
OEI-01-97-0
1-97-00196,
0196, June
2000
[86] American Psychological Association.
Association. Publication
Publication manual of the American Psychological
Association (5th
(5th ed.). Washington, DC: Author 2001.
[87] Aron AA,, Aron E
EN. Statistics for psychology (2nd
N. Statistics (2nd ed.). Upper Saddle River, NJ: Prentice
Hall 1999.
J. Statistical power analysis for the behavioral sciences (2nd
[88] Cohen J. Statistical (2nd ed.). Hillsdale, NJ:
Erlbaum 1988.
[89] Cohen J. Things I hav far). American Psychologist 1990;
havee learned (so far). 1990; 45: 1304-1312.
[90] Cohen J. A powe
powerr primer. Psychological Bulletin 1992;112:
primer. Psychological 1992;112: 155-159.
[91] Cohen J, Cohen P. Applied multiple regression/correlation analysis for the behavioral
sciences . Hillsdale, NJ: Erlbaum 1975.
[92] Comre
Comreyy AL, Lee HB. A first course in factor analysis (2nd
HB. A (2nd ed.). Hillsdale, NJ: Erlbaum
1992.
[93] Green SB. How many subjects does it take to do a regression analysis? Multivariate
Behavioral Research 1991;
1991; 26: 499-510.
[94] Guadag
Guadagnoli
noli E, Velic
elicer
er WF. Relat
Relation
ion of sampl
samplee size to the stabilit
stability
y of compone
component
nt pat-
terns. Psychological Bulletin 1988;
terns. Psychological 1988; 103: 265-275.
[95]] Har
[95 Harris RJ. A primer of multivariate statistics
ris RJ. statistics (2nd
(2nd ed.). New Y
York:
ork: Acade
Academic
mic Press
1985.
[96] How
Howell DC.. Statistical methods for psychology (4th
ell DC (4th ed.). Belmont, CA: Wadsworth 1997.
[97] Ho
[97] Hoyle
yle EH (Ed .). Statistical strategies for small sample research . Thousand Oaks, CA:
(Ed.).
Sage 1999.
[98] Kraem
Kraemer
er HC, Thiem
Thiemann
ann S. How many subje
S. How subjects?
cts? Statistic
Statistical
al power analysis in res
resear
earch
ch .
Newbury Park, CA: Sage 1987.
[99] Ped
Pedhazur
hazur EJ, Schmelk
Schmelkinin LP
LP.. Measur
Measuremen
ement,
t, desi
design,
gn, and analys
analysis:
is: An inte
integr
grate
atedd ap-
proach . Hillsdale, NJ: Erlbaum 1991.
[100] Tabac
abachnic
hnick
k BG, Fidell LS. Using multivariate statistics (3rd
LS. Using (3rd ed.). New York: Harper-
Collins
Collins 1996.
[101] Wilkinson L, Task F
Force
orce on Statistical Inference, AP
APA
A Board of Scientific Affairs. Sta-
tistical
tistical methods
chologist 1999; 54:in594-604.
1999; psyc
psychology
hology journal
journals:
s: Guide
Guidelines
lines and expla
explanation
nations. American Psy-
s. American
38
L. Research mistakes in the social and behavioral sciences . Ames: Iowa
[102] Wolins L. Iowa Stat
Statee
University Press 1982.
[103] Shus
Shuster JJ.. Handbook of Smaple Size Guidelines for Clinical Trials . CRC Press, Boca
ter JJ
Raton: Florida 1990
[104] Lemeshow S, Hosme
Hosmerr DW (Jr.)
(Jr.),, Klar J, Lwanga SK.
SK. Adequacy
Adequacy of Smaple Size in Health
Studies . World Health Organization, Wiley, New York: NY 1990.
[105] Hair JF, Anderson RE, T Tatham
atham RL, Black WC. Multivariate Data Analysis , 5th Edi-
WC. Multivariate
tion, Prentice Hall, New Jersey; 1998.
39