Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
VK Chadha*
55
For a quantitative variable, it is only the One of the factors influencing the width of
rough estimates of the proportion, means or rates the confidence interval is the sample size. The
that are required for estimating sample size. larger the sample size, the narrower is the
confidence interval.
Sampling error
The multiplying factor 1.96 is used when
An estimate of an outcome measure
calculating the 95 percent confidence interval.
calculated in an intervention study is subject to
In some circumstances, confidence intervals
sampling error, because it is based on a sample
other than 95 percent limits may be required and
of individuals and not on the whole population of
then values of the multiplying factor are as under:
interest. The term does not mean that the
sampling procedure was applied incorrectly, but Confidence interval (%) Multiplying factor
that when sampling is used to decide which
90 1.64
individuals are in which group, there will be an
element of random variation in the results. 95 1.96
Sampling error is reduced when the study size is
99 2.58
increased and vice versa.
99.9 3.29
Confidence Interval
Confidence intervals and their corresponding
The methods of statistical inference allow
multiplying factors, based on the Normal distribution
the investigator to draw conclusions about the
true value of the outcome measure on the basis Precision of effect measures – The
of the information in the sample. In general, the narrower the confidence interval, the greater the
observed value of the outcome measure gives precision of the estimate.
the best estimate of the true value. In addition,
Significance tests & P value
it is useful to have some indication of the precision
of this estimate, and this is done by attaching a In some instances, before calculating a
confidence interval to the estimate. The confidence interval to indicate a range of plausible
confidence interval is a range of plausible values values of the outcome measure of interest, it
for the true value of the outcome measure. It is may be appropriate to test a specific hypothesis
conventional to quote the 95 percent confidence about the outcome measure. In the context of
interval (also called 95 percent confidence limits). an intervention trial, this will often be the
This is calculated in such a way that there is a hypothesis that there is no true difference
95 percent probability that it includes the true between the outcomes in the groups under
value. comparison- null hypothesis. The objective is
If the outcome measure is a proportion thus to assess whether any observed difference
estimated from the sample data as ` . The 95 in outcomes between the study groups may have
percent confidence intervals to be presented here occurred just by chance due to sampling error.
are = ` ± 1.96 x SE, where SE denotes the The methods for testing the null hypothesis are
standard error of the estimate. Similarly, if the known as significance tests. The sample data
outcome measure is a mean, 95 percent of the are used to calculate a quantity (called a statistic)
values derived from different samples are which gives a measure of the difference between
expected to fall within 1.96 standard deviations the groups with respect to the outcome(s) of
of the mean. interest. Once the statistic has been calculated,
56
its value is referred to an appropriate set of obtaining a significant result of a study, even if
statistical tables, in order to determine the p – there is a real difference. It is necessary to
value (probability value) or ‘significance’ of the consider the probability of obtaining a statistically
results. significant result in a trial, and this probability is
called the power of the study. In other words, if
For example, suppose a difference in mean
the true difference exits, power of the study
PCV of 1.5 percent is observed at the end of the
indicates the probability of finding a statistically
malaria season between two groups of
significant difference between the two groups.
individuals, one of which was supplied with
Thus a power of 80 percent to detect a difference
mosquito-nets. A p-value of 0.03 would indicate
of a specified size means that if the study were
that, if nets had no true effect on PCV levels, (if
to be conducted repeatedly, a statistically
null hypothesis was true) there would only be a
significant result would be obtained four times
3 percent chance of obtaining an observed
out of five if the true difference was really of the
difference of 1.5 percent or greater.
specified size. When designing a study, the
The smaller the p-value, the less plausible objective is to ensure that the study size is large
the null hypothesis seems as an explanation of enough to give high power if the true effect of
the observed data. For example, a p-value of the intervention is large enough to be of practical
0.001 means that the null hypothesis is highly importance.
implausible, and this can be interpreted as very
The power of a study depends on:
strong evidence of a real difference between the
groups. On the other hand, a p-value of 0.20 1. The value of the true difference between
means that a difference of the observed the study groups; in other words, the true
magnitude could quite easily have occurred by effect of the intervention (effect size). The
chance, even if there was no real difference greater the effect, the higher the power to
between the groups. Conventionally, p-values of detect the effect as statistically significant
0.05 and below have been regarded as sufficiently for a study of a given size.
low to be taken as reasonable evidence against
2. The study size; The larger the study size,
the null hypothesis, and have been referred to
higher is the power.
as indicating a ‘statistically significant difference’.
3. The probability level at which a difference
While a small p-value can be interpreted
will be regarded as ‘statistically significant’.
as evidence for a real difference between the
groups, a larger ‘non-significant’ p-value must not The power also depends on whether a one-
be interpreted as indicating that there is no sided or two sided significance test is to be
difference. It merely indicates that there is performed and on the underlying variability of
insufficient evidence to reject the null hypothesis, the data.
so that there may be no true difference between One sided and Two sided tests
the groups.
If it is accepted that the null hypothesis is
Power of study false, that means alternate hypothesis is true.
The concept of power comes into play For example if the claim is about superiority of
when the focus of the study is to find out whether a new drug, this is a one-sided alternative. If
a significant difference exists between the two the claim is not of superiority or inferiority but
groups. Because of the variations resulting from only that they are different, the alternate
sampling error, we cannot always be certain of hypothesis is two sided. This means that when
57
the p-value is computed, it measures the needs to be guarded against. For this reason, P
probability (if the null hypothesis is true) of value is kept at a low level (<5%). When P
observing a difference as great as that actually value is small, it is safe to conclude that groups
observed in either direction (i.e. positive or are different. This threshold, 0.05 is the level of
negative). It is usual to assume that tests are significance.
two-sided.
The second type of error is failure to reject
Wrongly rejecting a true null hypothesis is null hypothesis when it is actually false. This
called type I error. The probability of this error is corresponds to missed diagnosis as also to
referred as P value as already discussed. The pronounce a criminal not guilty. The probability
maximum P value allowed in a problem is called of this error is denoted by β. In a clinical trial set
the level of significance (α). In a diagnostic set up, this is equivalent to declaring a drug
up, this is the probability of declaring a person ineffective when it is actually effective. The
sick when he is actually not. In a clinical trial set complimentary probability of type II error is the
up, P value is the probability that the drug is statistical power (1-β). Thus the power of a
declared effective or better when it is actually statistical test is a probability of correctly rejecting
not. a null hypothesis when it is false.
59
confidence, 1.645 for 90% confidence and 2.576 tuberculous infection among children of 0-9 years
for 99% confidence. of age in a locality, how many should be included
in the sample so that prevalence may be obtained
Example: For a survey aimed at estimating
within 10% of true value with 95% confidence.
vaccination coverage, P is usually anticipated at
The anticipated P is 5-10%
0.5, d=0.1 and the level of significance = 5%.
The estimated sample size using the above P= 0.05 (lower limit is to be taken to have
formula is 96 for random sampling. The larger sample & better precision)
estimated sample size is applicable only in case Confidence level is 95%
of simple random sampling (SRS). If another
Relative Precision (å) = 10% = 0.1
sampling method is used, a larger sample size
is likely to be needed because of design effect. The estimated sample size using formula
For cluster sampling strategy, the estimated 2 is 3457. For cluster sampling the sample size
sample size as above is multiplied by design will be 3457x D (design effect).
effect, which is defined as the ratio of variance III. Estimating the difference between two
obtained in cluster survey to the variance for the population proportions with specified
same sample size adopting SRS. absolute precision (Two – sample
situations)- equal n in the two groups
Sample design var iance
Design effect = Required information :-
var iance usin g SRS
- Anticipated population proportions = P1
In most sample surveys adopting cluster- & P2
sampling strategy, a design effect of 2 is taken.
- Confidence level = 95%
This means twice as many individuals would have
to be studied to obtain the same precision as - Absolute precision required on either
with SRS. Thus, the estimated sample size in side of the true value of the difference
the above example would be 96x2=192. However, between proportions = d
design effect should ideally be estimated from Sample size can be estimated using
previously available data or from pilot studies. following formula:
II. Estimating a population proportion: With
specified relative precision Z12 − αα [P1 (1 − P1 ) + P2 (1 − P2 )]
1−2
n = 2 ––– (3)
Required information :- d2
- Anticipated population proportion : P P1, P2 = anticipated value of the proportions
in the two populations.
- Confidence level
Example : What sample size to be selected
- Relative Precision: ε – The sample result
from each of two groups of people to estimate a
should fall within ε % of the true value.
risk difference to be within 3 percentage points
Sample size can be estimated using of true difference at 95% confidence when
following formula : anticipated P1 & P2 are 40% & 32% respectively.
- Power of the test: 100 (1-β) % VII. Estimating difference between means of
two populations with specified
Sample size can be estimated using
precision.
following formula :
Difference in means = µ1 − µ2
− −
{Z α 2P (1− P )] + Z1−β P1 (1− P1 ) + P2 (1− P2 ) { 2
Standard Deviation (SD) = s1, s2
1−
2
n= – (4) Absolute precision: d
(P1 − P2 )2
First calculate pooled variance and estimate
Values of Z for 90% power = 1.28 sample size using formula
80% power = 0.82
(n1 − 1)s12 − (n 2 − 1)s 22
Pooled variance (σ 2) =
V. Estimating a population mean: With (n1 − 1) + (n 2 − 1)
specified absolute precision
Z12 − αα [2σ 2 ]
1 −
Required information :- n = 22 ––––– (7)
d2
- Variance : σ , known or can be
2
estimated from a pilot study Example : Suppose you want to know the
sample sizes (equal groups) required for detecting
- Absolute precision : d
a mean difference of 0.3 mg / ml between well-
Sample size can be estimated using nourished and under-nourished children. This
formula: difference is considered to be clinically important.
Suppose the mean and SD of Hemoglobin (Hb)
Z12 − αα × σ 2 levels available from an earlier study in a random
1− 2
n = 2 ––––– (5) sample of well-nourished and under-nourished
d2 groups were as follows:
σ = standard deviation estimated from a Well-nourished (group-1) :
pilot study n1 = 100, x-1=10.1 and SD1= 0.9
VI. Estimating population mean: With Under-nourished (group-2) :
specified relative precision
n2 = 70, x-2= 9.7 and SD2= 1.1
Required information :- The SDs do not differ too much and we
- Population mean: µ can pool them. Thus,
Pooled variance (σ2 )
- Variance : σ2
= (99 x 0.92 +69 x 1.12) / (100 + 69 )
- Relative precision: (ε)
= 0.97
1.962 [2 × (0.97)2 ]
n = = 80
(0.3) 2
61
VIII. Hypothesis testing for two population would be used to determine the size.
means
SAMPLE SIZE ESTIMATION FOR OTHER
Required information :- SITUATIONS
Anticipated values of the population For more complex study designs, for
means : µ1 & µ2 example two groups of unequal size, comparison
Standard deviation: s1, s2 of more than two groups, incidence studies, and
interventions allocated to communities; the
Level of significance
methods of sample size estimation are more
Power of the test: 100 (1-β) % complex and may be referred to in a standard
Sample size can be estimated using statistical text book.
formula: There are computer programs available that
perform sample size calculations. In particular,
s12 − s 22 this facility is available in the package ‘Epi Info’,
Pooled variance (σ ) =
2
2 though it does not cover the full range of
possibilities.
2
SAMPLE SIZE IN QUALITATIVE RESEARCH
2σ Z α + Z1 − β
2
62