Sei sulla pagina 1di 3

30 SAMPLES. STANDARD, SUGGESTION, OR SUPERSTITION ?

If you’ve ever taken any applied statistics courses in college, you may have been exposed to the
mystique of 30 samples. Too many times I’ve heard statistician do-it-yourselfers tell me that
“you need 30 samples for statistical significance.” Maybe that’s what they were taught; maybe
that’s how they remember what they were taught. In either case, the statement merits more than a
little clarification, starting with the 30-samples part. Suffice it to say that if there were any way
to answer the how-many-samples-do-I-need question that simply, you would find it in every
textbook on statistics, not to mention TV quiz shows and fortune cookies. Still, if you do an
Internet search for “30 samples” you’ll get millions of hits.
Like many legends, there is some truth behind the myth. The 30-sample rule-of-thumb may have
originated with William Gosset, a statistician and Head Brewer for Guinness. In a 1908 article
published under the pseudonym Student (Student. 1908. Probable error of a correlation
coefficient. Biometrika 6, 2-3, 302–310.), he compared the variation associated with 750
correlation coefficients calculated from sets of 4 and 8 data pairs, and 100 correlation
coefficients calculated from sets of 30 data pairs, all drawn from a dataset of 3,000 data pairs.
Why did he pick 30 samples? He never said but he concluded, “with samples of 30 … the mean
value [of the correlation coefficient] approaches the real value [of the population] comparatively
rapidly,” (page 309). That seems to have been enough to get the notion brewing.
Since then, there have been two primary arguments put forward to support the belief that you
need 30 samples for a statistical analysis. The first argument is that the t-distribution becomes a
close fit for the Normal distribution when the number of samples reaches 30. (The t-distribution,
sometimes referred to as Student’s distribution, is also attributable to W. S. Gosset. The t-
distribution is used to represent a normally distributed population when there are only a limited
number of samples from the population.) That’s a matter of perspective.
This figure shows the difference between the Normal distribution and the t-distribution for 10 to
200 samples. The differences between the distributions are quite large for 10 samples but
decrease rapidly as the number of samples increases. The rate of the decrease, however, also
diminishes as the number of samples increases. At 30 samples, the difference between the
Normal distribution and the t-distribution (at 95% of the upper tail) is about 3½%. At 60
samples, the difference is about 1½%. At 120 samples, the difference is less than 1%. So from
this perspective, using 30 samples is better than 20 samples but not as good as 40 samples.
Clearly, there is no one magic number of samples that you should use based on this argument.
The second argument is based on the Law of Large Numbers, which in essence says that the
more samples you use the closer your estimates will be to the true population values. This sounds
a bit like what Gosset said in 1908, and in fact, the Law of Large numbers was 200 years old by
that time.

This figure shows how differences between means estimated from different numbers of samples
compare to the population mean. (These data were generated by creating a normally distributed
population of 10,000 values, then drawing at random 100 sets of values for each number of
samples from 2 to 100 (i.e., 100 datasets containing 2 samples, 100 datasets containing 3
samples, and so on up to 100 datasets containing 100 samples. Then, the mean of the datasets
was calculated for each number of samples. Unlike Gosset, I got to use a computer and some
expensive statistical software.) The small inset graph shows the largest and smallest means
calculated for datasets of each sample size. The large graph shows the difference between the
largest mean and the smallest mean calculated for each sample size. These graphs show that
estimates of the mean from a sampled population will become more precise as the sample size
increases (i.e., the Law of Large Numbers). The important thing to note is that the precision of
the estimated means increases very rapidly up to about ten samples then continues to increase,
albeit at a decreasing rate. Even with more than 70 or 80 samples, the spread of the estimates
continues to decrease. So again, there’s nothing extraordinary about using 30 samples.
So while Gosset may have inadvertently started the 30-samples tale, you have to give him a lot
of credit for doing all those calculations with pencil and paper. To William Gosset, I raise a pint
of Guinness.
Now we still have to deal with that how-many-samples-do-I-need question. As it turns out, the
number of samples you’ll need for a statistical analysis really all comes down to resolution.
Needless to say, that’s a very unsatisfying answer compared to … 30 samples.

Join the Stats With Cats group on Facebook.

http://statswithcats.wordpress.com/2010/07/11/30-samples-standard-suggestion-or-superstition/

Potrebbero piacerti anche