- Checking if a coin is fair
In
statistics , a fair coin is an idealized randomizing device with two states (usually named "heads" and "tails") which are equally likely to occur. It is based on the ubiquitouscoin flip used in sports and other situations where it is necessary to give two parties the same chance of winning. Depending on the occasion a specially designed chip or a simple currencycoin is used, which due to unequal weight distribution might be "unfair": one state might occur more frequently than the other, giving one party an unfair advantage. So it might be necessary to experimentally determine whether the coin is in fact "fair" – that is, if the probability of the coin falling on either side in the toss is approximately 50%. It is of course impossible to ever definitively rule out arbitrarily small deviations from fairness such as might be expected to affect only one flip in a lifetime of flipping, and it is always possible for an unfair (or "biased") coin to happen to turn up exactly 10 heads in 20 flips. As such, any fairness test must only establish a certain degree of confidence in a certain degree of fairness (a certain maximum bias). In more rigorous terminology, the problem is of determining the parameters of aBernoulli process , given only a limited sample ofBernoulli trial s.Preamble
This article describes experimental procedures for determining if a coin is fair. There are many statistical methods for analyzing such an experimental procedure. This article illustrates two of them.
Both methods prescribe an experiment (or trial) in which the coin is tossed many times and the result of each and every toss is recorded. A statistical analysis of the results can then be performed to decide if the coin is "fair" or "probably not fair".
* Posterior probability density function. This method assumes that the number of tosses is fixed and not under the experimenter's direct control. The true probability of obtaining a particular side when a fair coin is tossed (the "prior distribution") is already known. The probability that this particular coin is a "fair coin" can then be obtained by integrating the "posterior" PDF over the relevant interval.
* Estimator of true probability. This method assumes that the experimenter can decide and implement any number of coin tosses for this particular coin. The experimenter decides on the level of confidence required and the tolerable margin of error. These considerations determine the minimum number of tosses that must be performed to complete the experiment.
Posterior probability density function
One way of verifying this is to calculate the posterior
probability density function ofBayesian probability theory .A test is performed by tossing the coin "n" times and noting the number of heads "h" and tails "t":
:"H" = "h" (Total number of heads is "h"):"T" = "t" (Total number of tails is "t"):"N" = "n" = "h" + "t" (Total number of tosses is "n")
Next, let "r" be the actual probability of obtaining heads in a single toss of the coin. This is the value desired. Using
Bayes' theorem , posterior probability of "r" conditional on "H" and "T" is expressed as follows::
The prior summarizes what is known about the distribution of "r" in the absence of any observation. We will assume that the
prior distribution of "r" is uniform over the interval [0, 1] . That is, "f"("r") = 1. (In fact, we could use a prior distribution that reflects our experience with real coins.)The probability of obtaining "h" heads in "n" tosses of a coin with a probability of heads equal to "r" is given by a
binomial distribution ::
Putting it all together:
:
This is in fact a
beta distribution (theconjugate prior for the binomial distribution), whose denominator can be expressed in terms of thebeta function ::
If a uniform prior is assumed, and because "h" and "t" are integers, this can also be written in terms of
factorial s::
Example
For example, let "n"=10, "h"=7, i.e. the coin is tossed 10 times and 7 heads are obtained:
:
The graph on the right shows the
probability density function of "r" given that 7 heads were obtained in 10 tosses. (Note: "r" is the probability of obtaining heads when tossing the same coin once.)The probability for an unbiased coin
:
is small when compared with alternative hypothesis (a biased coin). However, it is not small enough to cause us to actually believe that the coin has a significant bias. Notice that this probability is slightly "higher" than our presupposition of the probability that the coin was fair corresponding to the uniform prior distribution, which was 10%.Using a prior distribution that reflects our prior knowledge of what a coin is and how it acts, the posterior distribution would not favor the hypothesis of bias. (But also notice that the number of trials done in this example is relatively small, and with more trials the choice of a prior distribution would be less relevant.)
Estimator of true probability
And hence the value of maximum error (E) is given by
Therefore, the final formula for the number of coin tosses for the estimator is
:
provided that and where to satisfy the
Central Limit Theorem .Example
1. If a maximum error of 0.01 is desired, how many times should the coin be tossed?
:
: at 68.27% level of confidence (Z=1): at 95.45% level of confidence (Z=2): at 99.90% level of confidence (Z=3.3)
2. If the coin is tossed 10000 times, what is the maximum error of the estimator on the value of (the actual probability of obtaining heads in a coin toss)?
::: at 68.27% level of confidence (Z=1): at 95.45% level of confidence (Z=2): at 99.90% level of confidence (Z=3.3)
3. The coin is tossed 12000 times with a result of 5961 heads (and 6039 tails). What interval does the value of (the true probability of obtaining heads) lie within if a confidence level of 99.999% is desired?
:
Now find the value of Z corresponding to 99.999% level of confidence. :
Now calculate E
:
The interval which contains r is thus:
:
:
Hence, 99.999% of the time, the interval above would contain which is the true value of obtaining heads in a single toss.
Other applications
The above mathematical analysis for determining if a coin is fair can also be applied to other uses. For example:
* Determining the product defective rates of a product when subjected to a particular (but well defined) condition. Sometimes a product can be very difficult or expensive to produce. Furthermore if testing such products will result in their destruction, a minimum amount of products should be tested. Using the same analysis the probability density function of the product defect rate can be found.
* Two party polling. If a small random sample poll is taken where the there are only two mutually exclusive choices, then this is equivalent to tossing a single coin multiple times using a bias coin. The same analysis can therefore be applied to determine actual voting ratio.
* Finding the proportion of females in an animal group. Determining the gender ratio in a large group of an animal species. Provided that a very small random sample is taken when performing the random sampling of the population, the analysis is similar to determining the probability of obtaining heads in a coin toss.
ee also
*
Binomial test
*Coin flipping
*Confidence interval
*Estimation theory
*Inferential statistics
*Margin of error
*Point estimation
*Statistical randomness References
*Guttman, Wilks, and Hunter: "Introductory Engineering Statistics", John Wiley & Sons, Inc. (1971) ISBN 0471337706
*Devinder Sivia: "Data Analysis, a Bayesian Tutorial", Oxford University Press (1996) ISBN 0198518897External links
* [http://www.statisticalengineering.com/frequentists_and_bayesians.htm frequentists and bayesians]
Wikimedia Foundation. 2010.