- Bias of an estimator
In

statistics , the difference between anestimator 'sexpected value and the true value of the parameter being estimated is called the**bias**. An estimator or decision rule having nonzero bias is said to be**biased**.Although the term "bias" sounds pejorative, it is not necessarily used in that way in statistics. Biased estimators may have desirable properties. Not only do they sometimes have a smaller

mean squared error than any unbiased estimator, but in some cases the only unbiased estimators are not even within theconvex hull of the parameter space, so their use is absurd.**Definition**Suppose we are trying to estimate the parameter $heta$ using an

estimator $widehat\{\; heta\}$ (that is, some function of the observed data). Then the**bias**of $widehat\{\; heta\}$ is defined to be:$operatorname\{E\}(widehat\{\; heta\})-\; heta.,$

In words, this would be "the expected value of the estimator $widehat\{\; heta\}$ minus the true value $heta$." This may be rewritten as

:$operatorname\{E\}(widehat\{\; heta\}-\; heta).,$

which would read "the expected value of the difference between the estimator and the true value" (the expected value of $heta$ is precisely $heta$ ).

**Examples****Estimating variance**Suppose "X"

_{1}, ..., "X"_{"n"}are independent and identically distributed normal random variables with expectation μ andvariance σ^{2}. Let:$overline\{X\}=(X\_1+cdots+X\_n)/n$

be the "sample average", and let

:$S^2=frac\{1\}\{n\}sum\_\{i=1\}^n(X\_i-overline\{X\},)^2$

be a "sample variance". We also know that the variance σ

^{2}is defined by::$\{\}sigma^2\; =\; frac\; 1N\; sum\_\{i=1\}^N\; left(x\_i\; -\; overline\{x\}\; ight)^\; 2\; ,$where N is the population size and "x"_{i}represents the member of the whole population.Then "S"

^{2}is a "biased estimator" of σ^{2}because:$operatorname\{E\}(S^2)=frac\{n-1\}\{n\}sigma^2\; eqsigma^2.$

In other words, the "sample variance" does not equal the "population variance", unless multiplied by the normalization factor.

Common sense would suggest to apply the population formula to the sample as well. The reason that it is biased is that the sample mean is generally somewhat closer to the observations in the sample than the population mean is, to these observations. This is so because the sample mean is, by definition, in the middle of the sample, while the population mean may even lie outside the sample. So the deviations to the sample mean will often be smaller than the deviations to the population mean, and so, if the same formula is applied to both, then this variance estimate will on average be somewhat smaller in the sample than in the population.

Note that when a transformation is applied to an unbiased estimator, the result is not necessarily itself an unbiased estimate of its corresponding population statistic. That is, for a non-linear function "f" and an unbiased estimator "U" of a parameter "p", "f"("U") is usually not an unbiased estimator of "f"("p"). For example the

square root of the unbiased estimator of the populationvariance is not an unbiased estimator of the populationstandard deviation .**Estimating a Poisson probability**A far more extreme case of a biased estimator being better than any unbiased estimator is well-known: Suppose "X" has a

Poisson distribution with expectation λ. It is desired to estimate:$operatorname\{P\}(X=0)^2=e^\{-2lambda\}.quad$

(For example, when incoming calls at a telephone switchboard are modeled as a Poisson process, and λ is the average number of calls per minute, then "e"

^{−2λ}is the probability that no calls arrive in the next two minutes.)Since the expectation of an unbiased estimator $delta(X)$ is equal to the estimand, i.e.:$E(delta(X))=sum\_\{x=0\}^infty\; delta(x)\; frac\{lambda^x\; e^\{-lambda\{x!\}=e^\{-2lambda\}$,

the only function of the data constituting an unbiased estimator is

:$delta(x)=(-1)^Xquad$.

If the observed value of "X" is 100, then the estimate is 1, although the true value of the quantity being estimated is obviously very likely to be near 0, which is the opposite extreme. And if "X" is observed to be 101, then the estimate is even more absurd: it is −1, although the quantity being estimated obviously must be positive.

The (biased) maximum likelihood estimator

:$e^\{-2X\}quad$

is far better than this unbiased estimator. Not only is its value always positive, but it is also more accurate in the sense that its

mean squared error (MSE):$e^\{-4lambda\}-2e^\{lambda(1/e^2-3)\}+e^\{lambda(1/e^4-1)\}$

is smaller; compare the unbiased estimator's MSE of

:$1-e^\{-4lambda\}$.

The MSEs are a functions of the true value λ. The bias of the maximum-likelihood estimator is:

:$e^\{-2lambda\}-e^\{lambda(1/e^2-1)\}$.

**Maximum of a discrete uniform distribution**The bias of maximum-likelihood estimators can be substantial. Consider a case where "n" tickets numbered from 1 through to "n" are placed in a box and one is selected at random, giving a value "X". If "n" is unknown, then the maximum-likelihood estimator of "n" is "X", even though the expectation of "X" is only ("n" + 1)/2; we can only be certain that "n" is at least "X" and is probably more. In this case, the natural unbiased estimator is 2"X" − 1.

**See also***

Omitted-variable bias

*Consistency (statistics) **External links*** [

*http://www.math.umn.edu/~hardy/An_Illuminating_Counterexample.pdf An Illuminating Counterexample*]

*Wikimedia Foundation.
2010.*