- Large deviations theory
In
Probability Theory , the Large Deviations Theory concerns the asymptotic behaviour of remote tails of sequences of probability distributions. Some basic ideas of the theory can be tracked back to Laplace and Cramér, although a clear unified formal definition was introduced in1966 by Varadhan [S.R.S. Varadhan, "Asymptotic probability and differential equations", Comm. Pure Appl. Math. 19 (1966),261-286.] . Large Deviations Theory formalizes the heuristic ideas of "concentration of measures" and widely generalizes the notion of convergence of probability measures.Roughly speaking, Large Deviation Theory concerns itself with the exponential decay of the probability measures of certain kinds of extreme or "tail" events, as the number of observations grows arbitrarily large.
Introductory examples
An elementary example
Consider a sequence of independent tosses of a faircoin. The possible outcomes could be head or tail. Let us denote the possible outcome of the i-th trial by, where we encode head as -1 and tail as 1. Now let denote the mean value after trials, namely
:
Then lies between -1 and 1. From the
law of large numbers (and alsofrom our experience) we know that as N become larger and larger, becomes closer and closer to with increasing probability. Let us make this statement more precise. For a given value , let us compute the probability that is greater than . By theChernoff inequality it can be shown that . This bound is rather sharp, in a suitable technical sense.In other words the probability is decaying exponentially rapidly as N grows large, at a rate depending on x.Large Deviations for sums of independent random variables
In the above mentioned example of coin-tossing we tacitly assumed that each toss is an independent trial. And for each toss, the probability of getting head or tail is always thesame. This makes the random numbers independent and identically distributed (i.i.d.). For i.i.d. variableswhose common distribution satisfies a certain growth condition, large deviation theory states that the following limit exists:
:
The function is called the "
rate function " or "Cramer function" or sometimes the "entropy function". Roughly speaking, the existence of this limit is what establishes the above mentioned exponential decay and allows us to conclude that for large , takes the form::
which is the basic result of Large Deviations Theory in this setting. Note that the inequality given in the first paragraph, as opposed to the asymptotic formula presented here, requires an additional argument.
If we know the probability distribution of , an explicit expression for the rate function can be obtained. This is given by a
Legendre transform :
where the function is called the
cumulant generating function (CGF), given by:
Here denotes expectation value with respect to the probability distribution function of and is any one of s. If follows a
Gaussian distribution ,the rate function becomes a parabola with its apex at the mean of the Gaussiandistribution.If the condition of Independent Identical Distribution is relaxed, particularly if the numbers are not independent but nevertheless satisfy the
Markov Property , the basic large deviations result stated above can be generalized.Formal Definition
Given a
Polish space let be a sequence of Borel probability measures on , let be a sequence of positive real numbers such that , and finally let be alower semicontinuous functional on . The sequence is said to satisfy aLarge deviation principle with "speed" and "rate" ,iff for each Borelmeasurable set :
where and denote respectively the closure and interior of .
Brief History
The first rigorous results concerning Large Deviations are due to the Swedish mathematician Harald Cramér, who applied them to model the insurance business. From the pointof view of an insurance company, the earning is at a constant rate per month(the monthly premium) but the claims come randomly. For the company to be successful over a certain period of time (preferably many months), the total earning shouldexceed the total claim. Thus to estimate the premium you have to ask the followingquestion : "What should we choose as the premium such that over months the total claim shouldbe less than ? " This is clearly the same question asked by the large deviations theory. Cramer gave a solution to this question for i.i.d.
gaussian random variable s, where the rate function is expressed as apower series .The results we have quoted above were later obtained by Chernoff, among other people. A very incomplete list of mathematicians who have made important advances would includeS.R.S. Varadhan (who has won the Abel prize),D. Ruelle andO.E. Lanford .Applications
Establishing Large Deviations Principles is one of the most effective ways to gather information out of a probabilistic model. Some of the best known applications of Large Deviation Theory rise in
Statistical Mechanics ,Quantum Mechanics ,Information Theory andRisk Management .Applications to Statistical Mechanics: Large Deviation and Entropy
The rate function is related to the
entropy in statistical mechanics. This can be heuristically seenin the following way. In statistical mechanics the entropy of a particular macro-state is relatedto the number of micro-states which corresponds to this macro-state. In our coin tossing example themean value could designate a particular macro-state. And the particular sequence ofheads and tails which gives rise to a particular value of constitutes a particularmicro-state. Loosely speaking a macro-state having more number of micro-states giving rise to it,has higher entropy. And a state with higher entropy has more chance of being realised in actualexperiments. The macro-state with mean value of zero (as many heads as tails) has the highest number micro-states giving rise to it and it is indeed the state with the highest entropy. And in most practical situationwe shall indeed obtain this macro-state for large number of trials. The "rate function" on the otherhand measures the probability of appearance of a particular macro-state. The smaller the rate functionthe higher is the chance of a macro-state appearing. In our coin-tossing the value of the "rate function" for mean valueequal to zero is zero. In this way one can see the "rate function" as the negative of the "entropy".References
Bibliography
* Entropy, Large Deviations and Statistical Mechanics by R.S. Ellis, Springer Publication. ISBN 3-540-29059-1
* Large Deviations for Performance Analysis by Alan Weiss and Adam Shwartz. Chapman and Hall ISBN 0-412-06311-5
* Large Deviations Techniques and Applications by Amir Dembo and Ofer Zeitouni. Springer ISBN 0-387-98406-2
* Random Perturbations of Dynamical Systems by M.I. Freidlin and A.D. Wentzell. Springer ISBN 0-387-98362-7See also
*
Chernoff's inequality
*Contraction principle (large deviations theory) , a result on how large deviations principles "push forward "
*Freidlin-Wentzell theorem , a large deviations principle forItō diffusion s
* Laplace principle, a large deviations principle in R"d"
*Schilder's theorem , a large deviations principle forBrownian motion
*Varadhan's lemma
*Extreme value theory External links
* [http://www.cl.cam.ac.uk/Research/SRG/netos/old-projects/measure/tutorial/rev-tutorial.ps.gz An elementary introduction to the Large Deviations Theory]
* [http://www.abelprisen.no/en/prisvinnere/2007/ Abel Prize 2007 awarded to S.R.S. Varadhan]
Wikimedia Foundation. 2010.