- Almost surely
In
probability theory , one says that an event happens almost surely (a.s.) if it happens with probability one. The concept is analogous to the concept of "almost everywhere " inmeasure theory . It is often encountered in questions that involveinfinite time, regularity properties or infinite-dimension al spaces such asfunction space s. Basic examples of use include thelaw of large numbers (strong form) or continuity of Brownian paths.Formal definition
Let ("Ω", "F", "P") be a
probability space . One says that an event "E" in "F" happens almost surely if "P"("E") = 1. Alternatively, an event "E" happens almost surely if the probability of "E" not occurring iszero .An alternate definition from a measure theoretic-perspective is that (since "P" is a measure over "Ω") "E" happens almost surely if "E" = "Ω"
almost everywhere ."Almost sure" versus "sure"
The difference between an event being "almost sure" and "sure" is the same as the subtle difference between something happening "with probability 1" and happening "always".
If an event is "sure", then it will always happen. No other event (even events with probability 0) can possibly occur. If an event is "almost sure", then other events are theoretically possible in a given sample space, however as the cardinality of the sample space increases, the probability of any other event asymptotically converges toward zero. Thus, one can never definitively say for any sample space that other events will never occur, but can in the general case assume this to be true. In this respect, the concept is similar to that of a mathematical limit.
Throwing a dart
For example, imagine throwing a dart at a unit square wherein the dart will impact exactly one point, and imagine that this square is the only thing in the universe. There is physically nowhere else for the dart to land. Then, the event that "the dart hits the square" is a sure event. No other alternative is imaginable.
Next, consider the event that "the dart hits the diagonal of the unit square exactly". The probability that the dart lands on any subregion of the square is equal to the area of that subregion. But, since the area of the diagonal of the square is zero, the probability that the dart lands exactly on the diagonal is zero. So, the dart will almost surely not land on the diagonal. Nonetheless the set of points on the diagonal is not empty and a point on the diagonal is no less possible than any other point.
The same may be said of any point on the square. Any such point "P" will contain zero area and so will have zero probability of being hit by the dart. However, the dart clearly must hit the square somewhere. Therefore, in this case, it is not only possible or imaginable that an event with zero probability will occur; one must occur. Thus, we would not want to say we were certain that a given such event would not occur, but rather "almost certain".
Tossing a coin
Suppose that an "ideal" (edgeless)
fair coin is flipped again and again. A coin has two sides, heads and tails, and therefore the event that "heads or tails is flipped" is a sure event. There can be no other result from such a coin.The infinite sequence of all heads ("H-H-H-H-H-H-…"), "ad infinitum", is possible in some sense (it does not violate any physical or mathematical laws to suppose that "tails" never appears) but it is very, very improbable. In fact, the probability of "tails" never being flipped in an infinite series is zero. Thus, though we cannot definitely say tails will be flipped at least once, we can say there will "almost surely" be at least a single tails flip in an infinite sequence of flips.
However, if instead of an infinite number of flips we stop flipping after some finite time, say a million flips, then the all-heads sequence has non-zero probability. The all-heads sequence has probability 2−1,000,000, thus the probability of getting a tails is 1 − 2−1,000,000 < 1, and the event is no longer "almost sure".
Asymptotically almost surely
In
asymptotic analysis , one says that a property holds asymptotically almost surely (a.a.s.) if, over a sequence of sets, the probability converges to 1. For instance, a large number is asymptotically almost surely composite, by theprime number theorem ; and in random graph theory, the statement "G"("n","p""n") is connected" (where "G"("n","p") denotes the graphs on "n" vertices with edge probability "p") is true a.a.s when "p"n > for any ε > 0.cite journal|last=Friedgut|first=Ehud|coauthors=Rödl, Vojtech; Rucinski, Andrzej; Tetali, Prasad|date=January 2006|title=A Sharp Threshold for Random Graphs with a Monochromatic Triangle in Every Edge Coloring|journal=Memoirs of the American Mathematical Society|publisher=AMS Bookstore|volume=179|issue=845|pages=pp. 3–4|issn=0065-9266|accessdate=2008-09-21]In
number theory this is referred to as "almost all ", as in "almost all numbers are composite". Similarly, in graph theory, this is sometimes referred to as "almost surely".cite book|last=Spencer|first=Joel H.|title=The Strange Logic of Random Graphs|publisher=Springer|date=2001|series=Algorithms and Combinatorics|pages=p. 4|chapter=0. Two Starting Examples|accessdate=2008-09-21]See also
*
Convergence of random variables , for "almost sure convergence"
*Constant random variable , for "almost surely constant"
*Almost everywhere , the corresponding concept in measure theoryNotes
References
*
*
Wikimedia Foundation. 2010.