Almost surely





In probability theory, one says that an event happens almost surely (sometimes abbreviated as a.s.) if it happens with probability one. In other words, the set of possible exceptions may be non-empty, but it has probability zero. The concept is precisely the same as the concept of "almost everywhere" in measure theory.


In probability experiments on a finite sample space, there is often no difference between almost surely and surely. However, the distinction becomes important when the sample space is an infinite set, because an infinite set can have non-empty subsets of probability zero.


Some examples of the use of this concept include the strong and uniform versions of the law of large numbers, and the continuity of the paths of Brownian motion.


The terms almost certainly (a.c.) and almost always (a.a.) are also used. Almost never describes the opposite of almost surely: an event that happens with probability zero happens almost never.[1]




Contents






  • 1 Formal definition


  • 2 Illustrative examples


    • 2.1 Throwing a dart


    • 2.2 Tossing a coin repeatedly




  • 3 Asymptotically almost surely


  • 4 See also


  • 5 Notes


  • 6 References





Formal definition


Let ,F,P){displaystyle (Omega ,{mathcal {F}},P)}(Omega ,{mathcal {F}},P) be a probability space. An event E∈F{displaystyle Ein {mathcal {F}}}Ein {mathcal {F}} happens almost surely if P(E)=1{displaystyle P(E)=1}P(E)=1. Equivalently, E{displaystyle E}E happens almost surely if the probability of E{displaystyle E}E not occurring is zero: P(EC)=0{displaystyle P(E^{C})=0}{displaystyle P(E^{C})=0}. More generally, any event E⊆Ω{displaystyle Esubseteq Omega }{displaystyle Esubseteq Omega } (not necessarily in F{displaystyle {mathcal {F}}}{mathcal {F}}) happens almost surely if EC{displaystyle E^{C}}E^C is contained in a null set: a subset of some N∈F{displaystyle Nin {mathcal {F}}}Ninmathcal F such that P(N)=0{displaystyle P(N)=0}{displaystyle P(N)=0}.[2] The notion of almost sureness depends on the probability measure P{displaystyle P}P. If it is necessary to emphasize this dependence, it is customary to say that the event E{displaystyle E}E occurs P-almost surely, or almost surely (P).



Illustrative examples


In general, an event can happen "almost surely" even if the probability space in question includes outcomes which do not belong to the event, as is illustrated in the examples below.



Throwing a dart


Imagine throwing a dart at a unit square (i.e. a square with area 1) so that the dart always hits exactly one point of the square, and so that each point in the square is equally likely to be hit.


Now, notice that since the square has area 1, the probability that the dart will hit any particular subregion of the square equals the area of that subregion. For example, the probability that the dart will hit the right half of the square is 0.5, since the right half has area 0.5.


Next, consider the event that "the dart hits a diagonal of the unit square exactly". Since the areas of the diagonals of the square are zero, the probability that the dart lands exactly on a diagonal is zero. So, the dart will almost never land on a diagonal (i.e. it will almost surely not land on a diagonal). Nonetheless the set of points on the diagonals is not empty and a point on a diagonal is no less possible than any other point: the diagonal does contain valid outcomes of the experiment.



Tossing a coin repeatedly


Consider the case where a (possibly biased) coin is tossed, corresponding to the probability space ({H,T},2{H,T},P){displaystyle ({H,T},2^{{H,T}},P)}{displaystyle ({H,T},2^{{H,T}},P)}, where the event {H}{displaystyle {H}}{displaystyle {H}} occurs if heads is flipped, and {T}{displaystyle {T}}{T} if tails. For this particular coin, assume the probability of flipping heads is P(H)=p∈(0,1){displaystyle P(H)=pin (0,1)}{displaystyle P(H)=pin (0,1)} from which it follows that the complement event, flipping tails, has P(T)=1−p{displaystyle P(T)=1-p}{displaystyle P(T)=1-p}.


Suppose we were to conduct an experiment where the coin is tossed repeatedly, with outcomes ω1,ω2,…{displaystyle omega _{1},omega _{2},ldots }{displaystyle omega _{1},omega _{2},ldots }, and it is assumed each flip's outcome is independent of all the others. That is, they are i.i.d.. Define the sequence of random variables on the coin toss space, (Xi)i∈N{displaystyle (X_{i})_{iin mathbb {N} }}{displaystyle (X_{i})_{iin mathbb {N} }} where Xi(ω)=ωi{displaystyle X_{i}(omega )=omega _{i}}X_i(omega)=omega_i. i.e. each Xi{displaystyle X_{i}}X_{i} records the outcome of the i{displaystyle i}i'th flip.


Any infinite sequence of heads and tails is a possible outcome of the experiment. However, any particular infinite sequence of heads and tails has probability zero of being the exact outcome of the (infinite) experiment. To see why, note that the i.i.d. assumption implies that the probability of flipping all heads over n{displaystyle n}n flips is simply P(Xi=H, i=1,2,…,n)=(P(X1=H))n=pn{displaystyle P(X_{i}=H, i=1,2,dots ,n)=left(P(X_{1}=H)right)^{n}=p^{n}}{displaystyle P(X_{i}=H, i=1,2,dots ,n)=left(P(X_{1}=H)right)^{n}=p^{n}}. Letting n→{displaystyle nrightarrow infty }nrightarrow infty yields zero, since p∈(0,1){displaystyle pin (0,1)}pin (0,1) by assumption. Note that the result is the same no matter how much we bias the coin towards heads, so long as we constrain p{displaystyle p}p to be greater than 0, and less than 1.


In particular, the event "the sequence contains at least one T{displaystyle T}T" happens almost surely (i.e., with probability 1).
However, if instead of an infinite number of flips we stop flipping after some finite time, say a million flips, then the all-heads sequence has non-zero probability. The all-heads sequence has probability p1,000,000≠0{displaystyle p^{1,000,000}neq 0}p^{1,000,000}neq 0, while the probability of getting at least one tails is 1−p1,000,000{displaystyle 1-p^{1,000,000}}1 - p^{1,000,000} and the event is no longer almost sure.



Asymptotically almost surely


In asymptotic analysis, one says that a property holds asymptotically almost surely (a.a.s.) if, over a sequence of sets, the probability converges to 1. For instance, a large number is asymptotically almost surely composite, by the prime number theorem; and in random graph theory, the statement "G(n,pn){displaystyle G(n,p_{n})}{displaystyle G(n,p_{n})} is connected" (where G(n,p){displaystyle G(n,p)}G(n,p) denotes the graphs on n{displaystyle n}n vertices with edge probability p{displaystyle p}p) is true a.a.s. when, for any ε>0{displaystyle varepsilon >0}varepsilon >0



pn<(1+ε)ln⁡nn{displaystyle p_{n}<{tfrac {(1+varepsilon )ln n}{n}}}{displaystyle p_{n}<{tfrac {(1+varepsilon )ln n}{n}}}.[3]

In number theory this is referred to as "almost all", as in "almost all numbers are composite". Similarly, in graph theory, this is sometimes referred to as "almost surely".[4]



See also





  • Convergence of random variables, for "almost sure convergence"


  • Degenerate distribution, for "almost surely constant"


  • Almost everywhere, the corresponding concept in measure theory


  • Infinite monkey theorem, a theorem using the aforementioned terms



Notes





  1. ^ Grädel, Erich; Kolaitis, Phokion G.; Libkin, Leonid; Marx, Maarten; Spencer, Joel; Vardi, Moshe Y.; Venema, Yde; Weinstein, Scott (2007). Finite Model Theory and Its Applications. Springer. p. 232. ISBN 978-3-540-00428-8..mw-parser-output cite.citation{font-style:inherit}.mw-parser-output q{quotes:"""""""'""'"}.mw-parser-output code.cs1-code{color:inherit;background:inherit;border:inherit;padding:inherit}.mw-parser-output .cs1-lock-free a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/6/65/Lock-green.svg/9px-Lock-green.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output .cs1-lock-limited a,.mw-parser-output .cs1-lock-registration a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/d/d6/Lock-gray-alt-2.svg/9px-Lock-gray-alt-2.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output .cs1-lock-subscription a{background:url("//upload.wikimedia.org/wikipedia/commons/thumb/a/aa/Lock-red-alt-2.svg/9px-Lock-red-alt-2.svg.png")no-repeat;background-position:right .1em center}.mw-parser-output .cs1-subscription,.mw-parser-output .cs1-registration{color:#555}.mw-parser-output .cs1-subscription span,.mw-parser-output .cs1-registration span{border-bottom:1px dotted;cursor:help}.mw-parser-output .cs1-hidden-error{display:none;font-size:100%}.mw-parser-output .cs1-visible-error{font-size:100%}.mw-parser-output .cs1-subscription,.mw-parser-output .cs1-registration,.mw-parser-output .cs1-format{font-size:95%}.mw-parser-output .cs1-kern-left,.mw-parser-output .cs1-kern-wl-left{padding-left:0.2em}.mw-parser-output .cs1-kern-right,.mw-parser-output .cs1-kern-wl-right{padding-right:0.2em}


  2. ^ Jacod, Jean; Protter, (2004). Probability Essentials. Springer. p. 37. ISBN 978-3-540-438717.


  3. ^ Friedgut, Ehud; Rödl, Vojtech; Rucinski, Andrzej; Tetali, Prasad (January 2006). "A Sharp Threshold for Random Graphs with a Monochromatic Triangle in Every Edge Coloring". Memoirs of the American Mathematical Society. AMS Bookstore. 179 (845): 3–4. ISSN 0065-9266.


  4. ^ Spencer, Joel H. (2001). "0. Two Starting Examples". The Strange Logic of Random Graphs. Algorithms and Combinatorics. 22. Springer. p. 4. ISBN 978-3540416548.




References




  • Rogers, L. C. G.; Williams, David (2000). Diffusions, Markov Processes, and Martingales. 1: Foundations. Cambridge University Press. ISBN 978-0521775946.


  • Williams, David (1991). Probability with Martingales. Cambridge Mathematical Textbooks. Cambridge University Press. ISBN 978-0521406055.




Comments

Popular posts from this blog

Information security

Volkswagen Group MQB platform

刘萌萌