Almost sure convergence: Difference between revisions
imported>Subpagination Bot m (Add {{subpages}} and remove any categories (details)) |
imported>Hendra I. Nurdin (Improvements) |
||
Line 1: | Line 1: | ||
{{subpages}} | {{subpages}} | ||
'''Almost sure convergence''' is one of the four main modes of [[stochastic convergence]]. | '''Almost sure convergence''' is one of the four main modes of [[stochastic convergence]]. It may be viewed as a notion of convergece for random variables that is similar to, but not the same as, the notion of [[pointwise convergence]] for real functions. | ||
==Examples== | <!--==Examples==--> | ||
===Basic example=== | <!--===Basic example===--> | ||
===Intermediate example=== | |||
<!--===Intermediate example===--> | |||
==Definition== | ==Definition== | ||
In this section, a formal definition of almost sure convergence will be given for complex vector-valued random variables, but it should be noted that a more general definition can also be given for random variables that take on values on more abstract sets. To this end, let <math>(\Omega,\mathcal{F},P)</math> be a [[measure space|probability space]] (in particular, <math>(\Omega,\mathcal{F}</math>) is a [[measurable space]]). A (<math>\mathbb{C}^n</math>-valued) '''random variable''' is defined to be any [[measurable function]] <math>X:(\Omega,\mathcal{F})\rightarrow (\mathbb{C}^n,\mathcal{B}(\mathbb{C}^n))</math>, where <math>\mathcal{B}(\mathbb{C}^n)</math> is the [[Borel set]] of <math>\mathbb{C}^n</math>. A formal definition of almost sure convergence can be stated as follows: | |||
A sequence <math>X_1,X_2,\ldots,X_n,\ldots</math> of random variables is said to be '''converge almost surely''' to a random variable <math>Y</math> if <math>\mathop{\lim}_{n \rightarrow \infty}X_k(\omega)=Y(\omega)</math> for all <math>\omega \in \Lambda</math>, where <math>\Lambda \subset \Omega</math> is some set satisfying <math>P(\Lambda)=1</math>. An equivalent definition is that the sequence <math>X_1,X_2,\ldots,X_n,\ldots</math> converge almost surely to <math>Y</math> if <math>\mathop{\lim}_{n \rightarrow \infty}X_k(\omega)=Y(\omega)</math> for all <math>\omega \in \Omega \backslash \Lambda'</math>, where <math>\Lambda'</math> is some set with <math>P(\Lambda')=0</math>. | |||
==Important cases of almost sure convergence== | ==Important cases of almost sure convergence== | ||
Line 34: | Line 39: | ||
==External links== | <!--==External links==--> |
Revision as of 07:15, 16 October 2007
Almost sure convergence is one of the four main modes of stochastic convergence. It may be viewed as a notion of convergece for random variables that is similar to, but not the same as, the notion of pointwise convergence for real functions.
Definition
In this section, a formal definition of almost sure convergence will be given for complex vector-valued random variables, but it should be noted that a more general definition can also be given for random variables that take on values on more abstract sets. To this end, let be a probability space (in particular, ) is a measurable space). A (-valued) random variable is defined to be any measurable function , where is the Borel set of . A formal definition of almost sure convergence can be stated as follows:
A sequence of random variables is said to be converge almost surely to a random variable if for all , where is some set satisfying . An equivalent definition is that the sequence converge almost surely to if for all , where is some set with .
Important cases of almost sure convergence
If we flip a coin n times and record the percentage of times it comes up heads, the result will almost surely approach 50% as .
This is an example of the strong law of large numbers.
References
See also
- Stochastic convergence
- Convergence in distribution
- Convergence in probability
- Convergence in rth order mean
Related topics