Definition:Infinitely Often (Probability Theory)
Jump to navigation
Jump to search
Definition
Let $\struct {\Omega, \Sigma, \Pr}$ be a probability space.
Let $\sequence {E_n}_{n \mathop \in \N}$ be a sequence in $\Sigma$.
We define:
\(\ds \set {E_n \text { infinitely often} }\) | \(=\) | \(\ds \set {\omega \in \Omega : \omega \in E_n \text { for infinitely many } n}\) | ||||||||||||
\(\ds \) | \(=\) | \(\ds \limsup_{n \mathop \to \infty} E_n\) |
where $\ds \limsup_{n \mathop \to \infty} E_n$ is the limit superior of $\sequence {E_n}_{n \mathop \in \N}$.
Notation
We may abbreviate "infinitely often" as "i.o." and write:
- $\set {E_n \text { infinitely often} } = \set {E_n \text { i.o.} }$
Sources
- 1991: David Williams: Probability with Martingales ... (previous) ... (next): $2.6$