|
| 1 | +--- |
| 2 | +layout: proof |
| 3 | +mathjax: true |
| 4 | + |
| 5 | +author: "Joram Soch" |
| 6 | +affiliation: "BCCN Berlin" |
| 7 | +e_mail: "joram.soch@bccn-berlin.de" |
| 8 | +date: 2024-09-13 11:02:05 |
| 9 | + |
| 10 | +title: "Weak law of large numbers" |
| 11 | +chapter: "General Theorems" |
| 12 | +section: "Probability theory" |
| 13 | +topic: "Expected value" |
| 14 | +theorem: "Weak law of large numbers" |
| 15 | + |
| 16 | +sources: |
| 17 | + - authors: "Ostwald, Dirk" |
| 18 | + year: 2023 |
| 19 | + title: "Ungleichungen und Grenzwerte" |
| 20 | + in: "Wahrscheinlichkeitstheorie und Frequentistische Inferenz" |
| 21 | + pages: "Einheit (6), Folie 19-20" |
| 22 | + url: "https://www.ipsy.ovgu.de/ipsy_media/Methodenlehre+I/Wintersemester+2324/Wahrscheinlichkeitstheorie+und+Frequentistische+Inferenz/6_Ungleichungen_und_Grenzwerte.pdf" |
| 23 | + |
| 24 | +proof_id: "P467" |
| 25 | +shortcut: "mean-wlln" |
| 26 | +username: "JoramSoch" |
| 27 | +--- |
| 28 | + |
| 29 | + |
| 30 | +**Theorem:** Let $X_1, \ldots, X_n$ be [independent and identically distributed](/D/iid) [random variables](/D/rvar) with [expected value](/D/mean) $\mathrm{E}(X_i) = \mu$ and finite [variance](/D/var) $\mathrm{Var}(X_i) < \infty$ for $i = 1,\ldots,n$. The [sample mean](/D/mean-samp) is defined as |
| 31 | + |
| 32 | +$$ \label{eq:mean-samp} |
| 33 | +\bar{X} = \frac{1}{n} \sum_{i=1}^{n} X_i \; . |
| 34 | +$$ |
| 35 | + |
| 36 | +Then, for any positive number $\epsilon > 0$, the probability that the absolute difference of the [sample mean](/D/mean-samp) from the [expected value](/D/mean) $\mu$ is smaller than $\epsilon$ will approach one, as $n$ goes to infinity: |
| 37 | + |
| 38 | +$$ \label{eq:mean-wlln} |
| 39 | +\lim_{n \rightarrow \infty} \mathrm{Pr}\left( \left| \bar{X} - \mu \right| < \epsilon \right) = 1 \; . |
| 40 | +$$ |
| 41 | + |
| 42 | + |
| 43 | +**Proof:** Since $X_1, \ldots, X_n$ are [independent and identically distributed](/D/iid), they have the same mean, denoted as $\mu$, and the same variance, denoted as $\sigma^2$. Using the [linearity of the expected value](/P/mean-lin), the expected value of the sample mean becomes: |
| 44 | + |
| 45 | +$$ \label{eq:mean-samp-mean} |
| 46 | +\begin{split} |
| 47 | +\mathrm{E}\left( \bar{X} \right) |
| 48 | +&= \mathrm{E}\left( \frac{1}{n} \sum_{i=1}^{n} X_i \right) \\ |
| 49 | +&= \frac{1}{n} \sum_{i=1}^{n} \mathrm{E}\left( X_i \right) \\ |
| 50 | +&= \frac{1}{n} n \mathrm{E}\left( X_i \right) \\ |
| 51 | +&= \mu \; . |
| 52 | +\end{split} |
| 53 | +$$ |
| 54 | + |
| 55 | +Moreover, with the [scaling of the variance upon multiplication](/P/var-scal) and the [additivity of the variance under independence](/P/var-add), the variance of the sample mean becomes: |
| 56 | + |
| 57 | +$$ \label{eq:mean-samp-var} |
| 58 | +\begin{split} |
| 59 | +\mathrm{Var}\left( \bar{X} \right) |
| 60 | +&= \mathrm{Var}\left( \frac{1}{n} \sum_{i=1}^{n} X_i \right) \\ |
| 61 | +&= \frac{1}{n^2} \mathrm{Var}\left( \sum_{i=1}^{n} X_i \right) \\ |
| 62 | +&= \frac{1}{n^2} \sum_{i=1}^{n} \mathrm{Var}\left( X_i \right) \\ |
| 63 | +&= \frac{1}{n^2} n \sigma^2 \\ |
| 64 | +&= \frac{\sigma^2}{n} \; . |
| 65 | +\end{split} |
| 66 | +$$ |
| 67 | + |
| 68 | +[Chebyshev's inequality](/P/cheb-ineq) makes a statement about a random variable $X$ in relation to its mean and variance for any positive number $x > 0$: |
| 69 | + |
| 70 | +$$ \label{eq:cheb-ineq} |
| 71 | +\mathrm{Pr}\left( \left| X - \mathrm{E}(\bar{X}) \right| \geq x \right) = \frac{\mathrm{Var}(X)}{x} \; . |
| 72 | +$$ |
| 73 | + |
| 74 | +Applying this inequality to the [random variable](/D/rvar) $\bar{X}$, we have: |
| 75 | + |
| 76 | +$$ \label{eq:mean-wlln-s1} |
| 77 | +\begin{split} |
| 78 | +\mathrm{Pr}\left( \left| \bar{X} - \mathrm{E}(\bar{X}) \right| \geq x \right) &= \frac{\mathrm{Var}(\bar{X})}{x} \\ |
| 79 | +\mathrm{Pr}\left( \left| \bar{X} - \mu \right| \geq \epsilon \right) &= \frac{\sigma^2}{n \epsilon} \; . |
| 80 | +\end{split} |
| 81 | +$$ |
| 82 | + |
| 83 | +Since [the cumulative distribution function can be used to relate probabilities of inverse events](/P/cdf-probexc), i.e. $\mathrm{Pr}\left( X \geq x \right) = 1 - \mathrm{Pr}\left( X < x \right)$, we have: |
| 84 | + |
| 85 | +$$ \label{eq:mean-wlln-s2} |
| 86 | +\begin{split} |
| 87 | +1 - \mathrm{Pr}\left( \left| \bar{X} - \mu \right| < \epsilon \right) &= \frac{\sigma^2}{n \epsilon} \\ |
| 88 | +\mathrm{Pr}\left( \left| \bar{X} - \mu \right| < \epsilon \right) &= 1 - \frac{\sigma^2}{n \epsilon} \; . |
| 89 | +\end{split} |
| 90 | +$$ |
| 91 | + |
| 92 | +Now taking the limit for $n \rightarrow \infty$ on both sides, while considering that $\epsilon$ and $\sigma^2$ are finite, gives: |
| 93 | + |
| 94 | +$$ \label{eq:mean-wlln-s3} |
| 95 | +\begin{split} |
| 96 | +\lim_{n \rightarrow \infty} \mathrm{Pr}\left( \left| \bar{X} - \mu \right| < \epsilon \right) |
| 97 | +&= \lim_{n \rightarrow \infty} \left( 1 - \frac{\sigma^2}{n \epsilon} \right) \\ |
| 98 | +&= 1 - \lim_{n \rightarrow \infty} \frac{\sigma^2 / \epsilon}{n} \\ |
| 99 | +&= 1 \; . |
| 100 | +\end{split} |
| 101 | +$$ |
0 commit comments