Probability
This page pairs well with Statistics.
Elements of Probability Theory
A probability space $(S,\mathcal{F},P)$ consists of a sample space $S$, a $\sigma$-algebra $\mathcal{F}\subseteq 2^S$, and a function $P:\mathcal{F}\to[0,1]$ such that:
- $P(\varnothing)=0$, $P(S)=1$;
- $P(E)\ge 0$ for all $E\in\mathcal{F}$;
- ($\sigma$-additivity) For any countable family of pairwise disjoint events $\{E_i\}$, $P\!\left(\bigcup_i E_i\right)=\sum_i P(E_i)$.
Random Variables
The CDF of $X$ is \[ F_X(x)=P(X\le x),\quad x\in\mathbb{R}. \] Key properties:
- non-decreasing;
- right-continuous;
- $\lim_{x\to-\infty}F_X(x)=0$;
- $\lim_{x\to+\infty}F_X(x)=1$.
$F_X$ completely characterises the distribution.
For a r.v. $X$ with CDF $F_X$, \[\mathbb{E}[X]=\int_{-\infty}^{\infty} x\,dF_X(x) =\begin{cases}\displaystyle \sum_{x\in S_X} x\,p_X(x), & \text{(discrete)}\\[0.5em] \displaystyle \int_{S_X} x\,f_X(x)\,dx, & \text{(continuous).}\end{cases}\]
Linearity: $\mathbb{E}[aX+b]=a\,\mathbb{E}[X]+b$.
Existence: $\mathbb{E}[X]$ is finite iff $\mathbb{E}[|X|]<\infty$.
For continuous $X$, the $100k\%$ quantile is $Q_X(k)=F_X^{-1}(k)$, $0<k<1$. A QQ-plot compares sample order statistics $\{x_{(k)}\}$ to theoretical quantiles $\{F^{-1}(p_k)\}$ (e.g.\ $p_k=(k-0.5)/n$); points should lie roughly on a straight line if the model fits.
\(X_n \xrightarrow{\text{a.s.}} X \;\Rightarrow\; X_n \xrightarrow{P} X \;\Rightarrow\; X_n \xrightarrow{d} X.\) Definitions:
- In distribution: \(F_{X_n}(x)\to F_X(x)\) at continuity points of \(F_X\).
- In probability: \(\forall\varepsilon>0,\ \mathbb P(|X_n-X|>\varepsilon)\to0.\)
- Almost surely: \(\mathbb P(\lim_{n\to\infty}X_n=X)=1.\)
Random Vectors
Marginals: $F_{X_1}(x_1)=\lim_{x_2\to +\infty}F_{X_1X_2}(x_1,x_2)$ (and symmetrically for $X_2$).
Independence: $X_1$ and $X_2$ are independent iff $F_{X_1X_2}(x_1,x_2)=F_{X_1}(x_1)F_{X_2}(x_2)$ for all $(x_1,x_2)$ (equivalently, $p_{X_1X_2}=p_{X_1}p_{X_2}$ in discrete case; $f_{X_1X_2}=f_{X_1}f_{X_2}$ in continuous case).
For $g:\mathbb{R}^2\to\mathbb{R}$, \[\mathbb{E}[g(X_1,X_2)]= \begin{cases}\displaystyle \sum_{x_1,x_2} g(x_1,x_2)\,p_{X_1X_2}(x_1,x_2), & \text{(discrete)}\\[0.5em] \displaystyle \iint g(x_1,x_2)\,f_{X_1X_2}(x_1,x_2)\,dx_2\,dx_1, & \text{(continuous).}\end{cases}\]
In particular, $\mathbb{E}[aX_1+bX_2]=a\,\mathbb{E}[X_1]+b\,\mathbb{E}[X_2]$.
Common Distributions
A Bernoulli r.v. \(X\sim \mathrm{Bern}(\pi)\) takes values in \(\{0,1\}\) with \(\mathbb P(X=1)=\pi,\ \mathbb P(X=0)=1-\pi\).
Moments: \(\mathbb E[X]=\pi,\ \mathrm{Var}(X)=\pi(1-\pi)\).
mgf: \(\varphi_X(s)=(1-\pi)+\pi e^{s}\).
If interarrival times of a Poisson process have rate \(\lambda>0\), then \(T\sim \mathrm{Exp}(\lambda)\) with cdf/pedf \[ F_T(t)=1-e^{-\lambda t}\ (t\ge0),\qquad f_T(t)=\lambda e^{-\lambda t}\ 1_{\{t\ge 0\}}. \]
Standard normal: \(Z\sim N(0,1)\) with \[ \varphi(z)=\frac{1}{\sqrt{2\pi}}e^{-z^2/2},\quad \Phi(z)=\int_{-\infty}^{z}\varphi(u)\,du. \] General normal: \(X\sim N(\mu,\sigma^2)\) has \[ f(x)=\frac{1}{\sqrt{2\pi}\,\sigma}\exp\!\left(-\frac{(x-\mu)^2}{2\sigma^2}\right). \] Standardisation: \(Z=(X-\mu)/\sigma\sim N(0,1)\).
Transformations of Random Variables
If \(Y=\phi(X)\) where \(X=(X_1,X_2)\), \(Y=(Y_1,Y_2)\), \(\phi\) is one-to-one and differentiable with inverse \(\psi=\phi^{-1}\), then with Jacobian \(J(y)=\det\big[\partial \psi_i/\partial y_j\big]\), \[ f_Y(y)=f_X\!\big(\psi(y)\big)\,|J(y)|. \]
If \(X_1\perp\!\!\!\perp X_2\):
- Discrete: \(p_{X_1+X_2}(y)=\displaystyle\sum_{x} p_{X_1}(y-x)\,p_{X_2}(x)\).
- Continuous: \(f_{X_1+X_2}(y)=\displaystyle\int f_{X_1}(y-x)\,f_{X_2}(x)\,dx\).
Backlinks
1. Statistics /wiki/mathematics/statistics/
This page pairs well with Probability.
Table of Distributions
| Distribution | mass/density function | $$S_X$$ | $$\mathbb{E}(X)$$ | $$\mathrm{Var}(X)$$ | $$\phi_X(s)$$ |
|---|---|---|---|---|---|
| Bernoulli $$\mathrm{Bern}(\pi)$$ | $$P(X=1)=\pi\\P(X=0)=1-\pi$$ | $$\{0,1\}$$ | $$\pi$$ | $$\pi(1-\pi)$$ | $$(1-\pi)+\pi e^{s}$$ |
| Binomial $$\mathrm{Bin}(n,\pi)$$ | $$p_X(x)=\binom{n}{x}\pi^{x}(1-\pi)^{n-x}$$ | $$\{0,1,\dots,n\}$$ | $$n\pi$$ | $$n\pi(1-\pi)$$ | $$(1-\pi+\pi e^{s})^{n}$$ |
| Geometric $$\mathrm{Geo}(\pi)$$ | $$p_X(x)=\pi(1-\pi)^{x-1}$$ | $$\{1,2,\dots\}$$ | $$\pi^{-1}$$ | $$(1-\pi)\pi^{-2}$$ | $$\frac{\pi}{e^{-s}-1+\pi}$$ |
| Poisson $$\mathcal{P}(\lambda)$$ | $$p_X(x)=e^{-\lambda}\lambda^{x}/x!$$ | $$\{0,1,\dots\}$$ | $$\lambda$$ | $$\lambda$$ | $$\exp\{\lambda(e^{s}-1)\}$$ |
| Uniform $$U[\alpha,\beta]$$ | $$f_X(x)=(\beta-\alpha)^{-1}$$ | $$[\alpha,\beta]$$ | $$\frac{1}{2}(\alpha+\beta)$$ | $$\frac{1}{12}(\beta-\alpha)^2$$ | $$\frac{e^{\beta s}-e^{\alpha s}}{s(\beta-\alpha)}$$ |
| Exponential $$\mathrm{Exp}(\lambda)$$ | $$f_X(x)=\lambda e^{-\lambda x}$$ | $$[0,\infty)$$ | $$\lambda^{-1}$$ | $$\lambda^{-2}$$ | $$\frac{\lambda}{\lambda-s}$$ |
| Gaussian $$\mathcal{N}(\mu,\sigma^{2})$$ | $$f_X(x)=\\\frac{1}{\sqrt{2\pi\sigma^{2}}}\exp\left\{-\frac{(x-\mu)^2}{2\sigma^{2}}\right\}$$ | $$\mathbb{R}$$ | $$\mu$$ | $$\sigma^{2}$$ | $$e^{\mu s+\frac{1}{2}\sigma^{2}s^{2}}$$ |
| Gamma $$\Gamma(\alpha,\lambda)$$ | $$f_X(x)=\\\frac{1}{\Gamma(\alpha)}\lambda^{\alpha}x^{\alpha-1}e^{-\lambda x}$$ | $$[0,\infty)$$ | $$\alpha\lambda^{-1}$$ | $$\alpha\lambda^{-2}$$ | $$\left(\frac{\lambda}{\lambda-s}\right)^{\alpha}$$ |
Statistical Inference
2. Wiki /wiki/