A continuous probability distribution is a function FX:R[0,1]F_X: \mathbb{R} \to [0, 1] which describes a random variable XX. When a sample of XX is taken, the probability of XX lying below a given value is governed by the distribution, according to the law P[Xx]=FX(x)\mathbb{P}[X \leqslant x] = F_X(x).

Specifically, this means that FXF_X satisfies:

This FXF_X is called the cumulative distribution function, or CDF. When its derivative FXF'_X exists, it is given the label fXf_X and called the probability density function, or PDF.

On this page, I catalogue every probability density function which is important or famous enough to deserve a name, and provide a set of summary statistics and properties.


The Uniform Distribution

The uniform distribution is extremely simple: heuristically, every same-sized subinterval within a finite support interval has an equal probability of occurring. It was first named by James Victor Uspensky in his 1937 book Introduction to Mathematical Probability, though it has been used for centuries beforehand.

PDF:f(x)={1baaxb0otherwiseCDF:F(x)={0xaxabaaxb1bx \begin{align*} \text{PDF}: f(x) &= \begin{cases} \frac{1}{b-a} & a \leqslant x \leqslant b \\ 0 & \text{otherwise} \end{cases} \\ \text{CDF}: F(x) &= \begin{cases} 0 & x \leqslant a \\ \frac{x-a}{b-a} & a \leqslant x \leqslant b \\ 1 & b \leqslant x \end{cases} \\ \end{align*}

The distribution is parameterised by two real numbers:

A standard uniform distribution (usually implemented as the basic “random” function on computers) takes a=0a = 0 and b=1b = 1.

The moments of the uniform distribution are:

Mean=12(a+b)Variance=112(ba)2Skewness=0Kurtosis=1.8E[(Xμσ)n]={3nn+1n even0n odd \begin{align*} \text{Mean} &= \frac{1}{2}(a+b) \\ \text{Variance} &= \frac{1}{12}(b-a)^2 \\ \text{Skewness} &= 0 \\ \text{Kurtosis} &= 1.8 \\ \mathbb{E}\left[ \left( \frac{X-\mu}{\sigma} \right)^n \, \right] &= \begin{cases} \frac{\sqrt{3}^{n}}{n+1} & n \text{ even} \\ 0 & n \text{ odd} \\ \end{cases} \\ \end{align*}

with the quantiles given by

Q(p)=F1(p)=a+p(ba)Median=Q(1/2)=12(a+b)Mode=any value in [a,b]Support=[a,b] \begin{align*} Q(p) = F^{-1}(p) &= a + p(b-a) \\ \text{Median} &= Q(1/2) = \frac{1}{2}(a + b) \\ \text{Mode} &= \text{any value in } [a, \, b] \\ \text{Support} &= [a, \, b] \end{align*}

The uniform distribution has a few basic properties:

Suppose X1,X2,X3,X_1, X_2, X_3, \ldots is an iid. sequence of standard uniform random variables. Then

The uniform distribution frequently arises in situations where there’s no bias towards any particular value. For example, in a well-mixed thermal bath, the starting phase of oscillating particles is uniformly distributed in the interval [0,2π][0, \, 2\pi].


The Exponential Distribution

The exponential distribution models the time between independent events occurring at a constant average rate. It was first derived by Agner Krarup Erlang in 1909 while studying telephone call arrivals, though similar work appeared earlier in kinetic theory. The distribution is characterized by the memoryless property: knowing an exponential process has lasted time tt doesn’t change the distribution of the remaining time.

PDF:f(x)={λeλxx00x<0CDF:F(x)={1eλxx00x<0 \begin{align*} \text{PDF}: f(x) &= \begin{cases} \lambda e^{-\lambda x} & x \geqslant 0 \\ 0 & x < 0 \end{cases} \\ \text{CDF}: F(x) &= \begin{cases} 1 - e^{-\lambda x} & x \geqslant 0 \\ 0 & x < 0 \end{cases} \\ \end{align*}

The distribution is parameterised by a single positive real number:

Sometimes the distribution is instead parameterised by β=1λ\beta = \frac{1}{\lambda}, called the scale parameter. The standard exponential distribution takes λ=1\lambda = 1.

The moments of the exponential distribution are:

Mean=1/λVariance=1/λ2Skewness=2Kurtosis=9E[Xn]=n!/λn \begin{align*} \text{Mean} &= 1/\lambda \\ \text{Variance} &= 1/\lambda^2 \\ \text{Skewness} &= 2 \\ \text{Kurtosis} &= 9 \\ \mathbb{E}[X^n] &= n!/\lambda^n \\ \end{align*}

with the quantiles given by

Q(p)=F1(p)=ln(1p)λMedian=ln(2)λMode=0Support=[0,) \begin{align*} Q(p) = F^{-1}(p) &= -\frac{\ln(1-p)}{\lambda} \\ \text{Median} &= \frac{\ln(2)}{\lambda} \\ \text{Mode} &= 0 \\ \text{Support} &= [0, \, \infty) \end{align*}

The exponential distribution has several important properties:

Suppose X1,X2,X3,X_1, X_2, X_3, \ldots is an iid. sequence of exponential random variables with rate λ\lambda. Then:

The exponential distribution models many real-world phenomena:


The Poisson Distribution

The Poisson distribution models the number of events occurring in a fixed interval when these events happen at a constant average rate and independently of each other. It was introduced by Siméon Denis Poisson in 1838 in his work “Research on the Probability of Judgments in Criminal and Civil Matters”.

Note that the Poisson distribution is discrete: it takes only integer values. This means it has a probability mass function, rather than a density.

PMF:f(k)=λkeλk!,kN0CDF:F(k)=eλi=0kλii!=γ(k+1,λ)k! \begin{align*} \text{PMF}: f(k) &= \frac{\lambda^k e^{-\lambda}}{k!}, \quad k \in \mathbb{N}_0 \\ \text{CDF}: F(k) &= e^{-\lambda} \sum_{i=0}^{\lfloor k \rfloor} \frac{\lambda^i}{i!} = \frac{\gamma(\lfloor k \rfloor + 1, \lambda)}{\lfloor k \rfloor!} \\ \end{align*}

The distribution is parameterised by a single positive real number:

A standard Poisson distribution takes λ=1\lambda = 1.

The moments of the Poisson distribution are:

Mean=λVariance=λSkewness=1/λKurtosis=1/λ+3E[Xn]=k=0knλkeλk!=Bn(λ) \begin{align*} \text{Mean} &= \lambda \\ \text{Variance} &= \lambda \\ \text{Skewness} &= 1/\sqrt{\lambda} \\ \text{Kurtosis} &= 1/\lambda + 3 \\ \mathbb{E}[X^n] &= \sum_{k=0}^{\infty} k^n \frac{\lambda^k e^{-\lambda}}{k!} = B_n(\lambda) \\ \end{align*}


where Bn(λ)B_n(\lambda) is the nthn^{\mathrm{th}} Bell polynomial evaluated at λ\lambda.

with the quantiles given by

Q(p)=F1(p) has no closed formMedian=λ+1/30.02/λ (approximation)Mode=λSupport=N0={0,1,2,} \begin{align*} Q(p) &= F^{-1}(p) \text{ has no closed form} \\ \text{Median} &= \lfloor \lambda + 1/3 - 0.02/\lambda \rfloor \text{ (approximation)} \\ \text{Mode} &= \lfloor \lambda \rfloor \\ \text{Support} &= \mathbb{N}_0 = \{0,1,2,\ldots\} \end{align*}

The Poisson distribution has several important properties:

Suppose we have a Poisson process with rate λ\lambda. Then:

The Poisson distribution models many real-world phenomena:


The Normal Distribution

The normal (or Gaussian) distribution is arguably the most important probability distribution in statistics. It was first introduced by Abraham de Moivre in 1733, and later popularised by Carl Friedrich Gauss in the early 1800s while studying astronomical measurement errors. The distribution’s ubiquity is explained by the Central Limit Theorem: the sum of many independent random variables almost always tends towards normality.

PDF:f(x)=1σ2πexp((xμ)22σ2),xRCDF:F(x)=12[1+erf(xμσ2)] \begin{align*} \text{PDF}: f(x) &= \frac{1}{\sigma\sqrt{2\pi}} \exp\left(-\frac{(x-\mu)^2}{2\sigma^2}\right), \quad x \in \mathbb{R} \\ \text{CDF}: F(x) &= \frac{1}{2}\left[1 + \text{erf}\left(\frac{x-\mu}{\sigma\sqrt{2}}\right)\right] \\ \end{align*}

The distribution is parameterised by two real numbers:

The standard normal distribution takes μ=0\mu = 0 and σ=1\sigma = 1, denoted N(0,1)\mathcal{N}(0,1).

The moments of the normal distribution are:

Mean=μVariance=σ2Skewness=0Kurtosis=3E[(Xμσ)n]={n!(n/2)!2n/2=n!!n even0n odd \begin{align*} \text{Mean} &= \mu \\ \text{Variance} &= \sigma^2 \\ \text{Skewness} &= 0 \\ \text{Kurtosis} &= 3 \\ \mathbb{E}\left[ \left( \frac{X-\mu}{\sigma} \right)^n \right] &= \begin{cases} \frac{n!}{(n/2)!2^{n/2}} = n!! & n \text{ even} \\ 0 & n \text{ odd} \end{cases} \\ \end{align*}

(note the double factorial) with the quantiles given by

Q(p)=F1(p)=μ+σ2erf1(2p1)Median=μMode=μSupport=R \begin{align*} Q(p) = F^{-1}(p) &= \mu + \sigma\sqrt{2}\,\text{erf}^{-1}(2p-1) \\ \text{Median} &= \mu \\ \text{Mode} &= \mu \\ \text{Support} &= \mathbb{R} \end{align*}

The normal distribution has several important properties:

Suppose X1,X2,X3,X_1, X_2, X_3, \ldots are iid. normal random variables. Then:

The normal distribution appears frequently in the real world: