Common Probability Distributions
Bernoulli Trials Family
- The Bernoulli distribution models a single trial with a binary outcome: outcome 1 (success) with probability
and outcome 0 (Failure) with probability . - The Binomial distribution is the sum of
independent Bernoulli trials. It models the total number of successes for a fixed trials. - The Geometric distribution models the number of trials required to get the first success.
- Discrete analog of the Exponential distribution.
- The Negative Binomial distribution models the number of trials required to get the r-th success. It can be viewed as the sum of
i.i.d Geometric variables.
| Bernoulli | Binomial | Geometric | Negative Binomial | |
|---|---|---|---|---|
| Notation | ||||
for |
for |
for |
for |
|
| (Sum of PMFs) | (Sum of PMFs) | |||
| Params |
Poisson Process Family
- The Poisson distribution models the number of events occurring in a fixed interval given a constant average rate
. - It is derived from the Binomial distribution by letting
and such that remains constant.
- It is derived from the Binomial distribution by letting
- The Exponential distribution models the waiting time until the 1st event occurs in a Poisson process, where the average waiting time is
. - It is the continuous analog of the Geometric distribution.
- The Gamma distribution models the waiting time until the
-th event occurs in a Poisson process. It is the sum of i.i.d Exponential variables. - The Weibull distribution generalizes the Exponential distribution to allow for changing failure rates over time.
| Poisson | Exponential | Gamma | Weibull | |
|---|---|---|---|---|
| Notation | ||||
for |
for |
for |
for |
|
| (Sum of PMFs) | for |
(Incomplete) | for |
|
for |
for |
(Complex) | ||
| Params |
Properties
- Memoryless Property: the probability of an event occurring after a duration
, given that it has not occurred by time , is the same as the initial probability that it would not occur after time . - If
are independent exponential random variables with rates , - Distribution of the Minimum: Then the random variable
is also exponentially distributed with a rate equal to the sum of the individual rates . - Competing Exponentials (Race Condition): The probability that a specific variable
is the first to occur (i.e., is the minimum) is the ratio of its rate to the total rate:
- Distribution of the Minimum: Then the random variable
The Gaussian Family
- The Normal (Gaussian) distribution models sums of independent random variables (via the Central Limit Theorem). It approximates the Binomial, Poisson, and Gamma distributions when samples are large.
- The Chi-Square distribution models the sum of the squares of
independent standard normal random variables. - It is mathematically a special case of the Gamma distribution (
).
- It is mathematically a special case of the Gamma distribution (
- The Student's t distribution models the mean of a normal population when the sample size is small and variance is unknown. It is the ratio of a Normal random variable to the square root of a Chi-Square variable. It converges to the Normal distribution as
.
| Component | Normal (Gaussian) | Chi-Square | Student's t |
|---|---|---|---|
| Notation | |||
for |
for |
||
| (Incomplete Gamma) | (Table) | ||
for |
Does not exist | ||
| Params |
The Bounded & Bayesian Family
- The Uniform distribution models a random variable where all intervals of the same length within the support
are equally probable. It represents maximum entropy (minimum information) over an interval. - The Beta distribution models random variables restricted to the interval
. It generalizes the Uniform distribution and serves as the conjugate prior for the Bernoulli, Binomial, and Geometric distributions in Bayesian statistics.
| Uniform | Beta | |
|---|---|---|
| Notation | ||
for |
for |
|
| (Complex) | ||
Other Concepts
TODO:
- Maybe do research into other common distributions that are relevant.
- Explain how we can sample from these distributions, with the uniform distribution as the base. Could maybe write a separate article on this.