Continuous Random Variable Distribution Families
Continuous distributions are used to describe random variables that have outcomes that fall anywhere within a given interval or intervals such as the lifetime of a certain insect, heights university students, or the amount of time kindergarteners spend watching TV in a given day. There are many families of continuous distributions; a few of the most common are discussed on this page.
The Uniform Distribution
The probability density function of a uniform random variable looks like a horizontal line segment over the support. This indicates that for any interval of a given length within the support, the probability that a value in that interval occurs is the same.
$f(x) = \frac{1}{b-a}$, for $a \leq x \leq b$
$E(X) = \frac{b+a}{2}$, $Var(X)=\frac{(b-a)^2}{12}$
The Exponential Distribution
This distribution is related to the Poisson distribution. The Poisson distribution counted
how many times an event would occur over a specified time or distance when the average
number of occurrences for that time or distance is known. The Exponential distribution
measures the amount of time or distance until the next occurrence of such an event, when we
know the average time or distance it takes to see another occurrence.
The parameter, λ, is the rate parameter. It is calculated by finding how
many times the event is observed (on average) in certain blocks of time or distance, such
as "about 2 energy spikes per 30 seconds" ($\lambda = \frac{1}{15}$), or "an average of 4
breakages per 70 feet of cable" ($\lambda = \frac{2}{35}$). Whatever units the random variable
(say, $X$) is measuring (e.g. seconds until the next spike, or feet of cable
until the next breakage) will be found on the bottom of the rate parameter.
This distribution is the only "memoryless" continous distribution. This means that,
if you have already waited 20 seconds without seeing the event in question, and you
want to know the probability that it will take 10 more seconds for the
event to be observed, it is the same as the probability that it would have taken
10 seconds from the beginning to see the event occur. It doesn't matter how long you
have already waited without seeing the event, whether 0 seconds or 100; the probability that
the event will happen in exactly 10 more seconds will always be the same.
$f(x)=\lambda e^{-\lambda x}$, for $0 \leq x$
$E(X) = \frac{1}{\lambda}$, $Var(X)=\frac{1}{\lambda^2}$
The Gamma Distribution
The Gamma Distribution is the maximum entropy probability distribution for a random
variable $X$ that has a positive expected value of $\frac{\alpha}{\lambda}$. We often
use this distribution when nothing is known about a distribution except that it describes
a measurement of waiting time, distance, decay, power loss, mass accumulation, or another
similar process. We use the "maximum entropy" distribution when we know very little
about the distribution because it minimizes the amount of prior information built into
the distribution, and because many physical systems tend to move toward maximal entropy
configurations over time.
In a Gamma[α,λ] distribution, α is the shape parameter, and
λ is the rate parameter. Whatever units the random variable (say, $X$) is
measuring (e.g. years until the next millimeter of cancer cells accumulates, or
loss of power in Joules within the next minute interval) will be found on
the bottom of the rate parameter
($\lambda=\frac{1\hspace{3 pt}millimeter}{6\hspace{3 pt}years}$ or
$\lambda=\frac{2\hspace{3 pt}minutes}{2235\hspace{3 pt}Joules}$).
The Exponential Distribution (see above) and the Chi-Square Distribution (see below) are
special cases of the Gamma Distribution. An exponential(λ) distribution is a
Gamma[1,λ] distribution, and a Chi-Square($\nu$) distribution is a
Gamma[$\frac{\nu}{2},\frac{1}{2}$] distribution.
$f(x)=\frac{\lambda^{\alpha}}{\Gamma(\alpha)}x^{\alpha-1}e^{-\lambda x}$, for $0 \leq x$
Where, for a positive integer $n$,
$\Gamma(n) = (n-1)!$,
or, more generally, for all complex numbers $z$ with a positive real part,
$\Gamma(z) = \int_0^{\infty}x^{z-1}e^{-x}dx$
$E(X) = \frac{\alpha}{\lambda}$, $Var(X)=\frac{\alpha}{\lambda^2}$
The Chi-Square Distribution
The Chi-Square Distribution is dependent on the parameter $\nu$, which is referred to as
the "degrees of freedom". It is the distribution of a sum of the squares of $\nu$ independent
standard normal random variables. It is most commonly used in Chi-Square Goodness of Fit
Tests -- which determine how well categorical data fit an expected set of proportions --
and in Chi-Square Tests of Independence, which determine the probability that data classified
across two categorical variables could have come from a population where the two categorical
variables are independent of one another.
The parameter $\nu$, or the degrees of freedom, is the minimum number of categories that need
data given before the rest can be found mathematically, if we are given the total. This is why
we call it the "degrees of freedom", because it denotes how many categories have the "freedom"
to take on any value before the rest of the categories' values are fixed numbers.
For instance, if we separate 100 Starburst candies into piles of red, pink, yellow, and orange,
and then count how many candies are in each pile, we only need to count three of the piles.
Once we have counted 30 red candies, 22 pink candies, and 20 orange candies, we know that there
must be 28 yellow candies in order for there to be 100 candies. Then, in this case, $\nu = 3$.
$f(x)= \sum^{\nu}_{i=1} Z^2_i$, for $0 \leq x$,
where $Z_1, Z_2, \ldots Z_{\nu}$ are independent, standard normal random variables.
$E(X) = \nu$, $Var(X)=\nu$
The t Distribution
"Student's t-distribution" is a family of probability distributions that estimate
the mean of a normally distributed population with a small-sized sample, where the population
standard deviation is unknown.
The t-distribution is dependent on the parameter $\nu$, which is referred to as
the "degrees of freedom". $\nu$ is the minimum number of subjects in the sample that we
need to get measurements for before the rest can be found mathematically, if we already know
the entire sample's average or sum. This is why we call it the "degrees of freedom", because
it denotes how many subjects have the "freedom" to take on any value before the rest of the
subjects' values are fixed numbers. Here, for any sample of size $n$ (meaning there are $n$
subjects in the sample), $\nu = n-1$.
For example, suppose we obtain a random sample of 5 values from a normally distributed population,
and the sample average is 20. Then, as we start recording the individual values, we get 23 for
the first subject, 15 for the second, 18 for the third, 25 for the fourth...and we don't have to
measure the fifth subject. The fifth subject's value is fixed, and can only be one value. Since
we know the average is 20, we can find the fifth subject's value (denoted by $x$) with the following
equation: $20=\frac{23+15+18+25+x}{5}$. Solving for $x$, we get $x=19$. Only the first 4 (or $n-1$)
values were free variables. After four subjects have been measured, the fifth subject's value is
automatically known.
$f(x) = \frac{\Gamma(\frac{\nu+1}{2})}{\sqrt{\nu\pi}\Gamma(\frac{\nu}{2})} (1+\frac{t^2}{\nu})^{-\frac{\nu+1}{2}}$, for $-\infty < x < \infty$
$E(X) = 0$
$Var(X) = \frac{\nu}{\nu-2}$ for $ \nu > 2, \infty$ for $1 < \nu \leq 2$, otherwise undefined.