notes-on-probability
note: for a basic course in probability, you may want to skip some notes on measure theory.
basics and combinatorics
the sample space is the set of all possible outcomes of an experiment. it can be finite or infinite.
an event is a subset of the sample space , or an element of the power set of the sample space .
the set of all observable events is denoted by , where .
usually, if is countable, . however, sometimes many events are excluded from since it is not possible for them to happen.
the set is called a -algebra if:
- ;
- , , then ; and
- , , then .
is a probability measure if it satisfies the following three axioms:
- ,
- , and
- ,
where are disjunct.
- ,
- ,
- if , then , and
- .
let be a set of events.
let be a set of events, then
where
if with where all have the same probability , is called laplace space and has a discrete uniform distribution. for some event , we have
the discrete uniform distribution exists only if is finite.
given two events and with , the probability of given is defined as
let be a set of disjunct events where , then, for any event ,
let be the set of disjunct event where with for all , then, for an event with , we have
a set of events are independent if, for all with , we have
the factorial function is defined by the product
for integer .
let with , the gamma function is defined via the following convergent improper integral:
- .
- .
- .
- , .
let be the number of total objects and be the number of objects we want to select. a permutation is an arrangement of elements where we care about ordering.
- repetition not allowed:
- repetition allowed:
let be the number of total objects and be the number of objects we want to select. a combination is an arrangement of elements where we do not care about ordering.
- repetition not allowed:
- repetition allowed:
repetition is the same as replacement.
- ,
- ,
- ,
- ,
- , and
- .
let , and .
- .
- .
- .
- .
- .
random variables
let be a probability space. a random variable on is a function
if the image is countable, is called a discrete random variable, otherwise it is called a continuous random variable.
the probability density function (pdf) of a random variable is a function defined as
with discrete, it is called probability mass function.
- if is a discrete random variable, then .
- if is a continuous random variable, then .
the cumulative distribution function (cdf) of a random variable is a function defined as
if the pdf is given, the cdf can be expressed with
- if , then (monotonicity).
- if , then .
- and .
- .
- .
- .
let be a random variable. the expected value is defined as
- if ,
- ,
- if ,
- ,
- ,
- , and
- for independent .
the expected value is a linear operator.
let . the th (raw) moment is defined as
the th central moment is defined as
let be a random variable and with , then
let be a random variable. the moment-generating function of is defined as
let be a random variable. the characteristic function of is defined as
where .
let be a random variable with . the variance of is defined as
- ,
- ,
- ,
- ,
- , and
- if , .
let be a random variable with . the standard deviation of is defined as
let and be random variables with finite expected value. the covariance of and is defined as
the covariance is a measure of correlation between two random variables. if tends to increase as increases. if tends to decrease as increases. if , then and are uncorrelated.
- ,
- , and
- .
let and be random variables with finite expected value. the correlation of and is defined as
correlation is the same as covariance but normalized with values between and .
the coefficient of variation is defined as the ratio of the standard deviation to the mean.
the indicator function for a set (event) is defined as
let be a continuous random variable with cumulative distribution function on the interval . its survival function or reliability function is defined as
suppose is a non-negative random variable. the probability distribution of is memoryless if for any , we have
let be a random variable and be an increasing function, then, for all with , we have
for practical uses usually .
let be a random variable with , then, if ,
if is a random variable and is a convex function, then
multivariate distributions
the joint probability density function with is a function defined as
the joint cumulative distribution function with is a function defined as
if the joint pdf is given, it can be expressed with
where and .
the marginal probability density function of given a joint pdf is defined as
where , and in the discrete case .
the idea of the marginal probability is to ignore all other random variables and consider only the one we're interested in.
the marginal cumulative distribution function of given a joint cdf is defined as
the conditional distribution is defined as
the random variables are independent if
similarly, if their pdf is absolutely continuous, they are independent if
if the random variables are independent and is a function with , then also are independent.
the random variables are independent if and only if, , we have
the joint expected value of a random variable is defined as
where , and in the discrete case .
the conditional expected value of random variables and is defined as
- ,
- if and are independent, and
- .
let .
where and .
when is a strictly increasing function, then
when is a strictly decreasing function, then
equivalently, .
in higher dimensions, the derivative generalizes to the determinant of the jacobian matrix --- the matrix with . thus, for real-valued vector and ,
let have a continuous cdf and define . then , i.e., for .
for and , if and exist, and for all in some neighbourhood of zero, then for all .
let be the integral of a function that is hard to evaluate, then
where is uniformly distributed. then, by the law of large numbers, we know that we can approximate by randomly sampling from .
let be independent random variables, then the sum has a pdf evaluated with a convolution between all pdfs
in the special case where , we have
often it is much easier to use properties of the random variables to find the sum instead of evaluating the convolution.
let and be independent random variables. to evaluate the pdf and the cdf of , we proceed as
where the pdf is
let and be independent random variables. to evaluate the pdf and the cdf of , we proceed as
where the pdf is
let be random variables. let and for every . the covariance and correlation matrices as defined respectively as
let and be random variables. if and only if there exist such that .