notes-on-statistics
note: this page is still in development.
a population is a set of similar items or events which is of interest for some question or experiment.
a sample is a set of individuals or objects collected or selected from a statistical population by a defined procedure.
finite sample distributions
a sequence of random variables is said to converge in distribution to a random variable if
for every .
for random vectors , we say that this sequence converges in distribution to a random -vector if
for every which is a continuity set of .
a sequence of random variables converges in probability towards the random variable if
for all .
given a real number , we say that the sequence converges in the th mean or in the norm towards the random variable if the th absolute moments of and exist, and
let be independent and identically distributed random variables with finite mean . let be the average of the first variables, then the law of large numbers establishes that:
- weak
- weak
- weak
- strong
suppose is a sequence of independent and identically distributed random variables with and . then, as approaches infinity, the random variables converge in distribution to a normal .
let be independent and identically distributed random variables drawn according to some distribution parametrized by , where is the set of all possible parameters for the selected distribution. the goal is to find the best estimator such that since the real cannot be known exactly from a finite sample.
an estimator from a parameter is a random variable that is symbolized as a function of the observed data.
an estimate is a realization of the estimator. it is a real value for the estimated parameter.
the bias of an estimator is defined as
we say that an estimator is unbiased if or .
the mean squared error of an estimator is defined as
a sequence of estimators of the parameter is called consistent if, for any ,
an estimator is consistent only if, as the sample data increases, the estimator approaches the real parameter.
the relative efficiency of two estimators is defined as
we say that is preferable if .
point estimation
the likelihood function is defined as
assuming , ,
for practical purposes, we often use the log-likelihood function since it is much easier to differentiate afterwards, and the maximum of is preserved for all .
the maximum likelihood estimator for is defined as
the score is the gradient the natural logarithm of the likelihood function with respect to an -dimensional parameter vector .
the score indicates the steepness of the log-likelihood function and thereby the sensitivity to infinitesimal changes to the parameter values.
let be the probability density function or probability mass function for conditioned on the value of . we define the fisher information as
the fisher information is a way of measuring the amount of information that an observable random variable carries about an unknown parameter upon which the probability of depends.
suppose is an unknown deterministic parameter which is to be estimated from independent observations of , each from a distribution according to some probability function . the variance of any unbiased estimator of is then bounded by the reciprocal of the fisher information . namely,
the efficiency of an unbiased estimator of a parameter is defined as
where is the fisher information of the sample.
- asymptotically unbiased. namely, .
- asymptotically efficient. namely, .
- consistency.
- .
interval estimation
note: this section is still in development.
parametric hypothesis testing
note: this section is still in development.
normality tests
note: this section is still in development.