2 Inferences
2.1 General theory
Cited from [1, Chapter 4].
2.1.1 Point estimation
Consider a random variable \(X\) with an unknown distribution. Our information about the distribution of \(X\) comes from a sample on \(X\). The sample ovservations have the same distribution as \(X\), and we denote them as the random variables \(X_1, X_2,\ldots,X_n\), where \(n\) denotes the sample size. When the sample is actually drawn, we use \(x_1,\ldots,x_n\) as the realizations of the sample.
Definition 2.1 (Random sample) If the random variables \(X_1,\ldots, X_n\) are iid, then these random variable constitute a random sample of size \(n\) from the common distribution.
Definition 2.2 (Statistics) Let \(X_1,\ldots,X_n\) denote a sample on a random variable \(X\). Let \(T=T(X_1,\ldots,X_n)\) be a function of the sample. \(T\) is called a statistic. Once the sample is drawn, \(t=T(x_1,\ldots,x_n)\) is called the realization of \(T\).
Assume that the distribution of \(X\) is known down to an unknown parameter \(\theta\) where \(\theta\) can be a vector. Then the pdf of \(X\) can be written as \(f(x;\theta)\). In this case we might find some statistic \(T\) to estimate \(\theta\). This is called a point estimator of \(\theta\). A realization \(t\) is called an estimate of \(\theta\).
Definition 2.3 (Unbiasedness) Let \(X_1,\ldots,X_n\) is a sample on a random varaible \(X\) with pdf \(f(x;\theta)\). Let \(T\) be a statistic. We say that \(T\) is an unbiased estimator of \(\theta\) if \(E(T)=\theta\).
Let \(X\) be a random variable, with mean \(\mu\) and variance \(\sigma^2\). Consider a sample \(\set{X_i}\) of size \(n\). By definition all \(X_i\)’s are iid. Therefore \(\Exp\qty(X_i)=\mu\), and \(\Var\qty(X_i)=\sigma^2\) for any \(i=1,\ldots, N\).
Consider the following statistics:
- \(\bar{\mu}=\dfrac1N\sum_{i=1}^NX_i\),
- \(\bar{\sigma}^2=\dfrac{1}{N-1}\sum_{i=1}^N(X_i-\bar{\mu})^2\).
Lemma 2.1
- \(\Exp(\bar{\mu})=\mu\).
- \(\Exp(\bar{\sigma}^2)=\sigma^2\).
Definition 2.4 The following are the unbiased estimators of \(\mu\) and \(\sigma^2\) of \(X\).
- \(\bar{\mu}=\dfrac1N\sum_{i=1}^NX_i\) is called the sample mean of the samples.
- \(\bar{\sigma}^2=\dfrac{1}{N-1}\sum_{i=1}^N(X_i-\bar{\mu})^2\) is called the sample variance of the samples.
2.1.2 Confidence intervals
Definition 2.5 (Confidence interval) Consider a sample of \(X\). Fix a number \(0<\alpha<1\). Let \(L\) and \(U\) be two statistics. We say the interval \((L,U)\) is a \((1-\alpha)100\%\) confidence interval for \(\theta\) if
\[ 1-\alpha=\Pr[\theta\in(L,U)]. \]