Statistics - PS1
โํ๋ฅ ๊ณผ ํต๊ณ(MATH230)โ ์์ ์์ ๋ฐฐ์ด ๊ฒ๊ณผ ๊ณต๋ถํ ๊ฒ์ ์ ๋ฆฌํ ํฌ์คํธ์ ๋๋ค. ์ ์ฒด ํฌ์คํธ๋ Probability and Statistics์์ ํ์ธํ์ค ์ ์์ต๋๋ค ๐ฒ
์ด ๊ธ์ โPoint Estimationโ ํฌ์คํธ์์ ์ ์ํ ์์ ๋ฌธ์ ๋ค์ ํ์ดํ ํฌ์คํธ์ ๋๋ค.
MSE is the sum of variance and square of bias
Theorem.
The <MSE; Mean Squared Error> of an estimator is defined as
\[\text{MSE} := E \left[ \left( \hat{\Theta} - \theta \right)^2 \right] = \text{Var}(\hat{\Theta}) + \left[ \text{Bias} \right]^2\]where $\text{Bias} := E [ \hat{\Theta} - \theta ]$.
Proof.
๋จผ์ , MSE(Mean Square Error)๋ ์๋์ ๊ฐ์ด ํํํ ์ ์๋ค.
\[\text{MSE} = E [(\hat{\Theta} - \theta)^2]\]์์ ์์ ๋ณํํ๋ฉด ์๋์ ๊ฐ๊ณ , ์ด๋ฅผ ์ ๊ฐํด๋ณด์.
\[\begin{aligned} E [(\hat{\Theta} - \theta)^2] &= E \left[ (\hat{\Theta} - E[\hat{\Theta}] + E[\hat{\Theta}] - \theta)^2 \right] \\ &= E \left[ \left((\hat{\Theta} - E[\hat{\Theta}]) + (E[\hat{\Theta}] - \theta)\right)^2 \right] \\ &= E \left[ \cancelto{\text{Var}(\hat{\Theta})}{(\hat{\Theta} - E[\hat{\Theta}])^2} + (\hat{\Theta} - E[\hat{\Theta}]) (E[\hat{\Theta}] - \theta) + (E[\hat{\Theta}] - \theta)^2 \right] \\ &= \text{Var}(\hat{\Theta}) + E \left[ (\hat{\Theta} - E[\hat{\Theta}]) (E[\hat{\Theta}] - \theta) + (E[\hat{\Theta}] - \theta)^2 \right] \\ &= \text{Var}(\hat{\Theta}) + E \left[ \hat{\Theta} E[\hat{\Theta}] - \cancel{E[\hat{\Theta}]^2} - \hat{\Theta} \theta + \theta E [\hat{\Theta}] + \cancel{E[\hat{\Theta}]^2} + 2 \theta E[\hat{\Theta}] + \theta^2\right] \\ &= \text{Var}(\hat{\Theta}) + E \left[ \hat{\Theta} E[\hat{\Theta}] - \hat{\Theta} \theta - \theta E[\hat{\Theta}] + \theta^2\right] \end{aligned}\]์์ ์์ ์ค๋ฅธํธ์ ํ ์ ์๋์ ๊ฐ์ด ๋ฌถ์ด์ค๋ค.
\[\begin{aligned} E \left[ \hat{\Theta} E[\hat{\Theta}] - \hat{\Theta} \theta - \theta E[\hat{\Theta}] + \theta^2\right] &= E \left[ (\hat{\Theta} - \theta) E[\hat{\Theta}] - (\hat{\Theta} - \theta )\theta\right] \\ &= E \left[ (\hat{\Theta} - \theta) (E[\hat{\Theta}] - \theta)\right] \\ &= E \left[ (\hat{\Theta} - \theta) \cdot \cancelto{\text{bias}}{E[\hat{\Theta} - \theta]}\right] \\ &= \text{bias} \cdot \cancelto{\text{bias}}{E \left[ \hat{\Theta} - \theta \right]} \\ &= \text{bias} \cdot \text{bias} = (\text{bias})^2 \end{aligned}\]๋ฐ๋ผ์, MSE๋ ์๋์ ๊ฐ๋ค.
\[\text{MSE} = \text{Var}(\hat{\Theta}) + (\text{bias})^2\]$\blacksquare$
Sample Variance is not the minimal variance estimator
Exercise.
Let $X_1, \dots, X_n$ be iid $N(\mu, \sigma^2)$.
Let $\displaystyle S^2 := \frac{1}{n-1} \sum^n_i (X_i - \bar{X})^2$ and $\displaystyle \hat{S}^2 := \frac{1}{n} \sum^n_i (X_i - \bar{X})^2$
Show that $\text{Var}(S^2) > \text{Var}(\hat{S}^2)$.
Solve.
๋ estimator์ Variance๋ฅผ ๊ตฌํด๋ณด์.
$S^2$์ ๋ํด $\dfrac{(n-1)S^2}{\sigma^2} \sim \chi^2(n-1)$๊ฐ ์ฑ๋ฆฝํ๋ฏ๋ก, ์๋์ $\chi^2$ ๋ถํฌ์ ์ฑ์ง์ ๋ฐ๋ผ ์๋์ ์์ด ์ฑ๋ฆฝํ๋ค.
\[\text{Var} \left(\frac{(n-1)S^2}{\sigma^2}\right) = \text{Var} \left(\chi^2(n-1)\right) = 2(n-1)\]์์ ์์ ์ ์ ๋ฆฌํ๋ฉด,
\[\begin{aligned} \text{Var} \left(\frac{(n-1)S^2}{\sigma^2}\right) &= 2(n-1) \\ \frac{(n-1)^2}{\sigma^4} \text{Var} (S^2) &= 2(n-1) \\ \text{Var} (S^2) &= \frac{2\sigma^4}{(n-1)} \end{aligned}\]์ด์ , $\hat{S}^2$์ Variance๋ฅผ ๊ตฌํด๋ณด์.
\[\begin{aligned} \text{Var}(\hat{S}^2) &= \text{Var} \left(\frac{n-1}{n}S^2\right) \\ &= \frac{(n-1)^2}{n^2}\text{Var}(S^2) \\ &= \frac{(n-1)^2}{n^2} \cdot \frac{2\sigma^4}{(n-1)} \\ &= \frac{2(n-1)\sigma^4}{n^2} \end{aligned}\]๋ estimator์ Variance๋ฅผ ๋น๊ตํด๋ณด๋ฉด,
\[\text{Var}(\hat{S}^2) = \frac{(n-1)^2}{n^2}\text{Var}(S^2) < \text{Var}(S^2)\]์ด๋ค. ์ฆ, $S^2$ ๋ณด๋ค ๋ฎ์ variance๋ฅผ ๊ฐ๋ estimator๊ฐ ์กด์ฌํ๋ค. $\blacksquare$
(๋จ, $\hat{S}^2$๋ biased estimator์!)
Which one is the most efficient variance estimator?
Exercise.
Compare $S^2$ and $\hat{S}^2$, which one is the most efficient estimator?
First, we know that the $S^2$ is an unbiased estimator, so $\text{bias}(S^2) = 0$.
Next, letโs find the bias of $\hat{S}^2$.
\[\begin{aligned} \text{bias}(\hat{S}^2) &= E \left[ \hat{S}^2 - \sigma^2 \right] \\ &= E \left[ \frac{1}{n} \sum_i^n (X_i - \bar{X})^2 - \sigma^2 \right] \\ &= E \left[ \frac{n-1}{n} \cdot \frac{1}{n-1} \sum_i^n (X_i - \bar{X})^2 - \sigma^2 \right] \\ &= \frac{n-1}{n} \cdot E \left[ \frac{1}{n-1} \sum_i^n (X_i - \bar{X})^2 \right] - \sigma^2 \\ &= \frac{n-1}{n} \cdot \sigma^2 - \sigma^2 \\ &= \frac{1}{n} \cdot \left( (n-1) \sigma^2 - n \sigma^2 \right) = - \frac{\sigma^2}{n} \end{aligned}\]Weโve already got the variance of two estimators. Then, the MSE for two estimators are:
\[\text{MSE}(S^2) = \frac{2\sigma^4}{(n-1)} + 0\] \[\text{MSE}(\hat{S}^2) = \frac{2(n-1)\sigma^4}{n^2} + \frac{\sigma^4}{n^2}\] \[\begin{aligned} \frac{2\sigma^4}{(n-1)} \quad &\text{vs.} \quad \frac{2(n-1)\sigma^4}{n^2} + \frac{\sigma^4}{n^2} \\ \frac{2}{(n-1)} \quad &\text{vs.} \quad \frac{2(n-1)}{n^2} + \frac{1}{n^2} \\ \frac{2}{(n-1)} \quad &\text{vs.} \quad \frac{2(n-1) + 1}{n^2} \\ \frac{2}{(n-1)} \quad &\text{vs.} \quad \frac{2n-1}{n^2} \\ 2 \cdot n^2 \quad &\text{vs.} \quad (2n-1) \cdot (n-1) \\ 2 n^2 \quad &\text{vs.} \quad 2n^2 - 3n + 1 \\ \end{aligned}\]์ฆ, $\text{MSE}(S^2) > \text{MSE}(\hat{S}^2)$์ด๋ฏ๋ก, $\hat{S}^2$๊ฐ โthe most efficient estimatorโ์ด๋ค! $\blacksquare$