โ€œํ™•๋ฅ ๊ณผ ํ†ต๊ณ„(MATH230)โ€ ์ˆ˜์—…์—์„œ ๋ฐฐ์šด ๊ฒƒ๊ณผ ๊ณต๋ถ€ํ•œ ๊ฒƒ์„ ์ •๋ฆฌํ•œ ํฌ์ŠคํŠธ์ž…๋‹ˆ๋‹ค. ์ „์ฒด ํฌ์ŠคํŠธ๋Š” Probability and Statistics์—์„œ ํ™•์ธํ•˜์‹ค ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค ๐ŸŽฒ

6 minute read

โ€œํ™•๋ฅ ๊ณผ ํ†ต๊ณ„(MATH230)โ€ ์ˆ˜์—…์—์„œ ๋ฐฐ์šด ๊ฒƒ๊ณผ ๊ณต๋ถ€ํ•œ ๊ฒƒ์„ ์ •๋ฆฌํ•œ ํฌ์ŠคํŠธ์ž…๋‹ˆ๋‹ค. ์ „์ฒด ํฌ์ŠคํŠธ๋Š” Probability and Statistics์—์„œ ํ™•์ธํ•˜์‹ค ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค ๐ŸŽฒ

์ด ๊ธ€์€ โ€œPoint Estimationโ€ ํฌ์ŠคํŠธ์—์„œ ์ œ์‹œํ•œ ์ˆ™์ œ ๋ฌธ์ œ๋“ค์„ ํ’€์ดํ•œ ํฌ์ŠคํŠธ์ž…๋‹ˆ๋‹ค.

MSE is the sum of variance and square of bias

Theorem.

The <MSE; Mean Squared Error> of an estimator is defined as

\[\text{MSE} := E \left[ \left( \hat{\Theta} - \theta \right)^2 \right] = \text{Var}(\hat{\Theta}) + \left[ \text{Bias} \right]^2\]

where $\text{Bias} := E [ \hat{\Theta} - \theta ]$.

Proof.

๋จผ์ €, MSE(Mean Square Error)๋Š” ์•„๋ž˜์™€ ๊ฐ™์ด ํ‘œํ˜„ํ•  ์ˆ˜ ์žˆ๋‹ค.

\[\text{MSE} = E [(\hat{\Theta} - \theta)^2]\]

์œ„์˜ ์‹์„ ๋ณ€ํ˜•ํ•˜๋ฉด ์•„๋ž˜์™€ ๊ฐ™๊ณ , ์ด๋ฅผ ์ „๊ฐœํ•ด๋ณด์ž.

\[\begin{aligned} E [(\hat{\Theta} - \theta)^2] &= E \left[ (\hat{\Theta} - E[\hat{\Theta}] + E[\hat{\Theta}] - \theta)^2 \right] \\ &= E \left[ \left((\hat{\Theta} - E[\hat{\Theta}]) + (E[\hat{\Theta}] - \theta)\right)^2 \right] \\ &= E \left[ \cancelto{\text{Var}(\hat{\Theta})}{(\hat{\Theta} - E[\hat{\Theta}])^2} + (\hat{\Theta} - E[\hat{\Theta}]) (E[\hat{\Theta}] - \theta) + (E[\hat{\Theta}] - \theta)^2 \right] \\ &= \text{Var}(\hat{\Theta}) + E \left[ (\hat{\Theta} - E[\hat{\Theta}]) (E[\hat{\Theta}] - \theta) + (E[\hat{\Theta}] - \theta)^2 \right] \\ &= \text{Var}(\hat{\Theta}) + E \left[ \hat{\Theta} E[\hat{\Theta}] - \cancel{E[\hat{\Theta}]^2} - \hat{\Theta} \theta + \theta E [\hat{\Theta}] + \cancel{E[\hat{\Theta}]^2} + 2 \theta E[\hat{\Theta}] + \theta^2\right] \\ &= \text{Var}(\hat{\Theta}) + E \left[ \hat{\Theta} E[\hat{\Theta}] - \hat{\Theta} \theta - \theta E[\hat{\Theta}] + \theta^2\right] \end{aligned}\]

์œ„์˜ ์‹์˜ ์˜ค๋ฅธํŽธ์˜ ํ…€์„ ์•„๋ž˜์™€ ๊ฐ™์ด ๋ฌถ์–ด์ค€๋‹ค.

\[\begin{aligned} E \left[ \hat{\Theta} E[\hat{\Theta}] - \hat{\Theta} \theta - \theta E[\hat{\Theta}] + \theta^2\right] &= E \left[ (\hat{\Theta} - \theta) E[\hat{\Theta}] - (\hat{\Theta} - \theta )\theta\right] \\ &= E \left[ (\hat{\Theta} - \theta) (E[\hat{\Theta}] - \theta)\right] \\ &= E \left[ (\hat{\Theta} - \theta) \cdot \cancelto{\text{bias}}{E[\hat{\Theta} - \theta]}\right] \\ &= \text{bias} \cdot \cancelto{\text{bias}}{E \left[ \hat{\Theta} - \theta \right]} \\ &= \text{bias} \cdot \text{bias} = (\text{bias})^2 \end{aligned}\]

๋”ฐ๋ผ์„œ, MSE๋Š” ์•„๋ž˜์™€ ๊ฐ™๋‹ค.

\[\text{MSE} = \text{Var}(\hat{\Theta}) + (\text{bias})^2\]

$\blacksquare$


Sample Variance is not the minimal variance estimator

Exercise.

Let $X_1, \dots, X_n$ be iid $N(\mu, \sigma^2)$.

Let $\displaystyle S^2 := \frac{1}{n-1} \sum^n_i (X_i - \bar{X})^2$ and $\displaystyle \hat{S}^2 := \frac{1}{n} \sum^n_i (X_i - \bar{X})^2$

Show that $\text{Var}(S^2) > \text{Var}(\hat{S}^2)$.

Solve.

๋‘ estimator์˜ Variance๋ฅผ ๊ตฌํ•ด๋ณด์ž.

$S^2$์— ๋Œ€ํ•ด $\dfrac{(n-1)S^2}{\sigma^2} \sim \chi^2(n-1)$๊ฐ€ ์„ฑ๋ฆฝํ•˜๋ฏ€๋กœ, ์•„๋ž˜์˜ $\chi^2$ ๋ถ„ํฌ์˜ ์„ฑ์งˆ์— ๋”ฐ๋ผ ์•„๋ž˜์˜ ์‹์ด ์„ฑ๋ฆฝํ•œ๋‹ค.

\[\text{Var} \left(\frac{(n-1)S^2}{\sigma^2}\right) = \text{Var} \left(\chi^2(n-1)\right) = 2(n-1)\]

์œ„์˜ ์‹์„ ์ž˜ ์ •๋ฆฌํ•˜๋ฉด,

\[\begin{aligned} \text{Var} \left(\frac{(n-1)S^2}{\sigma^2}\right) &= 2(n-1) \\ \frac{(n-1)^2}{\sigma^4} \text{Var} (S^2) &= 2(n-1) \\ \text{Var} (S^2) &= \frac{2\sigma^4}{(n-1)} \end{aligned}\]

์ด์ œ, $\hat{S}^2$์˜ Variance๋ฅผ ๊ตฌํ•ด๋ณด์ž.

\[\begin{aligned} \text{Var}(\hat{S}^2) &= \text{Var} \left(\frac{n-1}{n}S^2\right) \\ &= \frac{(n-1)^2}{n^2}\text{Var}(S^2) \\ &= \frac{(n-1)^2}{n^2} \cdot \frac{2\sigma^4}{(n-1)} \\ &= \frac{2(n-1)\sigma^4}{n^2} \end{aligned}\]

๋‘ estimator์˜ Variance๋ฅผ ๋น„๊ตํ•ด๋ณด๋ฉด,

\[\text{Var}(\hat{S}^2) = \frac{(n-1)^2}{n^2}\text{Var}(S^2) < \text{Var}(S^2)\]

์ด๋‹ค. ์ฆ‰, $S^2$ ๋ณด๋‹ค ๋‚ฎ์€ variance๋ฅผ ๊ฐ–๋Š” estimator๊ฐ€ ์กด์žฌํ•œ๋‹ค. $\blacksquare$
(๋‹จ, $\hat{S}^2$๋Š” biased estimator์ž„!)


Which one is the most efficient variance estimator?

Exercise.

Compare $S^2$ and $\hat{S}^2$, which one is the most efficient estimator?

First, we know that the $S^2$ is an unbiased estimator, so $\text{bias}(S^2) = 0$.

Next, letโ€™s find the bias of $\hat{S}^2$.

\[\begin{aligned} \text{bias}(\hat{S}^2) &= E \left[ \hat{S}^2 - \sigma^2 \right] \\ &= E \left[ \frac{1}{n} \sum_i^n (X_i - \bar{X})^2 - \sigma^2 \right] \\ &= E \left[ \frac{n-1}{n} \cdot \frac{1}{n-1} \sum_i^n (X_i - \bar{X})^2 - \sigma^2 \right] \\ &= \frac{n-1}{n} \cdot E \left[ \frac{1}{n-1} \sum_i^n (X_i - \bar{X})^2 \right] - \sigma^2 \\ &= \frac{n-1}{n} \cdot \sigma^2 - \sigma^2 \\ &= \frac{1}{n} \cdot \left( (n-1) \sigma^2 - n \sigma^2 \right) = - \frac{\sigma^2}{n} \end{aligned}\]

Weโ€™ve already got the variance of two estimators. Then, the MSE for two estimators are:

\[\text{MSE}(S^2) = \frac{2\sigma^4}{(n-1)} + 0\] \[\text{MSE}(\hat{S}^2) = \frac{2(n-1)\sigma^4}{n^2} + \frac{\sigma^4}{n^2}\] \[\begin{aligned} \frac{2\sigma^4}{(n-1)} \quad &\text{vs.} \quad \frac{2(n-1)\sigma^4}{n^2} + \frac{\sigma^4}{n^2} \\ \frac{2}{(n-1)} \quad &\text{vs.} \quad \frac{2(n-1)}{n^2} + \frac{1}{n^2} \\ \frac{2}{(n-1)} \quad &\text{vs.} \quad \frac{2(n-1) + 1}{n^2} \\ \frac{2}{(n-1)} \quad &\text{vs.} \quad \frac{2n-1}{n^2} \\ 2 \cdot n^2 \quad &\text{vs.} \quad (2n-1) \cdot (n-1) \\ 2 n^2 \quad &\text{vs.} \quad 2n^2 - 3n + 1 \\ \end{aligned}\]

์ฆ‰, $\text{MSE}(S^2) > \text{MSE}(\hat{S}^2)$์ด๋ฏ€๋กœ, $\hat{S}^2$๊ฐ€ โ€œthe most efficient estimatorโ€์ด๋‹ค! $\blacksquare$