0
$\begingroup$

Let $x_1, \cdots, x_n \sim f(x;\theta) = \frac{1}{2\theta}, -\theta < x < \theta$ and $\theta > 0$. I'm tasked with finding $\hat{\theta}$ using the moment method.

It's clear that $f(x;\theta) \sim \mathrm{UNIF}(-\theta, \theta)$, but $M_1 = \mathbb{E}(x) = 0$. What do I do?

On a related vein, what exactly is the benefit of using the moment method to find $\hat{\theta}$ vs. other methods? I don't mean to sound snooty but it just feels like I'm doing an unnecessary extra step.

$\endgroup$

1 Answer 1

1
$\begingroup$

The method of moments is not simply about matching the first moment. The general strategy is to use as many moments as is needed to obtain a system of equations that has a unique solution for the parameter. So in your case, because the first moment is zero, we use the second: $$\operatorname{E}[X^2] = \frac{\theta^2}{3},$$ hence we match the second raw sample moment to this: $$\frac{\hat \theta^2}{3} = \frac{1}{n} \sum_{i=1}^n X_i^2,$$ or $$\hat \theta = \sqrt{\frac{3}{n} \sum_{i=1}^n X_i^2}.$$ We could also match on the second central moment, also known as the variance: $$\operatorname{E}[(X - \operatorname{E}[X])^2] = \operatorname{Var}[X] = \frac{\theta^2}{3},$$ since the first moment is zero. This leads to a different estimator: $$\frac{\hat \theta^2}{3} = \frac{1}{n} \sum_{i=1}^n (X_i - \bar X)^2.$$ In practice, $\bar X$ will be "close" to zero, so these will be roughly the same, but notice that if $n = 1$, the first estimator is still defined, whereas the second will give $0$ for $\theta$, which is not sensible.


As a general approach, if we have, say, $k$ parameters to estimate, then the method of moments leads to a system in $k$ equations that we can (sometimes) recursively solve. This tends to be an easier way to obtain estimators for the parameters compared to, say, maximum likelihood estimation, but the problem with method of moment estimators is that they don't always make sense given the data.

$\endgroup$
5
  • $\begingroup$ Thank you so much! A very detailed yet simple response! $\endgroup$
    – JerBear
    Commented Mar 25, 2021 at 20:10
  • $\begingroup$ Actually I'm glad you brought that point up, @heropup. Here since there is only one parameter to estimate, is this an exception to the general rule? $\endgroup$
    – JerBear
    Commented Mar 25, 2021 at 20:12
  • 1
    $\begingroup$ @JerBear Well, the fact that the first moment is not informative of the parameter $\theta$ is a property of the distribution; but it doesn't stop us from proceeding to higher moments, and we still end up with one equation for one unknown: $k = 1$. The actual exception to this is when we might obtain a nonlinear system the unknown parameters that admits a unique solution with fewer or more equations than there are parameters to solve. $\endgroup$
    – heropup
    Commented Mar 25, 2021 at 20:16
  • $\begingroup$ Ah I think I remember learning about that in Linear Algebra, @heropup. Is that related to the "more rows than columns" part of linear dependence? $\endgroup$
    – JerBear
    Commented Mar 25, 2021 at 20:20
  • $\begingroup$ Also, would it be $\frac{1}{n-1}\sum_{i=1}^{n}{(X_i-\bar{X})^2}$ instead, since this is a sample estimation of the variance? $\endgroup$
    – JerBear
    Commented Mar 25, 2021 at 20:59

You must log in to answer this question.

Not the answer you're looking for? Browse other questions tagged .