Generalized Method of Moments¶
Setup¶
Let \(\smash{\boldsymbol{Y}_t}\) be an \(\smash{(n \times 1)}\) vector of random variables and \(\smash{\boldsymbol{\theta}}\) a \(\smash{(k \times 1)}\) vector of parameters governing the process \(\smash{\{\boldsymbol{Y}_{t}\}}\).
- Denote the true parameter vector as \(\smash{\boldsymbol{\theta}_{0}}\).
Moment Conditions¶
Suppose we can specify an \(\smash{(r \times 1)}\) vector valued function \(\smash{\boldsymbol{h}(\boldsymbol{\theta},\boldsymbol{Y}_{t}): (\mathbb{R}^{k} \times \mathbb{R}^{n}) \rightarrow \mathbb{R}^{r}}\) such that:
Define \(\smash{\boldsymbol{\mathcal{Y}}_t = (\boldsymbol{y}_1, \ldots, \boldsymbol{y}_t)}\) and
Note that \(\smash{\boldsymbol{g}_{T}(\boldsymbol{\theta}| \boldsymbol{\mathcal{Y}}_T): \mathbb{R}^{k} \rightarrow \mathbb{R}^{r}}\).
GMM Estimator¶
We want to choose \(\smash{\hat{\boldsymbol{\theta}}_{gmm}}\) such that the sample moments \(\smash{\boldsymbol{g}_{T}(\hat{\boldsymbol{\theta}}_{gmm}| \boldsymbol{\mathcal{Y}}_T)}\) are close to zero.
- If \(\smash{r = k}\), we can choose \(\smash{\hat{\boldsymbol{\theta}}_{gmm}}\) such that \(\smash{\boldsymbol{g}_{T}(\hat{\boldsymbol{\theta}}_{gmm}| \boldsymbol{\mathcal{Y}}_T) = 0}\) because we have \(\smash{k}\) equations and \(\smash{k}\) unknowns.
- If \(\smash{r > k}\) we have more equations than unknowns; in general there is no \(\smash{\hat{\boldsymbol{\theta}}_{gmm}}\) such that \(\smash{\boldsymbol{g}_{T}(\hat{\boldsymbol{\theta}}_{gmm}| \boldsymbol{\mathcal{Y}}_T) = 0}\).
GMM Estimator¶
If \(\smash{r > k}\), we minimize a quadratic form:
- The matrix \(\smash{W_{T}}\) places more weight on some moment conditions and less on others.
- We might have to use numerical optimization to minimze \(\smash{Q_{T}(\boldsymbol{\theta}|\boldsymbol{\mathcal{Y}}_T)}\).
Example: t-distribution¶
The method of moments estimator of the t-distribution is a special case of the GMM estimator.
- \(\smash{\boldsymbol{Y}_{t}} = Y_t\).
- \(\smash{\boldsymbol{\theta} = \nu}\)
- \(\smash{W_{T} = 1}\)
- \(\smash{\boldsymbol{h}(\boldsymbol{\theta},\boldsymbol{Y}_{t}) = Y_{t}^{2} - \frac{\nu}{\nu - 2}}\).
Note that
Example: t-distribution¶
In this case, \(\smash{r = k = 1}\), and
Since \(\smash{r = k = 1}\), \(\smash{\hat{\nu}_{gmm}}\) can be chosen such that \(\smash{Q_{T}(\boldsymbol{\theta}|\boldsymbol{\mathcal{Y}}_T) = 0}\).
Example: t-distribution with \(\smash{r = 2}\)¶
Suppose we add a moment condition for the t-distribution.
- If \(\smash{\nu > 4}\), then
- In this case, \(\smash{ r = 2 > 1 = k}\).
- We now have more moment conditions than parameters.
Example: t-distribution with \(\smash{r = 2}\)¶
We map this problem into GMM form in the following way:
- \(\smash{\boldsymbol{Y}_{t} = Y_t}\)
- \(\smash{\boldsymbol{\theta} = \nu}\)
Example: t-distribution with \(\smash{r = 2}\)¶
The weighting matrix \(\smash{W_{T} = I_2}\) places equal weight on the two moment conditions.
- We could alter this matrix to emphasize one condition more than another.
GMM Consistency¶
If \(\smash{\boldsymbol{Y}_{t}}\) is strictly stationary and \(\smash{\boldsymbol{h}}\) continuous, a law of large numbers will hold:
Under certain regularity conditions, it can be shown that