Maximum Likelihood Estimation for Conditional Variance Models

Innovation Distribution

For conditional variance models, the innovation process is εt=σtzt, where zt follows a standardized Gaussian or Student's t distribution with ν>2 degrees of freedom. Specify your distribution choice in the model property Distribution.

The innovation variance, σt2, can follow a GARCH, EGARCH, or GJR conditional variance process.

If the model includes a mean offset term, then

εt=ytμ.

The estimate function for garch, egarch, and gjr models estimates parameters using maximum likelihood estimation. estimate returns fitted values for any parameters in the input model equal to NaN. estimate honors any equality constraints in the input model, and does not return estimates for parameters with equality constraints.

Loglikelihood Functions

Given the history of a process, innovations are conditionally independent. Let Ht denote the history of a process available at time t, t = 1,...,N. The likelihood function for the innovation series is given by

f(ε1,ε2,,εN|HN1)=t=1Nf(εt|Ht1),

where f is a standardized Gaussian or t density function.

The exact form of the loglikelihood objective function depends on the parametric form of the innovation distribution.

  • If zt has a standard Gaussian distribution, then the loglikelihood function is

    LLF=N2log(2π)12t=1Nlogσt212t=1Nεt2σt2.

  • If zt has a standardized Student's t distribution with ν>2 degrees of freedom, then the loglikelihood function is

    LLF=Nlog[Γ(ν+12)π(ν2)Γ(ν2)]12t=1Nlogσt2ν+12t=1Nlog[1+εt2σt2(ν2)].

estimate performs covariance matrix estimation for maximum likelihood estimates using the outer product of gradients (OPG) method.

See Also

| | | | |

Related Examples

More About

Was this topic helpful?