Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
The MODEL Procedure

Minimization Methods

PROC MODEL currently supports two methods for minimizing the objective function. These methods are described in the following sections.

GAUSS

The Gauss-Newton parameter-change vector for a system with g equations, n non-missing observations, and p unknown parameters is

{{\Delta}}=(X'X)^{-1}X'r

where {\Delta} is the change vector, X is the stacked ng ×p Jacobian matrix of partial derivatives of the residuals with respect to the parameters, and r is an ng ×1vector of the stacked residuals. The components of X and r are weighted by the S-1 matrix. When instrumental methods are used, X and r are the projections of the Jacobian matrix and residuals vector in the instruments space and not the Jacobian and residuals themselves. In the preceding formula, S and W are suppressed. If instrumental variables are used, then the change vector becomes:

{{\Delta}}=(X'(S^{-1}{\otimes}W)X)^{-1}X'(S^{-1}{\otimes}W)r

This vector is computed at the end of each iteration. The objective function is then computed at the changed parameter values at the start of the next iteration. If the objective function is not improved by the change, the {\Delta} vector is reduced by one-half and the objective function is re-evaluated. The change vector will be halved up to MAXSUBITER= times until the objective function is improved. For FIML the X'X matrix is substituted with one of three choices for approximations to the Hessian. See the "FIML Estimation" section in this chapter.

MARQUARDT

The Marquardt-Levenberg parameter change vector is

{{\Delta}}=(X'X+{\lambda}\rm{diag}({X'X}))^{-1}{X'r}

where {\Delta} is the change vector, and X and r are the same as for the Gauss-Newton method, described in the preceding section. Before the iterations start, {\lambda} is set to a small value (1E-6). At each iteration, the objective function is evaluated at the parameters changed by {\Delta}. If the objective function is not improved, {\lambda} is increased to 10{\lambda} and the step is tried again. {\lambda} can be increased up to MAXSUBITER= times to a maximum of 1E15 (whichever comes first) until the objective function is improved. For the start of the next iteration, {\lambda} is reduced to max({\lambda}/10,1E-10).

Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
Top
Top

Copyright © 1999 by SAS Institute Inc., Cary, NC, USA. All rights reserved.