Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
The NLP Procedure

Hessian and CRP Jacobian Scaling

The rows and columns of the Hessian and crossproduct Jacobian matrix can be scaled when using the trust-region, Newton-Raphson, Double Dogleg, and Levenberg-Marquardt optimization techniques. Each element Gi,j, i,j = 1, ... ,n, is divided by the scaling factor di * dj, where the scaling vector d = (d1, ... ,dn) is iteratively updated in a way specified by the HESCAL=i option, as follows:
i = 0
: No scaling is done (equivalent to di=1).
i \neq 0
: First iteration and each restart iteration sets:
d_i^{(0)} = \sqrt{\max(| G^{(0)}_{i,i}|,\epsilon)}
i = 1
: refer to Mor\acute{e} (1978):
d_i^{(k+1)} = \max(d_i^{(k)},\sqrt{\max(| G^{(k)}_{i,i}|,\epsilon)})
i = 2
: refer to Dennis, Gay, and Welsch (1981):
d_i^{(k+1)} = \max(.6 * d_i^{(k)},\sqrt{\max(| G^{(k)}_{i,i}|,\epsilon)})
i = 3
: di is reset in each iteration:
d_i^{(k+1)} = \sqrt{\max(| G^{(k)}_{i,i}|,\epsilon)}
where \epsilon is the relative machine precision or, equivalently, the largest double precision value that when added to 1 results in 1.

Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
Top
Top

Copyright © 1999 by SAS Institute Inc., Cary, NC, USA. All rights reserved.