Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
The NLP Procedure

Computational Resources

Since nonlinear optimization is an iterative process that depends on many factors, it is difficult to estimate how much computer time is necessary to compute an optimal solution satisfying one of the termination criteria. The MAXTIME=, MAXITER=, and MAXFU= options can be used to restrict the amount of CPU time, the number of iterations, and the number of function calls in a single run of PROC NLP.

In each iteration k, the NRRIDG and LEVMAR techniques use symmetric Householder transformations to decompose the n ×n Hessian (crossproduct Jacobian) matrix G

G = V' T V     ,     V : orthogonal ,     T : tridiagonal
to compute the (Newton) search direction s
s(k) = - G(k)-1 g(k)     ,     k = 1,2,3, ... .
The QUADAS, TRUREG, NEWRAP, and HYQUAN techniques use the Cholesky decomposition to solve the same linear system while computing the search direction. The QUANEW, DBLDOG, CONGRA, and NMSIMP techniques do not need to invert or decompose a Hessian or crossproduct Jacobian matrix and thus require less computational resources then the first group of techniques.

The larger the problem, the more time is spent computing function values and derivatives. Therefore, many researchers compare optimization techniques by counting and comparing the respective numbers of function, gradient, and Hessian (crossproduct Jacobian) evaluations. You can save computer time and memory by specifying derivatives (using the GRADIENT, JACOBIAN, CRPJAC, or HESSIAN statement) since you will typically produce a more efficient representation than the internal derivative compiler.

Finite difference approximations of the derivatives are expensive since they require additional function or gradient calls:

Many applications need considerably more time for computing second-order derivatives (Hessian matrix) than for first-order derivatives (gradient). In such cases, a (dual) quasi-Newton or conjugate gradient technique is recommended, that does not require second-order derivatives.

The following table shows for each optimization technique which derivatives are needed (FOD: first-order derivatives; SOD: second-order derivatives), what kind of constraints are supported (BC: boundary constraints; LIC: linear constraints), and the minimal memory (number of double floating point numbers) required. For various reasons, there are additionally about 7n+m double floating point numbers needed.

Quadratic Programming FOD SOD BC LIC Memory
LICOMP--xx18n + 3nn
QUADAS--xx1n + 2nn/2
General OptimizationFODSODBCLICMemory
TRUREGxxxx4n + 2nn/2
NEWRAPxxxx2n + 2nn/2
NRRIDGxxxx6n + nn/2
QUANEWx-xx1n + nn/2
DBLDOGx-xx7n + nn/2
CONGRAx-xx3n
NMSIMP--xx4n + nn
Least-SquaresFODSODBCLICMemory
LEVMARx-xx6n + nn/2
HYQUANx-xx2n + nn/2 + 3m

Notes:

The total amount of memory needed to run an optimization technique consists of the technique-specific memory listed in the preceeding table, plus additional blocks of memory as shown in the following table:

  double int long 8byte
Basic Requirement7n+mn3nn+m
DATA= data setv--v
JACOBIANm(n+2)---
CRPJAC statementnn/2---
HESSIAN statementnn/2---
COV= statement(2*)nn/2 + n---
Scaling vectorn---
BOUNDS statement2nn--
Bounds in INEST=2n---
LINCON and TRUREGc(n+1)+nn+ nn/2+4n3c--
LINCON and otherc(n+1)+nn+2nn/2+4n3c--

Notes:

Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
Top
Top

Copyright © 1999 by SAS Institute Inc., Cary, NC, USA. All rights reserved.