Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
The NLMIXED Procedure

Modeling Assumptions and Notation

PROC NLMIXED operates under the following general framework for nonlinear mixed models. Assume that you have an observed data vector yi for each of i subjects, i = 1, ... ,s. The yi are assumed to be independent across i, but within-subject covariance is likely to exist because each of the elements of yi are measured on the same subject. As a statistical mechanism for modeling this within-subject covariance, assume that there exist latent random-effect vectors ui of small dimension (typically one or two) that are also independent across i. Assume also that an appropriate model linking yi and ui exists, leading to the joint probability density function
p(y_i | X_i, \phi, u_i) q(u_i | \xi)
where Xi is a matrix of observed explanatory variables and \phiand \xi are vectors of unknown parameters.

Let \theta = (\phi, \xi) and assume that it is of dimension n. Then inferences about \theta are based on the marginal likelihood function

m(\theta) = \prod_{i=1}^s \int p(y_i | X_i, \phi, u_i) 
 q(u_i | \xi) d u_i
In particular, the function
f(\theta) = - \log m(\theta)
is minimized over \theta numerically in order to estimate \theta, and the inverse Hessian (second derivative) matrix at the estimates provides an approximate variance-covariance matrix for the estimate of \theta. The function f(\theta) is referred to both as the negative log likelihood function and as the objective function for optimization.

As an example of the preceding general framework, consider the nonlinear growth curve example in the "Getting Started" section. Here, the conditional distribution p(y_i | X_i, \phi, u_i)is normal with mean

[(b1 + ui1)/(1 + exp[-(dij - b2)/ b3])]
and variance \sigma^2_e; thus \phi = (b_1,b_2,b_3,\sigma^2_e).Also, ui is a scalar and q(u_i | \xi) is normal with mean 0 and variance \sigma^2_u; thus \xi = \sigma^2_u.

The following additional notation is also found in this chapter. The quantity \theta^{(k)} refers to the parameter vector at the kth iteration, the function g(\theta) refers to the gradient vector \nabla f(\theta), and the matrix H(\theta) refers to the Hessian \nabla^2 f(\theta). Other symbols are used to denote various constants or option values.

Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
Top
Top

Copyright © 1999 by SAS Institute Inc., Cary, NC, USA. All rights reserved.