Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
The CALIS Procedure

Example 19.4: Linear Relations Among Factor Loadings

The correlation matrix from Kinzer and Kinzer (N=326) is used by Guttman (1957) as an example that yields an approximate simplex. McDonald (1980) uses this data set as an example of factor analysis where he supposes that the loadings of the second factor are a linear function of the loadings on the first factor, for example

b_{j2} = \alpha + \beta b_{j1} , j = 1,  ...  , n
This example is also discussed in Browne (1982). The matrix specification of the model is
C = F1F'1
with
F_1 = \pmatrix{b_{11}&\alpha + \beta b_{11}&u_{11}\cr
 b_{21}&\alpha + \beta b_{...
 ...beta b_{51}& & & & &u_{55}\cr
 b_{61}&\alpha + \beta b_{61}& & & & & &u_{66}\cr}
This example is recomputed by PROC CALIS to illustrate a simple application of the COSAN model statement combined with program statements. This example also serves to illustrate the identification problem.
   data Kinzer(TYPE=CORR);                                      
   Title "Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)";  
      _TYPE_ = 'CORR'; INPUT _NAME_ $ Obs1-Obs6;                
      datalines;                                                
   Obs1  1.00   .     .     .     .     .                       
   Obs2   .51  1.00   .     .     .     .                       
   Obs3   .46   .51  1.00   .     .     .                       
   Obs4   .46   .47   .54  1.00   .     .                       
   Obs5   .40   .39   .49   .57  1.00   .                       
   Obs6   .33   .39   .47   .45   .56  1.00                     
   ;

In a first test run of PROC CALIS, the same model is used as reported in McDonald (1980). Using the Levenberg-Marquardt optimization algorithm, this example specifies maximum likelihood estimation in the following code:

   proc calis data=Kinzer method=max outram=ram nobs=326;         
      Title2 "Linearly Related Factor Analysis, (Mcdonald,1980)";   
      Title3 "Identification Problem";                              
      Cosan F(8,Gen) * I(8,Ide);                                   
      Matrix F                                                     
             [ ,1]= X1-X6,                                        
             [ ,2]= X7-X12,                                       
             [1,3]= X13-X18;                                      
      Parms Alfa = .5 Beta = -.5;                                   
         X7  = Alfa + Beta * X1;                                      
         X8  = Alfa + Beta * X2;                                      
         X9  = Alfa + Beta * X3;                                      
         X10 = Alfa + Beta * X4;                                      
         X11 = Alfa + Beta * X5;                                      
         X12 = Alfa + Beta * X6;                                      
      Bounds X13-X18 >= 0.;                                         
      Vnames F Fact1 Fact2 Uvar1-Uvar6;                             
   run;

The pattern of the initial values is displayed in vector and in matrix form. You should always read this output very carefully, particularly when you use your own programming statements to constrain the matrix elements. The vector form shows the mapping of the model parameters to indices of the vector X that is optimized. The matrix form indicates parameter elements that are constrained by program statements by indices of X in angle brackets ( < > ). An asterisk trailing the iteration number in the displayed optimization history of the Levenberg-Marquardt algorithm indicates that the optimization process encountered a singular Hessian matrix. When this happens, especially in the last iterations, the model may not be properly identified. The computed \chi^2 value of 10.337 for 7 degrees of freedom and the computed unique loadings agree with those reported by McDonald (1980), but the maximum likelihood estimates for the common factor loadings differ to some degree. The common factor loadings can be subjected to transformations that do not increase the value of the optimization criterion because the problem is not identified. An estimation problem that is not fully identified can lead to different solutions caused only by different initial values, different optimization techniques, or computers with different machine precision or floating-point arithmetic.

To overcome the identification problem in the first model, restart PROC CALIS with a simple modification to the model in which the former parameter X1 is fixed to 0. This leads to 8 instead of 7 degrees of freedom. The following code produces results that are partially displayed in Output 19.4.1.

   data ram2(TYPE=RAM); set ram;                                 
      if _type_ = 'ESTIM' then                                    
         if _name_ = 'X1' then do;                                 
            _name_ = ' '; _estim_ = 0.;                            
         end;                                                      
   run;                                                          
                                                                 
   proc calis data=Kinzer method=max inram=ram2 nobs=326;        
      Title2 "Linearly Related Factor Analysis, (Mcdonald,1980)";  
      Title3 "Identified Model";                                   
      Parms Alfa = .5 Beta = -.5;                                  
         X7  = Alfa;                                                 
         X8  = Alfa + Beta * X2;                                     
         X9  = Alfa + Beta * X3;                                     
         X10 = Alfa + Beta * X4;                                     
         X11 = Alfa + Beta * X5;                                     
         X12 = Alfa + Beta * X6;                                     
      Bounds X13-X18 >= 0.;                                        
   run;

Output 19.4.1: Linearly Related Factor Analysis: Identification Problem
 
Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)
Linearly Related Factor Analysis, (Mcdonald,1980)
Identified Model

The CALIS Procedure
Covariance Structure Analysis: Pattern and Initial Values

COSAN Model Statement
  Matrix Rows Columns Matrix Type
Term 1 1 F 6 8 GENERAL  
  2 I 8 8 IDENTITY  

 


 
Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)
Linearly Related Factor Analysis, (Mcdonald,1980)
Identified Model

The CALIS Procedure
Covariance Structure Analysis: Maximum Likelihood Estimation

Parameter Estimates 13
Functions (Observations) 21
Lower Bounds 6
Upper Bounds 0
 
Optimization Start
Active Constraints 0 Objective Function 0.3234289189
Max Abs Gradient Element 2.2633860283 Radius 5.8468569273

 
Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)
Linearly Related Factor Analysis, (Mcdonald,1980)
Identified Model

The CALIS Procedure
Covariance Structure Analysis: Maximum Likelihood Estimation

Iteration   Restarts Function
Calls
Active
Constraints
  Objective
Function
Objective
Function
Change
Max Abs
Gradient
Element
Lambda Ratio
Between
Actual
and
Predicted
Change
1   0 2 0   0.07994 0.2435 0.3984 0 0.557
2   0 3 0   0.03334 0.0466 0.0672 0 1.202
3   0 4 0   0.03185 0.00150 0.00439 0 1.058
4   0 5 0   0.03181 0.000034 0.00236 0 0.811
5   0 6 0   0.03181 3.982E-6 0.000775 0 0.591
6   0 7 0   0.03181 9.275E-7 0.000490 0 0.543
7   0 8 0   0.03181 2.402E-7 0.000206 0 0.526
8   0 9 0   0.03181 6.336E-8 0.000129 0 0.514
9   0 10 0   0.03181 1.687E-8 0.000054 0 0.505
10   0 11 0   0.03181 4.521E-9 0.000034 0 0.498
11   0 12 0   0.03181 1.217E-9 0.000014 0 0.493
12   0 13 0   0.03181 3.29E-10 8.971E-6 0 0.489
 
Optimization Results
Iterations 12 Function Calls 14
Jacobian Calls 13 Active Constraints 0
Objective Function 0.0318073951 Max Abs Gradient Element 8.9711916E-6
Lambda 0 Actual Over Pred Change 0.4888109559
Radius 0.0002016088    
 
ABSGCONV convergence criterion satisfied.

 
Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)
Linearly Related Factor Analysis, (Mcdonald,1980)
Identified Model

The CALIS Procedure
Covariance Structure Analysis: Maximum Likelihood Estimation

Fit Function 0.0318
Goodness of Fit Index (GFI) 0.9897
GFI Adjusted for Degrees of Freedom (AGFI) 0.9730
Root Mean Square Residual (RMR) 0.0409
Parsimonious GFI (Mulaik, 1989) 0.5278
Chi-Square 10.3374
Chi-Square DF 8
Pr > Chi-Square 0.2421
Independence Model Chi-Square 682.87
Independence Model Chi-Square DF 15
RMSEA Estimate 0.0300
RMSEA 90% Lower Confidence Limit .
RMSEA 90% Upper Confidence Limit 0.0756
ECVI Estimate 0.1136
ECVI 90% Lower Confidence Limit .
ECVI 90% Upper Confidence Limit 0.1525
Probability of Close Fit 0.7137
Bentler's Comparative Fit Index 0.9965
Normal Theory Reweighted LS Chi-Square 10.1441
Akaike's Information Criterion -5.6626
Bozdogan's (1987) CAIC -43.9578
Schwarz's Bayesian Criterion -35.9578
McDonald's (1989) Centrality 0.9964
Bentler & Bonett's (1980) Non-normed Index 0.9934
Bentler & Bonett's (1980) NFI 0.9849
James, Mulaik, & Brett (1982) Parsimonious NFI 0.5253
Z-Test of Wilson & Hilferty (1931) 0.7019
Bollen (1986) Normed Index Rho1 0.9716
Bollen (1988) Non-normed Index Delta2 0.9965
Hoelter's (1983) Critical N 489

 
Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)
Linearly Related Factor Analysis, (Mcdonald,1980)
Identified Model

The CALIS Procedure
Covariance Structure Analysis: Maximum Likelihood Estimation

Estimated Parameter Matrix F[6:8]
Standard Errors and t Values
General Matrix
  Fact1 Fact2 Uvar1 Uvar2 Uvar3 Uvar4 Uvar5 Uvar6
Obs1 0
0
0
 
0.7151
0.0405
17.6382
<X7>
0.7283
0.0408
17.8276
[X13]
0
0
0
 
0
0
0
 
0
0
0
 
0
0
0
 
0
0
0
 
Obs2 -0.0543
0.1042
-0.5215
[X2]
0.7294
0.0438
16.6655
<X8>
0
0
0
 
0.6707
0.0472
14.2059
[X14]
0
0
0
 
0
0
0
 
0
0
0
 
0
0
0
 
Obs3 0.1710
0.0845
2.0249
[X3]
0.6703
0.0396
16.9077
<X9>
0
0
0
 
0
0
0
 
0.6983
0.0324
21.5473
[X15]
0
0
0
 
0
0
0
 
0
0
0
 
Obs4 0.2922
0.0829
3.5224
[X4]
0.6385
0.0462
13.8352
<X10>
0
0
0
 
0
0
0
 
0
0
0
 
0.6876
0.0319
21.5791
[X16]
0
0
0
 
0
0
0
 
Obs5 0.5987
0.1003
5.9665
[X5]
0.5582
0.0730
7.6504
<X11>
0
0
0
 
0
0
0
 
0
0
0
 
0
0
0
 
0.5579
0.0798
6.9938
[X17]
0
0
0
 
Obs6 0.4278
0.0913
4.6844
[X6]
0.6029
0.0586
10.2928
<X12>
0
0
0
 
0
0
0
 
0
0
0
 
0
0
0
 
0
0
0
 
0.7336
0.0400
18.3580
[X18]

 
Data Matrix of Kinzer & Kinzer, see GUTTMAN (1957)
Linearly Related Factor Analysis, (Mcdonald,1980)
Identified Model

The CALIS Procedure
Covariance Structure Analysis: Maximum Likelihood Estimation

Additional PARMS and Dependent Parameters
The Number of Dependent Parameters is 6
Parameter Estimate Standard
Error
t Value
Alfa 0.71511 0.04054 17.64
Beta -0.26217 0.12966 -2.02
X7 0.71511 0.04054 17.64
X8 0.72936 0.04376 16.67
X9 0.67027 0.03964 16.91
X10 0.63851 0.04615 13.84
X11 0.55815 0.07296 7.65
X12 0.60295 0.05858 10.29

The lambda value of the iteration history indicates that Newton steps can always be performed. Because no singular Hessian matrices (which can slow down the convergence rate considerably) are computed, this example needs just 12 iterations compared to the 17 needed in the previous example. Note that the number of iterations may be machine-dependent. The value of the fit funciton, the residuals, and the \chi^2 value agree with the values obtained in fitting the first model. This indicates that this second model is better identified than the first one. It is fully identified, as indicated by the fact that the Hessian matrix is nonsingular.

Chapter Contents
Chapter Contents
Previous
Previous
Next
Next
Top
Top

Copyright © 1999 by SAS Institute Inc., Cary, NC, USA. All rights reserved.