# Principal Components and Factor Analysis

This section covers principal components and factor analysis. The latter includes both exploratory and confirmatory methods.

## Principal Components

The **princomp( )** function produces an unrotated principal component analysis.

`# Pricipal Components Analysis`

# entering raw data and extracting PCs

#
from the correlation matrix

fit <- princomp(mydata, cor=TRUE)

summary(fit) # print variance accounted for

loadings(fit) # pc loadings

plot(fit,type="lines") # scree plot

fit$scores # the principal components

biplot(fit)

Use **cor=FALSE** to base the principal components on the covariance matrix. Use the **covmat=** option to enter a correlation or covariance matrix directly. If entering a covariance matrix, include the option** n.obs=.**

The **principal( )** function in the psych package can be used to extract and rotate principal components.

`# Varimax Rotated Principal Components`

# retaining 5 components

library(psych)

fit <- principal(mydata, nfactors=5, rotate="varimax")

fit # print results

*mydata* can be a raw data matrix or a covariance matrix. Pairwise deletion of missing data is used. **rotate **can "none", "varimax", "quatimax", "promax", "oblimin", "simplimax", or "cluster"

## Exploratory Factor Analysis

The **factanal( )** function produces maximum likelihood factor analysis.

`# Maximum Likelihood Factor Analysis`

# entering raw data and extracting 3 factors,

#
with varimax rotation

fit <- factanal(mydata, 3, rotation="varimax")

print(fit, digits=2, cutoff=.3, sort=TRUE)

# plot factor 1 by factor 2

load <- fit$loadings[,1:2]

plot(load,type="n") # set up plot

text(load,labels=names(mydata),cex=.7) # add variable names

The **rotation=** options include "varimax", "promax", and "none". Add the option **scores=**"regression" or "Bartlett" to produce factor scores. Use the **covmat=** option to enter a correlation or covariance matrix directly. If entering a covariance matrix, include the option** n.obs=.**

The ** factor.pa( ) function in the psych **package offers a number of factor analysis related functions, including principal axis factoring.

`# Principal Axis Factor Analysis`

library(psych)

fit <- factor.pa(mydata, nfactors=3, rotation="varimax")

fit # print results

*mydata* can be a raw data matrix or a covariance matrix. Pairwise deletion of missing data is used. Rotation can be "varimax" or "promax".

### Determining the Number of Factors to Extract

A crucial decision in exploratory factor analysis is how many factors to extract. The nFactors package offer a suite of functions to aid in this decision. Details on this methodology can be found in a PowerPoint presentation by Raiche, Riopel, and Blais. Of course, any factor solution must be interpretable to be useful.

`# Determine Number of Factors to Extract`

library(nFactors)

ev <- eigen(cor(mydata)) # get eigenvalues

ap <- parallel(subject=nrow(mydata),var=ncol(mydata),

rep=100,cent=.05)

nS <- nScree(x=ev$values, aparallel=ap$eigen$qevpea)

plotnScree(nS)

### Going Further

The FactoMineR package offers a large number of additional functions for exploratory factor analysis. This includes the use of both quantitative and qualitative variables, as well as the inclusion of supplimentary variables and observations. Here is an example of the types of graphs that you can create with this package.

`# PCA Variable Factor Map `

library(FactoMineR)

result <- PCA(mydata) # graphs generated automatically

Thye GPARotation package offers a wealth of rotation options beyond varimax and promax.

## Structual Equation Modeling

**Confirmatory Factor Analysis** **(CFA)** is a subset of the much wider **Structural Equation Modeling** **(SEM) **methodology. SEM is provided in R via the sem package. Models are entered via RAM specification (similar to PROC CALIS in SAS). While sem is a comprehensive package, my recommendation is that if you are doing significant SEM work, you spring for a copy of AMOS. It can be much more user-friendly and creates more attractive and publication ready output. Having said that, here is a CFA example using sem.

Assume that we have six observered variables (X1, X2, ..., X6). We hypothesize that there are two unobserved latent factors (F1, F2) that underly the observed variables as described in this diagram. X1, X2, and X3 load on F1 (with loadings lam1, lam2, and lam3). X4, X5, and X6 load on F2 (with loadings lam4, lam5, and lam6). The double headed arrow indicates the covariance between the two latent factors (F1F2). e1 thru e6 represent the residual variances (variance in the observed variables not accounted for by the two latent factors). We set the variances of F1 and F2 equal to one so that the parameters will have a scale. This will result in F1F2 representing the correlation between the two latent factors.

For **sem**, we need the covariance matrix of the observed variables - thus the **cov( )** statement in the code below. The CFA model is specified using the **specify.model( ) **function. The format is *arrow specification*, *parameter name*, *start value*. Choosing a start value of NA tells the program to choose a start value rather than supplying one yourself. Note that the variance of F1 and F2 are fixed at 1 (NA in the second column). The blank line is required to end the RAM specification.

`# Simple CFA Model`

library(sem)

mydata.cov <- cov(mydata)

model.mydata <- specify.model()

F1 -> X1, lam1, NA

F1 -> X2, lam2, NA

F1 -> X3, lam3, NA

F2 -> X4, lam4, NA

F2 -> X5, lam5, NA

F2 -> X6, lam6, NA

X1 <-> X1, e1, NA

X2 <-> X2, e2, NA

X3 <-> X3, e3, NA

X4 <-> X4, e4, NA

X5 <-> X5, e5, NA

X6 <-> X6, e6, NA

F1 <-> F1, NA, 1

F2 <-> F2, NA, 1

F1 <-> F2, F1F2, NA

mydata.sem <- sem(model.mydata, mydata.cov, nrow(mydata))

# print results (fit indices, paramters, hypothesis tests)

summary(mydata.sem)

# print standardized coefficients (loadings)

std.coef(mydata.sem)

You can use the **boot.sem( ) **function to bootstrap the structual equation model. See **help(boot.sem)** for details. Additionally, the function **mod.indices(** **)** will produce modification indices. Using modification indices to improve model fit by respecifying the parameters moves you from a confirmatory to an exploratory analysis.

For more information on **sem**, see Structural Equation Modeling with the sem Package in R, by John Fox.

## To Practice

To practice improving predictions, try the Kaggle R Tutorial on Machine Learning