Factor Analysis

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 22

Factor analysis

Factor analysis is a statistical method used to describe variability among observed, correlated variables in
terms of a potentially lower number of unobserved variables called factors. For example, it is possible that
variations in six observed variables mainly reflect the variations in two unobserved (underlying) variables.
Factor analysis searches for such joint variations in response to unobserved latent variables. The observed
variables are modelled as linear combinations of the potential factors plus "error" terms, hence factor
analysis can be thought of as a special case of errors-in-variables models.[1]

Simply put, the factor loading of a variable quantifies the extent to which the variable is related to a given
factor.[2]

A common rationale behind factor analytic methods is that the information gained about the
interdependencies between observed variables can be used later to reduce the set of variables in a dataset.
Factor analysis is commonly used in psychometrics, personality psychology, biology, marketing, product
management, operations research, finance, and machine learning. It may help to deal with data sets where
there are large numbers of observed variables that are thought to reflect a smaller number of
underlying/latent variables. It is one of the most commonly used inter-dependency techniques and is used
when the relevant set of variables shows a systematic inter-dependence and the objective is to find out the
latent factors that create a commonality.

Statistical model

Definition

The model attempts to explain a set of observations in each of individuals with a set of common
factors ( ) where there are fewer factors per unit than observations per unit ( ). Each individual has
of their own common factors, and these are related to the observations via the factor loading matrix (
), for a single observation, according to

where

is the value of the th observation of the th individual,


is the observation mean for the th observation,
is the loading for the th observation of the th factor,
is the value of the th factor of the th individual, and
is the th unobserved stochastic error term with mean zero and finite variance.

In matrix notation

where observation matrix , loading matrix , factor matrix , error term


matrix and mean matrix whereby the th element is simply .
Also we will impose the following assumptions on :

1. and are independent.


2. ; where is Expectation
3. where is the covariance matrix, to make sure that the factors are
uncorrelated, and is the identity matrix.

Suppose . Then

and therefore, from conditions 1 and 2 imposed on above, and


, giving

or, setting ,

Note that for any orthogonal matrix , if we set and , the criteria for being factors
and factor loadings still hold. Hence a set of factors and factor loadings is unique only up to an orthogonal
transformation.

Example

Suppose a psychologist has the hypothesis that there are two kinds of intelligence, "verbal intelligence" and
"mathematical intelligence", neither of which is directly observed.[note 1] Evidence for the hypothesis is
sought in the examination scores from each of 10 different academic fields of 1000 students. If each student
is chosen randomly from a large population, then each student's 10 scores are random variables. The
psychologist's hypothesis may say that for each of the 10 academic fields, the score averaged over the
group of all students who share some common pair of values for verbal and mathematical "intelligences" is
some constant times their level of verbal intelligence plus another constant times their level of mathematical
intelligence, i.e., it is a linear combination of those two "factors". The numbers for a particular subject, by
which the two kinds of intelligence are multiplied to obtain the expected score, are posited by the
hypothesis to be the same for all intelligence level pairs, and are called "factor loading" for this subject.
For example, the hypothesis may hold that the predicted average student's aptitude in the field of astronomy
is

{10 × the student's verbal intelligence} + {6 × the student's mathematical intelligence}.

The numbers 10 and 6 are the factor loadings associated with astronomy. Other academic subjects may
have different factor loadings.

Two students assumed to have identical degrees of verbal and mathematical intelligence may have different
measured aptitudes in astronomy because individual aptitudes differ from average aptitudes (predicted
above) and because of measurement error itself. Such differences make up what is collectively called the
"error" — a statistical term that means the amount by which an individual, as measured, differs from what
is average for or predicted by his or her levels of intelligence (see errors and residuals in statistics).
The observable data that go into factor analysis would be 10 scores of each of the 1000 students, a total of
10,000 numbers. The factor loadings and levels of the two kinds of intelligence of each student must be
inferred from the data.

Mathematical model of the same example

In the following, matrices will be indicated by indexed variables. "Subject" indices will be indicated using
letters , and , with values running from to which is equal to in the above example. "Factor"
indices will be indicated using letters , and , with values running from to which is equal to in the
above example. "Instance" or "sample" indices will be indicated using letters , and , with values running
from to . In the example above, if a sample of students participated in the exams,
the th student's score for the th exam is given by . The purpose of factor analysis is to characterize the
correlations between the variables of which the are a particular instance, or set of observations. In
order for the variables to be on equal footing, they are normalized into standard scores :

where the sample mean is:

and the sample variance is given by:

The factor analysis model for this particular sample is then:

or, more succinctly:

where

is the th student's "verbal intelligence",


is the th student's "mathematical intelligence",
are the factor loadings for the th subject, for .

In matrix notation, we have


Observe that by doubling the scale on which "verbal intelligence"—the first component in each column of
—is measured, and simultaneously halving the factor loadings for verbal intelligence makes no difference
to the model. Thus, no generality is lost by assuming that the standard deviation of the factors for verbal
intelligence is . Likewise for mathematical intelligence. Moreover, for similar reasons, no generality is lost
by assuming the two factors are uncorrelated with each other. In other words:

where is the Kronecker delta ( when and when ).The errors are assumed to be
independent of the factors:

Note that, since any rotation of a solution is also a solution, this makes interpreting the factors difficult. See
disadvantages below. In this particular example, if we do not know beforehand that the two types of
intelligence are uncorrelated, then we cannot interpret the two factors as the two different types of
intelligence. Even if they are uncorrelated, we cannot tell which factor corresponds to verbal intelligence
and which corresponds to mathematical intelligence without an outside argument.

The values of the loadings , the averages , and the variances of the "errors" must be estimated given
the observed data and (the assumption about the levels of the factors is fixed for a given ). The
"fundamental theorem" may be derived from the above conditions:

The term on the left is the -term of the correlation matrix (a matrix derived as the product of the
matrix of standardized observations with its transpose) of the observed data, and its diagonal
elements will be s. The second term on the right will be a diagonal matrix with terms less than unity. The
first term on the right is the "reduced correlation matrix" and will be equal to the correlation matrix except
for its diagonal values which will be less than unity. These diagonal elements of the reduced correlation
matrix are called "communalities" (which represent the fraction of the variance in the observed variable that
is accounted for by the factors):

The sample data will not exactly obey the fundamental equation given above due to sampling errors,
inadequacy of the model, etc. The goal of any analysis of the above model is to find the factors and
loadings which give a "best fit" to the data. In factor analysis, the best fit is defined as the minimum of
the mean square error in the off-diagonal residuals of the correlation matrix:[3]

This is equivalent to minimizing the off-diagonal components of the error covariance which, in the model
equations have expected values of zero. This is to be contrasted with principal component analysis which
seeks to minimize the mean square error of all residuals.[3] Before the advent of high-speed computers,
considerable effort was devoted to finding approximate solutions to the problem, particularly in estimating
the communalities by other means, which then simplifies the problem considerably by yielding a known
reduced correlation matrix. This was then used to estimate the factors and the loadings. With the advent of
high-speed computers, the minimization problem can be solved iteratively with adequate speed, and the
communalities are calculated in the process, rather than being needed beforehand. The MinRes algorithm is
particularly suited to this problem, but is hardly the only iterative means of finding a solution.

If the solution factors are allowed to be correlated (as in 'oblimin' rotation, for example), then the
corresponding mathematical model uses skew coordinates rather than orthogonal coordinates.

Geometric interpretation

The parameters and variables of factor analysis


can be given a geometrical interpretation. The data
( ), the factors ( ) and the errors ( ) can be
viewed as vectors in an -dimensional Euclidean
space (sample space), represented as , and
respectively. Since the data are standardized,
the data vectors are of unit length ( ).
The factor vectors define an -dimensional linear
subspace (i.e. a hyperplane) in this space, upon
which the data vectors are projected orthogonally.
This follows from the model equation
Geometric interpretation of Factor Analysis parameters
for 3 respondents to question "a". The "answer" is
represented by the unit vector , which is projected
onto a plane defined by two orthonormal vectors and
and the independence of the factors and the errors: . The projection vector is and the error is
. In the above example, the perpendicular to the plane, so that . The
hyperplane is just a 2-dimensional plane defined projection vector may be represented in terms of the
by the two factor vectors. The projection of the factor vectors as . The square of
data vectors onto the hyperplane is given by the length of the projection vector is the communality:
. If another data vector were plotted, the
cosine of the angle between and would be  :
the -entry in the correlation matrix. (Adapted from
Harman Fig. 4.3)[3]
and the errors are vectors from that projected point
to the data point and are perpendicular to the
hyperplane. The goal of factor analysis is to find a hyperplane which is a "best fit" to the data in some
sense, so it doesn't matter how the factor vectors which define this hyperplane are chosen, as long as they
are independent and lie in the hyperplane. We are free to specify them as both orthogonal and normal (
) with no loss of generality. After a suitable set of factors are found, they may also be
arbitrarily rotated within the hyperplane, so that any rotation of the factor vectors will define the same
hyperplane, and also be a solution. As a result, in the above example, in which the fitting hyperplane is two
dimensional, if we do not know beforehand that the two types of intelligence are uncorrelated, then we
cannot interpret the two factors as the two different types of intelligence. Even if they are uncorrelated, we
cannot tell which factor corresponds to verbal intelligence and which corresponds to mathematical
intelligence, or whether the factors are linear combinations of both, without an outside argument.

The data vectors have unit length. The entries of the correlation matrix for the data are given by
. The correlation matrix can be geometrically interpreted as the cosine of the angle between
the two data vectors and . The diagonal elements will clearly be s and the off diagonal elements will
have absolute values less than or equal to unity. The "reduced correlation matrix" is defined as

The goal of factor analysis is to choose the fitting hyperplane such that the reduced correlation matrix
reproduces the correlation matrix as nearly as possible, except for the diagonal elements of the correlation
matrix which are known to have unit value. In other words, the goal is to reproduce as accurately as
possible the cross-correlations in the data. Specifically, for the fitting hyperplane, the mean square error in
the off-diagonal components

is to be minimized, and this is accomplished by minimizing it with respect to a set of orthonormal factor
vectors. It can be seen that

The term on the right is just the covariance of the errors. In the model, the error covariance is stated to be a
diagonal matrix and so the above minimization problem will in fact yield a "best fit" to the model: It will
yield a sample estimate of the error covariance which has its off-diagonal components minimized in the
mean square sense. It can be seen that since the are orthogonal projections of the data vectors, their
length will be less than or equal to the length of the projected data vector, which is unity. The square of
these lengths are just the diagonal elements of the reduced correlation matrix. These diagonal elements of
the reduced correlation matrix are known as "communalities":

Large values of the communalities will indicate that the fitting hyperplane is rather accurately reproducing
the correlation matrix. The mean values of the factors must also be constrained to be zero, from which it
follows that the mean values of the errors will also be zero.

Practical implementation

Types of factor analysis

Exploratory factor analysis

Exploratory factor analysis (EFA) is used to identify complex interrelationships among items and group
items that are part of unified concepts.[4] The researcher makes no a priori assumptions about relationships
among factors.[4]

Confirmatory factor analysis

Confirmatory factor analysis (CFA) is a more complex approach that tests the hypothesis that the items are
associated with specific factors.[4] CFA uses structural equation modeling to test a measurement model
whereby loading on the factors allows for evaluation of relationships between observed variables and
unobserved variables.[4] Structural equation modeling approaches can accommodate measurement error
and are less restrictive than least-squares estimation.[4] Hypothesized models are tested against actual data,
and the analysis would demonstrate loadings of observed variables on the latent variables (factors), as well
as the correlation between the latent variables.[4]

Types of factor extraction

Principal component analysis (PCA) is a widely used method for factor extraction, which is the first phase
of EFA.[4] Factor weights are computed to extract the maximum possible variance, with successive
factoring continuing until there is no further meaningful variance left.[4] The factor model must then be
rotated for analysis.[4]

Canonical factor analysis, also called Rao's canonical factoring, is a different method of computing the
same model as PCA, which uses the principal axis method. Canonical factor analysis seeks factors that
have the highest canonical correlation with the observed variables. Canonical factor analysis is unaffected
by arbitrary rescaling of the data.

Common factor analysis, also called principal factor analysis (PFA) or principal axis factoring (PAF), seeks
the fewest factors which can account for the common variance (correlation) of a set of variables.

Image factoring is based on the correlation matrix of predicted variables rather than actual variables, where
each variable is predicted from the others using multiple regression.

Alpha factoring is based on maximizing the reliability of factors, assuming variables are randomly sampled
from a universe of variables. All other methods assume cases to be sampled and variables fixed.

Factor regression model is a combinatorial model of factor model and regression model; or alternatively, it
can be viewed as the hybrid factor model,[5] whose factors are partially known.

Terminology

Factor loadings
Communality is the square of the standardized outer loading of an item. Analogous to
Pearson's r-squared, the squared factor loading is the percent of variance in that indicator
variable explained by the factor. To get the percent of variance in all the variables
accounted for by each factor, add the sum of the squared factor loadings for that factor
(column) and divide by the number of variables. (Note the number of variables equals the
sum of their variances as the variance of a standardized variable is 1.) This is the same as
dividing the factor's eigenvalue by the number of variables.
When interpreting, by one rule of thumb in confirmatory factor analysis, factor loadings
should be .7 or higher to confirm that independent variables identified a priori are
represented by a particular factor, on the rationale that the .7 level corresponds to about
half of the variance in the indicator being explained by the factor. However, the .7 standard
is a high one and real-life data may well not meet this criterion, which is why some
researchers, particularly for exploratory purposes, will use a lower level such as .4 for the
central factor and .25 for other factors. In any event, factor loadings must be interpreted in
the light of theory, not by arbitrary cutoff levels.
In oblique rotation, one may examine both a pattern matrix and a structure matrix. The
structure matrix is simply the factor loading matrix as in orthogonal rotation, representing
the variance in a measured variable explained by a factor on both a unique and common
contributions basis. The pattern matrix, in contrast, contains coefficients which just
represent unique contributions. The more factors, the lower the pattern coefficients as a
rule since there will be more common contributions to variance explained. For oblique
rotation, the researcher looks at both the structure and pattern coefficients when attributing
a label to a factor. Principles of oblique rotation can be derived from both cross entropy
and its dual entropy.[6]
Communality
The sum of the squared factor loadings for all factors for a given variable (row) is the
variance in that variable accounted for by all the factors. The communality measures the
percent of variance in a given variable explained by all the factors jointly and may be
interpreted as the reliability of the indicator in the context of the factors being posited.
Spurious solutions
If the communality exceeds 1.0, there is a spurious solution, which may reflect too small a
sample or the choice to extract too many or too few factors.
Uniqueness of a variable
The variability of a variable minus its communality.
Eigenvalues/characteristic roots
Eigenvalues measure the amount of variation in the total sample accounted for by each
factor. The ratio of eigenvalues is the ratio of explanatory importance of the factors with
respect to the variables. If a factor has a low eigenvalue, then it is contributing little to the
explanation of variances in the variables and may be ignored as less important than the
factors with higher eigenvalues.
Extraction sums of squared loadings
Initial eigenvalues and eigenvalues after extraction (listed by SPSS as "Extraction Sums
of Squared Loadings") are the same for PCA extraction, but for other extraction methods,
eigenvalues after extraction will be lower than their initial counterparts. SPSS also prints
"Rotation Sums of Squared Loadings" and even for PCA, these eigenvalues will differ
from initial and extraction eigenvalues, though their total will be the same.
Factor scores
Component scores (in PCA)

The scores of each case (row) on each factor (column). To compute the factor score for a
given case for a given factor, one takes the case's standardized score on each variable,
multiplies by the corresponding loadings of the variable for the given factor, and sums
these products. Computing factor scores allows one to look for factor outliers. Also, factor
scores may be used as variables in subsequent modeling.

Criteria for determining the number of factors

Researchers wish to avoid such subjective or arbitrary criteria for factor retention as "it made sense to me".
A number of objective methods have been developed to solve this problem, allowing users to determine an
appropriate range of solutions to investigate.[7] However these different methods often disagree with one
another as to the number of factors that ought to be retained. For instance, the parallel analysis may suggest
5 factors while Velicer's MAP suggests 6, so the researcher may request both 5 and 6-factor solutions and
discuss each in terms of their relation to external data and theory.

Modern criteria

Horn's parallel analysis (PA):[8] A Monte-Carlo based simulation method that compares the observed
eigenvalues with those obtained from uncorrelated normal variables. A factor or component is retained if
the associated eigenvalue is bigger than the 95th percentile of the distribution of eigenvalues derived from
the random data. PA is among the more commonly recommended rules for determining the number of
components to retain,[7][9] but many programs fail to include this option (a notable exception being R).[10]
However, Formann provided both theoretical and empirical evidence that its application might not be
appropriate in many cases since its performance is considerably influenced by sample size, item
discrimination, and type of correlation coefficient.[11]

Velicer's (1976) MAP test[12] as described by Courtney (2013)[13] “involves a complete principal
components analysis followed by the examination of a series of matrices of partial correlations” (p.  397
(though note that this quote does not occur in Velicer (1976) and the cited page number is outside the pages
of the citation). The squared correlation for Step “0” (see Figure 4) is the average squared off-diagonal
correlation for the unpartialed correlation matrix. On Step 1, the first principal component and its associated
items are partialed out. Thereafter, the average squared off-diagonal correlation for the subsequent
correlation matrix is then computed for Step 1. On Step 2, the first two principal components are partialed
out and the resultant average squared off-diagonal correlation is again computed. The computations are
carried out for k minus one step (k representing the total number of variables in the matrix). Thereafter, all
of the average squared correlations for each step are lined up and the step number in the analyses that
resulted in the lowest average squared partial correlation determines the number of components or factors to
retain.[12] By this method, components are maintained as long as the variance in the correlation matrix
represents systematic variance, as opposed to residual or error variance. Although methodologically akin to
principal components analysis, the MAP technique has been shown to perform quite well in determining
the number of factors to retain in multiple simulation studies.[7][14][15][16] This procedure is made available
through SPSS's user interface,[13] as well as the psych package for the R programming language.[17][18]

Older methods

Kaiser criterion: The Kaiser rule is to drop all components with eigenvalues under 1.0 – this being the
eigenvalue equal to the information accounted for by an average single item.[19] The Kaiser criterion is the
default in SPSS and most statistical software but is not recommended when used as the sole cut-off criterion
for estimating the number of factors as it tends to over-extract factors.[20] A variation of this method has
been created where a researcher calculates confidence intervals for each eigenvalue and retains only factors
which have the entire confidence interval greater than 1.0.[14][21]

Scree plot:[22] The Cattell scree test plots the components as the X-axis and the corresponding eigenvalues
as the Y-axis. As one moves to the right, toward later components, the eigenvalues drop. When the drop
ceases and the curve makes an elbow toward less steep decline, Cattell's scree test says to drop all further
components after the one starting at the elbow. This rule is sometimes criticised for being amenable to
researcher-controlled "fudging". That is, as picking the "elbow" can be subjective because the curve has
multiple elbows or is a smooth curve, the researcher may be tempted to set the cut-off at the number of
factors desired by their research agenda.

Variance explained criteria: Some researchers simply use the rule of keeping enough factors to account for
90% (sometimes 80%) of the variation. Where the researcher's goal emphasizes parsimony (explaining
variance with as few factors as possible), the criterion could be as low as 50%.

Bayesian method

A Bayesian approach based on the Indian buffet process returns a probability distribution over the plausible
number of latent factors.[23]

Rotation methods
The unrotated output maximizes the variance accounted for by the first factor first, then the second factor,
etc. The unrotated solution is orthogonal. This means that the correlation between the factors is zero. A
disadvantage of using the unrotated solution is that usually most items load on the early factors and many
items load substantially on more than one factor.

Rotation serves to make the output easier to interpret by rotating the axes of the coordinate system to form a
pattern of loadings where each item loads strongly on only one of the factors and more weakly on the other
factors. Rotations can be orthogonal or oblique. Oblique rotation allows the factors to correlate.[24]

Varimax is the most commonly used rotation method. Varimax is an orthogonal rotation of the factor axes
that maximizes the variance of the squared loadings of a factor (column) on all the variables (rows) in a
factor loadings matrix. Each factor tends to have only few variables with large loadings by the factor.
Varimax simplifies columns of the loadings matrix. This makes it as easy as possible to identify each
variable with a single factor.

Quartimax rotation is an orthogonal rotation that minimizes the number of factors needed to explain a
variable. It simplifies the rows of the loadings matrix rather than the columns. Quartimax often produces a
general factor that have loadings for many of the variables. This is close to the unrotated solution.
Quartimax is useful if many of the variables are correlated so that a dominating factor can be expected.[25]

Equimax rotation is a compromise between varimax and quartimax.

In many practical applications, it is unrealistic to assume that the factors are uncorrelated. Oblique rotations
are preferred in this situation. Allowing for factors that are correlated with one another is especially
applicable in psychometric research, since attitudes, opinions, and intellectual abilities tend to be correlated
and it would be unrealistic to assume otherwise.[26]

Oblimin rotation is the standard method when one wishes an oblique (non-orthogonal) solution.

Promax rotation is an alternative oblique rotation method that is computationally faster than the oblimin
method and therefore is sometimes used for very large datasets.

Problems with factor rotation

It can be difficult to interpret a factor structure when each variable is loading on multiple factors. Small
changes in the data can sometimes tip a balance in the factor rotation criterion so that a completely different
factor rotation is produced. This can make it difficult to compare the results of different experiments. This
problem is illustrated by a comparison of different studies of world-wide cultural differences. Each study
has used different measures of cultural variables and produced a differently rotated factor analysis result.
The authors of each study believed that they had discovered something new, and invented new names for
the factors they found. A later comparison of the studies found that the results were rather similar when the
unrotated results were compared. The common practice of factor rotation has obscured the similarity
between the results of the different studies.[27]

Higher order factor analysis

Higher-order factor analysis is a statistical method consisting of repeating steps factor analysis – oblique
rotation – factor analysis of rotated factors. Its merit is to enable the researcher to see the hierarchical
structure of studied phenomena. To interpret the results, one proceeds either by post-multiplying the
primary factor pattern matrix by the higher-order factor pattern matrices (Gorsuch, 1983) and perhaps
applying a Varimax rotation to the result (Thompson, 1990) or by using a Schmid-Leiman solution (SLS,
Schmid & Leiman, 1957, also known as Schmid-Leiman transformation) which attributes the variation
from the primary factors to the second-order factors.

Exploratory factor analysis (EFA) versus principal components


analysis (PCA)
Factor analysis is related to principal component analysis (PCA), but the two are not identical.[28] There
has been significant controversy in the field over differences between the two techniques. PCA can be
considered as a more basic version of exploratory factor analysis (EFA) that was developed in the early
days prior to the advent of high-speed computers. Both PCA and factor analysis aim to reduce the
dimensionality of a set of data, but the approaches taken to do so are different for the two techniques.
Factor analysis is clearly designed with the objective to identify certain unobservable factors from the
observed variables, whereas PCA does not directly address this objective; at best, PCA provides an
approximation to the required factors.[29] From the point of view of exploratory analysis, the eigenvalues of
PCA are inflated component loadings, i.e., contaminated with error variance.[30][31][32][33][34][35]

Whilst EFA and PCA are treated as synonymous techniques in some fields of statistics, this has been
criticised.[36][37] Factor analysis "deals with the assumption of an underlying causal structure: [it] assumes
that the covariation in the observed variables is due to the presence of one or more latent variables (factors)
that exert causal influence on these observed variables".[38] In contrast, PCA neither assumes nor depends
on such an underlying causal relationship. Researchers have argued that the distinctions between the two
techniques may mean that there are objective benefits for preferring one over the other based on the analytic
goal. If the factor model is incorrectly formulated or the assumptions are not met, then factor analysis will
give erroneous results. Factor analysis has been used successfully where adequate understanding of the
system permits good initial model formulations. PCA employs a mathematical transformation to the original
data with no assumptions about the form of the covariance matrix. The objective of PCA is to determine
linear combinations of the original variables and select a few that can be used to summarize the data set
without losing much information.[39]

Arguments contrasting PCA and EFA

Fabrigar et al. (1999)[36] address a number of reasons used to suggest that PCA is not equivalent to factor
analysis:

1. It is sometimes suggested that PCA is computationally quicker and requires fewer resources
than factor analysis. Fabrigar et al. suggest that readily available computer resources have
rendered this practical concern irrelevant.
2. PCA and factor analysis can produce similar results. This point is also addressed by
Fabrigar et al.; in certain cases, whereby the communalities are low (e.g. 0.4), the two
techniques produce divergent results. In fact, Fabrigar et al. argue that in cases where the
data correspond to assumptions of the common factor model, the results of PCA are
inaccurate results.
3. There are certain cases where factor analysis leads to 'Heywood cases'. These encompass
situations whereby 100% or more of the variance in a measured variable is estimated to be
accounted for by the model. Fabrigar et al. suggest that these cases are actually informative
to the researcher, indicating an incorrectly specified model or a violation of the common
factor model. The lack of Heywood cases in the PCA approach may mean that such issues
pass unnoticed.
4. Researchers gain extra information from a PCA approach, such as an individual's score on a
certain component; such information is not yielded from factor analysis. However, as
Fabrigar et al. contend, the typical aim of factor analysis – i.e. to determine the factors
accounting for the structure of the correlations between measured variables – does not
require knowledge of factor scores and thus this advantage is negated. It is also possible to
compute factor scores from a factor analysis.

Variance versus covariance

Factor analysis takes into account the random error that is inherent in measurement, whereas PCA fails to
do so. This point is exemplified by Brown (2009),[40] who indicated that, in respect to the correlation
matrices involved in the calculations:

"In PCA, 1.00s are put in the diagonal meaning that all of the variance in the matrix is to be
accounted for (including variance unique to each variable, variance common among variables,
and error variance). That would, therefore, by definition, include all of the variance in the
variables. In contrast, in EFA, the communalities are put in the diagonal meaning that only the
variance shared with other variables is to be accounted for (excluding variance unique to each
variable and error variance). That would, therefore, by definition, include only variance that is
common among the variables."

— Brown (2009), Principal components analysis and exploratory factor analysis –


Definitions, differences and choices

For this reason, Brown (2009) recommends using factor analysis when theoretical ideas about relationships
between variables exist, whereas PCA should be used if the goal of the researcher is to explore patterns in
their data.

Differences in procedure and results

The differences between PCA and factor analysis (FA) are further illustrated by Suhr (2009):[37]

PCA results in principal components that account for a maximal amount of variance for
observed variables; FA accounts for common variance in the data.
PCA inserts ones on the diagonals of the correlation matrix; FA adjusts the diagonals of the
correlation matrix with the unique factors.
PCA minimizes the sum of squared perpendicular distance to the component axis; FA
estimates factors that influence responses on observed variables.
The component scores in PCA represent a linear combination of the observed variables
weighted by eigenvectors; the observed variables in FA are linear combinations of the
underlying and unique factors.
In PCA, the components yielded are uninterpretable, i.e. they do not represent underlying
‘constructs’; in FA, the underlying constructs can be labelled and readily interpreted, given
an accurate model specification.
In psychometrics

History

Charles Spearman was the first psychologist to discuss common factor analysis[41] and did so in his 1904
paper.[42] It provided few details about his methods and was concerned with single-factor models.[43] He
discovered that school children's scores on a wide variety of seemingly unrelated subjects were positively
correlated, which led him to postulate that a single general mental ability, or g, underlies and shapes human
cognitive performance.

The initial development of common factor analysis with multiple factors was given by Louis Thurstone in
two papers in the early 1930s,[44][45] summarized in his 1935 book, The Vector of Mind.[46] Thurstone
introduced several important factor analysis concepts, including communality, uniqueness, and rotation.[47]
He advocated for "simple structure", and developed methods of rotation that could be used as a way to
achieve such structure.[41]

In Q methodology, William Stephenson, a student of Spearman, distinguish between R factor analysis,


oriented toward the study of inter-individual differences, and Q factor analysis oriented toward subjective
intra-individual differences.[48][49]

Raymond Cattell was a strong advocate of factor analysis and psychometrics and used Thurstone's multi-
factor theory to explain intelligence. Cattell also developed the scree test and similarity coefficients.

Applications in psychology

Factor analysis is used to identify "factors" that explain a variety of results on different tests. For example,
intelligence research found that people who get a high score on a test of verbal ability are also good on
other tests that require verbal abilities. Researchers explained this by using factor analysis to isolate one
factor, often called verbal intelligence, which represents the degree to which someone is able to solve
problems involving verbal skills.

Factor analysis in psychology is most often associated with intelligence research. However, it also has been
used to find factors in a broad range of domains such as personality, attitudes, beliefs, etc. It is linked to
psychometrics, as it can assess the validity of an instrument by finding if the instrument indeed measures the
postulated factors.

Advantages
Reduction of number of variables, by combining two or more variables into a single factor.
For example, performance at running, ball throwing, batting, jumping and weight lifting could
be combined into a single factor such as general athletic ability. Usually, in an item by
people matrix, factors are selected by grouping related items. In the Q factor analysis
technique, the matrix is transposed and factors are created by grouping related people. For
example, liberals, libertarians, conservatives, and socialists might form into separate groups.
Identification of groups of inter-related variables, to see how they are related to each other.
For example, Carroll used factor analysis to build his Three Stratum Theory. He found that a
factor called "broad visual perception" relates to how good an individual is at visual tasks.
He also found a "broad auditory perception" factor, relating to auditory task capability.
Furthermore, he found a global factor, called "g" or general intelligence, that relates to both
"broad visual perception" and "broad auditory perception". This means someone with a high
"g" is likely to have both a high "visual perception" capability and a high "auditory
perception" capability, and that "g" therefore explains a good part of why someone is good or
bad in both of those domains.

Disadvantages
"...each orientation is equally acceptable mathematically. But different factorial theories
proved to differ as much in terms of the orientations of factorial axes for a given solution as in
terms of anything else, so that model fitting did not prove to be useful in distinguishing
among theories." (Sternberg, 1977[50]). This means all rotations represent different
underlying processes, but all rotations are equally valid outcomes of standard factor analysis
optimization. Therefore, it is impossible to pick the proper rotation using factor analysis
alone.
Factor analysis can be only as good as the data allows. In psychology, where researchers
often have to rely on less valid and reliable measures such as self-reports, this can be
problematic.
Interpreting factor analysis is based on using a "heuristic", which is a solution that is
"convenient even if not absolutely true".[51] More than one interpretation can be made of the
same data factored the same way, and factor analysis cannot identify causality.

In cross-cultural research
Factor analysis is a frequently used technique in cross-cultural research. It serves the purpose of extracting
cultural dimensions. The best known cultural dimensions models are those elaborated by Geert Hofstede,
Ronald Inglehart, Christian Welzel, Shalom Schwartz and Michael Minkov. A popular visualization is
Inglehart and Welzel's cultural map of the world.[27]

In political science
In an early 1965 study, political systems around the world are examined via factor analysis to construct
related theoretical models and research, compare political systems, and create typological categories.[52] For
these purposes, in this study seven basic political dimensions are identified, which are related to a wide
variety of political behaviour: these dimensions are Access, Differentiation, Consensus, Sectionalism,
Legitimation, Interest, and Leadership Theory and Research.

Other political scientists explore the measurement of internal political efficacy using four new questions
added to the 1988 National Election Study. Factor analysis is here used to find that these items measure a
single concept distinct from external efficacy and political trust, and that these four questions provided the
best measure of internal political efficacy up to that point in time.[53]
In order to study the impact of important campaign events such as the presidential debates, rallies, and
Hillary Clinton’s email controversy, factor analysis is employed to construct measures of popularity for US
presidential candidates such as Donald Trump in 2016 and Obama in 2012. Popularity factors are
synthesized from data collected from Twitter, Facebook, YouTube, Instagram, FiveThirtyEight, and
prediction markets.[54]

In marketing
The basic steps are:

Identify the salient attributes consumers use to evaluate products in this category.
Use quantitative marketing research techniques (such as surveys) to collect data from a
sample of potential customers concerning their ratings of all the product attributes.
Input the data into a statistical program and run the factor analysis procedure. The computer
will yield a set of underlying attributes (or factors).
Use these factors to construct perceptual maps and other product positioning devices.

Information collection

The data collection stage is usually done by marketing research professionals. Survey questions ask the
respondent to rate a product sample or descriptions of product concepts on a range of attributes. Anywhere
from five to twenty attributes are chosen. They could include things like: ease of use, weight, accuracy,
durability, colourfulness, price, or size. The attributes chosen will vary depending on the product being
studied. The same question is asked about all the products in the study. The data for multiple products is
coded and input into a statistical program such as R, SPSS, SAS, Stata, STATISTICA, JMP, and SYSTAT.

Analysis

The analysis will isolate the underlying factors that explain the data using a matrix of associations.[55]
Factor analysis is an interdependence technique. The complete set of interdependent relationships is
examined. There is no specification of dependent variables, independent variables, or causality. Factor
analysis assumes that all the rating data on different attributes can be reduced down to a few important
dimensions. This reduction is possible because some attributes may be related to each other. The rating
given to any one attribute is partially the result of the influence of other attributes. The statistical algorithm
deconstructs the rating (called a raw score) into its various components and reconstructs the partial scores
into underlying factor scores. The degree of correlation between the initial raw score and the final factor
score is called a factor loading.

Advantages
Both objective and subjective attributes can be used provided the subjective attributes can
be converted into scores.
Factor analysis can identify latent dimensions or constructs that direct analysis may not.
It is easy and inexpensive.

Disadvantages
Usefulness depends on the researchers' ability to collect a sufficient set of product attributes.
If important attributes are excluded or neglected, the value of the procedure is reduced.
If sets of observed variables are highly similar to each other and distinct from other items,
factor analysis will assign a single factor to them. This may obscure factors that represent
more interesting relationships.
Naming factors may require knowledge of theory because seemingly dissimilar attributes
can correlate strongly for unknown reasons.

In physical and biological sciences


Factor analysis has also been widely used in physical sciences such as geochemistry, hydrochemistry,[56]
astrophysics and cosmology, as well as biological sciences, such as ecology, molecular biology,
neuroscience and biochemistry.

In groundwater quality management, it is important to relate the spatial distribution of different chemical
parameters to different possible sources, which have different chemical signatures. For example, a sulfide
mine is likely to be associated with high levels of acidity, dissolved sulfates and transition metals. These
signatures can be identified as factors through R-mode factor analysis, and the location of possible sources
can be suggested by contouring the factor scores.[57]

In geochemistry, different factors can correspond to different mineral associations, and thus to
mineralisation.[58]

In microarray analysis
Factor analysis can be used for summarizing high-density oligonucleotide DNA microarrays data at probe
level for Affymetrix GeneChips. In this case, the latent variable corresponds to the RNA concentration in a
sample.[59]

Implementation
Factor analysis has been implemented in several statistical analysis programs since the 1980s:

BMDP
JMP (statistical software)
Mplus (statistical software)]
Python: module scikit-learn[60]
R (with the base function factanal or fa function in package psych). Rotations are
implemented in the GPArotation R package.
SAS (using PROC FACTOR or PROC CALIS)
SPSS[61]
Stata

Stand-alone
Factor [1] (https://psico.fcep.urv.cat/utilitats/factor/Download.html) - free factor analysis
software developed by the Rovira i Virgili University

See also
Mathematics
portal

Confirmatory factor analysis Non-negative matrix factorization


Exploratory factor analysis Q methodology
Design of experiments Recommendation system
Formal concept analysis Root cause analysis
Independent component analysis Facet theory

Notes
1. In this example, "verbal intelligence" and "mathematical intelligence" are latent variables.
The fact that they're not directly observed is what makes them latent.

References
1. Jöreskog, Karl G. (1983). "Factor Analysis as an Errors-in-Variables Model". Principals of
Modern Psychological Measurement. Hillsdale: Erlbaum. pp. 185–196. ISBN 0-89859-277-
1.
2. Bandalos, Deborah L. (2017). Measurement Theory and Applications for the Social
Sciences. The Guilford Press.
3. Harman, Harry H. (1976). Modern Factor Analysis. University of Chicago Press. pp. 175,
176. ISBN 978-0-226-31652-9.
4. Polit DF Beck CT (2012). Nursing Research: Generating and Assessing Evidence for
Nursing Practice, 9th ed. Philadelphia, USA: Wolters Klower Health, Lippincott Williams &
Wilkins.
5. Meng, J. (2011). "Uncover cooperative gene regulations by microRNAs and transcription
factors in glioblastoma using a nonnegative hybrid factor model" (https://web.archive.org/we
b/20111123144133/http://www.cmsworldwide.com/ICASSP2011/Papers/ViewPapers.asp?
PaperNum=4439). International Conference on Acoustics, Speech and Signal Processing.
Archived from the original (http://www.cmsworldwide.com/ICASSP2011/Papers/ViewPaper
s.asp?PaperNum=4439) on 2011-11-23.
6. Liou, C.-Y.; Musicus, B.R. (2008). "Cross Entropy Approximation of Structured Gaussian
Covariance Matrices" (http://ntur.lib.ntu.edu.tw/bitstream/246246/155199/1/23.pdf) (PDF).
IEEE Transactions on Signal Processing. 56 (7): 3362–3367. Bibcode:2008ITSP...56.3362L
(https://ui.adsabs.harvard.edu/abs/2008ITSP...56.3362L). doi:10.1109/TSP.2008.917878 (htt
ps://doi.org/10.1109%2FTSP.2008.917878). S2CID 15255630 (https://api.semanticscholar.o
rg/CorpusID:15255630).
7. Zwick, William R.; Velicer, Wayne F. (1986). "Comparison of five rules for determining the
number of components to retain". Psychological Bulletin. 99 (3): 432–442.
doi:10.1037/0033-2909.99.3.432 (https://doi.org/10.1037%2F0033-2909.99.3.432).
8. Horn, John L. (June 1965). "A rationale and test for the number of factors in factor analysis".
Psychometrika. 30 (2): 179–185. doi:10.1007/BF02289447 (https://doi.org/10.1007%2FBF0
2289447). PMID 14306381 (https://pubmed.ncbi.nlm.nih.gov/14306381). S2CID 19663974
(https://api.semanticscholar.org/CorpusID:19663974).
9. Dobriban, Edgar (2017-10-02). "Permutation methods for factor analysis and PCA".
arXiv:1710.00479v2 (https://arxiv.org/abs/1710.00479v2) [math.ST (https://arxiv.org/archive/
math.ST)].
10. * Ledesma, R.D.; Valero-Mora, P. (2007). "Determining the Number of Factors to Retain in
EFA: An easy-to-use computer program for carrying out Parallel Analysis" (http://pareonline.
net/getvn.asp?v=12&n=2). Practical Assessment Research & Evaluation. 12 (2): 1–11.
11. Tran, U. S., & Formann, A. K. (2009). Performance of parallel analysis in retrieving
unidimensionality in the presence of binary data. Educational and Psychological
Measurement, 69, 50-61.
12. Velicer, W.F. (1976). "Determining the number of components from the matrix of partial
correlations". Psychometrika. 41 (3): 321–327. doi:10.1007/bf02293557 (https://doi.org/10.10
07%2Fbf02293557). S2CID 122907389 (https://api.semanticscholar.org/CorpusID:1229073
89).
13. Courtney, M. G. R. (2013). Determining the number of factors to retain in EFA: Using the
SPSS R-Menu v2.0 to make more judicious estimations. Practical Assessment, Research
and Evaluation, 18(8). Available online: http://pareonline.net/getvn.asp?v=18&n=8
14. Warne, R. T.; Larsen, R. (2014). "Evaluating a proposed modification of the Guttman rule for
determining the number of factors in an exploratory factor analysis". Psychological Test and
Assessment Modeling. 56: 104–123.
15. Ruscio, John; Roche, B. (2012). "Determining the number of factors to retain in an
exploratory factor analysis using comparison data of known factorial structure".
Psychological Assessment. 24 (2): 282–292. doi:10.1037/a0025697 (https://doi.org/10.103
7%2Fa0025697). PMID 21966933 (https://pubmed.ncbi.nlm.nih.gov/21966933).
16. Garrido, L. E., & Abad, F. J., & Ponsoda, V. (2012). A new look at Horn's parallel analysis
with ordinal variables. Psychological Methods. Advance online publication.
doi:10.1037/a0030005 (https://doi.org/10.1037%2Fa0030005)
17. Revelle, William (2007). "Determining the number of factors: the example of the NEO-PI-R"
(http://www.personality-project.org/r/book/numberoffactors.pdf) (PDF).
18. Revelle, William (8 January 2020). "psych: Procedures for Psychological, Psychometric, and
PersonalityResearch" (https://cran.r-project.org/web/packages/psych/).
19. Kaiser, Henry F. (April 1960). "The Application of Electronic Computers to Factor Analysis".
Educational and Psychological Measurement. 20 (1): 141–151.
doi:10.1177/001316446002000116 (https://doi.org/10.1177%2F001316446002000116).
S2CID 146138712 (https://api.semanticscholar.org/CorpusID:146138712).
20. Bandalos, D.L.; Boehm-Kaufman, M.R. (2008). "Four common misconceptions in exploratory
factor analysis" (https://books.google.com/books?id=KFAnkvqD8CgC&pg=PA61). In Lance,
Charles E.; Vandenberg, Robert J. (eds.). Statistical and Methodological Myths and Urban
Legends: Doctrine, Verity and Fable in the Organizational and Social Sciences. Taylor &
Francis. pp. 61–87. ISBN 978-0-8058-6237-9.
21. Larsen, R.; Warne, R. T. (2010). "Estimating confidence intervals for eigenvalues in
exploratory factor analysis" (https://doi.org/10.3758%2FBRM.42.3.871). Behavior Research
Methods. 42 (3): 871–876. doi:10.3758/BRM.42.3.871 (https://doi.org/10.3758%2FBRM.42.
3.871). PMID 20805609 (https://pubmed.ncbi.nlm.nih.gov/20805609).
22. Cattell, Raymond (1966). "The scree test for the number of factors". Multivariate Behavioral
Research. 1 (2): 245–76. doi:10.1207/s15327906mbr0102_10 (https://doi.org/10.1207%2Fs
15327906mbr0102_10). PMID 26828106 (https://pubmed.ncbi.nlm.nih.gov/26828106).
23. Alpaydin (2020). Introduction to Machine Learning (5th ed.). pp. 528–9.
24. "Factor rotation methods" (https://stats.stackexchange.com/q/185216). Stack Exchange.
Retrieved 7 November 2022.
25. Neuhaus, Jack O; Wrigley, C. (1954). "The Quartimax Method". British Journal of Statistical
Psychology. 7 (2): 81–91. doi:10.1111/j.2044-8317.1954.tb00147.x (https://doi.org/10.1111%
2Fj.2044-8317.1954.tb00147.x).
26. Russell, D.W. (December 2002). "In search of underlying dimensions: The use (and abuse)
of factor analysis in Personality and Social Psychology Bulletin". Personality and Social
Psychology Bulletin. 28 (12): 1629–46. doi:10.1177/014616702237645 (https://doi.org/10.11
77%2F014616702237645). S2CID 143687603 (https://api.semanticscholar.org/CorpusID:14
3687603).
27. Fog, A (2022). "Two-Dimensional Models of Cultural Differences: Statistical and Theoretical
Analysis" (https://backend.orbit.dtu.dk/ws/files/292673942/Two_dimensional_models_of_cul
ture.pdf) (PDF). Cross-Cultural Research. 57 (2–3): 115–165.
doi:10.1177/10693971221135703 (https://doi.org/10.1177%2F10693971221135703).
S2CID 253153619 (https://api.semanticscholar.org/CorpusID:253153619).
28. Bartholomew, D.J.; Steele, F.; Galbraith, J.; Moustaki, I. (2008). Analysis of Multivariate
Social Science Data. Statistics in the Social and Behavioral Sciences Series (2nd ed.).
Taylor & Francis. ISBN 978-1584889601.
29. Jolliffe I.T. Principal Component Analysis, Series: Springer Series in Statistics, 2nd ed.,
Springer, NY, 2002, XXIX, 487 p. 28 illus. ISBN 978-0-387-95442-4
30. Cattell, R. B. (1952). Factor analysis. New York: Harper.
31. Fruchter, B. (1954). Introduction to Factor Analysis. Van Nostrand.
32. Cattell, R. B. (1978). Use of Factor Analysis in Behavioral and Life Sciences. New York:
Plenum.
33. Child, D. (2006). The Essentials of Factor Analysis, 3rd edition. Bloomsbury Academic
Press.
34. Gorsuch, R. L. (1983). Factor Analysis, 2nd edition. Hillsdale, NJ: Erlbaum.
35. McDonald, R. P. (1985). Factor Analysis and Related Methods. Hillsdale, NJ: Erlbaum.
36. Fabrigar; et al. (1999). "Evaluating the use of exploratory factor analysis in psychological
research" (http://www.statpower.net/Content/312/Handout/Fabrigar1999.pdf) (PDF).
Psychological Methods.
37. Suhr, Diane (2009). "Principal component analysis vs. exploratory factor analysis" (http://ww
w2.sas.com/proceedings/sugi30/203-30.pdf) (PDF). SUGI 30 Proceedings. Retrieved 5 April
2012.
38. SAS Statistics. "Principal Components Analysis" (http://support.sas.com/publishing/pubcat/c
haps/55129.pdf) (PDF). SAS Support Textbook.
39. Meglen, R.R. (1991). "Examining Large Databases: A Chemometric Approach Using
Principal Component Analysis". Journal of Chemometrics. 5 (3): 163–179.
doi:10.1002/cem.1180050305 (https://doi.org/10.1002%2Fcem.1180050305).
S2CID 120886184 (https://api.semanticscholar.org/CorpusID:120886184).
40. Brown, J. D. (January 2009). "Principal components analysis and exploratory factor analysis
– Definitions, differences and choices" (http://jalt.org/test/PDF/Brown29.pdf) (PDF). Shiken:
JALT Testing & Evaluation SIG Newsletter. Retrieved 16 April 2012.
41. Mulaik, Stanley A (2010). Foundations of Factor Analysis. Second Edition. Boca Raton,
Florida: CRC Press. p. 6. ISBN 978-1-4200-9961-4.
42. Spearman, Charles (1904). "General intelligence objectively determined and measured".
American Journal of Psychology. 15 (2): 201–293. doi:10.2307/1412107 (https://doi.org/10.2
307%2F1412107). JSTOR 1412107 (https://www.jstor.org/stable/1412107).
43. Bartholomew, D. J. (1995). "Spearman and the origin and development of factor analysis".
British Journal of Mathematical and Statistical Psychology. 48 (2): 211–220.
doi:10.1111/j.2044-8317.1995.tb01060.x (https://doi.org/10.1111%2Fj.2044-8317.1995.tb01
060.x).
44. Thurstone, Louis (1931). "Multiple factor analysis". Psychological Review. 38 (5): 406–427.
doi:10.1037/h0069792 (https://doi.org/10.1037%2Fh0069792).
45. Thurstone, Louis (1934). "The Vectors of Mind". The Psychological Review. 41: 1–32.
doi:10.1037/h0075959 (https://doi.org/10.1037%2Fh0075959).
46. Thurstone, L. L. (1935). The Vectors of Mind. Multiple-Factor Analysis for the Isolation of
Primary Traits. Chicago, Illinois: University of Chicago Press.
47. Bock, Robert (2007). "Rethinking Thurstone". In Cudeck, Robert; MacCallum, Robert C.
(eds.). Factor Analysis at 100. Mahwah, New Jersey: Lawrence Erlbaum Associates. p. 37.
ISBN 978-0-8058-6212-6.
48. Mckeown, Bruce (2013-06-21). Q Methodology. ISBN 9781452242194. OCLC 841672556
(https://www.worldcat.org/oclc/841672556).
49. Stephenson, W. (August 1935). "Technique of Factor Analysis" (https://doi.org/10.1038%2F1
36297b0). Nature. 136 (3434): 297. Bibcode:1935Natur.136..297S (https://ui.adsabs.harvar
d.edu/abs/1935Natur.136..297S). doi:10.1038/136297b0 (https://doi.org/10.1038%2F136297
b0). ISSN 0028-0836 (https://www.worldcat.org/issn/0028-0836). S2CID 26952603 (https://a
pi.semanticscholar.org/CorpusID:26952603).
50. Sternberg, R. J. (1977). Metaphors of Mind: Conceptions of the Nature of Intelligence. New
York: Cambridge University Press. pp. 85–111.
51. "Factor Analysis" (https://web.archive.org/web/20040818062948/http://comp9.psych.cornell.
edu/Darlington/factor.htm). Archived from the original (http://comp9.psych.cornell.edu/Darling
ton/factor.htm) on August 18, 2004. Retrieved July 22, 2004.
52. Gregg, Phillip M.; Banks, Arthur S. (1965). "Dimensions of political systems: Factor analysis
of a cross-polity survey". American Political Science Review. 59 (3): 602–614.
doi:10.2307/1953171 (https://doi.org/10.2307%2F1953171).
53. Niemi, Richard G.; Craig, Stephen C.; Mattei, Franco (December 1991). "Measuring Internal
Political Efficacy in the 1988 National Election Study" (https://doi.org/10.2307/1963953).
American Political Science Review. 85 (4): 1407–1413. doi:10.2307/1963953 (https://doi.or
g/10.2307%2F1963953). ISSN 0003-0554 (https://www.worldcat.org/issn/0003-0554).
54. Franch, Fabio (May 2021). "Political preferences nowcasting with factor analysis and
internet data: The 2012 and 2016 US presidential elections" (https://doi.org/10.1016/j.techfor
e.2021.120667). Technological Forecasting and Social Change. 166: 120667.
doi:10.1016/j.techfore.2021.120667 (https://doi.org/10.1016%2Fj.techfore.2021.120667).
ISSN 0040-1625 (https://www.worldcat.org/issn/0040-1625).
55. Ritter, N. (2012). A comparison of distribution-free and non-distribution free methods in factor
analysis. Paper presented at Southwestern Educational Research Association (SERA)
Conference 2012, New Orleans, LA (ED529153).
56. Subbarao, C.; Subbarao, N.V.; Chandu, S.N. (December 1996). "Characterisation of
groundwater contamination using factor analysis". Environmental Geology. 28 (4): 175–180.
Bibcode:1996EnGeo..28..175S (https://ui.adsabs.harvard.edu/abs/1996EnGeo..28..175S).
doi:10.1007/s002540050091 (https://doi.org/10.1007%2Fs002540050091).
S2CID 129655232 (https://api.semanticscholar.org/CorpusID:129655232).
57. Love, D.; Hallbauer, D.K.; Amos, A.; Hranova, R.K. (2004). "Factor analysis as a tool in
groundwater quality management: two southern African case studies". Physics and
Chemistry of the Earth. 29 (15–18): 1135–43. Bibcode:2004PCE....29.1135L (https://ui.adsa
bs.harvard.edu/abs/2004PCE....29.1135L). doi:10.1016/j.pce.2004.09.027 (https://doi.org/10.
1016%2Fj.pce.2004.09.027).
58. Barton, E.S.; Hallbauer, D.K. (1996). "Trace-element and U—Pb isotope compositions of
pyrite types in the Proterozoic Black Reef, Transvaal Sequence, South Africa: Implications
on genesis and age". Chemical Geology. 133 (1–4): 173–199. doi:10.1016/S0009-
2541(96)00075-7 (https://doi.org/10.1016%2FS0009-2541%2896%2900075-7).
59. Hochreiter, Sepp; Clevert, Djork-Arné; Obermayer, Klaus (2006). "A new summarization
method for affymetrix probe level data" (https://doi.org/10.1093%2Fbioinformatics%2Fbtl03
3). Bioinformatics. 22 (8): 943–9. doi:10.1093/bioinformatics/btl033 (https://doi.org/10.1093%
2Fbioinformatics%2Fbtl033). PMID 16473874 (https://pubmed.ncbi.nlm.nih.gov/16473874).
60. "sklearn.decomposition.FactorAnalysis — scikit-learn 0.23.2 documentation" (https://scikit-le
arn.org/stable/modules/generated/sklearn.decomposition.FactorAnalysis.html). scikit-
learn.org.
61. MacCallum, Robert (June 1983). "A comparison of factor analysis programs in SPSS,
BMDP, and SAS". Psychometrika. 48 (2): 223–231. doi:10.1007/BF02294017 (https://doi.or
g/10.1007%2FBF02294017). S2CID 120770421 (https://api.semanticscholar.org/CorpusID:1
20770421).

Further reading
Child, Dennis (2006), The Essentials of Factor Analysis (https://books.google.com/books?id
=rQ2vdJgohH0C) (3rd ed.), Continuum International, ISBN 978-0-8264-8000-2.
Fabrigar, L.R.; Wegener, D.T.; MacCallum, R.C.; Strahan, E.J. (September 1999).
"Evaluating the use of exploratory factor analysis in psychological research". Psychological
Methods. 4 (3): 272–299. doi:10.1037/1082-989X.4.3.272 (https://doi.org/10.1037%2F1082-
989X.4.3.272).
B.T. Gray (1997) Higher-Order Factor Analysis (http://eric.ed.gov/ERICWebPortal/contentdeli
very/servlet/ERICServlet?accno=ED407418) (Conference paper)
Jennrich, Robert I., "Rotation to Simple Loadings Using Component Loss Function: The
Oblique Case," Psychometrika, Vol. 71, No. 1, pp. 173–191, March 2006.
Katz, Jeffrey Owen, and Rohlf, F. James. Primary product functionplane: An oblique rotation
to simple structure. Multivariate Behavioral Research, April 1975, Vol. 10, pp. 219–232.
Katz, Jeffrey Owen, and Rohlf, F. James. Functionplane: A new approach to simple structure
rotation. Psychometrika, March 1974, Vol. 39, No. 1, pp. 37–51.
Katz, Jeffrey Owen, and Rohlf, F. James. Function-point cluster analysis. Systematic
Zoology, September 1973, Vol. 22, No. 3, pp. 295–301.
Mulaik, S. A. (2010), Foundations of Factor Analysis, Chapman & Hall.
Preacher, K.J.; MacCallum, R.C. (2003). "Repairing Tom Swift's Electric Factor Analysis
Machine" (https://kuscholarworks.ku.edu/bitstream/1808/1492/1/preacher_maccallum_2003.
pdf) (PDF). Understanding Statistics. 2 (1): 13–43. doi:10.1207/S15328031US0201_02 (http
s://doi.org/10.1207%2FS15328031US0201_02). hdl:1808/1492 (https://hdl.handle.net/180
8%2F1492).
J.Schmid and J. M. Leiman (1957). The development of hierarchical factor solutions.
Psychometrika, 22(1), 53–61.
Thompson, B. (2004), Exploratory and Confirmatory Factor Analysis: Understanding
concepts and applications (https://archive.org/details/exploratoryconfi0000thom),
Washington DC: American Psychological Association, ISBN 978-1591470939.
Hans-Georg Wolff, Katja Preising (2005)Exploring item and higher order factor structure with
the schmid-leiman solution : Syntax codes for SPSS and SAS (https://web.archive.org/web/2
0120523230030/http://cat.inist.fr/?aModele=afficheN&cpsidt=16877522)Behavior research
methods, instruments & computers, 37 (1), 48-58

External links
A Beginner's Guide to Factor Analysis (http://www.tqmp.org/RegularArticles/vol09-2/p079/p0
79.pdf)
Exploratory Factor Analysis. A Book Manuscript by Tucker, L. & MacCallum R. (1993).
Retrieved June 8, 2006, from: [2] (http://www.unc.edu/~rcm/book/factornew.htm) Archived (htt
ps://web.archive.org/web/20130523100027/http://www.unc.edu/~rcm/book/factornew.htm)
2013-05-23 at the Wayback Machine
Garson, G. David, "Factor Analysis," from Statnotes: Topics in Multivariate Analysis.
Retrieved on April 13, 2009 from StatNotes: Topics in Multivariate Analysis, from G. David
Garson at North Carolina State University, Public Administration Program (https://archive.tod
ay/20121214194305/http://www2.chass.ncsu.edu/garson/pa765/statnote.htm)
Factor Analysis at 100 (http://www.fa100.info/index.html) — conference material
FARMS — Factor Analysis for Robust Microarray Summarization, an R package (http://www.
bioinf.jku.at/software/farms/farms.html)

Retrieved from "https://en.wikipedia.org/w/index.php?title=Factor_analysis&oldid=1160800553"

You might also like