Professional Documents
Culture Documents
Cronbach's α (Reliability of data) and Factor Analysis (Construct Validity)
Cronbach's α (Reliability of data) and Factor Analysis (Construct Validity)
and
Factor Analysis (Construct
Validity)
Reliability of Data
• Reliability, like validity, is a way of assessing
the quality of the measurement procedure
used to collect data in a research.
• In order for the results from a study to be
considered valid, the measurement procedure
must first be reliable.
• Cronbach's α is a measure of reliability
Good
Reliability
If interested…
Helps improve reliability,
P a value of greater than
0.3 is ok.
P
https://stats.stackexchange.com/questions/1576/what-are-the-differences-between-factor-
analysis-and-principal-component-analysi
Assumptions
• Large enough sample to yield reliable estimates
of the correlations among the variables
• Statistical inference is improved if the variables
are multivariate normal
• Relationships among the pairs of variables are
linear
• Absence of outliers among the cases
• Some degree of collinearity among the variables
but not an extreme degree or singularity among
the variables
An intercorrelation Matrix
• Basic assumption:
Variables that significantly correlate with
each other do so because they are
measuring the same "thing".
• The problem:
What is the "thing" that correlated
variables are measuring in common?
Given nine metric variable …
X1 X2 X3 X4 X5 X6 X7 X8 X9
Oblique rotations
To supress output with
smaller coefficient in the
component matrix
(explained later)
Kaiser Meyer Olkin (KMO)
Good,
anything
above 0.7
is ok, 0.8
is good
and 0.9 is
great
Bartlett's Test of Sphericity - This tests the
null hypothesis that the correlation matrix is
an identity matrix. An identity matrix is
matrix in which all of the diagonal elements
are 1 and all off diagonal elements are 0.
You want to reject this null hypothesis. Here
we are rejecting the null (otherwise it means something is really
terrible with the data)
Communalities
• Table of communalities shows the proportion
of each variable’s variance that can be
explained by the factors.
• The communality value which should be more
than 0.5 (till 0.3 is ok too) to be considered for
further analysis. Else these variables are to be
removed from further steps factor analysis.
• Example: For ‘Brand image’ 59.1 of the
variance is explained or accounted for
Any value below 0.3 or
0.4 means that this
item has issues in
correlating with other
items
Total variance explained
• The Eigenvalue table has been divided
into three sub-sections, i.e. Initial Eigen
Values, Extracted Sums of Squared
Loadings and Rotation of Sums of
Squared Loadings.
• Eigenvalue actually reflects the number
of extracted factors whose sum should
be equal to number of items which are
subjected to factor analysis.
Remember when we were giving the Eigen value we said SPSS to retain only those
with greater than 1. So here all the variable with Eigen value less than one is
excluded.
The results indicate that the 20 variables are reduced to 4 components ie these four
FACTORS do a good job in explaining the 20 variables.
• Initial Eigen Values column shows the % of
variance explained by each individual items in
the questionnaire.
• For analysis and interpretation purpose we
are only concerned with Extracted Sums of
Squared Loadings. This shows the % of
variance explained by each component in the
final factors.
• The factors all together explain only 63.204%
of variance – meaning rest of the variance is
unexplained (meaning - the variables in our study is
inadequate to that level).
Eigen value
• Eigen values are calculated and used in
deciding how many factors to extract in the
overall factor analysis.
• In matrix algebra, under certain conditions,
matrices can be diagonalized.
• Matrices are often diagonalized in multivariate
analyses. In that process, eigenvalues are used
to consolidate the variance.
• In factor analysis, eigenvalues are used to
condense the variance in a correlation matrix.
• "The factor with the largest eigenvalue has the
most variance and so on, down to factors with
small or negative eigenvalues that are usually
omitted from solutions" (Tabachnick and
Fidell, 1996, p. 646). From the analyst's
perspective, only variables with eigenvalues of
1.00 or higher are traditionally considered
worth analyzing.
Scree Plot
• Scree plot and Total variance explained Both
deal with factor extraction method and can be
used to determine how many factors to retain.
We want to retain the
components above
the scree
The output without supressing small
coefficients looks like…
Since we are
calling for
rotated Matrix
preference
should be given
for rotated
matrix only
ü
0.3