StudyDriver in your
Smartphone!

What is Factor Analysis?

 Check out more papers on Analysis Data Analysis Scientific Method

1. DEFINITION OF FACTOR ANALYSIS

Factor analysis (FA) refers to a latent structure approach that can be used to analyze interrelationships among a large number of variables by explaining the underlying unobservable variables (latent variables) that are reflected in the observed variables (manifest variables) known as factors. With FA, the researcher can first identify the separate dimensions of the structure and then determine the extent to which each variable is explained by each dimension.A  Once these dimensions and the explanation of each variable are determined, the summarization and reduction of data can be achieved. In summarizing the data, FA describes the underlying dimensions of data in a much smaller number of items than the original variables. It examines the pattern of correlations (or covariances) between the observed measures. Data reduction can be achieved by calculating scores for each underlying dimension and substituting them for the original variables. FA is an interdependence technique where variates (factors) are formed to maximize their explanation of the entire variable set. These groups of variable would represent dimensions within the data which the researcher needs to label them. Basically, there are two types of FA, exploratory and confirmatory. The first analysis is used to discover the nature of the construct that influence a set of response and latter, test a specified set of constructs is influencing responses in a predicted way.

Data summarization

The goal of data summarization is achieved by defining a small number of factors that adequately represent the original set of variables.

Data reduction

Data reduction is achieved by identifying representative variables from a much larger set of variables for use in subsequent multivariate analyses or creating an entirely new set of variables whilst retaining the nature and character of the original variables. Data reduction relies on the factor loadings and uses them a s a basis for either identifying variables for subsequent analysis with other techniques or making estimates of the factor themselves (factor scores or summated scales), which then replace the original variables in subsequent analysis.A Factor analytic technique is run according to their purpose either an exploratory or confirmatory perspective. Many researchers consider using the Exploratory Factor Analysis (EFA) when they are searching for structure among a set of variables or as a data reduction technique. EFA technique does not set any a priori constraints on the estimation of the components or the number of components to be extracted compared to the Confirmatory Factor Analysis (CFA). CFA is used to confirm what is expected on the basis of pre-established theory.

2. PURPOSE OF FACTOR ANALYSIS

The primary purpose of FA is to discover simple patterns in the pattern of relationships amongst variables by defining the underlying structure in a data matrix. This could be done by data summarization) and reduction.

3. HISTORY OF FACTOR ANALYSIS

FA was pioneered in 1904 by psychologist, Charles Spearman, who hypothesized that the enormous variety of test of mental ability (measures of mathematical skill, vocabulary, verbal skills and others) could be explained by one underlying factor of general intelligence he called g.FA was developed to analyze the test scores of g so as to determine if g is made up of a single underlying general factor or of several more limited factors measuring attributes like mathematical ability. Raymond Cattell expanded the Spearman g test by using a multi factor theory to explain intelligence. He also developed several mathematical methods such as Cree Test and similarity coefficient.A  His statistical methods led to an improved version of factor analyses by statistician.

4. PRINCIPAL COMPONENT (PCA) VERSUS FACTOR ANALYSIS (FA)

There are many debates amongst statistician on the different of Principal Component and FA. A distinct different is Principal Component assumes that responses are measured based on the underlying factors whist the latter are based on the measured responses. Principal component analysis is used when the objective is to summarize most of the original information (variance) in a minimum number of factors for prediction purposes. In contrast, FA is used primarily to identify underlying factors or dimension that reflect what the variables share in common. Principal components are defined as linear combinations of measurement, that contain small proportions of unique variance and in some instances, error variance whilst FA considers only the common or shared variance, assuming that both the unique and the error variance are not of interest in defining the structure of the variables. PCA produces an orthogonal transformation of the variables without taking into consideration of underlying model whilst FA is based on a proper statistical model and is more concern with explaining the covariance structure of the variables than with explaining the variance (Chatfield, 1980). The calculation of PC scores is straightforward whilst the calculation of factor scores is more complex and a variety of methods can be used. Looking at the practical perspective, principal component analysis is most appropriate when the primary concern is data reduction focusing on the minimum number of factors needed to account for the maximum portion of the total variance represented in the original set of variables. FA is most appropriate when the primary objective is to identify the latent dimension or construct represented in the original variables.

5. STEPS IN FACTOR ANALYSIS

5.1.1 FA is robust to assumptions of normality

If the variables are normally distributed, then the solution is enhanced. To check normality, .....

5.1.2 Measure the sampling adequacy of sample size

There are many proposed sample size for FA. Guilford (1954) recommended that the sample size should be at least 200 whilst Hair, Black, Babin & Anderson (2010) stated that the minimum is to have at least five times as many observation as the number of variables to be analyzed and the more acceptable size would have 10:1 ratio. Comrey and Lee (1992) provided the following guidance in determining the adequacy of sample size: Table 1: Determining the Adequacy of Sample Size

 Sample Size Indication 100 Poor 200 Fair 300 Good 500 Very good 1,000 or more Excellent

5.1.3 All variables must be must be suitable for correlational analysis.

The sample is identified homogeneous with the respect to the underlying factor structure. It is inappropriate to treat a subset of items as a set of items known to differ in FA such as gender, where it will mislead the representation of the unique structure of each group. There are various ways to quantify the degree of intercorrelations amongst the variables such as the Measure of Sampling Adequacy (MSA). The index ranges from 0 to 1 when each variable is perfectly predicted without error by other variables.A  If MSA value falls below 0.50, researcher should identify variable for deletion to achieve an overall value of 0.50. According to Hair et al. (2010) can be interpreted as the followings: Table 2: Measure of Sampling Adequacy (MSA).

 Measure Of Sampling Adequacy Indication 0.8 or above Meritorious 0.7 or above Middling 0.60 or above Mediocre 0.5 or above Miserable Below 0.5 Unacceptable

Another method of determining the appropriateness of FA is the Bartlet test of sphericity and Kaiser-Myer-Oikin (KMO), a statistical test for the presence of correlations among the variables that indicates the significant status of the correlation matrix among at least some of the variables. KMO should indicates more than 0.5. The factor analyst must ensure that the data matrix has sufficient correlations to justify the application of FA. The anti image correlation matrix can be used to indicate whether the data matrix is suitable for FA. It is based on the correlation matrix of unpredicted variables using multiple regression. FA should not be performed when anti image correlation is less than 0.5 due to the lack of sufficient correlation with other variables.

6. SELECT TYPE OF ANALYSIS

6.2.1 EXTRACTION

In FA, the researchers group variables by their correlations, such that in a group (factor) have high correlations with each other. It is important to understand how much variable's variance is shared with olther variables in that factor versus what cannot be shared. The total variance of any variable os composed of its common, unique and error variances. As a variable is more highly correlated with one of more variables, the commune variable known as communalities increases.

6.2.2 ROTATION

This important tools refers to the movement of the reference axes of the factors from the origin to some other position. The ultimate effect of rotating the factor matrix is to redistribute the variance from earlier factors to later ones to achieve a simpler, theoretically more meaningful pattern. There are two ways of rotation, either orthogonal factor rotation or oblique factor rotation. In orthogonal factor rotation, the axes rotation is maintain at 90 degrees compared to oblique factor rotation. The major orthogonal approaches are Varimax, Quartimax and Equimax. The Varimax method encourages the detection of factors each of which is related to few variables, on the other hand, Quartimax seeks to maximize the variance of the squared loadings for each variables and tend to produce factors with high loadings for all variables. Equimax is a solution of compromise between Varimax and Quartimax. For Oblique factor rotation, Oblimin, Promax, Orthoblique, Dquart, Doblimin and Orthoblique has been developed. Oblimin allows factors to covary and to correlate with each other. The researcher need to choose either orthogonal or oblique factor rotation based on the particular needs of a given research problem. However, Hair et al (2010) suggested that Orthogonal Rotation method is preferred when the research goal is data reduction to either a smaller number of variables or a set of uncorrelated measures for subsequent use in other multivariate techniques. Where as the oblique rotation methods are best suited to the goal of obtaining several theoretical meaningful factors or construct.

Factor loadings indicatehow strongly a measured variable is correlated with a factor. A 0.30 loadings translates to approximately 10 percent explanation and a 0.50 loadings indicates that 25 percent of the variance is accounted for by the factor. Using practical significance of factor loadings, Hair et al. (2010) proposed the followings (for sample size of 100 or above):

 Factor Loadings Indication A± 0.30 to 0.49 Meets the minimal level for interpretation of structure A± 0.50 or greater Practically significant Exceed 1.7 Indicative of well defined structure

Comrey & Lee (1992) also proposed practical significance of factor loading as below:

 Factor Loadings Indication More than 0.70 Excellent Less than 0.63 Very good Less than 0.55 Good Less than 0.45 Fair Less than 0.32 Poor

In relation to the table above, Hair et al (2010) provide guidelines for identifying significant factor loandings based on sample size as below:

Table 5: Guidelines for Identifying Significant Factor Loadings Based on Sample Size

 Factor Loadings Sample Size Needed for Significant a 0.30 350 0.35 250 0.40 200 0.45 150 0.50 120 0.55 100 0.60 85 0.65 70 0.70 60 0.75 50

Significance is based on a 0.5 significance level (Î±), a power level of 80 percent, and standard errors assumed to be twice those conventional correlation coefficients. Source: Computation made with SOLO Power Analysis, BDMP Statistical Software, Inc. 1993 Assess the Communalities of Variable Communalities measures the percent of variance in a given variable explained by all the factors joint and may be interpreted as the reliability of the indicator. Communalities is used to indicate any variables that are not adequately accounted for by the factor solution. Variables with communalities less than 0.50 are considered of not having an acceptable level of explanation and researchers may then need to extract more factors to explain the variance.

6.3 DETERMINE NUMBER OF FACTORS

There are number of methods to determine the optimal number of factors.A Latent root Criterion/Kaiser Criterion. The latent root criterion or also known as Kaiser Criterion states that factors having latent roots or eigenvalues of the correlation matrix that are greater than 1 are considered significant. Eigenvalue refers to amount of variance explained by each principal component to each factor. Hair et all (2010) suggested that using eigenvalue for establishing a cut off is most reliable when the number of variables is between 20 and 50. Scree Test Criterion. The Cattell scree test is derived by plotting the latent roots against the number of factors in their order of extraction and the shape of the resulting curve is used to evaluate the cutoff point. From the Scree test, as one moves to the right, toward later components, the eigenvalues drop, The Cartell Scree test states to drop all other components after the one starting the elbow (a point after which the remaining eigenvalues decline in approximately linear fashion. Variance Criterion Variance Criterion is an approach to ensure practical significance for the derived factors in which the cumulative percentages of the variance extracted by successive factors. Hair (2010) proposed that it is uncommon to accept a solution that accounts for 60 percent of the total variance as a satisfactory solution.

6.4 NAME AND DEFINE FACTORS

As the variables become correlated and group together, the researchers need to label the group that can represent each group of variables as accurate as possible.

6.5 ANALYSE INTERNAL RELIABILITY

Reliability is an indicator to measure internal reliability. The rationale for internal consistency is that the individual items or indicators of the scale should all be measuring the same construct and highly correlated. There are two diagnostic measures of reliabilities, either to look at the item-to-total correlation and inter item correlation or the reliability coefficient. If the researcher choose the first method, the item-to-total correlations should exceed 0.50 and inter item correlation exceed 0.30. Using reliabilities coefficient, Zikmund, Babin, Carr & Griffin (2010) provide guideline in determining reliabilities as in Table below:

Table 6: Coefficient alpha (Î±) to Determine Reliabilities

 Coefficient alpha (Î±) Indication Between 0.80 to 0.95 Very good Between 0.70 to 0.80 Good Between 0.60 to 0.70 Fair Below 0.60 Poor

7.0 EXPLANATORY FACTOR ANALYSIS USING STATISTICAL PACKAGE FOR SOCIAL SCIENCE (SPSS)

Correlation Matrix
att1 att2 att3 att4 att5 att6 att7 att8 att9 att10 att11 att12 att13 att14 att15 att16
Correlation att1 1.000 .664 .250 .435 .490 .315 .378 .328 .574 .336 .575 .338 .176 .436 .379 .560
att2 .664 1.000 .383 .506 .444 .456 .345 .260 .525 .316 .468 .414 .320 .533 .480 .674
att3 .250 .383 1.000 .457 .210 .321 .216 .054 .217 .206 .231 .225 .429 .425 .314 .296
att4 .435 .506 .457 1.000 .351 .352 .336 .240 .415 .352 .405 .416 .331 .558 .439 .529
att5 .490 .444 .210 .351 1.000 .210 .318 .194 .303 .216 .603 .330 .188 .296 .238 .352
att6 .315 .456 .321 .352 .210 1.000 .358 .128 .379 .475 .329 .290 .276 .421 .311 .486
att7 .378 .345 .216 .336 .318 .358 1.000 .256 .373 .344 .332 .320 .175 .333 .265 .397
att8 .328 .260 .054 .240 .194 .128 .256 1.000 .348 .209 .215 .128 .128 .200 .231 .265
att9 .574 .525 .217 .415 .303 .379 .373 .348 1.000 .437 .368 .383 .203 .492 .398 .609
att10 .336 .316 .206 .352 .216 .475 .344 .209 .437 1.000 .366 .296 .181 .325 .289 .419
att11 .575 .468 .231 .405 .603 .329 .332 .215 .368 .366 1.000 .338 .176 .382 .333 .445
att12 .338 .414 .225 .416 .330 .290 .320 .128 .383 .296 .338 1.000 .186 .377 .266 .386
att13 .176 .320 .429 .331 .188 .276 .175 .128 .203 .181 .176 .186 1.000 .391 .233 .318
att14 .436 .533 .425 .558 .296 .421 .333 .200 .492 .325 .382 .377 .391 1.000 .428 .579
att15 .379 .480 .314 .439 .238 .311 .265 .231 .398 .289 .333 .266 .233 .428 1.000 .559
att16 .560 .674 .296 .529 .352 .486 .397 .265 .609 .419 .445 .386 .318 .579 .559 1.000
Anti-image Matrices
att1 att2 att3 att4 att5 att6 att7 att8 att9 att10 att11 att12 att13 att14 att15 att16
Anti-image Covariance att1 .399 -.141 -.001 -.012 -.051 .047 -.042 -.058 -.112 -.005 -.120 .029 .048 -.002 .014 -.016
att2 -.141 .366 -.057 -.013 -.054 -.080 .030 -.010 -.004 .051 .016 -.060 -.030 -.024 -.045 -.102
att3 -.001 -.057 .647 -.135 -.004 -.062 -.023 .075 .021 -.001 .006 .016 -.189 -.074 -.066 .063
att4 -.012 -.013 -.135 .519 -.027 .021 -.017 -.052 .011 -.046 -.024 -.097 -.019 -.107 -.057 -.047
att5 -.051 -.054 -.004 -.027 .570 .040 -.068 -.010 .010 .034 -.224 -.063 -.038 .021 .028 .009
att6 .047 -.080 -.062 .021 .040 .605 -.090 .041 -.010 -.183 -.039 -.004 -.033 -.046 .017 -.060
att7 -.042 .030 -.023 -.017 -.068 -.090 .719 -.089 -.024 -.061 -.002 -.075 .008 -.015 .001 -.035
att8 -.058 -.010 .075 -.052 -.010 .041 -.089 .817 -.102 -.032 .002 .053 -.055 .016 -.054 .017
att9 -.112 -.004 .021 .011 .010 -.010 -.024 -.102 .482 -.099 .039 -.071 .022 -.072 -.013 -.095
att10 -.005 .051 -.001 -.046 .034 -.183 -.061 -.032 -.099 .647 -.082 -.040 -.006 .023 -.014 -.029
att11 -.120 .016 .006 -.024 -.224 -.039 -.002 .002 .039 -.082 .491 -.025 .020 -.029 -.034 -.015
att12 .029 -.060 .016 -.097 -.063 -.004 -.075 .053 -.071 -.040 -.025 .711 .006 -.038 .011 .004
att13 .048 -.030 -.189 -.019 -.038 -.033 .008 -.055 .022 -.006 .020 .006 .737 -.094 .016 -.040
att14 -.002 -.024 -.074 -.107 .021 -.046 -.015 .016 -.072 .023 -.029 -.038 -.094 .501 -.026 -.064
att15 .014 -.045 -.066 -.057 .028 .017 .001 -.054 -.013 -.014 -.034 .011 .016 -.026 .632 -.125
att16 -.016 -.102 .063 -.047 .009 -.060 -.035 .017 -.095 -.029 -.015 .004 -.040 -.064 -.125 .362
Anti-image Correlation att1 .897a -.369 -.002 -.026 -.107 .095 -.079 -.101 -.256 -.009 -.271 .055 .089 -.005 .028 -.042
att2 -.369 .911a -.118 -.030 -.118 -.169 .059 -.019 -.008 .105 .037 -.118 -.058 -.055 -.095 -.280
att3 -.002 -.118 .865a -.234 -.006 -.099 -.034 .104 .037 -.001 .011 .024 -.274 -.130 -.104 .131
att4 -.026 -.030 -.234 .939a -.050 .037 -.027 -.079 .023 -.079 -.047 -.159 -.031 -.209 -.100 -.107
att5 -.107 -.118 -.006 -.050 .875a .069 -.106 -.015 .018 .056 -.423 -.098 -.058 .039 .047 .021
att6 .095 -.169 -.099 .037 .069 .907a -.136 .058 -.019 -.293 -.072 -.007 -.049 -.083 .027 -.127
att7 -.079 .059 -.034 -.027 -.106 -.136 .950a -.117 -.041 -.089 -.003 -.106 .011 -.025 .002 -.070
att8 -.101 -.019 .104 -.079 -.015 .058 -.117 .894a -.162 -.044 .003 .069 -.071 .024 -.075 .032
att9 -.256 -.008 .037 .023 .018 -.019 -.041 -.162 .921a -.177 .080 -.121 .036 -.147 -.023 -.229
att10 -.009 .105 -.001 -.079 .056 -.293 -.089 -.044 -.177 .901a -.145 -.059 -.008 .041 -.022 -.061
att11 -.271 .037 .011 -.047 -.423 -.072 -.003 .003 .080 -.145 .883a -.042 .034 -.057 -.060 -.036
att12 .055 -.118 .024 -.159 -.098 -.007 -.106 .069 -.121 -.059 -.042 .944a .009 -.063 .016 .008
att13 .089 -.058 -.274 -.031 -.058 -.049 .011 -.071 .036 -.008 .034 .009 .887a -.154 .023 -.078
att14 -.005 -.055 -.130 -.209 .039 -.083 -.025 .024 -.147 .041 -.057 -.063 -.154 .946a -.047 -.150
att15 .028 -.095 -.104 -.100 .047 .027 .002 -.075 -.023 -.022 -.060 .016 .023 -.047 .943a -.262
att16 -.042 -.280 .131 -.107 .021 -.127 -.070 .032 -.229 -.061 -.036 .008 -.078 -.150 -.262 .922a

KMO and Bartlett's Test

Kaiser-Meyer-Olkin Measure of Sampling Adequacy .914
Bartlett's Test of Sphericity Approx. Chi-Square 2491.010
df 120
Sig. .000

Communalities

Initial Extraction
att1 .601 .617
att2 .634 .606
att3 .353 .526
att4 .481 .514
att5 .430 .645
att6 .395 .360
att7 .281 .278
att8 .183 .164
att9 .518 .598
att10 .353 .308
att11 .509 .576
att12 .289 .274
att13 .263 .320
att14 .499 .550
att15 .368 .356
att16 .638 .682
Extraction Method: Principal Axis Factoring.

Total Variance Explained

Total % of Variance Cumulative % Total % of Variance Cumulative % Total % of Variance Cumulative %
dimension0 1 6.452 40.324 40.324 5.959 37.243 37.243 3.346 20.915 20.915
2 1.340 8.373 48.697 .833 5.206 42.449 2.150 13.438 34.353
3 1.062 6.639 55.336 .582 3.637 46.086 1.877 11.733 46.086
4 .951 5.942 61.278
5 .841 5.253 66.531
6 .756 4.727 71.257
7 .656 4.101 75.359
8 .643 4.017 79.376
9 .577 3.608 82.985
10 .528 3.298 86.283
11 .499 3.118 89.401
12 .421 2.633 92.033
13 .389 2.431 94.464
14 .348 2.176 96.640
15 .302 1.889 98.529
16 .235 1.471 100.000
Extraction Method: Principal Axis Factoring.

Factor Matrixa

Factor
1 2 3
att16 .797
att2 .778
att1 .725 -.301
att14 .702
att9 .696   -.324
att4 .688
att11 .643 -.333
att15 .581
att6 .569
att5 .562 -.401 .410
att10 .526
att12 .522
att7 .519
att3 .487 .441 .308
att13 .412 .345
att8 .352
Extraction Method: Principal Axis Factoring.
a. 3 factors extracted. 18 iterations required.

Rotated Factor Matrixa

Factor
1 2 3
att9 .732
att16 .710 .359
att1 .567   .525
att2 .555 .391 .382
att10 .498
att6 .464 .366
att15 .462 .344
att7 .424
att8 .370
att12 .361
att3   .709
att13   .545
att14 .470 .544
att4 .403 .527
att5     .770
att11 .341   .655
Extraction Method: Principal Axis Factoring. Rotation Method: Varimax with Kaiser Normalization.
a. Rotation converged in 6 iterations.

Factor Transformation Matrix

Factor 1 2 3
dimension0 1 .717 .515 .470
2 -.113 .751 -.650
3 -.688 .412 .597
Extraction Method: Principal Axis Factoring.A Rotation Method: Varimax with Kaiser Normalization.
Did you like this example?

What is factor analysis?. (2017, Jun 26). Retrieved July 22, 2024 , from
https://studydriver.com/what-is-factor-analysis/

Save time with Studydriver!

Get in touch with our top writers for a non-plagiarized essays written to satisfy your needs

Get custom essay

Stuck on ideas? Struggling with a concept?

A professional writer will make a clear, mistake-free paper for you!

Leave your email and we will send a sample to you.
Stop wasting your time searching for samples!
You can find a skilled professional who can write any paper for you.
Get unique paper

Hi!
I'm Amy :)