代写范文

留学资讯

写作技巧

论文代写专题

服务承诺

资金托管
原创保证
实力保障
24小时客服
使命必达

51Due提供Essay,Paper,Report,Assignment等学科作业的代写与辅导,同时涵盖Personal Statement,转学申请等留学文书代写。

51Due将让你达成学业目标
51Due将让你达成学业目标
51Due将让你达成学业目标
51Due将让你达成学业目标

私人订制你的未来职场 世界名企,高端行业岗位等 在新的起点上实现更高水平的发展

积累工作经验
多元化文化交流
专业实操技能
建立人际资源圈

Stats_File

2013-11-13 来源: 类别: 更多范文

Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis Latent variable (i.e. factor) Research questions answered by factor analysis Applications of factor analysis Exploratory applications Confirmatory applications R factor analysis Q factor analysis Factor loadings Steps in factor analysis Initial v final solution Factorability of an intercorrelation matrix Bartlett's test of sphericity and its interpretation Kaiser-Meyer-Olkin measure of sampling adequacy (KMO) and its interpretation Identity matrix and the determinant of an identity matrix Methods for extracting factors Principal components Maximum likelihood method Principal axis method Unwieghted least squares Generalized least squares Alpha method Image factoring Criteria for determining the number of factors Eigenvalue greater than 1.0 Cattell's scree plot Percent and cumulative percent of variance explained by the factors extracted Component matrix and factor loadings Communality of a variable Determining what a factor measures and naming a factor Factor rotation and its purpose Varimax Quartimax Equimax Orthogonal v oblique rotation Reproduced correlation matrix Computing factor scores Factor score coefficient matrix Using factor score in multivariate dependency models Lecture Outline ✓ Identifying patterns of intercorrelation ✓ Factors v correlations ✓ Steps in the factor analysis process ✓ Testing for "factorability" ✓ Initial v final factor solutions ✓ Naming factors ✓ Factor rotation ✓ Computing factor scores ✓ Using factors scores in multivariate dependency models Factor Analysis Interdependency Technique Seeks to find the latent factors that account for the patterns of collinearity among multiple metric variables Assumptions Large enough sample to yield reliable estimates of the correlations among the variables Statistical inference is improved if the variables are multivariate normal Relationships among the pairs of variables are linear Absence of outliers among the cases Some degree of collinearity among the variables but not an extreme degree or singularity among the variables Large ratio of N / k An intercorrelation Matrix Basic assumption Variables that significantly correlate with each other do so because they are measuring the same "thing". The problem What is the "thing" that correlated variables are measuring in common' Given nine metric variable … | | | | | | | | | | | | |X1 |X2 |X3 |X4 |X5 |X6 |X7 |X8 |X9 | | | | | | | | | | | | |X1 |1.00 |0.80 |0.70 |0.95 |0.01 |0.20 |0.18 |0.16 |0.03 | |X5 | | | | |1.00 |0.93 |0.02 |0.05 |0.03 | |X7 | | | | | | |1.00 |0.95 |0.90 | An intercorrelation Matrix (cont.) Notice the patterns of intercorrelation ✓ Variables 1, 2, 3 & 4 correlate highly with each other, but not with the rest of the variables ✓ Variables 5 & 6 correlate highly with each other, but not with the rest of the variables ✓ Variables 7, 8, & 9 correlate highly with each other, but not with the rest of the variables Deduction The nine variables seem to be measuring 3 "things" or underlying factors. Q What are these three factors' Q To what extent does each variable measure each of these three factors' Research Questions That Lend Themselves to Factor Analysis A 20 item attitudinal survey of citizen attitudes about the problems of crime and the administration of justice. Q Does the survey measure 20 different independent attitudinal dimensions or do the survey items only measure a few underlying attitudes' A pre-sentence investigation Q Are the individual items in a pre-sentence investigation measuring as many independent background factors, or do they measure a few underlying background dimensions; e.g. social, educational, criminal, etc.' The purpose of factor analysis is to reduce multiple variables to a lesser number of underlying factors that are being measured by the variables. Applications of Factor Analysis Exploratory factor analysis A non-theoretical application. Given a set of variables, what are the underlying dimensions (factors), if any, that account for the patterns of collinearity among the variables' Example Given the multiple items of information gathered on applicants applying for admission to a police academy, how many independent factors are actually being measured by these items' Confirmatory factor analysis Given a theory with four concepts that purport to explain some behavior, do multiple measures of the behavior reduce to these four factors' Example Given a theory that attributes delinquency to four independent factors, do multiple measures on delinquents reduce to measuring these four factors' Applications of Factor Analysis (cont.) R and Q Factor Analysis R factor analysis involves extracting latent factors from among the variables Q factor analysis involves factoring the subjects vis-à-vis the variables. The result is a "clustering" of the subjects into independent groups based upon factors extracted from the data. This application is not used much today since a variety of clustering techniques have been developed that are designed specifically for the purpose of grouping multiple subjects into independent groups. The Logic of Factor Analysis Given an N by k database … | | | |Subjects |Variables | | |X1 |X2 |X3 |… |Xk | |1 |2 |12 |0 |… |113 | |2 |5 |16 |2 |… |116 | |3 |7 |8 |1 |… |214 | |… |… |… |… |… |… | |N |12 |23 |0 |… |168 | Compute a k x k intercorrelation matrix … | |X1 |X2 |X3 |… |Xk | |X1 |1.00 |0.26 |0.84 |… |0.72 | |X2 | |1.00 |0.54 |… |0.63 | |X3 | | |1.00 |… |0.47 | |… | | | |… |… | |Xk | | | | |1.00 | Reduce the intercorrelation matrix to a k x F matrix of factor loadings … + |Variables |Factor I |Factor II |Factor III | |X1 |0.932 |0.013 |0.250 | |X2 |0.851 |0.426 |0.211 | |X3 |0.134 |0.651 |0.231 | |… |… |… |0.293 | |Xk |0.725 |0.344 |0.293 | What is a Factor Loading' A factor loading is the correlation between a variable and a factor that has been extracted from the data. Example Note the factor loadings for variable X1. |Variables |Factor I |Factor II |Factor III | |X1 |0.932 |0.013 |0.250 | Interpretation Variable X1 is highly correlated with Factor I, but negligibly correlated with Factors II and III Q How much of the variance in variable X1 is measured or accounted for by the three factors that were extracted' Simply square the factor loadings and add them together (0.9322 + 0.0132 + 0.2502) = 0.93129 This is called the communality of the variable. Steps in Factor Analysis Step 1 Compute a k by k intercorrelation matrix. Compute the factorability of the matrix. Step 2 Extract an initial solution Step 3 From the initial solution, determine the appropriate number of factors to be extracted in the final solution Step 4 If necessary, rotate the factors to clarify the factor pattern in order to better interpret the nature of the factors Step 5 Depending upon subsequent applications, compute a factor score for each subject on each factor. An Eleven Variable Example The variables and their code names • Sentence (sentence) • Number of prior convictions(pr_conv) • Intelligence (iq) • Drug dependency (dr_score) • Chronological age (age) • Age at 1st arrest (age_firs) • Time to case disposition (tm_disp) • Pre-trial jail time (jail_tm) • Time served on sentence (tm_serv) • Educational equivalency (educ_eqv) • Level of work skill (skl_indx) Intercorrelation Among the Variables [pic] Q How much collinearity or common variance exits among the variables' Q Is the intercorrelation matrix "factorable"' Ways to Determine the Factorability of an Intercorrelation Matrix Two Tests Bartlett's Test of Sphericity Kaiser-Meyer-Olkin Measure of Sampling Adequacy (KMO) Consider the intercorrelation matrix below, which is called an identity matrix. | |X1 |X2 |X3 |X4 |X5 | |X1 |1.00 |0.00 |0.00 |0.00 |0.00 | |X2 | |1.00 |0.00 |0.00 |0.00 | |X3 | | |1.00 |0.00 |0.00 | |X4 | | | |1.00 |0.00 | |X5 | | | | |1.00 | The variables are totally noncollinear. If this matrix was factor analyzed … It would extract as many factors as variables, since each variable would be its own factor. It is totally non-factorable Bartlett's Test of Sphericity In matrix algebra, the determinate of an identity matrix is equal to 1.0. For example … 1.0 0.0 I = 0.0 1.0 1.0 0.0 I = 0.0 1.0 I = (1.0 x 1.0) - (0.0 x 0.0) = 1.0 Example Given the intercorrelation matrix below, what is its determinate' 1.0 0.63 R = 63. 1.00 R = (1.0 x 1.0) - (0.63 x 0.63) = 0.6031 Bartlett's Test of Sphericity (cont.) Bartlett's test of sphericity Calculates the determinate of the matrix of the sums of products and cross-products (S) from which the intercorrelation matrix is derived. The determinant of the matrix S is converted to a chi-square statistic and tested for significance. The null hypothesis is that the intercorrelation matrix comes from a population in which the variables are noncollinear (i.e. an identity matrix) And that the non-zero correlations in the sample matrix are due to sampling error. Chi-square (2 = - [(n-1) - 1/6 (2p+1+2/p)] [ln S + pln(1/p) (lj ] p = number of variables, k = number of components, lj = jth eigenvalue of S df = (p - 1) (p - 2) / 2 Results of Bartlett's Test of Sphericity [pic] Test Results (2 = 496.536 df = 55 p ( 0.001 Statistical Decision The sample intercorrelation matrix did not come from a population in which the intercorrelation matrix is an identity matrix. Kaiser-Meyer-Olkin Measure of Sampling Adequacy (KMO) If two variables share a common factor with other variables, their partial correlation (aij) will be small, indicating the unique variance they share. aij = (rij (1, 2, 3, …k ) KMO = ((( r2ij ) / ((( r2ij + ((( a2ij ) If aij ( 0.0 The variables are measuring a common factor, and KMO ( 1.0 If aij ( 1.0 The variables are not measuring a common factor, and KMO ( 0.0 Kaiser-Meyer-Olkin Measure of Sampling Adequacy (KMO) (cont.) Interpretation of the KMO as characterized by Kaiser, Meyer, and Olkin … | | | |KMO Value |Degree of Common Variance | | | | |0.90 to 1.00 |Marvelous | | | | |0.80 to 0.89 |Meritorious | | | | |0.70 to 0.79 |Middling | | | | |0.60 to 0.69 |Mediocre | | | | |0.50 to 0.59 |Miserable | | | | |0.00 to 0.49 |Don't Factor | Results of the KMO [pic] The KMO = 0.698 Interpretation The degree of common variance among the eleven variables is "mediocre" bordering on "middling" If a factor analysis is conducted, the factors extracted will account for fare amount of variance but not a substantial amount. Extracting an Initial Solution A variety of methods have been developed to extract factors from an intercorrelation matrix. SPSS offers the following methods … ✓ Principle components method (probably the most commonly used method) ✓ Maximum likelihood method (a commonly used method) ✓ Principal axis method also know as common factor analysis ✓ Unweighted least-squares method ✓ Generalized least squares method ✓ Alpha method ✓ Image factoring An Initial Solution Using the Principal Components Method In the initial solution, each variable is standardized to have a mean of 0.0 and a standard deviation of (1.0. Thus … The variance of each variable = 1.0 And the total variance to be explained is 11, i.e. 11 variables, each with a variance = 1.0 Since a single variable can account for 1.0 unit of variance … A useful factor must account for more than 1.0 unit of variance, or have an eigenvalue ( ( 1.0 Otherwise the factor extracted explains no more variance than a single variable. Remember the goal of factor analysis is to explain multiple variables by a lesser number of factors. The Results of the Initial Solution [pic] 11 factors (components) were extracted, the same as the number of variables factored. Factor I The 1st factor has an eigenvalue = 3.698. Since this is greater than 1.0, it explains more variance than a single variable, in fact 3.698 times as much. The percent a variance explained … (3.698 / 11 units of variance) (100) = 33.617% The Results of the Initial Solution (cont.) Factor II The 2nd factor has an eigenvalue = 2.484. It is also greater than 1.0, and therefore explains more variance than a single variable The percent a variance explained (2.484 / 11 units of variance) (100) = 22.580% Factor III The 3rd factor has an eigenvalue = 1.237. Like Factors I & II it is greater than 1.0, and therefore explains more variance than a single variable. The percent a variance explained (1.237 / 11 units of variance) (100) = 11.242% The remaining factors Factors 4 through 11 have eigenvalues less that 1, and therefore explain less variance that a single variable. The Results of the Initial Solution (cont.) Nota Bene ✓ The sum of the eigenvalues associated with each factor (component) sums to 11. (3.698 + 2.484 + 1.237 + 0.952 + … + 2.437 E-02) = 11 ✓ The cumulative % of variance explained by the first three factors is 67.439% In other words, 67.439% of the common variance shared by the 11 variables can be accounted for by the 3 factors. This is reflective of the KMO of 0.698, a "mediocre" to "middling % of variance ✓ This initial solution suggests that the final solution should extract not more than 3 factors. Cattell's Scree Plot Another way to determine the number of factors to extract in the final solution is Cattell's scree plot. This is a plot of the eigenvalues associated with each of the factors extracted, against each factor. At the point that the plot begins to level off, the additional factors explain less variance than a single variable. Raymond B. Cattell (1952) Factor Analysis New York: Harper & Bros. Factor Loadings The component matrix indicates the correlation of each variable with each factor. [pic] The variable sentence Correlates 0.933 with Factor I Correlates 0.104 with Factor II Correlates -0.190 with Factor II Factor Loadings (cont.) The total proportion of the variance in sentence explained by the three factors is simply the sum of its squared factor loadings. (0.9332 + 0.1042 - 0.1902) = 0.917 This is called the communality of the variable sentence The communalities of the 11 variables are as follows: (cf. column headed Extraction) [pic] As is evident from the table, the proportion of variance in each variable accounted for by the three factors is not the same. What Do the Three Factors Measure' The key to determining what the factors measure is the factor loadings For example Which variables load (correlate) highest on Factor I and low on the other two factors' [pic] Factor I sentence (.933) tm_serv (.907) age (.853) jail_tm (.659) age_firs (-.581) pr_conv (.548) dr_score (.404) What Do the Three Factors Measure' (cont.) Naming Factor I What do these seven variables have in common, particularly the first few that have the highest loading on Factor I' Degree of criminality' Career criminal history' You name it … Factor II educ_equ (.935) skl_indx (.887) iq (.808) Naming Factor II Educational level or job skill level' Factor III Tm_disp (.896) Naming Factor III This factor is difficult to understand since only one variable loaded highest on it. Could this be a measure of criminal case processing dynamics' We would need more variables loading on this factor to know. Summary of Results ✓ The 11 variables were reduced to 3 factors ✓ These three factors account for 67.44% of the covariance among the variables ✓ Factors I appears to measure criminal history ✓ Factor II appears to measure educational/skill level ✓ Factor III is ambiguous at best ✓ Three of the variables that load highest on Factor I do not load too high; i.e. age_firs, pr_conv, and dr_score. Compare their loadings and communalities. | | | |Loading on Factor I |Communality | | | | |age_firs -.581 |.483 | |pr_conv .548 |.303 | |dr_score .404 |.252 | The communality is the proportion of the variance in a variable accounted for by the three factors. What Can Be Done if the Factor Pattern Is Not Clear' Sometimes one or more variables may load about the same on more than one factor, making the interpretation of the factors ambiguous. Ideally, the analyst would like to find that each variable loads high (( 1.0) on one factor and approximately zero on all the others (( 0.0). An Ideal Component Matrix |Variables |Factors (f) | | |I |II |… |f | |X1 |1.0 |0.0 |… |0.0 | |X2 |1.0 |0.0 |… |0.0 | |X3 |0.0 |0.0 |… |1.0 | |X4 |0.0 |1.0 |… |0.0 | |… |… |… |… |… | |Xk |0.0 |0.0 |… |1.0 | The values in the matrix are factor loadings, the correlations between each variable and each factor. Factor Rotation Sometimes the factor pattern can be clarified by "rotating" the factors in F-dimensional space. Consider the following hypothetical two-factor solution involving eight variables. F I X8 X1 X4 X6 F II X3 X5 X7 X2 Variables 1 & 2 load on Factor I, while variables 3, 5, & 6 load on Factor II. Variables 4, 7, & 8 load about the same on both factors. What happens if the axes are rotated' Factor Rotation (cont.) Orthogonal rotation The axes remain 90( apart F I X8 X1 X4 X6 F II X3 X5 X7 X2 Variables 1, 2, 5, 7, & 8 load on Factor II, while variables 3, 4, & 6 load on Factor I. Notice that relative to variables 4, 7, & 8, the rotated factor pattern is clearer than the previously unrotated pattern. What Criterion is Used in Factor Rotation' There are various methods that can be used in factor rotation … Varimax Rotation Attempts to achieve loadings of ones and zeros in the columns of the component matrix (1.0 & 0.0). Quartimax Rotation Attempts to achieve loadings of ones and zeros in the rows of the component matrix (1.0 & 0.0). Equimax Rotation Combines the objectives of both varimax and quartimax rotations Orthogonal Rotation Preserves the independence of the factors, geometrically they remain 90( apart. Oblique Rotation Will produce factors that are not independent, geometrically not 90( apart. Rotation of the Eleven Variable Case Study Method Varimax rotation. Below are the unrotated and rotated component matrices. [pic] [pic] Did the Rotation Improve the Factor Pattern' | | |Effect on the | |Variable |Factor |Loading | | | | | |Sentence |I: no change |Increased | | | |(.933 to .956) | |Pr_conv |I: no change |Decreased | | | |(.548 to .539) | |IQ |II: no change |Increased | | | |(.808 to .823) | |Dr_score |I: no change |Increased | | | |(.404 to .431) | |Tm_disp |III: no change |Slight decrease | | | |(.896 to .892) | |Jail_tm |I: no change |Decreased | | | |(.659 to .579) | |Tm_serv |I: no change |Increased | | | |(.907 to .945) | |Educ_eqv |II: no change |Increased | | | |(.935 to .942) | |Skl_indx |II: no change |Increased | | | |(.887 to .891) | |Age |I: no change |Decrease | | | |(.853 to .844 | |Age_firs |I: no change |Decreased | | | |(-.581 to -.536) | Interpretation Not much change. The rotated pattern is not a substantial improvement over the unrotated pattern. How Good a Fit is The Three Factor Solution' Measures of goodness-of-fit KMO (0.698) mediocre to middling Percent of variance accounted for 67.44%, the same for both the unrotated and rotated solutions Communalities (the proportion of the variability in each variable accounted for by the three factors) Ranges from a high of 0.917 for sentence to a low of 0.252 for dr_score Factor patter Fairly clear for Factors I and II, ambiguous for Factor III Reproduced correlation matrix One measure of the goodness-of-fit is whether the factor solution can reproduce the original intercorrelation matrix among the eleven variables. Reproduced Intercorrelation Matrix [pic] The upper half of the table presents the reproduce bivariate correlations. Compare these with the lower half of the table that presents the residuals. Residual = (observed - reproduced correlation) The diagonal elements in the upper half of the table are the communalities associated with each variable. Over half of the residuals (52%) are greater than 0.05 Computing Factor Scores A useful byproduct of factor analysis is factor scores. Factor scores are composite measures that can be computed for each subject on each factor. They are standardized measures with a mean = 0.0 and a standard deviation of 1.0, computed from the factor score coefficient matrix. Application Suppose the eleven variables in this case study were to be used in a multiple regression equation to predict the seriousness of the offense committed by parole violators. Clearly, the eleven predictor variables are collinear, a problem in the interpretation of the extent to which each variable effects outcome. Instead, for each subject, computed a factor score for each factor, and use the factors as the predictor variables in a multiple regression analysis. Recall that the factors are noncollinear. Regression Analysis of Crime Seriousness With the Eleven Variables [pic] [pic] Regression Analysis of Crime Seriousness With the Eleven Variables (cont.) [pic] Interpretation R2 = 0.696 Significant predictor variables: sentence on the previous offence, prior convictions, and previous jail time The other predictors were not entered into the equation, not because they are unrelated to crime seriousness, but because they are collinear with the variables in the equation. Regression Analysis with Factor Scores [pic] [pic] [pic] Interpretation of the Regression Analysis With Factor Scores R2 = 0.643 (previous regression model = 0.696) ser_indx = 3.829 + 1.587 (F I) + 0.303 (F III) F = Factor Factors I and III were found to be significant predictors. Factor II was not significant. Factor I is assumed to be a latent factor that measures criminal history, while Factor II appears to measure educational/skill level. Since only one variable (tm_disp) loaded on Factor III, it is difficult to theorize about the nature of the latent variable it measures other than it measures time to disposition. Comparison of the beta weights indicates that the criminal history factor (Factor I) has a greater effect in explaining the seriousness of the offense than does Factor III (tm_disp).
上一篇:Strategic_Management 下一篇:Soc120_Week2_Assignment