ࡱ> wyv Mbjbj }xxD@`````$PL .At@   @@@@@@@$BXE@`     @``4@ |``@ @n:?0L>/b= @@0.A=ErE@??`?   @@   .A    E         * &:    Exploratory Factor Analysis Factor Analysis and Test Validity Exploratory Factor Analysis Factor Analysis is a multivariate statistical method whose primary purpose is to define the underlying structure for a group of related variables. This technique addresses the problem of analyzing the structure of the interrelationships (correlations) among a large number of variables (e.g. test scores, test items, questionnaire responses etc.) by defining a set of common underlying dimensions, known as factors. The two primary uses of factor analysis are summarization and data reduction. It allows us to take a large set of data and make it more interpretable. Exploratory Factor Analysis is used to determine to traits or factors that comprise a set of data. Confirmatory Factor Analysis attempts to validate hypothesized factors that one expects to emerge from data collected from test questions, questionnaire responses, etc. This process is used to validate questionnaires, personality inventories and many other types of psychological tests such the Wechsler Intelligence Scale for Children 4th Ed. (WISC- IV), MMPI-II, etc. A Measure for Assessing Anxiety Associated With Using SPSS Factor analysis is frequently used to develop questionnaires. The primary use of factor analysis in the development of questionnaires is to ensure that questions designed to measure an ability or trait are in fact related to the construct that you intend to measure. The example that we will use to demonstrate factor analysis is taken from a research study conducted by Field (2000) to study student anxiety related to the use of SPSS. Field generated a twenty-three item questionnaire based on interviews with anxious and non-anxious students. Each question was a statement followed by a five point Likert Scale ranging from Strongly Disagree through Neither Agree or Disagree to Strongly Agree. The questionnaire is attached. Research question Do the twenty three items comprising the SPSS Anxiety Scale represent one unitary trait ( or factor) to explain how students experience anxiety with SPSS? There are four basic steps to the factor analysis process: Data screening: calculate the descriptive statistics and a correlation matrix of all variables to be used in the analysis. Extract factors Rotate factors to create a more understandable factor structure Interpret results The Major Assumption Underlying Factor Analysis The relationship between measurement variables is linear (i.e. correlated to one another). Scatterplots of variables can be examined to determine if variables are linearly related. Sample Size As a general rule, the minimum sample size number at least five times as many observations as there are variables to be analyzed, and a more acceptable size would reflect a 10-1 ratio. At a minimum, the sample size should be at least 100 however 200 or more would be better. In this case the sample consists of 2,571 completed questionnaires. Conducting the Factor Analysis Using SPSS The first part of the analysis consists of determining the number of extracted factors. Click Analyze, Click Data Reduction and Click Factor. You should see the Factor Analysis dialogue box. Holding down the control key, click the 23 SPSS anxiety variables (items a through j). The click on > to move the items to the Variables box in the Factor Analysis dialog box. At the bottom of this box, click on Descriptives. In the Descriptives dialog box make sure that Univariate descriptives, Coefficients, Significance Levels, Determinant, KMO, Barletts test of Sphericity and Reproduced are checked. Now click on Continue. Click Extraction. You will see the Factor Analysis: Extraction dialog box. In the Extraction box you will notice that the default Method is Principal Components. This is the most commonly used method for exploratory purposes. Continuing in Extraction box, we also want to make sure that the boxes Correlation Matrix, Unrotated Factor Solution, and Eignevalues over: 1 and Scree plot are checked. Now click on Continue. Rotating Factors When back in the Factor Analysis dialogue box, click Rotation. You will see the Rotation Dialog box. In the Rotation dialog box make sure that you check Varimax. Varimax is a type of orthogonal rotation method. Make sure that the Rotated Solution and Loading plots boxes are checked. Change the maximum iterations to 30. (Normally 25 is sufficient but we have an unusually large dataset to work with in this example). Now click on Continue Now back in the Factor Analysis dialogue box, click Options. In the Factor Analysis: Options box under Missing Values select Exclude Cases Pairwise. Now under Coefficient Display Format: Make sure that you check Sorted by Size. Click on Continue. From the Factor Analysis dialogue box click on Scores. In the Factor Analysis: Scores dialogue box check Save as variables and under Save as Variables select Anderson-Rubin. Next check Display Factor Score Coefficient Matrix. Click on Continue. In the Factor Analysis dialogue box, click Continue and OK. Data Screening During the data screening procedure, we examine the descriptive statistics and correlation matrix to determine if the relationship between variables satisfies the assumptions required to conduct a factor analysis. Low Correlations The first step is to examine the correlation matrix (refer to output) between variables (items) to examine how well they relate to one another. If we find that there are variables that do not correlate well with any other variables (or very few) then we should consider excluding these variables before the factor analysis is conducted. We would like to see our correlation coefficients exceed .30. Multicollinearity (Singularity) The opposite problem of low correlations is variable that correlate too highly. It is important to avoid extreme multicollinearity (i.e. variable that are very highly correlated) or singularity (variables that are perfectly correlated). As with regression, singularity causes problems in factor analysis because it becomes impossible to determine the unique contribution of a variable that is highly correlated with another variable. Bottom Line for Treating Collinearity At this early stage we look to eliminate any variables that show no relationship (do not correlate) with any other variables or that correlate too highly with other variables (i.e. r > .90). Evaluating Variables and Examining Singularity To evaluate the issues of low correlations and Singularity refer to the Correlation Matrix in addition to KMO (Kaiser-Meyer-Olkin) and Barletts Test of Sphericity sections of the SPSS output. Examine the correlations between variables Notice a Determinant value is listed at the bottom of the correlation matrix. The value for the determinant is an important test for multicollinearity or singularity. The determinant of the correlation matrix should be greater than .00001. If the Determinant value is less than this value, it would be important to attempt to identify pairs of variables where r > .8 and consider eliminating them from the analysis. The Barletts test is designed to determine if the correlation matrix is an identity matrix (where all correlation coefficients are 0). A significant value (less than .05) indicates that the data do not produce an identity matrix indicating there are adequate relationships between variables to conduct the factor analysis. Results from this test also indicate that the correlations among variables overall are not so strong suggesting multicollinearity. The KMO test is a measure of whether the distribution of values based on the sample is adequate for conducting a factor analysis. This test indicates the amount of overlap or shared variance between pairs of variables (remember we are trying to identify items that are related but yet provide unique information to the factors we are attempting to identify). Values should be greater than .5. Factor Extraction Refer to the output entitled Communalities. Communalities are estimates of shared or common variance among the variables after extraction has taken place. Communalities for each variable can also be interpreted as the squared multiple correlation (R2) of the variable predicted from the combination of extracted factors. The goal of factor analysis is to identify groups of variables (items in this case) that are related to one another and derive a description of the underlying traits that best represent the data structure. Refer to the SPSS output table entitled Total Variance Explained from the SPSS output: The underlying objective in principal component analysis is to obtain uncorrelated linear combinations of the original variables that account for as much of the total variance in the original variables as possible. These uncorrelated linear combinations are referred to as linear components. The first principal component is the linear combination of variables that accounts for the maximum amount of variance. The analysis then proceeds to find the second linear combination - uncorrelated with the first linear combination that accounts for the next largest amount of variance (after that which has been attributed to the first component has been removed). This process continues until all variance accounted for is represented by principal components or factors. The total amount of variance for a component or factor is represented as an Eigenvalue. The eigenvalue for the first component is 3.730 and accounts for 16.219% of the variability or variance of the total data structure .Components or factors with eigenvalues of 1 or greater are considered to contribute significantly to the data structure. SPSS by default extracts only components or factors with eigenvalues of 1 or greater. In the case of factor analysis we would like to explain at least 60-70% of the variance in the data structure by the factors that are extracted. How much total variance in the data structure is explained by this example (refer to your Total Variance Explained table). Criteria for Number of Factors to Extract Eigenvalue Eigenvalue The portion of the total variance of a correlation matrix that is explained by a linear combination of items in a factor. Components with eigenvalues greater than 1 should be retained. This criteria is reliable when the number of variables is < 30 and the communalities are > .70, or the number of individuals is > 250 and the mean communality for all variables is > .60 2. Variance Retain components that account for at least 70% of the total variability. 3. Scree plot Retain all components with the sharp descent, before eigenvalues level off. 4. Consider the residuals provided by the reproduced correlation matrix That is consider the residuals or difference between the actual correlations and reproduced correlations that stem from the factor analysis model based on the data analyzed. When considering the Reproduced Correlation Matrix retain the components generated by the model if only a few residuals (the difference between the empirical and reproduced correlations represented in the lower portion of the Reproduced Correlation Matrix output) exceed a difference of .05 between actual correlations and reproduced correlations. If several reproduced correlations differ, you may want to include more components. A condition of the data of concern is when more that 50% of reproduced and actual correlations differ by more than .05 Rotation of Factors An important analysis for interpreting factors is factor rotation. Rotation is the process by which a factor solution is made more interpretable without altering the underlying mathematical structure. The reference axes of factors are turned about their origin until the axes of each factor are better aligned with the variables they represent. Orthogonal Rotation methods such as Varimax, Equamax, and Quartimax are rotation methods designed to identify factors that are independent of one another. The whole notion of factor analysis is to identify groups of variables that can explain independent underlying traits in the data structure. However, there may be cases where for theoretical reasons (based on research and hypotheses) that factors may be related to one another. When this is the case Oblique rotation methods would be used. In SPSS, Direct Oblim and Promax are the oblique rotation methods available. Refer to the Total Variance Explained Table of SPSS. Notice that as a result of rotation the percentage of variance in the data explained by each factor changes. This is because the factors have been rotated to better represent the proportion of variance that they are responsible for explaining when considering all the variables collectively. Rotated Component (Factor) Matrix The Rotated Component (Factor) Matrix table in SPSS provides the Factor Loadings for each variable (in this case item) for each factor. A Factor Loading is the Pearson correlation (r) coefficient between the original variable with a factor. For example, if we consider question 6, we can see that it loads on or correlates .800 with Component 1 (Factor 1), -.010 with Component 2 (Factor 2) and .097 with Component 3 (Factor 3) and -.072 with Component 4 (Factor4). So, which factor does Question 6 load heaviest on? A factor loading of + .50 or greater is consider practically significant with a sample size of approximately 100. Of course the larger the sample size the smaller the loadings need to be for practical significance. We will continue to identify which items are highly correlated with factors and attempt to find the patterns of items which represent factors. The next step, is to examine the items to determine if they represent some type of attribute or trait that can be use to name the factor. In our example, four factors have emerged. Questions 6, 18, 13, 7, 14, 10, and 15 load on Factor 1. Questions 20, 21, 3, 12, 4, 16, 1, 5 load on Factor 2. Questions 8, 17, 11 load on Factor 3 and Questions 9, 22, 23, 2, and 19 load on Factor 4. It appears that Factor I (questions 6, 18, 13, 7, 14, 10, and 15) would be indicative of perceptions of students toward Computer Use. Factor 2 comprised of 20, 21, 3, 12, 4, 16, 1, 5 seem to be related to fear or stress related to statistical concepts and could be named Fear of Statistics. Questions 8, 17, 11 load on Factor 3 and appear to be related to students self-perceptions of their mathematical skills and could be named Mathematics Self-Concept. The fourth and final factor composed of questions 9, 22, 23, 2, and 19 seem to be related to student perceptions of the use of SPSS and could be labeled SPSS Self-Concept. Factor Scores The purpose of factor analysis is to reduce a large set of data into a smaller subset of measurement variables. The factor scores tell us an individuals score on this subset of measures. Therefore, any further analysis can be done using factor scores rather than the original data. Secondly, factor scores may be appropriate to use for Multiple Regression analysis because they are produced from uncorrelated factors. Thus these scores reduce or eliminate multicollinearity that can cause problems with multiple regression analysis. Statistics Anxiety Questionnaire* By Andy Field Please answer the following questions SD = Strongly Disagree; D = Disagree; N = neither; A = Agree; SA = Strongly Agree SD D N A SA Statistics make me cry ___ ___ ___ ___ ___ My friends will think I am stupid for not being able to cope with SPSS. ___ ___ ___ ___ ___ Standard deviations excite me. ___ ___ ___ ___ ___ I dream that Pearson is attacking me with correlation coefficients. ___ ___ ___ ___ ___ I dont understand statistics. ___ ___ ___ ___ ___ I have little experience with computers. ___ ___ ___ ___ ___ All computers hate me. ___ ___ ___ ___ ___ I have never been good a mathematics. ___ ___ ___ ___ ___ My friends are better at statistics than me. ___ ___ ___ ___ ___ Computers are useful for only playing games. ___ ___ ___ ___ ___ I did badly at mathematics at school. ___ ___ ___ ___ ___ People try to tell you that SPSS makes Statistics easier to understand but it doesnt. ___ ___ ___ ___ ___ I worry that I will cause irreparable damage Because of my incompetence with computers. ___ ___ ___ ___ ___ Computers have minds of their own and Deliberately go wrong whenever I use them. ___ ___ ___ ___ ___ Computers are out to get me. ___ ___ ___ ___ ___ SPSS Anxiety Scale (continued) SD D N A SA I weep openly at the mention of central tendency. ___ ___ ___ ___ ___ 17. I slip into a coma whenever I see an equation ___ ___ ___ ___ ___ 18. SPSS always crashes when I try to use it. ___ ___ ___ ___ ___ Everybody looks at me when I use SPSS. ___ ___ ___ ___ ___ I cant sleep for thoughts of eigenvectors. ___ ___ ___ ___ ___ I wake up under my duvet thinking that I am trapped under a normal distribution. ___ ___ ___ ___ ___ My friends are better at SPSS than I am. ___ ___ ___ ___ ___ If I am good at statistics people will think I ___ ___ ___ ___ ___ am a nerd. *Field, A. (2000). Discovering statistics using SPSS for windows. Sage: Newbury Park, CA. s so that I could work to solve problems     PAGE  PAGE 319 PAGE   PAGE 321  ()EGVz   > `    Q Y  " A P Q [ w   $ C ^ _ e f p мդhrhX6hIhr hrhXhrhX56 h)sh)sh)sh)s5h)sh)s5CJ(aJ( h)s5hrhX5 hr5 hX5hrhXCJ$aJ$hrCJ$aJ$ hraJ:    ()*+,-./01234567$a$gdXgdX789:;<=>?@ABCDEFGHIJKLMNOPQR $h^ha$gd)s$a$gdXRSTUVxyz{|}~ $h^ha$gd)s[ \ Z [  & Fgdr & FgdX & FgdrgdX & FgdX $h^ha$gd)s % DVrny"#%&'()GRSTyz'>@LNacnpOUYnĻ h+t9CJhrhXCJ h0_CJ hrh+t9h+t9 hrhu<9hWDhrhX5hXhr hrhXhrhXH*G & ' $a$gdXgdXq"#%&'()STgdX$a$gdXZ^ZgdX8^8gdX & FgdX & FgdXgdXTstuDE= $h^ha$gdXh^hgdX & FgdXgdXgdXn  &,@CNitu",klp;CDEUdy(*;GVms|)4;S_hlnprhrhX5>*hrhX>*hrhr5 hX5 hr5hrhX5 hrhXQ=>34pqrst[\mn  !!!$a$gdXh^hgdX & FgdXgdXrst\n '-7AikLZ !!!!!!!<"N"^"""""#.#/#G#f##J$K$$$$$ޤޚhrhX56 h0_5h$thX6 h$thXhrh$t5 h$t5hI hrhU hrhu<9 hrhXhrhX5hrhX5CJaJhrhr5 hX5^`>gd!u$a$gdX & F^`gdXgdXgdPJ+i+++6,C,^,},- ----- /00I0V00001W1[1`1v1{11111111112222:3;3333334<4M444445%5:5;5I5X5ϽϵϵϵϽϫhTIhTI5h$th!uhX5>*h!uhX5 h!uh$th!u h!uh!u h!uhXhrh 5hI hrh hrhX5hWD hrhXh0_<4222G5H555555555767_8`8t9u9w9:::::<== & F^`gdX$a$gdXgdXX5d5t5555555555555567@7Z7y7g88888909<9A9H9u9w999:::;8;G;;;;;===H='@*@dBvBDDD hIo5 hrhIhrhX5>*hrhX6 hrh{ hrh hrhXhrhX5 hP5hrh 5hrhTI5 hX5hTIh$t5hTIhX5hTIhTI5 hTI56==??eBfBgBuBvBDDDDDDDDDDDDDDDDDDD$a$gdIo$a$gdXgdXDDDDDDDDDDDDDDDDDDDEEWEXEYEmEnEE & F gdIogdIo$a$gdIoDDDDDDD/E6E5FSF_HHIIIXLLLLLLLLLLLLLLLLLLLLLLLLLLžᶲwqwqmwqwheL heL0JjheL0JUhD jhD Uh\&h\&5)hP5B*CJOJQJ\^JaJphh;T/h=|hyghIhIo5 hIo56hIhIo56hYhP5 hn5 hI5 hIo5 hh5 heL5 hP5 h0_5 hr5)EEEEE4F5F`FFFFFGG6G7GrGsGGGGG3H4H\HHHH & F gdIoh^hgdIoHII8IxIyIIIIIIII J0J7JJJJJK KJKKKwK & F gdIo^gdIogdIogdP & F gdIoh^hgdIowKKKKK7LCLDLELLLLLLLLLLLLLLLLLL & F gd\& 7$8$H$gd=| & F gdIoh^hgdIoLLLLLLLLLLLLLMMMMMMM & F gd\&$a$gd^Sh]hgdr &`#$gdrh]hgd^ &`#$gd^SLLLLLLMMMMMMMMMMMMMh\&h\&5hD h^SheLheL heL0JjheL0JUhh0JmHnHuC0P5&P1h:p^S/ =!"#$% Dp56&P1h:pXj/ =!"#$% *~666666666vvvvvvvvv666666>6666666666666666666666666666666666666666666666666hH6666666666666666666666666666666666666666666666666666666666666666662 0@P`p2( 0@P`p 0@P`p 0@P`p 0@P`p 0@P`p 0@P`p8XV~8XV~ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@ 0@66_HmH nH sH tH @`@ XNormalCJ_HaJmH sH tH B@B X Heading 1$$@&a$5CJDA`D Default Paragraph FontRiR  Table Normal4 l4a (k (No List 4>@4 XTitle$a$5CJ2B@2 X Body TextCJ4@4 ^Header  !.)@!. ^ Page Number4 24 SFooter  !jCj r Table Grid7:V0PK!pO[Content_Types].xmlj0Eжr(΢]yl#!MB;.n̨̽\A1&ҫ QWKvUbOX#&1`RT9<l#$>r `С-;c=1g3=KjK爉ŬʉaqHwo?x}>6!N J?6v?w_Z3tç$y/,YVϲYLcDt~q"9⧣E"g 36AxgKA,ƉwV=&D-䧘a"FEG<+)/騳2YfEA5i5r{F2rhnz*kU̘aby&Z4MtoKnwu[K@Y5Fm3AM2ޕaŨ; uۭ=: ުv|TV_>%^/*!C!=I.pŭG2#=u{A*?xMVfȯ׆h,"N~e /w$wmBTRUzcn5fwЪtq:nT=q k:V=*^&whv3om <"^PK! ѐ'theme/theme/_rels/themeManager.xml.relsM 0wooӺ&݈Э5 6?$Q ,.aic21h:qm@RN;d`o7gK(M&$R(.1r'JЊT8V"AȻHu}|$b{P8g/]QAsم(#L[PK-!pO[Content_Types].xmlPK-!֧6 -_rels/.relsPK-!kytheme/theme/themeManager.xmlPK-!Atheme/theme/theme1.xmlPK-! ѐ' theme/theme/_rels/themeManager.xml.relsPK] DEE &&&&&1?????B nr$J+X5DLM',02468;@7R T=!'42=DEHwKLM()*+-./13579:<=>? "&-18<B!!!!@ @H 0(  0( V@W@ B S  ?<<DDDDDDDDDDDDDEEEEEsw +&@&"'*')*)r5{5Q9Y9<<<<DDDDDDDDDEE33333333333I(I(V((")")v)v)))))<<DDDDDDDDDEEEE R<Hn0`-p=+& 957277 AAbUB~pk-O4(HEW!V1jt'w{fe0Dx}rtR ^`OJQJo( 8^8`OJQJo( ^`OJQJo(o  p^ `OJQJo(  @ ^ `OJQJo( x^x`OJQJo( H^H`OJQJo(o ^`OJQJo( ^`OJQJo( 88^8`hH. ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH. 808^8`05hH. ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH. ^`hH. ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH. ^`hH. ^`hH. pL^p`LhH. @ ^@ `hH. ^`hH. L^`LhH. ^`hH. ^`hH. PL^P`LhH.%p^`p56CJOJQJo(phhH ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH.h ^`hH.h ^`hH.h pL^p`LhH.h @ ^@ `hH.h ^`hH.h L^`LhH.h ^`hH.h ^`hH.h PL^P`LhH. 88^8`hH. ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH.h ^`hH.h ^`hH.h pL^p`LhH.h @ ^@ `hH.h ^`hH.h L^`LhH.h ^`hH.h ^`hH.h PL^P`LhH.h(p^`p56B*CJOJQJo(phhH ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH.^`o(. ^`hH. pLp^p`LhH. @ @ ^@ `hH. ^`hH. L^`LhH. ^`hH. ^`hH. PLP^P`LhH.h(p^`p56B*CJOJQJo(phhH ^`hH. pp^p`hH. @ @ ^@ `hH. ^`hH. ^`hH. ^`hH. ^`hH. PP^P`hH.^`o(. ^`hH. pLp^p`LhH. @ @ ^@ `hH. ^`hH. L^`LhH. ^`hH. ^`hH. PLP^P`LhH.HEW'w{Hn UB -p +& 27 27-px}1jk-O95AAP?T0V؝PT0UV ؝PT0U ؝PT0؝@ӣT0&؝ <                           :                          :                 :                 1\m9:9D U&yg-9#\&@);T/u<9+t9T?GTIP0_\7a/eIoqr$t!u=|0Y^^S6 & ! n|q5.xs)sr~FIWD3dJSh{ eLi !XjXDD@<<<<E@UnknownG*Ax Times New Roman5Symbol3 *Cx Arial? *Cx Courier New;WingdingsA$BCambria Math"1h<[G['lFF : #|F : #|x4DD2QHP ?yg2!xxdͣ' ,Factor Analysis Art BangertMicrosoft Office User@          Oh+'0 $0 T ` l x'Factor Analysis Art Bangert Normal.dotmMicrosoft Office User4Microsoft Macintosh Word@^в@u@\?@ACDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdeghijklmopqrstuxRoot Entry FL>/bz1TableB FWordDocument}SummaryInformation(fDocumentSummaryInformation8nCompObjr  F Microsoft Word 97-2004 Document MSWordDocWord.Document.89q