Science method
Exploratory Factor Analysis - Science method
Explore the latest questions and answers in Exploratory Factor Analysis, and find Exploratory Factor Analysis experts.
Questions related to Exploratory Factor Analysis
In EFA, is it okay that each of the participants will answer all the items the questionnaires? For example, I have 50 items and I decided to have 500 participants. Is it okay if each of the participant will answer all the items?
Thanks.
After mapping all survey questions to KPI,Using Exploratory Factor Analysis, I got main loaded factors , which can be said as KPI, can than low loaded factors be considered as influencing factors to these KPI.
Performing Exploratory Factor Analysis (EFA) and Confirmatory Factor Analysis (CFA) involves several systematic steps. Here's a detailed guide for each:
How to Perform Exploratory Factor Analysis (EFA)
- Prepare the Data:Collect Data: Ensure you have a sufficiently large dataset. Check for Suitability: Use the Kaiser-Meyer-Olkin (KMO) test and Bartlett’s Test of Sphericity to determine if your data is suitable for factor analysis.
- Select the Number of Factors:Eigenvalues: Use the Kaiser criterion (eigenvalues > 1). Scree Plot: Examine the scree plot to identify the point where the eigenvalues level off. Parallel Analysis: Compare the eigenvalues with those obtained from random data.
- Extract the Factors:Methods: Common methods include Principal Axis Factoring (PAF) or Maximum Likelihood (ML). Software: Use statistical software like SPSS, R (using the psych package), or SAS.
- Rotate the Factors:Orthogonal Rotation: Varimax rotation for uncorrelated factors. Oblique Rotation: Promax or Oblimin for correlated factors. Interpretation: Identify and label the factors based on the rotated factor loadings.
- Interpret the Results:Factor Loadings: Examine loadings to understand which variables correlate strongly with which factors. Communalities: Check communalities to understand the variance explained by the factors.
How to Perform Confirmatory Factor Analysis (CFA)
- Specify the Model:Theory-Driven: Define the number of factors and the loadings of observed variables on these factors based on theoretical expectations or previous research. Path Diagram: Draw a path diagram to visualize the model.
- Collect and Prepare Data:Sample Size: Ensure a large enough sample, typically a minimum of 200 observations or 5-10 times the number of parameters. Data Cleaning: Check for missing data and outliers.
- Estimate the Model:Software: Use software like AMOS, LISREL, Mplus, or R (using the lavaan package). Fit the Model: Use the software to estimate the parameters of the specified model.
- Assess Model Fit:Fit Indices: Evaluate fit indices such as Chi-Square (χ²), Root Mean Square Error of Approximation (RMSEA), Comparative Fit Index (CFI), and Tucker-Lewis Index (TLI). Modification Indices: Check modification indices to identify potential improvements to the model.
- Refine the Model:Modification: Make theoretical justifiable adjustments based on modification indices. Re-Estimation: Re-estimate the model after modifications.
- Interpret the Results:Factor Loadings: Examine standardized factor loadings to understand relationships between observed variables and latent factors. Residuals: Analyze residuals to ensure the model is capturing the data well.
Example using R for EFA and CFA
EFA Example (Using psych package)
# Install and load the psych package
install.packages("psych")
library(psych)
# Sample data
data <- read.csv("your_data.csv")
# Perform EFA
efa_result <- fa(data, nfactors = 3, rotate = "varimax")
# Print the result
print(efa_result)
CFA Example (Using lavaan package)
# Install and load the lavaan package
install.packages("lavaan")
library(lavaan)
# Define the CFA model
model <- '
Factor1 =~ X1 + X2 + X3
Factor2 =~ X4 + X5 + X6
Factor3 =~ X7 + X8 + X9
'
# Sample data
data <- read.csv("your_data.csv")
# Fit the CFA model
cfa_result <- cfa(model, data = data)
# Summary of the results
summary(cfa_result, fit.measures = TRUE)
By following these steps, you can perform EFA to explore the underlying factor structure of your data and CFA to confirm the hypothesized structure.
To give reference
Singha, R. (2024). How to perform Exploratory Factor Analysis (EFA) and Confirmatory Factor Analysis (CFA)? Retrieved from https://www.researchgate.net/post/How_to_perform_Exploratory_Factor_Analysis_EFA_and_Confirmatory_Factor_Analysis_CFA?_init=1
I am running a PCA in JASP and SPSS with the same settings, however, the PCA in SPSS shows some factors with negative value, while in JASP all of them are positive.
In addition, when running a EFA in JASP, it allows me to provide results with Maximum Loading, whle SPSS does not. JASP goes so far with the EFA that I can choose to extract 3 factors and somehow get the results that one would have expected from previous researches. However, SPSS does not run under Maximum Loading setting, regardless of setting it to 3 factors or Eigenvalue.
Has anyone come across the same problem?
UPDATE: Screenshots were updated. EFA also shows results on SPSS, just without cumulative values, because value(s) are over 1. But why the difference in positive and negative factor loadings between JASP and SPSS.
I am running an Exploratory Factor Analysis (EFA) for a variable.
EFA showed that the first six initial eigenvalues were more than 1, so I needed to retain them according to Larsen & Warne (2010) and Shrestha (2021) stated that researchers retain all factors for which the eigenvalue is greater than 1.0 following the rule of Guttman (1954).
The seventh component was also more than one but less than 1.2. How can I justify that the seventh component can be removed? Because an initial eigenvalue less than 1.2 is typically considered insignificant and may justify the removal of a factor.
Therefore, I have been trying to find a reference to justify my action. Please provide me with references that mention that an initial eigenvalue less than 1.2 is typically considered insignificant and may justify removing a factor.
Thank you.
My research investigates the influence of training and development and performance appraisal on the competitive advantage of Nigerian private schools. It also explores the potential role of ethical climate as a mediating factor in this relationship.
In this study, a quantitative research approach was utilized to collect and analyze numerical data and to identify patterns and relationships. The analysis includes a variety of techniques, such as descriptive analysis, exploratory factor analysis, confirmatory factor analysis, and structural equation modeling. Widely recognized statistical software programs, SPSS and AMOS, have been used to perform these advanced analytical methods.
However, the reviewer comments that the variable construct is still based on theory, not empirically in Nigerian Private Schools. How can I do it or respond to it?
I am running an exploratory factor analysis for six constructs. Each of them has eight items. The first six factors were more than 1, so I needed to retain them, but the seventh component was more than one but less than 1.2. How can I justify that the seventh component can be removed because an initial eigenvalue less than 1.2 is typically considered insignificant and may justify the removal of a factor? I have been trying to find a reference to justify my action. Please provide me the references that mentioned that an initial eigenvalue less than 1.2 is typically considered insignificant and may justify the removal of a factor.
I appreciate any help you can provide.
I am running an exploratory factor analysis for six constructs. Each of them has eight items. The first six factors were more than 1, so I needed to retain them, but the seventh component was more than one but less than 1.2. How can I justify that the seventh component can be removed because an initial eigenvalue less than 1.2 is typically considered insignificant and may justify the removal of a factor? I have been trying to find a reference to justify my action. Please provide me the references that mentioned that an initial eigenvalue less than 1.2 is typically considered insignificant and may justify the removal of a factor.
I appreciate any help you can provide.
I need to access and read academic studies that use the exploratory factor analysis (EFA) as a analysis method with no more than 10 initial factors. I want to explore if this is academically possible and if there are studies that show the feasibility. If you can help me out with that, i would highly appreciate your efforts. Thank you very much guys :)
with three factors, the total variance explained is 53%. With two factors, the total variance is 45%. Can I keep one item for a given factor if I think that item is important to be kept in the tool? That item will not be loaded into other factors when I select the two-factor model.
Hi!
This might be a bit of a stupid question, but I am currently writing my master thesis. One of the things I am doing is a factor analysis on a scale developed in Canada. This scale has only been validated on the Canadian workforce (the developers have one time done a exploratory factor analysis and two times done a confirmatory factor analysis). I am doing an exploratory and a confirmatory factor analysis in the Norwegian workforce to see what factor structure I would find here, and if it is the same as in Canada. As this this only one of three things I am doing in my masters I have hypothesis for all the other findings, so my supervisor would like me to have hypothesis for the factor structure as well. Whenever I try to come up with some arguments, I always feel like I am just arguing for the same attitudes in both countries, rather than the factor structure.
My question is: How do you make a hypothesis for this where you argue for the same/a different factor structure without arguing for the same/different attitudes?
Thank you in advance! :)
I did EFA on a set of Likert Scale data.
I got 3 factors based on parallel analysis.
Factor 1 - 15 items
Factor 2 - 5 items
Factor 3 - 3 items
Didi CFA and got good fit.
Is it possible to reduce the number of factors to say 4 or 5?
I tried reducing it, did cronbach's alpha ..before was like 0.93, after is 0.92-ish
BTW, factor loadings of all 14 of the 15 items are also quite high - 0.69 to 0.92
I'm trying to reduce the number because in SEM, the 15 items is a lot of freeparameters resulting in identification issues.
If you have suggestions, let me know. Thanks!
Hi,
I did an exploratory factor analysis. Direct Oblimin and Principal Axis factoring. The 12 variables loaded on 2 factors. The factor loadings on factor 1 (column one in pattern matrix) and factor loadings on factor 2 (column two in pattern matrix) are highly negatively correlated (R2=0.991).
Is this normal or problematic? Thank you!
Can I use exploratory factor analysis on the same data after an unfavorable result of confirmatory factor analysis? In the procedure where I first apply exploratory factor analysis on a data set, then it is no longer correct (and there are methodological and statistical reasons for this) to use confirmatory factor analysis on the same data set. But what about the first case (from CFA to EFA on the same data)?
I am compiling factors from previous literature review for exploratory factor analysis.
Q1) How to extract the data into csv file?
Q2) is it more reliable to use SMARTPLS 4 to do EFA?
Q3) is it possible to use jamovi to do EFA?
Q4) how do I prove the reliability and validity of using jamovi as the tool?
Thank you
I would like to find out if there is any impact of social media on the creation of entrepreneurial opportunity. I have two groups (Online business entrpreneurs and traditional business entrepreneurs) and 20 statements in total that I have measured using 5 point likert scale (strongly agree, agree, neutral, disagree, strongly disagree and Not applicable(suggested by my Supervisor)). Those who run traditional business chose "not applicable" for 8 items and I have ignored those 8 items and proceed with rest of the 12 items for further analysis e.g. Mann-Whitney U test, Kruskal-Walis test and correlation analysis.
Now I have 3 questions:
1. Is it okay to ignore those items having "not applicable"?
2. I have tried exploratory factor analysis for both groups and got 3 factors for each, however, one factor among them has only two items under it. Is two items under a factor acceptable?
3. What other tests or analysis I can do other than the above mentioned tests?
I conducted an Exploratory Factor Analysis using Principal Axis Factoring with Promax rotation, resulting in the identification of 8 factors. Now, I aim to examine potential associations between these extracted factor scores (treated as continuous variables) and other variables in my research. However, I am uncertain about the most appropriate statistical approach to compute these factor scores.
Should I compute the average sum of each factor, employ regression, or utilize the Bartlett method, Anderson-Rubin, any others….?
Any insights or alternative approaches are highly appreciated.
Thank you!
Hello,
I have a data set with N = 369 individuals measured at a single time point. The goal of the study is to create an assessment of psychological safety (PS). The assessment is a self-report measure asking participants to indicate how psychologically safe they feel using a unipolar 5-point Likert scale ranging from 1 (not at all) to 5 (extremely).
In addition to the assessment I am creating, I also measured a number of demographic variables (e.g., age, salary) and a few additional measures of team environment for validation (e.g., an existing measure of PS, level of team interdependence).
My primarily goal is to run exploratory factor analysis (EFA). This is the first time anyone has conceptualized PS as multidimensional, so one of the primary goals is to uncover the potential factor structure of PS. Also, to identify candidate items for deletion.
In order to prepare for the EFA analyses, I am cleaning the data by following recommendations in (the excellent) Tabachnik & Fidell (2013, 6th ed).
I am currently at the point where I am checking the data for multivariate outliers, starting with Mahalanobis distance. And I cannot find explicit guidelines regarding which variables I should be including as "IVs" in the analysis.
QUESTION: Which variables should I be including in my search for multivariate outliers? Do I include all variables, or only my target variables?
Specifically, do I include only the variables that represent the item pool for my forthcoming PS assessment? Or do I include all the PS items AND demographic variables, the existing PS assessment, interdependence measure, etc.??
I ran the Mahalanobis distance analyses 2 times using both approaches, and found substantial differences:
- TIME 1 - With just the PS assessment variables --> I identified n = 28 multivariate outliers.
- TIME 2 - With PS items + demographics, etc. --> I identified n = 10 multivariate outliers (all identified as outliers in the TIME 1 analysis).
Syntax I am using - the bolded variables are the ones I am questioning if I should include or not:
REGRESSION
/MISSING LISTWISE
/STATISTICS COEFF OUTS R ANOVA COLLIN TOL
/CRITERIA=PIN(.05) POUT(.10)
/NOORIGIN
/DEPENDENT Subjno
/METHOD=ENTER Age Salary Edu WorkStructure TeamSize Tenure_OnTeam JapaneseBizEnviron EdmondsonPS_TOT Interdep_TOT Valued_TOT PS_1 PS_48 PS_141 PS_163 PS_43 PS_53 PS_73 PS_133 PS_135 PS_19 PS_60_xl26 PS_93 PS_106_xl26 PS_143 PS_58 PS_86 PS_182 PS_56 PS_69 PS_103 PS_164 PS_22 PS_35 PS_91 PS_30 PS_59 PS_63 PS_90 PS_131 PS_140 (**Note, PS assessment var list is truncated b/c large number)
/RESIDUALS = OUTLIERS(MAHAL)
/SAVE MAHAL.
i’m trying to run polychotic correlation with Stata v13, but I’m confused.
Hey guys,
I have found two multi-item scales in my previous research regarding my master thesis. I want to know if I can compute an EFA for the dependent and for the independent variable?
I'm doing a validation study and for construct validity, I'll analyze my data through EFA and CFA. So, the same data should be used in both processes or should we divide the data? And if we divide the data what percent of the data should be included in EFA and what percent in CFA?
Can you please help me on how to use Exploratory Factor analysis as a Method of my research, actually my research is all about the Dimension of illegal activities commission. Please -I'm new researcher
What is meant by collecting data separatly for newly developed questions to run EFA in PLS-SEM? Like do i need to make to separte questionnaire sets one consisting of questions adopted from prior literatures and the other for the self developed questions?
Is it possible to run EFA from the data collected by pilot study and for the empirical work only running the CFA?
Please suggest.
Thanks in advance.
Dear RG community,
I have been trying to help my PhD student with an EFA. We have ended up with a very nice stabilised solution (using principal axis factoring, varimax rotation and extracting a fixed number of factors) apart from the determinant of the correlation matrix being too low. We had initially checked for correlations > 0.8 in absolute value but did not find any so we went back and removed one item from each pair where the bivariate correlation was > 0.7 in absolute value. This improved the determinant but it was still considerably lower than the recommended threshold of 0.00001. What do you suggest we should do: continue to reduce the bivariate correlation threshold and remove more items, just live with the low determinant value as what we are really interested in is the scales themselves derived from the EFA, or is there another way to detect and remove multicollinearity?
Thanks very much for your expert advice.
Hello, all!
I have translated and heavily edited a validated survey for use in a new context. This involved changing language, context, and cutting items from 56 to 30.
I'm now trying to find the underlying factor structure of this data, which I assume will differ from the 8 factors in the original 56-item scale.
For the purposes of my research, I tried to have respondents compare their expectations with their preferences for a given item. So, an item would present a statement such as (hypothetically) "During class, students make connections between content from different subjects." This would be followed by two Likert-type questions: "I would like this to be true" and "I expect this will be true". Repsondents answered each of these sub-questions for every item, on a 6-point scale (no neutral).
I'm a bit stuck, conceptually, on how to approach factor analysis with these paired questions. Conceptually, I'd assume that it shouldn't matter whether I incuded both "Q_a" and "Q_b", since those should load onto the same factor. But then, I think, the nature of the questions might confound such an analysis.
Does anyone have any wisdom or literature on how this type of paired response factor analysis has been done?
Thank you!
Dear researchers, can both EFA and CFA be applied to the database obtained at the end of the scale development process at the same time?
1- I determined exploratory factor analysis (eight dimensions and 37 questions were finalized).
2- Then I checked the measurement model. In the measurement model, only 13 questions had factor loadings greater than 0.5.
3- In this case, the previous dimensions taken from the exploratory factor analysis cannot be used, because one has been removed and the three dimensions have only 2 questions.
4- In this new mode, it is not possible to define a model with hidden variables, because we only have the indicators. Because the number of indicators for each dimension in this new mode is not enough, so the dimensions cannot be considered. Because according to Hair (2010), at least three questions are needed for each variable.
Question:
What do you recommend?
It is welcome any study from any field, as much as possible :)
Thank in you in advance
Good morning,
I am not an expert on Factor Analysis, so I hope the explanation of my problem makes sense.
My current task is to perform analyses on an older data set from experiments my lab conducted a couple of years ago. We have 212 individual items, consisting of a dozen or so demographic questions and items from a total of 24 different scales that measure separate constructs. Given the large number of items and constructs, I would like to reduce the number of dimensions to achieve a more clear starting point for theory development. Obviously, an exploratory factor analysis is a good choice for this.
My question is whether I have to input the 200 or so individual non-demographic items in the EFA, or whether I can instead just use the 24 composite variables/constructs and reduce the number of dimensions from there. My hesitation using all the individual items is that an EFA would simply return something very similar to the composite variables, as the constructs generally have a quite high internal consistency and are fairly distinct from each other based on theory. The obvious caveat with using the composite variables is that it is not something I have seen done much, and as I am not an expert on EFA, I am unsure as to whether there is a major methodological road block to using composite variables that I am unaware of.
Thank you for your help!
Best wishes,
Pascal
The result of the exploratory factor analysis for the measurement tool I used is not compatible with the original scale. For example, in the original study, item 3 is in factor 1, but in the study I conducted, it is in factor 3. In general, all items in the scale show factor loadings in different sub-dimensions than the original ones. How should I proceed?
Hi
I'm going to do a cluster analysis, after conducting an exploratory factor analysis for a series of Likert scale questions.
The cluster analysis will of course include other variables such as demographics, together with the factor analysis products.
My question is, do I use Factor Score (regression based or Bartlett based), for those 'Factors' from EFA or do I use mean score (average of the Likert scale questions that manifest individual factor)?
I know factor score has many advantage than mean score, but factor scores have a standard deviation of 1, and mean of 0. It's hard to interpret the later on cluster analysis based on these. While mean score seems to make more sense to interpret.
So, which score shall I use for the later on cluster analysis? mean score or factor score, or something else?
Thank you.
Hi all,
Kindly help me to know when can we use Exploratory Factor Analysis(EFA) and Confirmatory Factor analysis(CFA)?
Regards,
In the context of regression, methods for detecting collinearity are well described in the literature. In the context of exploratory factor analysis (EFA), however, I am facing a situation where I get highly unstable factor loadings and factor correlations from different boot-strapped samples (despite a sample size of more than 700), which I assume is due to multicollinearity. I am not sure how to explore its source. The common methods rely on the intercorrelation among explanatory variables (i.e., the latent factors in EFA), but the intercorrelation its-self is highly unstable.
My search for finding references on this topic was not much successful. Any explanation and/or reference on multicollinearity in EFA would be appreciated.
Thank you in advance,
Ali
Hello,
I applied exploratory factor analysis with network analysis to data from healthy and diseased patients. The analysis shows different clusters of parameters; some are similar in both groups, and some groups cluster differently. For instance, the parameters IleValLeu are similarly clustered in both groups whereas the parameters Pro Ala are not (see figure).
How shall I interpret the data? For instance, in the Pro/Ala case, I expected to see some differences between the groups, but they look pretty much the same to me.
Is the differences about correlation? The scatterplot of the data shows slightly different regression models, but not something compelling.
Is it the value itself? But again there is no real difference in the value distribution between the two groups.
So, what is the actual outcome of the network analysis?
Thank you
Hypothetically, if I would like to validate a scale and I need to explore its latent factors first with EFA followed by a CFA to validate its structure, do I need a new data set for the CFA? Some argued that randomly spliting one dataset into two is also acceptable. My question is that can I apply EFA on the dataset (with all data included) and then randomly select half of the dataset to conduct a CFA?
Dear Research Scholars, you are doing well!
I am a Ph.D. scholar in Education. Now I am working on my thesis. kindly guide me that when to perform the EFA whether it use on pilot studying data or actual research data.
Regards
Muhammad Nadeem
Ph.D. In Education , faculty of Education,
University of Sindh, Jamshoro
I am conducting an exploratory factor analysis and to determine the number of factors I used a paired analysis.
How can I generate the number of factors correctly in stata? Or other tool?
When using parallel analysis in stata, for example, if you proceed with Principal Axis Factoring all my Eigenvalues from the Parallel Analysis using a Principal Axis Factoring lower than 1. (Suggests in this case to retain all factors)
when out of curiosity, I use principal component factors, or even principal component analysis (I know this is not EFA), it suggests retaining 3 factors (which satisfies me)
When designing the questionnaire for EFA, what do I need to keep in mind when it comes to the order of the questions?
More specifically, does the order of the questions need to be completely randomized or is it generally allowed to still ask questions in topic blocks according to potential factors/constructs I have in mind?
Thanks everyone!
For my research project I am adding new items to a previously validated scale. in the previous research they performed an exploratory factor analysis revealing a two-factor structure, but the internal consistency of one of the scales was quite poor so the aim of my study was to add new words to improve the internal consistency. so do i need to do another exploratory factor analysis as the scale will now have new words or can i do a confirmatory factor analysis because i'm still using the same scale?
I did the confirmatory factor analysis of the questionnaire. If in each subscale I select the questions that have the highest factor load and I consider these questions as a short version and do other steps to verify the validity and reliability of the questionnaire, is this the correct way?
by doing exploratory factor analysis, I have less sub scale in comparison with sub scales in the long-form questionnaire.
In exploratory factor analysis, to name the dimensions, some items are different and heterogeneous from other items of the same dimension, and they make it difficult to name each category. For example, in the first category "personality type of the manager", in the 3rd category, two items "level of analysis of the spouse" and "wife's perception of the existence of social justice", in the 5th category, the item "organizational level of the manager (senior, middle, operational manager) ", in category 6, the item "Education level of wife" and in category 7, the item "Job transition of wife".
Regardless of the heterogeneous items above with other items in the same category, the naming of each category has been done as follows. The question is: what to do with heterogeneous objects? Are they deleted?
Category 1. Family intimacy: spouse's relationship with family; years of married life; wife's compassionate analysis of the company events recounted by the manager; Manager's personality type
Category 3. Manager's attitude towards his wife: manager's mental health; The amount of analysis of the wife; Manager's leadership style; The manager's level of trust in his wife; Wife's perception of the existence of social justice
Category 5. Spouse's power of influence on the manager: manager's organizational level (senior, middle, operational manager); The wife's effort to show support to the manager; Spouse's ability to regulate manager's emotions; The extent of the spouse's influence on the manager's decision-making; wife being a housewife
Category 6. Environment: the level of stability of the environment; Accompanying men (in the role of manager) with their working wives; Manager's income; spouse's level of education; The degree of balance between work and family by the manager
Category 7. Spouse's physical health Question: Spouse's job transfer; Wife's physical health; Manager's physical health
Here is the code I'm using:
[fa(vars, 4, rotate="oblimin", fm = "pa")]
Hi. I recently developed new survey items to measure extrinsic motivation. In the exploratory factor analysis, four items intended to measure extrinsic motivation loaded onto the same factor with negative scores (-.822; -.813; -.808; -.553). However, in the subsequent confirmatory factor analysis, the same items produced positive scores (.78; .80; .67; .56). I do not have a theory about why this could be and I would appreciate any help with explaining this finding.
Thank you.
I used confirmatory factor analysis without using the exploratory factor analysis first. The assumption that supported this was the clear presence of the factors in the literature. Is it correct not to go for the exploratory factor analysis and jump directly to CFA when you have clearly established factors in the literature and theory?
Can i split a factor that has been identified through EFA.
N=102 4 factors have been identified.
However, one of the 4 actually has two different ideas that obviously are factoring together. I am working on trying to explain how they go together but it is very easy to explain them as two separate factors.
When I conduct a Conformatory analysis the model fit is better for them separate. . . but running a confirmatory analysis on the same population of subjects that I conducted the Exploratory analysis on appears to be a frowned upon behavior.
Exploratory Factor Analysis and Confirmatory Factor Analysis are used in scale development studies. Rasch Analysis method can also be used in scale development. There are some researchers who consider the Rasch Analysis as up-to-date analysis. Frankly, I don't think so, but is there a feature that makes EFA, CFA or Rasch superior to each other in Likert type scale development?
I am conducting an EFA for a big sample and nearly 100 variables, but no matter what I do, the determinant keeps its ~0 value.
What should I do now?
I plan to develop a semi-structured assessment tool and further validate it on a relatively small sample of below 50 (clinical sample). I have been asked by the research committee to consider factor analysis.
So in this context, I wanted to know if anyone has used regularized factor analysis for tool validation which is recommended for small sample sizes?
I am aware that a high degree of normality in the data is desirable when maximum likelihood (ML) is chosen as the extraction method in EFA and that the constraint of normality is less important if principal axis factoring (PAF) is used as the method of extraction.
However, we have a couple of items in which the data are highly skewed to the left (i.e., there are very few responses at the low end of the response continuum). Does that put the validity of our EFAs at risk even if we use PAF?
This is a salient issue in some current research I'm involved in because the two items are among a very small number of items that we would like, if possible, to load on one of our anticipated factors.
Dear all,
I am conducting research on the impact of blockchain traceability for charitable donations on donation intentions (experimental design with multiple conditions, i.e., no traceability vs. blockchain traceability).
One scale/factor measures “likelihood to donate” consisting of 3 items (dependent variable).
Another ”trust” factor, consisting of 4 items (potential mediator).
Furthermore, a “perception of quality” consisting of 2 items (control).
And a scale “prior blockchain knowledge” consisting of 4 items (control).
My question is: since all these scales are taken from prior research, is CFA sufficient? Or, since the factors are from different studies (and thus have never been used together in one survey/model) should I start out with an EFA?
For instance, I am concerned that one (or perhaps both) items of ”perception of charity quality” might also load on the “trust”-scale. e.g., the item “I am confident that this charity uses money wisely”
Curious to hear your opinions on this, thank you in advance!
Greetings,
I am a DBA student conducting a study about "Factors Impacting Employee Turnover in the Medical Device Industry in the UAE."
My research model consists of 7 variables, out of which:
- 5 Variables measured using multi-item scales adapted from literature ex. Perceived External Prestige (6 items), Location (4 items), Flextime (4 items),.. etc.
- 2 are nominal variables
I want to conduct a reliability analysis using SPSS & I thought I need to do the below?
- Conduct reliability test using SPSS Cronbach's alpha for each construct (except for nominal variables)
- Deal with low alpha coefficients (how to do so?)
- Conduct Exploratory Factor Analysis to test for discriminant validity
Am I thinking right? Attached are my results up to now..
Thank you
Hi everyone,
I have longitudinal data for the same set of 300 subjects over seven years. Can I use '''year' as a control variable? Initially, I used one way ANOVA and found no significant different across seven years in each construct.
Which approach is more appropriate?. Pooling time series after ANOVA (if not significant) or using 'year' as a control variable?
Hello everyone,
As the title suggests, I am trying to figure out how to compute a reduced correlation matrix in R. I am running an Exploratory Factor Analysis using Maximum Likelihood as my extraction method, and am first creating a scree plot as one method to help me determine how many factors to extract. I read in Fabrigar and Wegener's (2012) Exploratory Factor Analysis, from their Understanding Statistics collection, that using a reduced correlation matrix when creating a scree plot for EFA is preferable compared to the unreduced correlation matrix. Any help is appreciated!
Thanks,
Alex
I`m conducting the translation of a very short scale of 12 ítems to assess therapeutic alliance in children. I have 61 answers and I wonder if that number of subjects it`s acceptable to run Exploratory Factor Analysis. I know that there is a suggestion of 5 participants for item to do EFA and 10 participants for item to do CFA. However, the number of participants here seem to be very smal for these analysys. What it´s your opinion?
Exploratory factor analysis was conducted to determine the underlying constructs of the questionnaire. The results show the % variance explained by each factor. What does % variance mean in EFA? How to interpret the results? How can I explain the % variance to a non-technical person in simple non-statistical language?
I have good result with low variance explanation (Less than 50%) in exploratory factor analysis, and read some discussions about the acceptable for total variance explanation < 50% in social sciences. Please recommend papers to support this issue or give me your suggestion.
Thanks in advance.
I am using the Environmental Motives Scale in a new population. My sample size is 263.
The results of my exploratory factor analysis showed 2 factors (Eigenvalue>1 and with loadings >0.3) - Biospheric Factor and Human Factor
Cronbach alpha was high for both factors (>0.8)
However, unexpectedly, confirmatory factor analysis showed that the model did not fit well:
RMSEA= 0.126, TLI =0.872 and SRMR = 0.063, AIC = 6786
After a long time on Youtube, I then checked the residual matrix and found that the standardized covariance residuals between two of the items in the Biospheric factor was 7.480. From what I understand if values are >3, it indicates that there may be additional factor/s that are accounting for correlation besides the named factor. I therefore tried covarying the error terms of those two items and rechecked the model fit using CFA.
Results of this model show much better model fit.
RMSEA = 0.083, TLI = 0.945, SRMR = 0.043, AIC = 6731 (not as much difference as I thought there would be)
The questions I am now left with (which google does not seem to have the answer to) are:
1. Is it acceptable to covary the error terms to improve model fit?
2. How does covarying error terms impact on the scoring of the individual scales? Can I still add up the items to measure biospheric vs human scales as I would have without the covarying terms?
I would be so grateful for any insight or assistance.
Thank you
Tabitha
I used exploratory factor analysis for 4 latent variables, the result in the table called "total of variance explained" has "% of variance". is it similar to average variance extracted?
What steps to do discriminant validity in SPSS? I run the factor analysis, then compute the latent variable to become observed variables, after that I run the correlation. is it the correct process?
Thanks for your attention
Hi, I am working on a project about ethical dilemmas. This project requires development of a new questionnaire that should be valid and reliable. We started with collecting the items from the literature (n= 57), performed content validity where irrelevant items were removed (n=46), and piloted it to get the level of internal consistency. Results showed that the questionnaire has a high level of content validity and internal consistency. We were requested to perform exploratory factor analysis to confirm convergent and discriminant validity.
Extraction PCA
rotation varimax
Results: the items' communalities were higher than 0.6.
kMO 70%
Barttlett's test is significant.
Number of extracted factors 11with total explained variance 60%.
My issue is 6 factors contain only 2 items. Should I remove all these items?
With notice that the items are varied, each one describes a different situation, and only they share in that they are ethical dilemmas and deleting them will affect the overall questionnaire ability to assess participants' level of difficulty and frequency of such situations.
EFA is new concept for me; I am really confused by this data.
Do you know any renowned article which has been published in Scopus journal describing that for conducting the Exploratory Factor Analysis (EFA), which method is the best, 'Principal component' or 'Principal Axis Factoring ' in SPSS?
I collected 109 responses for 60 indicators to measure the status of urban sustainability as a pilot study. So far I know, I cannot run EFA as 1 indicator required at least 5 responses, but I do not know whether I can run PCA with limited responses? Would you please suggest to me the applicability of PCA or any other possible analysis?
Query1)
Can mirt exploratory factor analysis method be used for factor structure for marketing/management research studies because most of the studies that I have gone through are related to education test studies?
My objective is to extract factors to be used in subsequent analysis (Regression/SEM)
My data is comprised of questions like:
Data sample for Rasch Factors
Thinking about your general shopping habits, do you ever:
a. Buy something online
b. Use your cell phone to buy something online
c. Watch product review videos online
RESPONSE CATEGORIES:
Yes = 1
No = 0
Data sample for graded Factors
Thinking about ride-hailing services such as Uber or Lyft, do you think the following statements describe them well?
a. Are less expensive than taking a taxi
c. Use drivers who you would feel safe riding with
d. Save their users time and stress
e. Are more reliable than taking a taxi or public transportation
RESPONSE CATEGORIES:
Yes =3
Not sure = 2
No = 1
Query2) If we use mirt exploratory factor analysis using rasch model for dichotomous and graded for polytomous, do these models by default contain tetrachoric correlation for rash model and polychoric correlation for graded models?
My objective is to extract factors to be used in subsequent analysis (Regression/SEM)
Note: I am using R for data analysis
Hi,
I used a self-efficacy tool for my sample. According to the original article, there is only one factor in the tool. However, in the Exploratory factor analysis for my sample, two factors were found. How can I interpret this result?
Thank you so much for your answer in advance!
Hi, I have run an exploratory factor analysis (principal axis factoring, oblique rotation) on 16 items using a 0.4 threshold. This yielded two factors, which I had anticipated as the survey was constructed to measure two constructs. Two items had factor loadings <0.4 (from Factor 1) so I removed them, leaving 14. However, upon closer inspection, one of the items from Factor 2 loaded on to Factor 1 (|B| = 0.460).
The distinction between the two constructs is very clear so there should not be any misunderstanding on the part of the participants (n = 104). I'm unsure of what to do. I checked the Cronbach's alpha for each factor: Factor 1 (a = .835 with the problematic item, a = .839 without). Factor 2 is a = .791).
Do I remove the item? Any advice would be very much appreciated. Thank you!
Can anyone help me with the sample size calculation for the exploratory factor analysis? Do you know how to calculate it and with which statistical program? Thank you.
Can mirt exploratory factor analysis method be used for factor structure for marketing/management research studies because most of the studies that I have gone through are related to education test studies?
My objective is to extract factors to be used in subsequent analysis (Regression/SEM)
My data is comprised of questions like:
Data sample for Rasch Factors
Thinking about your general shopping habits, do you ever:
a. Buy something online
b. Watch product review videos online
RESPONSE CATEGORIES:
1 Yes
2 No
Data sample for graded Factors
Thinking about ride-hailing services such as Uber or Lyft, do you think the following statements describe them well?
a. Are less expensive than taking a taxi
b. Save their users time and stress
RESPONSE CATEGORIES:
1 Yes
2 No
3 Not sure
I have extracted factor scores after EFA and then used K-means to identify clusters.
But there are confusion in validating the number of cluster in SPSS, as it does not show any AIC or BIC value on the basis i can differentiate and finalise the number of clusters.
I have adopted my questionnaire from previous literature. I want to know if I still need to carry out EFA before carrying out PLS-SEM for my thesis?
Hi,
I am working on exploratory factor analysis in SPSS using promax rotation.
Upon checking pattern matrix, there is one question which has factor loading greater than 1.0? should I need to ignore if loading greater than 1.0?
Also i realised there are few negative loadings in pattern matrix? Can negative loadings be still considered for the analysis?
Thanks much.
We can say that these factorial analysis approach are generally used for two main purposes:
1) a more purely psychometric approach in which the objectives tend to verify the plausibility of a specific measurement model; and,
2) with a more ambitious aim in speculative terms, applied to represent the functioning of a psychological construct or domain, which is supposed to be reflected in the measurement model.
What do you think of these generals' use?
The opinion given for Wes Bonifay and colleagues can be useful for the present discussion: