Science topic

SPSS - Science topic

Explore the latest questions and answers in SPSS, and find SPSS experts.
Questions related to SPSS
  • asked a question related to SPSS
Question
3 answers
In a causal model (such as multiple IVs and single DV) with presence of a mediator or moderator, do we have to consider such mediator or moderator when assessing the parametric assumptions or do we have ignore them and consider only the IV/s and DV in the model?
Relevant answer
Answer
Since you are going to involve a third variable that will eventually impact your results, you need to take that third variable into account and check for normality and other assumptions before you carry out your final analysis. However, while analysing the IV and DV, if the data is not found to be normally distributed, then a mediator or moderator is less likely to help ensure normality. In such a scenario, you could simply opt for non-parametric tests.
  • asked a question related to SPSS
Question
2 answers
Hello everyone,
I am conducting analysis on my data and have mostly already figured it out. However, there is still one problem I haven't been able to master.
Statistical program: SPSS 29.0
Cross-over study design with NINE (9) subjects and TWO (2) treatments:
Subjects were given treatment 1 or 2 on two separate study dates. At the end of the study all subjects had received both treatments.
After administering the treatments, the patients were monitored for blood parameter changes on NINE (9) separate measuring time points.
What I aim to do, is to compare the two different treatments and have used RM analysis to do so. Initially I defined Treatment(2) and Time(9) as Within-subject-factors and have been able to gain an answer to most of my questions.
But what I haven't understood yet is how can I compare individual measured time points (for example treatment 1 blood parameter at 30 mins compared to treatment 2 blood parameter value at 30 min) between the treatments through the RM ANOVA - to my knowledge and understanding, the output does not provide this and I haven't been able to figure out how to get it out of the analysis? Or can I? Or do I have to go about it with a different analysis completely?
Thank you!
Best Regards, Isa
Relevant answer
Answer
Hello Isa,
I suspect you could elicit specific tests (e.g., treat 1 vs. treat 2 for the respective 30-min time point) in whatever software you're using. However, you could always just run (as many as 9) dependent t-tests to do the same thing, with some appropriate adjustment for multiple tests.
The sample size (9 paired score sets) won't offer a lot of statistical power, however, unless the treatment differences are pretty strong, no matter what kind of analysis you elect to use.
Good luck with your work.
  • asked a question related to SPSS
Question
1 answer
Hello Experts,
My outcome is a scale variable (3 separate domains-continuous)
Covariates are mixed (categorical and continuous)
I am developing a 3 level model (care aides, unit and facility), unfortunately i found a warning message (iteration was terminated but convergence has not been achieved. the mixed procedure continues despite this warning. subsequent results produced are based on the last iteration. validity of the model fit is uncertain). In this scenario, R2 and ICC was very poor.
Now, I moved to 2 level model (care aides, Unit). here the same problem, I can not put my important covariates in this model. If i exclude some of my important covariates, there are no warning message.
Same models works in STATA but SPSS showed warning message.
Q1> Is that the limitations for SPSS?
Q2> Is there any options that If I can change, the problem will be resolved?
Relevant answer
Answer
If you are using the default settings in both packages, the SPSS MIXED command is using restricted maximum likelihood (REML), whereas Stata's -mixed- command is using maximum likelihood estimation (option ml). That might make a difference in what you are seeing. HTH.
  • asked a question related to SPSS
Question
5 answers
If iI use EFA on SPSS to explore the factors and later I intend to check the reliability and validity of these explored factors using SmartPLS rather than CFA on AMOS, will it be valid approach?
Relevant answer
Answer
Alamir Louro, thank you for responding to my request - and for having done so with so much detail. You are certainly right about SPSS being limited by providing the default of eigenvalues > 1 criterion for suggesting - and producing - the number of factors in a data set. At least, researchers can over-ride that and specify the number of factors they want to be extracted, perhaps on the basis of parallel analysis, scree plot, etc. However, it would be nice if a package produced parallel analysis as a matter of course.
It was generous of you to provide your email address. I might get in touch with you some day to have a bit of "joint experience" with factor analysis.
Thanks again.
  • asked a question related to SPSS
Question
3 answers
Hello,
I want to perform a single mediation (model 4, hayes) with Process in SPSS.
Before I can do that, I have to check the assumption for linearity. I created a matrix scatter-plot, but because my independent variable is binary it doesn't look right. I have 1 dependent variable, which is metric.
Does anyone know, how I can check for the linearity assumption, if my independent variable is binary and single level (Intervention group vs. control group)?
Relevant answer
Answer
There is no need to check linearity for a binary independent variable. You can always draw a straight line through two points (the two group means on your binary variable). More relevant would be homoskedasticity (the assumption of equal variances/error variances in the two groups in the population).
  • asked a question related to SPSS
Question
5 answers
I am using SPSS to perform binary logistic regression. One of the parameters generated is the prediction probability. Is there a simple mathematical formula that could be used to calculate it manually? e.g. based on the B values generated for each variable in model?
Relevant answer
Answer
People have certainly done that, Nasir Al-Allawi. A Google search on <logistic regression scoring system> turns up lots of resources. Good luck with your work.
  • asked a question related to SPSS
Question
2 answers
Since I found out that there is a correlation between Timeliness and Semantic Accuracy (I'm studying linked data quality dimensions assessment, trying to evaluate a dimension quality -in this case Timeliness- from another dimension (Semantic Accuracy)), I presumed that regression analysis is the next step in this matter.
-the Semantic accuracy formula I used is: msemTriple = |G ∧ S| / |G|
msemTriple measures the extent to which the triples in the repository G (the original LOD dataset) and in the gold standard S have the same values.
-the Timeliness formula I used is:
Timeliness((de)) = 1-max⁡{1-Currency/Volatility,0}
where :
Currency((de)) = (1-(lastmodificationTime(de )-lastmodificationTime(pe ))/(currentTime-startTime))*Ratio (the Ratio measures the extent to which the triples in the the LOD dataset (in my case wikidata) and in the gold standard (wikipedia) have the same values.)
and
Volatility((de)) = (ExpiryTime(de )-InputTime(de ))/(ExpiryTime(pe )-InputTime(pe ) )
(de is the entity document of the datum in the linked data dataset and pe is the correspondent entity document in the gold standard).
NB: I worked on Covid-19 statistics per country as a dataset sample, precisely Number of cases, recoveries and deaths
Relevant answer
Answer
  • asked a question related to SPSS
Question
3 answers
I choose Model 1 to test 3 iv,1 dv and 1 moderator. I want to know the benefits/advantages of using Model 1. Can anyone help?
Relevant answer
Answer
I still dont understand your question. It doesnt make sense to me. You choose your model, which fits to your hypotheses and your data generating process. If you have only 1 moderator and all assumptions are met, model 1 is fine, but if you have a mediation hypothesis this wont make any sense for example. Thefore, model 1 is not better than others.
  • asked a question related to SPSS
Question
3 answers
Aloha Everyone,
My university recently bought SPSS 29 Faculty Premium Package, the vendor is thinkEDU, the said AMOS is part of it, that is one of the reason we bought, but there is no AMOS inside it, i could not locate? How to confirm if it has or not, if it has where can I find and use AMOS? Please advice.
Thank you
Rojan Baniya
Relevant answer
Answer
I looked at the SPSS website and found this: "Amos is included in the Premium edition of SPSS Statistics (except in Campus Edition, where it is sold separately)."
It sounds like you will need to contact SPSS support to get a definitive answer.
  • asked a question related to SPSS
Question
10 answers
These are few questions for your reference,
How much did you learn about managing your money from your parents?
· None
· Hardly at all
· Little
· Some
A lot
How often were you influenced by or did you discuss about finances with your parents?
· Never
· Once a year
· Every few months
· Twice a month
Weekly
What is your current investment amount in stocks/shares? (Portfolio value)
· 1 - 90,000
· 90,000–170,000
· 170,000–260,000
· 260,000–340,000
· More than 340,000
The above questions are allocated weights from 1 to 5.
Relevant answer
Answer
thanks your answer You're right, but the questions asked were only selected as rankings, there is no degree. Cannot be weighted
  • asked a question related to SPSS
Question
8 answers
I have four multivariate linear regression models which differ in the level of data aggregation. Now, I would like to compare them based on the AIC and BIC. For this, I need the log-likelihood as a measure of the model fit. Can I get that value for each model through SPSS, and if yes, how?
Thanks for your help and best regards,
Isabel
Relevant answer
Answer
Sorry for being late to the thread and not being an SPSS user, but is the problem that the log-likelihood is not produced, or something that could easily produce it?
  • asked a question related to SPSS
Question
6 answers
We are recoding our data and transforming the data values into the values we won't according to the questionaries we have used and its scores.
The first couples of recodes went fine, and we can see the values in the Data view after we ran the syntax. But then all of a sudden after we ran a recode of the next variable in our syntax, the value did not show in the Data view, and instead of the value, the columns just have a dot; .
We don't have any missing values, and the original data, we are recoding, has values. So what are we doing wrong, can any body help us?
We have been checking the syntax over and over again, to see if we are doing something diffrent form the first couples of recodes, where nothing is wrong.. but we can't find any differences.
Relevant answer
Answer
I tried to recode a string to a numeric variable twice, once by typing in the syntax and then by using the drop-down menu. Neither one worked. The hand-entered syntax kept generating different errors even though it's identical to the drop-down version. The drop-down version ran correctly (no error messages), but the string variable didn't get recoded.
  • asked a question related to SPSS
Question
4 answers
Hi folks,
i want to survey employee in three waves and need to connect the three timepoints to the individual. Furthermore I need to survey one leader of every employee at each timepoint for causal inferences.
Do you have any suggestions for software which is easy to use for this purpose?
Thank you
Relevant answer
Answer
Its very simple, cheap and works very well.
  • asked a question related to SPSS
Question
3 answers
Hi all, please assist me on what statistical analysis I should use. I have 3 DVs (severity of violence, justification of violence and severity of punishment) and 1 IV (gender of perpetrator). I have trouble running my data on SPSS as I cannot select the DVs on it.
Relevant answer
Answer
I have several questions.
  1. How are your DVs measured? Are they quantitative variables that can fairly be described using means and SDs? Or are they ordinal variables for which means & SDs would not be very defensible?
  2. How many gender categories have you got, and what are the sample sizes for each category?
  3. Do you have one multivariate research question or 3 univariate research questions?
With 3 DVs, you may be considering MANOVA. My question 3 above is a question that Huberty & Morris (1989) raised in their classic article on MANOVA:
If you are considering MANOVA, see also this article:
Another option to consider, particularly if your DVs are not really quantitative variables (for which means & SDs are defensible) would be to turn things on their head by estimating a logistic regression model with gender as the outcome and with your 3 DVs as the explanatory variables. Frank Harrell described this approach in his well-known textbook on regression modeling: See the excerpt in the attached PDF. But note that it requires having a sufficient number of observations in each gender category. By the 20:1 rule of thumb Harrell describes in the Datamethods Author Checklist (link below), one would need at least 60 observations in each gender category with 3 explanatory variables in the model.
HTH.
  • asked a question related to SPSS
Question
3 answers
Is it possible to perform EFA by SPSS Amos?
Relevant answer
Answer
It is possible to run an EFA in Amos (see https://www.ibm.com/support/pages/exploratory-factor-analysis-efa-amos#:~:text=You%20can%20do%20that%20in,with%20a%20few%20parameter%20changes.) but it seems to me that it would be a lot easier to simply use SPSS for EFA.
  • asked a question related to SPSS
Question
9 answers
Does anyone know how to run Mediation analysis in SPSS version 25 software????
My research is a case-control study, the exposure being: dietary pattern; outcome: gastric cancer; mediator: energy (kcal).
Relevant answer
Answer
Thanks for clarifying. Why do you think you need to carve g/day into tertiles? Carving quantitative variables into categories can cause many problems. See this page for more info:
  • asked a question related to SPSS
Question
2 answers
I am currently conducting research into the dark triad and attitudes to cheating and was looking for some guidance in regards to the analysis. I want to analyse the Dark Triad traits as separate conducts as well as having an overall impact of the Dark Triad on attitudes to cheating, how would I go around actually making this possible on SPSS?
Thank you
Relevant answer
Answer
How about using multiple regression with the Dark Triad traits as predictor (independent) variables and attitude as the dependent variable?
  • asked a question related to SPSS
Question
2 answers
Hi! I weighted 1 case and then performed Compare Mean with other groups of variables in spss, but the N (number of cases) is increased, which is greater than the unweighted compare mean's N and the numbers are over the total cases we have, why this happened, and how can I figure it out? thank you.
Relevant answer
Answer
Hello Yy,
In SPSS, using the weight cases command causes the program to treat a case as being replicated k times, where k is the value of the weighting variable chosen. So, it is as if your data set has been changed by adding (k - 1) additional instances of a given case, each having the same data points as the original case.
So, yes, the apparent N can exceed the number of non-missing cases (usually, rows) in the data set when you use the weight cases option.
Good luck with your work.
  • asked a question related to SPSS
Question
6 answers
In plant breeding, what are uses discrimination function.
Relevant answer
Answer
Discriminant function technique involves the development of selection criteria on a combination of various characters and aids the breeder in indirect selection for genetic improvement in yield. In plant breeding, the selection index refers to a linear combination of characters associated with yield.
  • asked a question related to SPSS
Question
5 answers
Dear community,
I spend the last days reading many questions and responses as well as several journal articles and research papers on the topic of time series analysis and the individual steps. Unfortunately, the more I read the more confused I became...now I hope that you can help me out.
I am writing my master thesis about FDI determinants in Ethiopia. I have selected FDI per GDP as dependent variable and several independent variables (exchange rate change, inflation rate, labor force participation rate, telephone subscriptions per 100 citizens, a governance index (ranking from 0 to 100), and exports + imports as ratio of GDP). I am also thinking about adding a dummy variable to reflect the effects of the Growth and Transformation Plan in 2010.
The next step is to analyse the data empirically in order to investigate the impact of the explanatory variables on the FDI inflow. Due to data availability I only cover a period of 20 years (1996-2016). I read several papers on this topic but somehow everyone performed/applied different steps/tests/models. Also the order of the test performed varies in the papers.
I am really confused by now with regards to the differences between OLS, ECM, ARDL, and VAR and what is the most appropriate method in my case.
In addition, authors performed (some didn't) different tests for unit root/stationary, for co-integration, for multicollinearity, for autocorrelation, for heteroskedasticity. Also in a different order.
Besides, I am confused about the lag selection process and the meaning/difference of AR(1) and I(1).
Moreover, many authors transformed the variables first with log. I cannot do that as I have to negative observations (inflation rate).
Earlier I also read something about trend and difference stationary and depending on this different unit root test.
Like I said, I am just so confused by now that I don't even know who and where to start.
I am working mainly with SPSS but will perform the unit root tests in Eviews as SPSS do not have this function.
I really hope that you can help me by providing a little guideline on what I need to do and in which order. Thank you so much!
Relevant answer
Answer
Here are some steps you can follow to perform time series analysis for a period of 20 years:
  1. Collect the data and clean it.
  2. Prepare visualization with respect to time vs. key feature.
  3. Observe the stationarity of the series.
  4. Develop charts to understand its nature.
  5. Build the model – AR, MA, ARMA, and ARIMA.
  6. Evaluate the model using statistical measures such as RMSE, MAPE, etc.
  • asked a question related to SPSS
Question
4 answers
After running the McDonald Omega in SPSS 28, I regularly get this statement "Omega cannot be estimated due to negative or zero item covariances. This may be due to items needing to be reverse scored, or to violations of model assumption." Because of the violation assumption, we went from Alpha to Omega. Kindly help!!
Relevant answer
Answer
It could simply be that some of the items are bad. If their intercorrelations are low or negative, something must be wrong with either your data or the items. Perhaps some of the items are unreliable or invalid; or the items measure multiple dimensions (factors). Have you examined the factor structure using EFA or CFA?
  • asked a question related to SPSS
Question
8 answers
There are 2 groups: one has an average of 305.15 and standard deviation of 241.83 while the second group has an average of 198.1 and a standard deviation of 98.1. Given the large standard deviation of the first group, its mean should not be significantly different from the second group. But when I conducted the independent sample t-test, it was which doesn't make sense. Is there any another test I can conduct to analyze the date (quantitative)?
The data is about solid waste generation on a monthly basis (averages). And I am comparing March and April data with that of other months. Also, the sample sizes are not equal i.e. less days in Feb as compared December for example.
Relevant answer
Answer
It is difficult to give a comprehensive answer without more details on 1) the data, 2) the experimental design, 3) your question and 4) what exactly you did to perform the Student's T-test.
Note however that usual Student's T-test assumes equal theoretical variances between the two populations, which is probably wrong considering your two values. You should use corrected T-tests for this case, like the Aspin-Welch test (default in R / t.test).
Note also that the T-test compares the means, that can be very precise even if the sample SD is very high if your sample size is big enough, so you cannot conclude just « by eye » based on means and sample SD. You should look at sample SEM to do that and better understand what's going on.
Never forget that T-test assumes Gaussian distributions, which may well be wrong also in your samples, and this all the more matter, than sample sizes are small and different (just like the hypothesis of variance equality is more important for unequal sized samples).
Last, don't confuse « statically significant » and « of practical interest », results of your test should be interpreted along with effect size / confidence intervals for the difference and so on.
  • asked a question related to SPSS
Question
3 answers
Can I do regression analysis for such kind of variables?
Relevant answer
Answer
In any case a Spearman correlation coefficient tells you the basic information
  • asked a question related to SPSS
Question
5 answers
I have been performed the multinomial logistic regression model in SPSS. The goodness of fit shows that the model fits the data. Based on my literature study, there are several methods can be performed to validate the model, but SPSS 23's window of performing Logistic Regression doesn't show the options. Kindly help me to inform that what particular method I can use to validate the model in SPSS.
Relevant answer
Answer
Hello Tanmoy Basu In a multinomial logistic regression model that you do using SPSS software, there are many tools and outputs to check the appropriateness of the model, I have sent some of them in the attached images. As far as I know, SPSS is not lacking in this field and it is a good software.
  • asked a question related to SPSS
Question
6 answers
I'm writing my master's thesis on the investigating of delay factors on building construction projects. My independent variables (8) is the delay factors (materials-related, manpower-related, equipment-related, project management-related, consultant-related, owner-related, governmental-related and external-related). The dependent variables is delay in construction for frequency in occurence and degree of severity. which statistical test will be appropriate to apply in SPSS?
Null Hypothesis (H0): There isn’t a significant statistical differences in frequency and severity between causes of delays caused by delay group.
Alternative Hypothesis (H1): There is significant statistical differences in frequency and severity between causes of delays caused by delay group.
Is One-Way ANOVA is applicable for degree of severity and frequency of occurence?
Relevant answer
Answer
Before getting into possible statistical procedures (and Bruce Weaver 's point about what the variables are and if there are two DVs), I have some questions about the null hypothesis. "Null Hypothesis (H0): There isn’t a significant statistical differences in frequency and severity between causes of delays caused by delay group." First, the null is usually NOT whether there is a significant difference, but that there is absolutely no difference in the population. Can you verify this is what you mean. Further, please clarify what you mean "between causes of delays caused by ..." Is your null that in the population there are no differences in both the frequency and the severity variables? And frequency is presumably a count variable, but you also talk about delay which is presumably a time variable. As Bruce says, other aspects (e.g., are these 8 predictor variables and how so they relate) could also be made clearer if you wish to useful feedback.
  • asked a question related to SPSS
Question
4 answers
I am using SEM for my studies. My questionnaire involve several adapted questionnaire and one newly established questionnaire. I am sure to run EFA after Pilot study. Do I still run CFA using SPSS or is it already within the SMART PLS I will utilise later?
Relevant answer
Answer
Hi Gira,
the most important thing ist to consider whethe a factor model (which the CFA applies) is at least theoretically plausible. "Scales" are often heterogeneous sets of measures that "tap" the "domain" of a (multidimensional) construct (note my marks), not reflective indicators of ONE and the same underyling dimensions/latent variable.
In this thread, you will find more:
In addition, as you started with some conceptional ideas, I would skip the EFA and go direct to CFA (if this is reasonable).
Best,
Holger
  • asked a question related to SPSS
Question
5 answers
I run a multinomial logistic regression. In the SPSS output, under the table "Parameter Estimates", there is a message "Floating point overflow occurred while computing this statistic. Its value is therefore set to system missing." How should I deal with this problem? Thank you.
Relevant answer
Answer
Hello Atikhom,
Which particular statistic in the output was omitted? Which version of spss are you using?
Are you willing to post your data and proposed model, so that others could attempt to recreate the condition?
Some obvious points to consider about your data set whenever unexpected results such as the one you report occur:
1. Are your data for any continuous variables suitably scaled? (so that the leading significant digit is not many orders of magnitude away from the decimal point)
2. For your categorical variables, do you have cases in each possible cell/level? (check frequencies for each such variable)
3. Do you have any instances of perfect relationships among categorical variables (perhaps due to empty cells)? (check cross-tabulations for variable pairs and sets)
4. Is one of the IVs redundant with another variable in the data set?
5. Do you have missing data (and are attempting some sort of imputation process)?
That may not cover the waterfront, but at least it may give you some ideas when checking your data.
Good luck with your work.
  • asked a question related to SPSS
Question
5 answers
I would like to know if the mean score on financial knowledge of graduate students is greater than the mean score of subgraduate students. I already did an independent sample T-test in SPSS, but I'm not sure if I should proceed with the analysis of the results because the sample of subgraduate students appears not be normally distributed.
The p-value of both normality tests (Kolmogorov and Shapiro) for subgraduate students are less than the level of significance of 0.05. So, there is not a normal distribution for subgraduate students.
I have to mention that the sample size for subgraduate students is 127 and the sample size for graduate students is 29. For graduate students the Kolmogorov-Smirnov and Shapiro-Wilk p-value is .200 and .180, respectively (greater than the 0.05 level of significance), so for graduate students (althoug she sample size is less than subgraduate students) the assumption of normality is met.
What do you recommend? Should I select another test?
Thanks in advance.
Relevant answer
Answer
If one of your independent samples for a t-test is not normally distributed, you can use the Mann-Whitney U test instead of a t-test. The Mann-Whitney U test is a nonparametric test that does not require normality
  • asked a question related to SPSS
Question
1 answer
can a DMRT( Duncan's Multiple Range Test) result in 11 columns for which i get compact lettering till 'k' (starting from 'a')?
Software used : SPSS v25.0
After i did the post-Hoc with Duncan, i got an output for a specific dependent factor with 18 treatments (3 transcripts of each) as 11 columns.
To put this DMRT result as compact letter display in my table, i'm getting letters 'a' to 'k' for the respective columns.
Is this possible/correct?
significance values for few columns exceeds 0.05. and few are below 0.05.
Relevant answer
Answer
Hello Ushnish,
If your intention is to show a summary of all possible pairwise comparisons among 18 conditions (= 153 comparisons), that will be a very full tabular display no matter what you do!
If only a few of the comparisons meet your criterion for statistical significance, you might restrict your display and summary to just those comparisons, and note that all others were nonsignificant.
One relevant point is that the Duncan test is likely not the best choice for several technical reasons. Consider something along the lines of the Benjamini-Hochberg method or the Ryan-Einot-Gabriel-Welsh method.
Finally, to address your question about letters, if the Duncan output showed 11 "homogeneous subsets," then yes, you may attach letters a-k as appropriate to show instances of nonsignificant differences (subject to the implied concern above about density of your display).
Good luck with your work.
  • asked a question related to SPSS
Question
3 answers
I conducted an Exploratory Factor Analysis using Principal Axis Factoring with Promax rotation, resulting in the identification of 8 factors. Now, I aim to examine potential associations between these extracted factor scores (treated as continuous variables) and other variables in my research. However, I am uncertain about the most appropriate statistical approach to compute these factor scores.
Should I compute the average sum of each factor, employ regression, or utilize the Bartlett method, Anderson-Rubin, any others….?
Any insights or alternative approaches are highly appreciated.
Thank you!
Relevant answer
Answer
Hello Christian-Joseph,
Here's a link to a very readable introduction to some of the considerations associated with different alternatives to the computation of factor scores:
Good luck with your work.
  • asked a question related to SPSS
Question
4 answers
I am working on new research to identify the reliability of the different sets of questionnaires corresponding to different learning styles. I collected and cleaned the data and loaded it into the SPSS software. I have data for around 600 learners from different backgrounds of learning. For questionnaire set 1, I have only 38 variables that measure the questions' uncertainty and use values 0 for uncertain, 1 for false, and 2 for true. I am not getting the results I want. The total factors should be 5 but I am getting around 14 whose eigen values are 1 or greater than 1 in the software. What am I doing wrong here? Please help.
Relevant answer
Answer
Hello Ankita Jiyani Mangtani. Like other respondents, I am not really sure what you want to do. But I can focus on one aspect of what you wrote:
The total factors should be 5 but I am getting around 14 whose eigen values are 1 or greater than 1 in the software.
The eigenvalues > 1 rule of thumb is not recommended these days, as it tends to produce models with too many factors. See the following article for a nice overview of that issue and several others that may be relevant for you.
HTH.
  • asked a question related to SPSS
Question
1 answer
I am using SPSS 23 version on Windows 10. Recently it mentioned a pop-up window at launching: Serialization scheme was not recognized inet: Local computer 0.
How can I fix this problem?
  • asked a question related to SPSS
Question
3 answers
I currently running SPSS and PLS to test my study hypothesis. the case that there are two variables that are associated it (theoretically) and they are independent variable. Thus, i need to combine that two independent variables to become one independent variable. Furthermore, that one independent variable would like to test the relationship with one dependent variable.
Please enlighten me
Relevant answer
Answer
As Jokin stated, you cannot combine two independent variables to create one independent variable. If it is theoretically related, you can create a latent variable of these two independents variables and use that latent score as one variable. Ideally, you need three independent variables for creating one latent variable, but two would be sufficient for testing your hypothesis.
  • asked a question related to SPSS
Question
2 answers
Relevant answer
Answer
As of my knowledge, SPSS and Minitab did not have direct integrations with pandas, a Python library. However, both SPSS and Minitab provide various interfaces and methods to interact with Python.
IBM's SPSS software has a Python plug-in that allows you to execute Python scripts directly within the SPSS interface. This means you can write a script that uses pandas to manipulate your data, then use the SPSS Python integration to execute that script within SPSS.
Minitab, on the other hand, does not have as direct a connection to Python as SPSS. However, you could certainly write a Python script that uses pandas to prepare your data, then import the result into Minitab for further analysis.
Both SPSS and Minitab are primarily used for statistical analysis rather than machine learning, although they do have some machine learning capabilities. Python libraries like scikit-learn and TensorFlow are more commonly used for machine learning, and pandas is often used to prepare the data for these libraries.
This status might have changed recently, so I recommend checking the latest documentations or official websites of these software tools for the most recent updates.
  • asked a question related to SPSS
Question
3 answers
Hello everyone :)
For my study, I would like to conduct a factor analysis followed by a reliability test for one of my concepts "Attitude towards unemployed" measured according to 5 items with a Likert scale as answers. I've conducted my factor analysis in SPSS and 2 factors were extracted: 3 items loaded well on factor 1, while the other 2 loaded well with factor 2.
Since I want to measure the reliability of my scale, should I do it for each factor and their items or should I do it with all of the 5 items? Indeed, I don't really understand why I ended up with 2 factors since my items are measuring the same concept... And I was intended to establish Cronbach’s α for the 5-item scale.
Also, the 2 items that loaded well with my factor 2 were reversely coded before the analysis.
Thank you in advance for your answers and have a nice day :)
Manon.
Relevant answer
Answer
Several pubs on reverse scoring (e.g., Woods, C. M. (2006) ‘Careless responding to reverse-worded items: Implications for confirmatory factory analysis’, Journal of Psychopathology and Behavioral Assessment, 28(3), 189-194. Available at: https://doi.org/10.1007/s10862-005-9004-7). The biggest issue comes when people just add negations to otherwise positive statements. This increases RTs and lowers reliability (e.g., ). Nowadays we usually have reverse and non-reverse scores as two highly correlated factors in a CFA. Just there are other approaches.
  • asked a question related to SPSS
Question
4 answers
I am using 1 Iv and 4 DVs, each of the variable has five questions with five Likert scale. What method of test should I use to analyze my data?
Relevant answer
Answer
What are the 4 DVs? Do you have one multivariate question, or 4 univariate questions? I have Huberty & Morris (1989) in mind as I ask that second question. See also Huang (2020).
  • asked a question related to SPSS
Question
5 answers
We need a new program for statistical analysis that provides results different from what is provided by the usual programs in the analysis
Relevant answer
Answer
Hello Mohammed,
You can try Python and R.
Best wishes.
  • asked a question related to SPSS
Question
10 answers
I’ve got a data set and I want to calculate R2 for linear regression equation from another study.
For example, I have derived an equation from my data (with R2) and I want to test how other equations perform on my data (and thus calculate R2 for them). Then, I want to compare R2 from my data set with R2 from derivation studies.
Do you have any software for this? Any common statistical software could cope with this task (e.g. SPSS or SAS)? Maybe you have any tutorials on YouTube for this?
Relevant answer
Answer
Hello Pavel Makovsky. In that case, modify the code I posted earlier. With your dataset open and active:
REGRESSION
/STATISTICS COEFF OUTS CI(95) R ANOVA
/DEPENDENT YourDV
/METHOD=ENTER your explanatory variables
/SAVE PRED(yhat1).
* Now use another set of coefficients to generate yhat2.
COMPUTE yhat2 = use the other regression coefficients here but with the variables in your dataset.
REGRESSION
/STATISTICS R
/DEPENDENT YourDV
/METHOD=ENTER yhat1.
REGRESSION
/STATISTICS R
/DEPENDENT YourDV
/METHOD=ENTER yhat2.
  • asked a question related to SPSS
Question
12 answers
Hello!
I'm currently working on the LDQ detection for my thesis. I have read so many things about missing values and how to deal with them that unfortunately I am now even more confused than ever before.
I am working with SPSS 26 and I have several categorical as well as continuous variables which are important for my research question.
Initially I followed the procedure Analyze menu > Missing Value Analysis
The range of my n is between 145 and 158 (so not a really huge sample size). In the univariate statistics I get percent missing around 7.6% to 8.2%.
I ran Little's MCAR test which became non-significant with Chi²(15) = 1.390 and p = 1.000
I made a mistake in the construction of my questionnaire because participants were not able to skip questions - looking at the data set most of my missing values is simply because after a certain amount of time people simple dropped out and every question that would have followed was considered "missing". The "initial" sample size without the drop outs was around 350 participants.
I ready that missing data analysis would be important especially to look what type of missing it is, that you're dealing with in your data set.
To me (or my really limited understanding of those things compared to experts) it makes complete sense that I have higher amounts than the recommended 5% of missings where some sort of imputation or replacement of the missing values should be considered because mine are mostly due to drop outs. Also the p = 1.000 at the Little's test is highly irritating for me.
I have read a ton of articles the past few days but even though they all give really good explanations for missing data and multiple imputation and which method is best - I found no answer what I should do in my specific case where the missing data is because people stopped at one point to answer the questionnaire and closed it (I already stated in my limitations that preventing them from skipping questions should be handled in a different way in the future).
Can someone please help me out and give me a recommendation?
I found an article from D. A. Bennett (1999) that is called "How can I deal with missing data in my study?" and gives a cut off of 10% of missing values but gives no information if it is considered for each variable or the overall data set and I found no way to calculate the missing values for the complete data set instead of "just" columns or rows.
I hope I somehow was able to explain my issue with that missing values due to drop out and my confusion how to handle them and especially how to argue in my thesis WHY I handled them the way I did.
I really hope that some expert finds time to give me a recommendation - I'm happy to answer any further questions. I'm simply overwhelmed atm with the amount of information I read and suddenly nothing makes sense anymore.
Thanks everyone in advance!
Relevant answer
Answer
I have not yet had time to read that Enders (2023) article in its entirety. But after reading his explanation of the MCAR, MAR, and MNAR missingness mechanisms, I cobbled together this slide, which some readers may find helpful.
  • asked a question related to SPSS
Question
3 answers
What is endogeneity and how can we run it in SPSS/AMOS? what is the step-by-step procedure, particularly survey data/cross-sectional research?
Relevant answer
Answer
Endogeneity refers to a situation in which a variable in a statistical model is correlated with the error term or disturbance term. This correlation can lead to biased and inconsistent parameter estimates, making it challenging to establish causal relationships between variables. Endogeneity often arises in observational or cross-sectional research designs, where the researcher cannot control for all possible confounding factors.
To address endogeneity in your analysis using survey data in SPSS/AMOS, you can consider implementing the two-stage least squares (2SLS) method. This method is commonly used in econometrics to address endogeneity by using instrumental variables. Here is a step-by-step procedure:
  1. Identify potential endogenous variables: Identify the variables in your model that you suspect may be endogenous. These are the variables that are potentially correlated with the error term.
  2. Find suitable instrumental variables: Instrumental variables are variables that are correlated with the endogenous variable but not directly correlated with the error term. Look for variables that satisfy this condition and can be used as instruments.
  3. Run the first-stage regression: In SPSS, run a regression model where you predict the endogenous variable(s) using the instrumental variables. Obtain the predicted values (fitted values) from this regression.
  4. Run the second-stage regression: Use the predicted values obtained from the first-stage regression as a new independent variable in your main regression model. This is done to address the endogeneity issue. Run your main regression model using the predicted values as a control variable.
  5. Interpret the results: Examine the coefficient estimates and their significance in the second-stage regression. These estimates represent the effects of the exogenous variables on the dependent variable, taking into account the endogeneity issue.
In AMOS, which is a software primarily used for structural equation modeling (SEM), you can also address endogeneity using a similar two-stage approach. You would specify a measurement model for your latent variables and a structural model for the relationships between the latent variables. You can include instrumental variables as exogenous variables in your structural model to account for endogeneity.
It's important to note that implementing the 2SLS method or addressing endogeneity, in general, requires careful consideration of the specific research context, availability of instrumental variables, and the assumptions underlying the method. It is recommended to consult with a statistician or econometrician who can guide you through the process and ensure its appropriate application to your specific research design.
  • asked a question related to SPSS
Question
2 answers
The variable physical environment effect, is only a subset of the independent variable ( environmental factors) in my research, there are social and cultural environment effects as well. They are measured in my questionnaire with five questions and the responses are; ( never, rarely, often and always). The dependent variable, student performance, was also measured in the same format as the environmental factors(i.e with five questions and Never, rarely...being the responses). I have coded them into SPSS with the measure; Ordinal. I want to answer the research question; 1. How physical environment affect student performance? 2. How social environment affect student performance? 3. To what extent does cultural environment influence student performance? I've computed the composite score(mean) for the questions, can I use these scores in the ordinal regression analysis? Or is there any other way to compute the questions into a single variable, for both the independent and dependent variables?
Relevant answer
Answer
In your study, where you have measured the effects of physical, social, and cultural environments on student performance using ordinal scales, you can use ordinal regression analysis to answer your research questions.
To conduct the ordinal regression analysis, you do not necessarily need to compute a composite score for the questions. Instead, you can use each question as a separate predictor in the analysis. Each question represents a different aspect of the environment (physical, social, or cultural), and using them as separate predictors allow you to assess the unique contribution of each aspect to student performance.
Here's a general outline of the steps you can follow:
  1. Prepare your data: Ensure that your data is properly coded and formatted in SPSS. Make sure the independent and dependent variables are coded as ordinal variables.
  2. Run ordinal regression analysis: In SPSS, go to "Analyze" -> "Regression" -> "Ordinal..." and select the dependent variable (student performance) and the independent variables (physical environment, social environment, and cultural environment). Specify the appropriate link function (e.g., logit or probit) based on the distributional assumptions of your data.
  3. Interpret the results: Examine the coefficient estimates, their significance levels, and the odds ratios associated with each independent variable. These results can provide insights into the effects of physical, social, and cultural environments on student performance.
By using each question as a separate predictor in the ordinal regression analysis, you retain the specificity and granularity of the different aspects of the environment being studied. This approach allows you to explore the individual effects of physical, social, and cultural environments on student performance.
Keep in mind that ordinal regression assumes proportional odds, which means it assumes that the relationship between the predictors and the outcome is consistent across different levels of the outcome variable. It's important to assess this assumption, for example, by conducting tests of parallel lines.
Additionally, be cautious when interpreting the results as causal relationships. While ordinal regression can help identify associations between variables, establishing causal relationships requires rigorous experimental designs or the consideration of other potential confounding factors.
Overall, by using ordinal regression analysis and examining the effects of different environmental factors on student performance separately, you can address your research questions and gain insights into the influence of physical, social, and cultural environments on student performance.
  • asked a question related to SPSS
Question
3 answers
Hi, I'm doing CFA using SPSS Amos and I have two variables with single items each. After reviewing previous discussions, I've fix the loadings and error variance to 1,1. If I put error variance to 0, the variances will be zero. However, the variance is not that important to me as I would like to find the factor loadings between the single item and variable but I'm not sure how to do it.
Or the factor loading for single item is the value between e25 --> single item?
If so, the value between e26 --> Pay_R is more than 1. How should I interpret this?
I'm not familiar with SPSS Amos and would appreciate all guidance on this. Thank you.
Relevant answer
Answer
In SEM, a single indicator latent variable is a latent variable with only one indicator. The factor loading for a single indicator latent variable is equal to the square root of the variance of the indicator. You can find more information on SEM in Helmut Lütkepohl and Markus Krätzig’s book “Structural Vector Autoregressive Modeling and Impulse Responses”.
  • asked a question related to SPSS
Question
14 answers
I am running 6 separate binomial logistic regression models with all dependent variables having two categories, either 'no' which is coded as 0 and 'yes' coded as 1.
4/6 models are running fine, however, 2 of them have this error message.
I am not sure what is going wrong as each dependent variable have the same 2 values on the cases being processed, either 0 or 1.
Any suggestions what to do?
Relevant answer
Answer
Sure Bruce Weaver I was running logistic regression as a part of the propensity score matching technique. While watching the tutorial video on Youtube (https://youtu.be/2ubNZ9V8WKw) I realized I am including variables in the equation that I shouldn't have. So, I excluded them and magically, the error did not appear anymore. It was that simple.(: Good luck to everyone who's facing this error!
  • asked a question related to SPSS
Question
13 answers
there are many statistics software that researchers usually use in their works. In your opinion, which one is better? which one do you offer to start?
Your opinions and experience can help others in particular younger researchers in selection.
Sincerely
Relevant answer
Answer
SPSS, as my results necessitate analysis using SPSS
  • asked a question related to SPSS
Question
4 answers
When I try to upload data from an Excel file to SPSS as usual, it says that the file contains no data and that the command has been stopped. I have searched the internet for an answer to why this occurs, but I haven't found a solution. Has anyone come across this before and can advise?
Relevant answer
Answer
If you're experiencing an issue where SPSS is unable to recognize data in your Excel file and the command is being stopped, there could be a few potential causes for this problem. Here are some common reasons and solutions to consider:
  1. Compatibility issues: Ensure that the version of SPSS you are using is compatible with the version of Excel you are using. It's possible that newer versions of Excel may have changes in file formats that SPSS doesn't support. Try saving the Excel file in an older format (e.g., .xls instead of .xlsx) and see if SPSS can recognize the data.
  2. Incorrect file format: Verify that you are selecting the correct file format when importing the data into SPSS. Make sure you are choosing the appropriate Excel format option (.xls or .xlsx) that matches your file.
  3. Empty worksheet or wrong sheet selection: Check if the Excel file you're trying to import contains data in the selected worksheet. SPSS may display an error if the selected worksheet is empty. Ensure that the worksheet you're selecting contains the actual data.
  4. Data range selection: Double-check that you are specifying the correct range of cells containing the data in the Excel file. It's important to accurately define the range to import the data properly.
  5. Special characters or formatting: Remove any special characters or formatting from your Excel file, such as merged cells, formulas, or unusual data types. SPSS may struggle with such elements and could fail to recognize the data.
  6. Software or system issues: If none of the above solutions work, there could be an issue with your SPSS installation or your system. Try reinstalling or updating SPSS to the latest version. Additionally, ensure that your system meets the necessary requirements for running SPSS.
If none of these solutions resolve the problem, it might be worth reaching out to SPSS support or consulting with a technical expert who can analyze the specific error message or provide further assistance based on your system and software configuration.
  • asked a question related to SPSS
Question
1 answer
I tried to do the survival analyse in SPSS. After getting the survival table, i do not know how to draw and computing the median survival time from Kaplan-Meier using SPSS Statistics ? and How do I get 1-, 3-, and 5-year survival with survival analysis in SPSS? please help me. thanks you
Relevant answer
Answer
You may find some relevant examples in Chapter 2 here:
You could also try posting your question to the SPSSX-L discussion forum. It is an old-school (listserv) mailing list, but you can also access it via this website:
Good luck.
  • asked a question related to SPSS
Question
3 answers
Hello everyone! I am currently doing moderation/mediation analyses with Hayes Process.
As you can see the model 3 is significant with R2=.48
The independent variables have no sig. direct effect on the dependent variable, but significant interaction effects. The curious thing is: toptMEAN does not correlate with any of the variables, but still fits into the regression model. Should I take this as confirmation that toptMEAN has an effect on the dependent variable even though it does not correlate? Or am I missing something in the interpretation of these results?
(Maybe you could also suggest a different model for me to run. model 3 is currently the one with the highest r2 i found)
Relevant answer
Answer
It is very well possible to have significant interaction effects but no significant "main" effects. It is also possible to have significant "main" effects and no significant interactions. And it is possible that variables have significant zero-order correlations with the dependent variable but no significant regression coefficients (e.g., due to redundancies and/or overfitting of the regression model). Finally, it possible to find non-significant zero-order correlations between IVs and DV but significant regression coefficients (e.g., due to suppression effects).
Many of these effects cannot easily be identified from the zero-order correlations. That's one reason why we run multiple regression analyses--to identify redundancies, interactions, suppression effects, etc. that are not easy to see in a bivariate correlation analysis.
Note also that the "main" (lower-order) effects (and their significance) in the moderated regression model depend on whether you centered the predictors or not. This can make a huge difference for the interpretation. Especially when predictor variables do not have a meaningful zero point, centering prior to calculating the interaction terms is recommended (e.g., Aiken & West, 1991; Cohen et al., 2003). Otherwise, the lower-order terms (and their significance) may not be interpretable at all.
In any case, it would be a good idea for you to plot the effects to get a better understanding of what is going on. That is, look at the regression lines for different values of your moderators to understand the meaning of the interaction effects.
Aiken, L. S., & West, S. G. (1991). Multiple regression: Testing and interpreting interactions. Newbury Park: Sage.
Cohen, J., Cohen, P., West, S. G. & Aiken, L. S. (2003). Applied multiple regression/correlation analysis for the behavioral sciences. Mahwah, NJ: Erlbaum. (Chapters 7 and 9)
  • asked a question related to SPSS
Question
1 answer
Hey there, I‘m using SPSS for the statistical synthesis of my Meta-Analysis and I can only choose log(OR), log(RR) etc as an Effect sizes. Therefore my forest plot is also on a logarithmic scale… I want my forest plot with the measure size RR or OR, not log(RR) and log(OR). Does anybody have experience with this topic and does know how I can change this?
Relevant answer
Answer
Disclaimer: I have not yet tried the (relatively new) meta-analysis commands in SPSS. But I have just taken a quick look at the GUI dialog for it. If you click on Plot, and then check the box for Display exponentiated form, do you get what you're looking for? HTH.
  • asked a question related to SPSS
Question
8 answers
I'm a self-taught SPSS user but need your help at this juncture.
I have 7 elevation (m) data each for 7 hamlets and I want to know which of the elevation value is statistically significant by running a statistical test on SPSS to obtain a p-value.
Relevant answer
Answer
If the values are a random sample from some (statistical) population (with some specific but unknown distribution), I find it legitimate to test the hypothesis that a particular value is from this population. For a "usual" test one calculated a test statistic (which is a simple value), and finds the tail probability of this value from a sampling distribution as the p-value. This is not different here: the "test statistic" is the observed value itself, and the distribution is the population distribution. This is nothing but an outlier test (Grubbs test and alikes work similarily, but are based on the assumtion that the distribution is normal).
To make that clear: I don't see how the question is meaningful for the 7 elevation values. I don't think that this is a random sample and even if it was I don't see how it could be contain "outliers". This would require to either have a (possible) inconsistency in the selection process or a possible difference in data generation process (different geological histories? - and how is this related to the sampling?).
  • asked a question related to SPSS
Question
3 answers
what is the difference between a covariate and a confounder?
Relevant answer
Answer
A covariate is an independent variable of your study. For example, if you want to assess the association of cholera outbreak against eating raw vegetables, eat out in restaurants and drinking untreated water. These are covariates. But a confounder variable is the one which has not been collected during your data collection but could affect the finding of your study either in the positive or negative direction. For example, in the above study if you found that those who ate raw vegetables were more likely to have cholera illness but if you did not collect information on their travel history your finding could have been confounded with their contact with cholera patients during their travel. Eating raw vegetables might not be really the risk but rather their contact history.
  • asked a question related to SPSS
Question
3 answers
Hi everybody,
I have a questionnaire with 4 point Likert scale questions on participant experience. I want to score the results into low/moderate/high experience categories. How do I do this on SPSS?
I have assigned the following values on SPSS.
1-never
2-sometimes
3-frequently
4-always
Appreciate any help. Thank you
Relevant answer
Answer
Why do you want to create a high, medium, low categorization? In general, scale variables are created by summing the individual items, so that you have a continuous variables which is characterized by a mean and standardization. Reducing that continuous variable to a three-category variable would be a serious loss of information.
  • asked a question related to SPSS
Question
13 answers
I was preparing a seminar and accidentally stumbled upon an odd situation. When there is a negative relationship between the predictor (Extraversion) and the dependent variable (Depression measured by BDI-II), the calculated predicted value showed a negative perfect correlation (-1) with the predictor in SPSS. This is a non-sense. So, I checked the same issue using jamovi. The result was the same. In both programs, when I draw the scatterplot between the IV and DV, it clearly showed a strong negative trend. But, when I draw the scatterplot in regression menu in SPSS using the predicted value on X and the DV on Y, the scatterplot showed a strong positive trend.
I guessed it might be deliberate and tested the same with variables that are positively related to the DV. The predicted values and the predictor in this case showed a perfect positive correlation (+1), which makes sense.
Does anyone notice this issue? Or is there any reason why the same issue happened in two programs?
Relevant answer
Answer
The issue you encountered with the calculated predicted value showing a negative perfect correlation with the predictor in both SPSS and jamovi is indeed puzzling. It seems contradictory to have a negative relationship between the predictor (Extraversion) and the dependent variable (Depression), yet observe a positive trend when plotting the predicted values against the DV in the regression menu of SPSS. It's difficult to pinpoint the exact cause of this issue without more information, but it could potentially be attributed to a software bug or a misinterpretation of the data. It would be advisable to double-check your data and analysis procedures to ensure accuracy and consider consulting with experts or the support teams of SPSS and jamovi to address this discrepancy.
  • asked a question related to SPSS
Question
3 answers
I'm carrying out a longitudinal study,
For data analysis we have to collect data several times for each variable and for the same patient
I use SPSS to draw up statistical analysis and I found a difficult how I can enter data for each patient?
any suggestion to resolve this problem please ?
Relevant answer
Answer
To efficiently enter data for each patient in SPSS for your longitudinal study, consider organizing your dataset in a wide format, where each patient is represented by a single row and each variable is recorded in multiple columns corresponding to different time points. This format allows you to easily enter data for each patient across time, maintaining a clear structure for analysis. Alternatively, you can use a long format where each patient has multiple rows, with one row for each observation at different time points. In this case, you would need to assign a unique identifier to each patient and use a variable to indicate the time point of each observation. Both wide and long formats have their advantages, so choose the one that best suits your research goals and analytical needs.
  • asked a question related to SPSS
Question
4 answers
...
Relevant answer
Answer
Hello again Nico,
DId you have your pilot study respondents rate each of the purported utilitarian products on the hedonic dimensions/attributes, and vice versa?
Otherwise, how could you be certain that an out-of-category product wouldn't rate higher than the in-category products? (Even if you thought that such an outcome was unlikely.)
If you must reduce each set, then likely you would opt for utilitarian products which had highest average ratings across the five utilitarian dimensions _and_ (ideally) lowest average ratings across the five hedonic dimensions; the opposite would be the target for hedonic products.
I don't think that PCA would be productive here, since any observed coalition of products might be due to some completely different attribute, and not necessarily the 5 (utilitarian) and 5 (hedonic) you've identified.
Good luck with your work.
  • asked a question related to SPSS
Question
3 answers
I have 414 data set for developing a motivation scale. I used EFA and CFA. So I split the data to two groups. 207 for EFA and 207 left for CFA. AFter I finished the analyze with SPSS for EFA and SPSS AMOS for CFA, I knew that some of questions are not valid in EFA while it is valid in CFA and vice-versa. what should I do with this? this is for my thesis project. should I just report it like that? and when I report it, some questions that are not valid in EFA and CFA at the same time. Which one should I call as valid? Thank you in advance.
Relevant answer
Answer
What do you mean when you say some of the questions are not "valid" in EFA but are in CFA?
  • asked a question related to SPSS
Question
11 answers
In my model the indirect effect is significant and the direct effect is not significant. However, the estimate of the direct effect is much higher than the one of the indirect path. Should I also focus on the interpretation of the estimate and if yes, what does it mean?
Relevant answer
Answer
Since the indirect effect is the product of two or more regression coefficients, it is not impossible or uncommon for it to be smaller than the direct effect yet statistically significant.
  • asked a question related to SPSS
Question
4 answers
Hey there, I‘m currently working on a surivival meta-analysis and I use SPSS for my statistics. The problem is that I have a lot of data regarding mean OS / PFS, but no belonging standard deviations… does anyone have an idea how to solve this problem? I mean I can‘t make a meta-analysis with only the „mean“ data… thank you already very much!
Relevant answer
Answer
Hello Jana von Holzen. I have some questions.
  1. What does "mean OS / PFS" mean?
  2. What is the effect size you wish to compute for each study?
Re question 2, it sounds like it might be a (possibly standardized) mean difference. But you mentioned "survival", which makes me wonder if it might not be a hazard rate ratio.
Thanks for clarifying.
  • asked a question related to SPSS
Question
2 answers
In an earlier version of SPSS I was able to get visual means for each group (see picture below) when creating scatterplots with one categorical and one scaled variable.
Now I have tried for hours to find out in my updated SPSS 26.0 how to do it, but I can't figure it out.
I have tried do use the same syntax as earlier, but that does not create mean-dots, which lead me to think I used some modification. Still can't figure it out?
Any suggestions?
Relevant answer
Answer
hi! i am just seeing this question. yes, i have solution for this using graphs / legacy dialogues. (i teach spss statistics online at university) i see it was a few years ago, so, please reply if you still have this challenge! dean
  • asked a question related to SPSS
Question
4 answers
Hi everyone,
For my studies on the IRR I have binary data and three raters. But on one variable they all score a 0 for all subjects. This means that everything is rated as a 0. But SPSS then gives me the warning that every outcome is the same and that the command stopped. What do I do with it and what do I have to report in my studies?
Hopefully you can help me!
Relevant answer
Answer
I would delete that item from the kappa calculation, and then when you report your coefficient, note that one item was excluded due to 100% agreement.
  • asked a question related to SPSS
Question
4 answers
This is what I know so please let me know if I miss anything:
1. We do bivariate and collect all the p-values
2. Choose p-values <0.25
3. Run logistic analysis for all the variables whose p-value is <0.25
4. Eliminate the variables one-by-one, starting by the highest number of p-value in the new model
5. But when to stop eliminating them all?
And to look for adjusted Odds Ratio:
1. Collect all variables into one model
2. Run a log reg analysis and look up for the Exp B
3. Compare them with the previous odds ratio from the previous bivariate analysis
p.s: I use complex sample models
Relevant answer
Answer
Version 23
  • asked a question related to SPSS
Question
1 answer
PCA analysis on the covariance matrix performed in SPSS gives as output the raw components and the components rescaled by a constant factor for each variable. Does anybody know how this factor is computed? and which one is selected during the analysis?
Relevant answer
Answer
Principal Component Analysis (PCA) is a statistical technique used to identify patterns and reduce the dimensionality of a dataset. When performing PCA on the covariance matrix in SPSS, the results typically provide information on the principal components, eigenvalues, eigenvectors, and variance explained. Here's how to interpret the output:
  1. Principal Components (PCs): The principal components are linear combinations of the original variables that capture the maximum amount of variance in the data. Each PC represents a different pattern or structure in the data. They are usually ordered based on the amount of variance they explain, with the first PC explaining the most variance, followed by the second PC, and so on.
  2. Eigenvalues: Eigenvalues represent the amount of variance explained by each principal component. They indicate the importance of each PC in accounting for the variation in the data. Larger eigenvalues indicate that the corresponding principal component explains more variance. Eigenvalues can help you determine the number of PCs to retain for further analysis. A common rule of thumb is to retain PCs with eigenvalues greater than 1 or based on the scree plot, where a sharp drop in eigenvalues occurs.
  3. Eigenvectors: Eigenvectors represent the weights or coefficients assigned to each original variable in the linear combination that forms each principal component. They indicate the direction and strength of the relationship between the original variables and the PCs. The signs and magnitudes of the coefficients can help you understand the relative importance and contribution of each variable to the corresponding PC.
  4. Variance Explained: SPSS provides information on the percentage of variance explained by each principal component. This indicates how much of the total variability in the data is accounted for by each PC. It helps you assess the cumulative contribution of the retained PCs to the overall variation. You can use this information to decide how many PCs to retain based on the desired amount of variance explained.
To interpret the results of PCA, you typically examine the eigenvalues and the variance explained to determine the number of PCs to retain. You can also explore the eigenvectors to understand the underlying patterns or structures in the data and how variables contribute to these patterns.
It is important to note that the interpretation of PCA results should be based on the specific context and objectives of your analysis. Additionally, it is often helpful to combine PCA results with domain knowledge and further analyses, such as factor loadings, variable contributions, and biplots, to gain a comprehensive understanding of the data patterns and relationships.
  • asked a question related to SPSS
Question
6 answers
I am a PhD scholar. When importing the rotated component matrix in Amos 23, I'm having trouble. My factor loads while conducting CFA have different values than the loadings I get in SPSS after performing factor analysis and transferring them to Amos 23 using the pattern matrix plugin.
I have an acceptable model fit, but because of fluctuation in cfa factor loadings, my AVE falls below the.5 cutoff. Is there a way to fix this?
Relevant answer
Answer
Hi,
just to be correct: The chisquare ist 1 thousand, 6 hundreds etc. (asking because of your use of the decimal points). If yes, you model is highly misspecified. The huge degrees of freedom suggests that you have a very large number of indicators. Can you tell how many indicators and how many factors?
I would not interpret the loadings and go back to the drawing board. Select 2-3 indicators which are crystal clear (from the perspective that these should reflect the same underlying latent).
Here is a paper, in which we also had a misspecified model and it gives some clues how we dealt with that:
Rosman, T., Kerwer, M., Steinmetz, H., Chasiotis, A., Wedderhoff, O., Betsch, C., & Bosnjak, M. (2021). Will COVID-19-related economic worries superimpose virus-related worries, reducing nonpharmaceutical intervention acceptance in Germany? A prospective pre-registered study. International Journal of Psychology. https://doi.org/10.1002/ijop.12753
Rcode
Good luck
Holger
  • asked a question related to SPSS
Question
3 answers
I've recently completed a data analysis for my thesis using SPSS and I'm now considering the best way to visualize my results. While SPSS does offer visualization tools, I'm contemplating whether to use a dedicated visualization app instead. Can I use a different application for data visualization after performing the analysis in SPSS? If so, are there any recommended practices to follow or potential issues to avoid? I'd appreciate any advice on this matter
Relevant answer
Answer
After conducting your data analysis in SPSS for your thesis, there are several data visualization tools you can consider to present your findings effectively. Here are some popular options:
  1. Microsoft Excel: Excel is a widely used tool for data visualization. It offers various chart types, such as bar charts, line graphs, scatter plots, and more. You can create visually appealing graphs and customize them according to your needs. Excel is user-friendly and accessible to most researchers.
  2. Tableau: Tableau is a powerful data visualization software that allows you to create interactive and dynamic visualizations. It offers a wide range of chart types, dashboards, and filtering options. Tableau provides advanced features for exploring and presenting your data effectively, making it suitable for complex data analysis.
  3. R and ggplot2: R is a statistical programming language that provides extensive capabilities for data analysis and visualization. The ggplot2 package in R is particularly popular for creating high-quality and customizable graphs. R allows for more flexibility and advanced statistical graphics compared to other software options, but it may require some programming skills.
  4. Python and Matplotlib/Seaborn: Python is a versatile programming language widely used in data analysis and visualization. The Matplotlib and Seaborn libraries in Python offer a wide range of plotting options. Matplotlib provides flexibility, while Seaborn offers more visually appealing and statistically-oriented visualizations.
  5. SPSS Visualization Designer: If you prefer to stick with SPSS for visualization, SPSS Visualization Designer allows you to create custom visualizations with advanced features. It provides a drag-and-drop interface and offers numerous options for customizing and styling your graphs.
  6. Datawrapper: Datawrapper is a web-based tool that focuses on creating simple and interactive visualizations, including charts and maps. It offers easy-to-use templates and customization options. Datawrapper is particularly useful for creating visualizations for online publications or websites.
Consider the complexity of your data, the type of visualization you want to create, and your comfort level with different tools when selecting the most suitable option. Additionally, it's important to ensure that the chosen tool aligns with any specific guidelines or requirements set by your institution or intended publication platform.
Remember that the purpose of data visualization is to effectively communicate your findings, so choose tools and techniques that best represent and highlight the key insights from your analysis.
  • asked a question related to SPSS
Question
6 answers
I am running mediation analysis using the PROCESS macro for SPSS and I have binary independent variables. It is my understanding that Haynes (2017) recommends that you report unstandardized beta coefficients. Can anyone help my explain why we would report these instead of standardized?
Relevant answer
Answer
For a binary independent (X) variable, it makes little sense to say "for a one-standard deviation unit change in X, we expect an xx standard deviation change in Y." Instead, we interpret the unstandardized coefficient, which gives the mean difference between the two categories/groups on X.
  • asked a question related to SPSS
Question
2 answers
I am currently at a Japanese university and working on a project with my co-workers.
We use DropBox to share an SPSS file (.sav) to run some analysis in AMOS. My co-workers (using the Japanese version of Windows 11) can open the file from Dropbox and conduct analysis without any problem. However, in my case (I am using an English version of Windows 11), my AMOS (no matter set in Japanese or English) cannot read the file.
Does anyone here have similar experiences? Is it because of my OS issue?
Relevant answer
Answer
Dear Christian Geiser,
If your AMOS software is unable to read your SPSS file, there are a few steps you can take to troubleshoot the issue:
1. Check the file format: Ensure that your SPSS file is saved in a compatible format that AMOS can read. AMOS typically supports the .sav format, which is the standard format for SPSS data files. If your file is saved in a different format, try saving it as a .sav file and then attempt to open it in AMOS again.
2. Verify the version compatibility: Confirm that your version of AMOS is compatible with the version of SPSS used to create the file. Sometimes, compatibility issues arise if you are using an older version of AMOS that does not support the file format created by a newer version of SPSS. Make sure you have the latest updates for both AMOS and SPSS.
3. File integrity: Check if the SPSS file itself is intact and not corrupted. Try opening the file in SPSS to ensure it can be accessed without any issues. If SPSS can read the file successfully, the problem might be specific to the AMOS software.
4. Reinstall AMOS: If none of the above steps resolve the issue, try uninstalling and reinstalling AMOS. This can help resolve any potential software conflicts or glitches that may be causing the problem.
5. Contact support: If the problem persists after attempting these troubleshooting steps, consider reaching out to the technical support team for both AMOS and SPSS. They can provide further assistance and guidance tailored to your specific situation.
By following these steps, you should be able to diagnose and resolve the issue with AMOS not being able to read your SPSS file.
  • asked a question related to SPSS
Question
4 answers
I am doing a mixed ANOVA with one within-subjects factor and two between-subjects factors. To prevent order effects, the order in which participants received each level of the within-subjects factor has been randomized. Is it possible to include this order randomization in a repeated measures mixed ANOVA in SPSS? And if so, should it be included as a covariate?
Relevant answer
Answer
I have some questions.
  1. How many subjects are there?
  2. How many levels does the within-Ss factor have?
  3. How exactly was the order "randomized" to use your word?
  4. How many different orders were used?
Thanks for clarifying.
  • asked a question related to SPSS
Question
5 answers
Hello everyone, I did a survey with a 4 item 3 point-likert scales ( for example; not true, sometimes and true) and a 5 item 4 point-likert scales ( for example; strongly disagree, disagree, agree, strongly agree). I conducted a cronbach's alpha, but it was too low due to a low number of items.
Would there be an alternative to report the reliability analysis that would result in a higher score? Ideally, I would like to create scales from these items.
I also thought of reporting only the correlation coefficient of these scales.
Could you help me?
Relevant answer
Answer
Cronbach's Alpha in SPSS Statistics - procedure, output and interpretation of the output using a relevant example | Laerd Statistics.
  • asked a question related to SPSS
Question
5 answers
Hello researchers,
I am facing problem to do a regression analysis with three independent variables, one mediating variable, and one independent variable. How ca I do this in SPSS? Any one please can you help me?
Relevant answer
Answer
Hello again Md.,
I would recommend you try either path analysis (available in any SEM package) or the associated multiple linear regression models.
The SEM model would have these proposed relationships:
1. IV1 -> Med
2. IV2 -> Med
3. IV3 -> Med
4. Med -> DV
Good luck with your work.
  • asked a question related to SPSS
Question
7 answers
Can I perform Principal component analysis on SPSS and then move to Smartpls to test hypotheses? (Higher and lower order constructs are all reflective).
Relevant answer
Answer
Hello Tarek,
Almost any set of measures involving human responses would be better treated via common factor analysis than PCA.
When you're starting off with possible indicators, it is ok to evaluate proposed factors individually. However, unless you simultaneously evaluated the multi-factor model, you'll have no evidence for convergent validity (across factors which ought to be correlated, by your framework) or discriminant validity (across factors which ought to have little or no correlation, according to your framework).
So, even in EFA, evaluating multiple factors (and using some sort of oblique rotation, such as promax), is a better way to go. And yes, principal axis factoring is fine. Note that, were you using covariance-based method for your confirmatory work, maximum likelihood extraction would make more sense.
Good luck with your work.
  • asked a question related to SPSS
Question
4 answers
I have a dataset of results obtained from a technique of XRF. However, I only received the results with the mean and the variation of three replicas. Can I add the dataset with mean + standard deviation in SPSS?
Relevant answer