Science topic
Scale Development - Science topic
Explore the latest questions and answers in Scale Development, and find Scale Development experts.
Questions related to Scale Development
I hope this email finds you well. My name is Areeba Shafique and I am a bachelor's student at Fatima Jinnah Women University. I am conducting a research study on Impact of neuroticism and social media exposure on climate change anxiety among young adults ;role of environmental concern. . As part of my research, I am interested in using the Environmental attitude inventory (EAI-24) scale developed by John Duckitt
Could you please provide me with the contact information of the author of the EAI-24 Scale, specifically John Duckitt, so I can request permission to use their scale in my research? Alternatively, if you could forward my request to the author, I would greatly appreciate it or give me Email .
Thank you for your time and assistance.
Best regards,
Areeba Shafique
Hello.
The panel advised me to use a sequential exploratory design for scale development. However, the construct that I will be studying is from the 70s. Also, there are existing scales about it.
My question is: should I write an RRL before conducting interviews? Aside from that, if RRL is advisable, how should it guide my interview questions?
Thank you.
Hello. I'm currently writing my thesis proposal about innovative work behavior. Can I just ask if anyone know what is the scoring and interpretation mechanics of Kleysen & Street's innovative work behavior scale? Does it have a global score? I attached the original study and a subsequent study below. Thank you.
Hello,
I am looking for a reference that specifies acceptable thresholds for McDonald's Omega. However, upon reviewing the existing cited works, I have noticed that no specific threshold has been determined for McDonald's Omega. Instead, the values generally attributed to Cronbach’s Alpha are applied to McDonald's Omega.
I have also examined studies that suggest there is no difference in cutoff points between McDonald's Omega and Cronbach’s Alpha.
As I am currently conducting a scale development study, could you recommend a reference that supports the acceptability of my obtained value?
Thank you for your assistance.
what book to be refer for formative scale development and validation?
I used 21st century skills scale developed by Ravitz in my study. One of the comments I received from one of the professors was that "you cannot simply use an instrument developed for another country and apply the same instrument to your country. You have to contextualize the instrument".
Hi,
I have developed a semi-structured interview assessment tool for a clinical population which gives scores - 1,2,3 (each score is qualitatively described on ordinal scale), using a mixed methodology. The tool has 31 questions.
The content validation was done in the Phase 1&2 and the tool was administered on a small sample (N=50) in Phase 3 to establish psychometric properties. The interview was re-administered on 30 individuals for retest reliability. The conceptual framework on which the tool is developed is multidimensional.
When I ran cronbach alpha for the entire interview it is .75 but for subscales it is coming between 0.4-0.5. The inter-item correlation is pretty low and some are negative and many are non-signifcant. The item-total correlation for many is below 0.2 and also has a negative correlation- based on the established criteria we will have to remove many items.
I am hitting a roadblock in analysis and was wondering if there is any other way in whcih we can establish reliablity of a qualitative tool with low sample, or other ways to interpret the data where a mixed methodology has been used (QUAL-quan).
Since the sample is small I will be unable to do factor analysis, but will be establishing convergent/divergent validity with other self-report scales.
Thanks in advance
I have seen many comments implying if a newly developed scale has a solid model background, EFA can (or better, should) be skipped. In a cognitive scale that I have recently developed, I had a clear design on my items, based on the previous theory. However, after administrating it to my study population, I ran a WLSMV CFA with two first-order factors and saw that some items (out of a total of 50) have weak (<0.30) or very weak (<0.10) loadings and possible cross-loadings.
My fit indices improved to an excellent range after deleting some of the lowest-loading items. Even after that, I have items with factor loadings of ~0.20. I have good reliability when they stay. And they don't look bad, theoretically. After pruning them to have a minimum loading of 0.3, not only my already good fit indices don't improve much, but my reliability gets lower. And I lose a good chunk of items. You don't want to assess cognitive skills with 15 items since almost all batteries have 30-40 items minimum. Should I keep them?
Also, some of the items with ceiling effect (98% correct responses) stay in the CFA model with good loadings. Should I keep them?
There are clear guidelines on item-deleting strategies for EFA. What about CFA?
I am glad to announce the 5th Annual Research Faculty Development Program (FDP) title, 'Scale Development and Data Analysis with PLS-SEM in SmartPLS 4.0' (23 – 29 December 2024), organized by the Jaipuria Institute of Management, Noida.
PROGRAM OUTCOMES
· The participants will be able to conduct cross-sectional research in social science.
· The participants will be able to frame theoretical frameworks in social science research.
· The participants will be able to develop correct scales for conducting effective research in the
industry.
· The participants can use advanced statistical techniques of cross-sectional research to forecast and predict consumer preferences.
*The participants will receive a Free Two-Month Professional License Key worth 56 Euros (INR 5,393).
Link to register: Registration Link: https://lnkd.in/gp3-hGfX
A certificate will be provided on successful completion of the program.
Seats: 50 (First-cum-first-serve basis).
Platform: Offline at Jaipuria Institute of Management, Sector 62, Noida, India.
For more details, Please contact MS. SHUCHİTA TEWARİ (9811050643) or email to annualresearch.fdpnoida@jaipuria.ac.in
Hello
I recently developed a tool to Generate the item pool for psychological assessment based on the definition and literature review of the variable of the interest.
can you review the it a write your thoughts about it.
you can use the tool here : https://scales.arabpsychology.com/item-pool-generator/
Any input is appreciated
Thank You
Hi researchers. I am developing a scale on Psychosocial well-being. I have done the Factor analysis. I have used principal component analysis (PCA) to analyse the data in order to obtain the minimum number of factors required to represent the available data set. Two of the dimensions have shown Eigen values more than 1. Should I remove these two Eigenvalues? What does it mean in a simple language (not the statistical explanation)? Thank you in advance.
Suppose X is a variable and it has 4 sub construct/ dimensions and a researcher adds two constructs based on literature that hold valid in current research. Is it one of the ways to proceed with EFA and CFA to establish the validity and reliability of variable in the current context without conducting interviews and focus groups, which is a part of scale development? But here only dimension is added.
please give me the scoring system of Academic stress scale developed by Rajendra and Kalikapan?
I am working on scale development in behavioral finance by undertaking a mixed-method approach using the exploratory sequential design. The phenomenon has diverse meanings in existing literature (some measuring it in terms of behavior while others use combinations of dimensions such as knowledge and access). I am unclear about its definition, so I want to explore the perspectives about the concept and what components participants feel it includes by taking a phenomenological approach.
My notation for research design is qual→ QUAN→ QUAN. Please guide me as if my approach is right. Do I need to go in so much depth as my main aim is to develop and validate the scale and not undertake a qualitative approach? I just need to take a qualitative viewpoint to support my framework or to guide the initial items bank and dimensions I created using a literature review. Second, if not then is it still remains a mixed-method design? and if yes, then guide me as to how much the sample size should be? I referred to Creswell & Poth's (2018) Qualitative inquiry research design, which says 3-10 (Dukes, 1984) or 5-25 (Polkinghorne, 1989). I am confused, so please if you have any concrete reference, suggest here.
Thanks in advance to the reviewers!
please give me the scoring system of Academic stress scale developed by Rajendra and Kalikappan?
It is interesting to know about an exact guide on reporting scale development methods? What should we actually report in the results?
Does anyone have information about the validity of the diabetes self-efficacy scale developed by the Stanford Patient Education Research Center? There are eight items on the instrument.
I am asking this for a scale development study. We have mutual/same items asked to both groups, and some items are asked only to a specific group (e.g., people of immigrant descent). Is there a way to conduct EFA for both groups at the same time using the mutual/same items asked to the participants of immigrant and non-immigrant descent?
Hello everyone,
Thank you for your time and patience in advance! I attempt to develop a pedagogical framework and a competence scale for my PhD project. I have developed the drafts of both. I adapted several theories and previous similar frameworks for the current pedagogical framework development. I wonder if the expert panel is an indispensable step before I design teaching materials based on this framework. (I totally agree that having an expert panel can without any doubt benefit my design)
I asked this question because organizing the expert panel discussion might be challenging for me due to the time and resources limit.
Many thanks for your answers!
Best.
Bonnie
I am currently encountering some issues with a specific scale in my research. Could anyone here help me?
In my study, I am employing the "Servant Leadership Scale" developed by Barbuto and Wheeler in 2006. This scale comprises 15 items, such as: “I encourage my followers to be hopeful about the company; I can help my followers overcome negative emotions," etc.
I have two questions regarding this scale:
(1) Can I use scale developed in corporate setting to conduct research by using the same scale in eduational setting?
Specifically, Barbuto and Wheeler developed this scale within a corporate setting. Hence, I wonder if it is suitable for me to use this scale in an academic context to assess “servant leadership” among university professors in the higher education setting. If so, should I acknowledge in the manuscript that this study utilized the scale developed by Barbuto and Wheeler with revisions. For instance, items such as "I encourage my followers to be hopeful about the company" was adapted to "I encourage my followers to be hopeful about the university" to explicitly illustrate the revisions made to the original scale and to underscore the rigor of the paper to editors and reviewers?
(2) If the Scale was designed for managers to take part in the survey, could employee take part in the survey instead?
Specifically, the "Servant Leadership Scale " developed by Barbuto and Wheeler (2006) expects leaders, rather than followers, to take part in the survey to assess leaders' servant leadership. However, If I intend to distribute this survey to students (students are regarded as followers of professors) and invite them to evaluate the servant leadership of their professors ? Should I explicitly state in the paper that this study employed the scale developed by Barbuto and Wheeler(2006) with revisions. For instance, items such as "I can help my followers overcome negative emotion" was adapted to "my professor can help me overcome negative emotion”?
Because this is really urgent, and looking forward to your earliest help!
I am working in the area of digital finance, and i want to develop a scale, but there is no pre existing scale for the concerned variable. All the existing measures are taken from national surveys items. Moreover, there are varied conceptualizations of the concept. I have developed my own conceptual definition using two theoretical models. Now, i am developing my initial pool of items but i facing problems as the concept has not been studied in the digital context. Please suggest some way forward, and any reference. Also, i want to know is it valid if i am developing a scale using theoretical background, and on the basis of those theories i am adding dimensions into my proposed scale. Because usually papers on scale development does not mention any theory. Please guide me.
can anybody know how i can get permission for use Bisht battery stress scale developed by Dr. Abha Rani Bisht?
Dear researchers,
I am looking for psychometric scales to use in my survey, measuring depression, anxiety, resilience, and social desirability. There are so many scales available, but it is hard to find scales that consider the public's opinion in the development process using focus groups and cognitive interviews. Can you tell me if you happen to know any? It would be even better if extra steps were taken in the scale development process, such as reviews of existing scales, expert opinion, and theoretical construction.
Thanks a lot!
Sincerely,
Siying
I have a multi-part question that is a little ill-formed but the gist of it is this: I have 15 or so dichotomous items that divide into drug use behaviors that put one at risk for an overdose and protective behaviors that potentially reduce risk. I also have 3 socioeconomic indicators - also dichotomous.
My goal is to analyze these variables to obtain a scale that assesses risk for an overdose. I do have past-year overdoses experienced against which I could use any scale to determine predictive validity (sensitivity, specificity).
Prior analyses I did on the risk items only indicated they could be represented by two factors. Preliminary analyses that included the protective items found, not surprisingly, they formed a third factor.
So it seems I have a preliminary "scale" that includes 3 subscales. And I am stuck at this point with a few questions about it:
1) I have read some work and watched videos about assessing the structure of multiple subscale scales that leads to assessing bifactor or higher order factor structure. It's interesting but does that kind of analyses really add anything to scale construction as opposed to conceptual/theoretical clarity?
2) Also read about the use of multidimensional item response theory (mIRT) (as opposed to factor analyses as a means of scale development. I appreciate the more detailed information IRT provides and scale items but again, is there any advantage to using mIRT for scale development versus EFA/CFA? And even so, what is the path from mIRT to setting up an actual scale, determining how to score the items and then testing the predictive validity of the scale and scoring?
Any insights would be much appreciated.
Hello everyone,
My interest is scale development and psychometrics evaluation; and until now I was working based on the classical test theory. I really liked to evaluate the psychometric properties of the tools with item response theory. I would appreciate it if someone could analyze the data collected in this way as a colleague. The data is related to the field of nursing.
Thanks in advance,
Sincerely,
Reza Ghanei Gheshlagh
+98 9144050284
https://scholar.google.com/citations?user=DA1hhmIAAAAJ&hl=en
Hi I am developing and validating a clinical interview based tool which will be administered by clinicians.
To assess its qualitative face validity, I have done cognitive interviews with the target population of individuals diagnosed with addiction
However, I want to know if there is any way to assess face validity through quantitative scores. I have come across "impact scores" but they are mostly used for self-report scales.
Is it suggested to apply the same process for a clinician administered measure?
In this case, can some patients read the interview and score the importance of the items to assess face validity?
Thank you!
Hi,
I am about to develop a scale that could measure sustainability literacy. For this i would like to consider different types of questionnaire in this scale such as:
- Questions with yes/no responses.
- Questions with multiple choice options "where one is correct".
- Question with 5 point likert-scale "strongly disagree - strongly agree"
My main question is will i be able to conduct factor analysis (exploratory and confirmatory) on these different types of questionnaire? or do i need to stick to one type of it (most researches chose likert-scale)?
To cut a long story short, I'm measuring perceptions regarding organisational change across schools in SEAsia.
Having developed my own framework based upon existing frameworks and a scale developed and adapted from existing scales, I have designed and conducted a mixed-method Likert-scale and open ended survey.
Whilst I have a matrix framework which shows how the scales align to each factor on the framework, received personal positive feedback from published researchers working in field recognising its validity and had positive feedback from educators working within this context, I still think it might benefit from extra analysis...
Would there be any real benefit including EFA since I am quite confident attributing items to certain factors?
Ps I was going to opt for Confirmatory Factor Analysis, but my sample size is only 71. So is this just something for me to write about in my limitations section?
Any thoughts or feedback appreciated!
I've started running an analysis for a new outcome measure. I was not part of the face validity design. The scale started with 13 items. I was wondering if there is any literature or references that guide the ideal number of items to have at the start of scale development, or references that discusses number of items?
My objective is scale development. but the indicators in the construct also have associations between them, so I want to know whether it is valid to study the interactions between a construct's indicators. my reasoning for the same is that studying interactions could help me understand if the effects of specific indicators differ depending on the levels or presence of other indicators.
I'm not able to reach Professor Reynolds directly and none of the authors of the papers that have used his scale have replied to my emails. If anyone has used this scale then please let me know if I can use it with relevant citation. Thank you
Hi all,
I'm currently in the analysis phase for a program evaluation/scale development project where most of the questions follow a standardized option response format:
Strongly agree
Agree
Neither agree nor disagree
Disagree
Strongly disagree
However, there is one item whose response options are:
Strongly agree
Agree
Neutral
Disagree
Strongly disagree
My question is, for the type of analysis I'm doing (year over year differences, some validity/reliability using R), would this item for which the mid-point response option is "Neutral" as opposed to Neither agree nor disagree" make a significant difference in my analysis results? Could I interpret the results with confidence given the change in wording for the mid-point option, or could I move forward with analysis with the inclusion of this item? I would omit it entirely, but it belongs to a scale comprising 5 items.
Thank you in advance for any guidance!
Hi everyone,
This may be bit of a stretch, but if you have used the following scale before, could you please tell me what higher (and lower) scores indicate? I cannot find the scale in English.
Fragebogen zur Erfassung von Einstellungen gegenüber übergewichtigen Menschen [Questionnaire for Measurement of Attitudes toward Obese People], Degner 2006)
For a research on a psychotherapeutic intervention for health care employees in the face of COVID-19 I am looking for the Nursing Stress Scale or a similar instrument to measure stress in nursing-staff. It would be great if anyone could help me.
Thank you :)
Now that EFA shows uni-dimensionality with only one factor extracted and CFA conducted ;then what shall be next steps for the following in scale development/construction.
i) convergent validity
ii)discriminant validity
We are working on a project investigating the impact of electricity reduction/loadshedding on the small business sector. We would appreciate any similar studies in other contexts. Including scales used to measure impact of electricity reduction/loadshedding.
#electricityreduction #loadshedding
Dear researchers, can both EFA and CFA be applied to the database obtained at the end of the scale development process at the same time?
SSQ (Simulator Sickness Questionnaire) is known to have a complex factor structure, with items loading on multiple dimensions.
In the original study (Kennedy et al., 1993), it is stated that "The N, O, and D scores are then calculated from the weighted totals using the conversion formulas given at the bottom of the table."
Those formulas are:
Nausea = [ Sum obtained by adding symptom scores ] x 9.54
Oculomotor = [ Sum obtained by adding symptom scores ] x 7.58
Disorientation = [ Sum obtained by adding symptom scores ] x 13.92
Total Severity = (Nausea + Oculomotor + Disorientation) x 3.74
It is not clear in the article that how those multipliers, 9.54, 7.58, 13.92 and 3.74 were derived.
Question A: How did they derive those multipliers?
I am working on a Turkish translation of SSQ, and my results are promising. However, it looks like I need to remove some items, and make some changes in scoring.
Attached file contains a comparison of factor weights of my results and Kennedy et al's. original work, besides Bark et al.'s (2013) results on some driving simulator experiments. My results are more similar to Kennedy et al. study, compared to Balk et al study.
The data is collected through 84 participants who had 2 different VR game sessions. SSQ-TR factor analysis is done using Principal Components with Varimax rotation and 3 factors emerged based on eigenvalue>1 assumption.
Question B: I seek for suggestions for factoring the SSQ-TR.
I have some ideas on removing some items and re-adjusting item/load structure, indicated on the shared spreadsheet.
References
Kennedy, R. S., Lane, N. E., Berbaum, K. S., & Lilienthal, M. G. (1993). Simulator Sickness Questionnaire: An Enhanced Method for Quantifying Simulator Sickness. The International Journal of Aviation Psychology, 3(3), 203–220. doi:10.1207/s15327108ijap0303_3
Balk, S. A., Bertola, M. A., & Inman, V. W. (2013). Simulator sickness questionnaire: Twenty years later.
These items are taken from WVS7 questionnaire to find religious tolerance :
- Do you trust people of other religion?
- Whenever science and religion conflict, religion is always right
- The only acceptable religion is my religion
- Mention if you would not have people of a different religion as your neighbor
All of them are in 4 point scale. Higher value would indicate higher tolerance. Alpha value is below 0.2.
What should be done? Should I carry on ignoring the alpha? Is alpha even appropriate in this case?
I have found only one valid 40-item scale developed by De Vries, et al. in 2012. I need a shorter one. I am glad one can provide me with one shorter scale.
Tnx
Hi researchers
I would like to use seidman zager's teacher burnout scale developed in 1987. Any one with a copy of this questionnaire?
I am conducting a research which involves scale development for emotional experiences with Wanghong (Internet-famous) restaurant dining consumption. With reference to the steps in prior literature, I have already done interview and expert review on the measurement scales. It is interesting to see that the emotional experiences may be categorized into three stages, pre-, during, and post-dining experience. I have conducted the first study, with the objective to purify the scale. I have done one analysis on all the measurement items using EFA without the consideration of three stages, and four factors emerged. In order to reflect the finding that emotional experiences are different in the three stages, I think three EFAs should be conducted? It seems to me that the first way is more methodological correct, while the second way is more theoretically or conceptually correct. Would appreciate if anyone may give me some advices on this! Thanks a lot!
Dear Colleagues,
I am looking for cross-cultural studies on identity development. I have found various studies using the Utrecht-Management of Identity Commitments Scale. However, I struggle with finding any studies using the Dimensions of Identity Development Scale developed by Koen Luyckx. I've tried to find something in Google Scholar, but the outcomes of my research were poor. If you know about this kind of study, let me know.
Kind regards,
Kamil Janowicz
Dear researchers,
I've been working on two different scales for two separate studies. I'm not getting the expected results from Exploratory Factor Analysis (EFA). Confirmatory Factor Analysis, on the other hand, provides me with results that confirm the construct's validity. Given this, do I need to perform EFA in the following cases?
1. A study of scale development based on Bandura's (1977) self-efficacy model. Bandura theorizes that efficacy expectations vary on several dimensions, including magnitude, generality, and strength. As a result, I accepted these dimensions as scale sub-factors and wrote some items about them.
2. A scale development study based on curriculum learning objectives. There are four units in this curriculum. Each unit has its own set of learning objectives. I wrote some items around these learning objectives. I wanted to learn students' perspectives on how well they met these learning objectives, so I accepted each unit as a dimension (sub-factor) of the scale in CFA...
Source: Bandura, A. (1977). Self-efficacy: toward a unifying theory of behavioral change. Psychological review, 84(2), 191.
Hi everyone,
As part of my PhD I'm validating a patient-reported disease severity scale for patients with a rare condition. It assesses the severity across 5 symptom groups using a 0-5 likert rating. It's been adapted from a previously validated clinician-reported version to form a lay-reported version so that patients can report their own disease severity. The symptom groups are the same but the ways in which the response options are worded are different between the two questionnaires, which means this version needs validating. Initial testing done on the questionnaire suggests the isn't much differentiation across the response options on most of the items. I was thinking about interviewing patients, amending the questionnaire and then running some quantitative analyses to validate the scale.
I'm looking at using IRT, as the scale is not to be utilized in clinical settings, as there is already a validated clinician-reported tool to measure disease severity in the population. However the main problem I face is the patient population is incredibly small and I'm unlikely to get more than say 100 participants, all the stuff I'm reading on scale development says I need a lot more data otherwise the analysis won't have sufficient power.
Has anyone got any experience validating questionnaires using small sample sizes or has any advice regarding different validation strategies?
Many thanks!
For convenience, I collected data from a single large sample for scale development.
and then I randomly split into two samples for EFA and CFA.
In this case, I wondering which sample (total? or CFA sample?) should be evaluated for the criterion validity or reliability of the newly developed scale.
Hello!
I am in the process of adapting the questionnaire for my research. The data on one of the scales has a strong deviation from the normal distribution - 64% of the answers have the lowest possible value (the sample consisted of 282 people). The rest of the responses were also distributed towards the low values. There were 3 items on the scale. A Likert scale of 1 to 7 was used for responses.
However, it seems to me that the construct being measured implies such a result in this sample.
Can you tell me whether such a scale can be considered usable? Or is it too heavily skewed? Are there any formal criteria for this?
I've already done a CFA, and I've checked the test-retest reliability and convergent validity (the results are acceptable).
Thank you very much in advance for your answers.
Could you please elaborate on the specific differences between scale development and index development (based on formative measurement) in the context of management research? Is it essential to use only the pre-defined or pre-tested scales to develop an index, such as brand equity index, brand relationship quality index? Suggest some relevant references.
I have "value co-creation behavior" variable (Yi and Gong, 2013) in my research model. It is the product of a scale development paper (Yi and Gong, 2013) and a third-order construct. When I test it as a second-order construct, there is no any problem. But in its original paper, it is specified as a third-order construct. So, I would like to test it as it is but do not know how to make it on AMOS. If you have experience in testing third-order constructs on AMOS, Could you please help me with it together with the images from the program?
Thanks in advance.
Yi, Y. and Gong, T. (2013). Customer value co-creation behavior: Scale development and validation. Journal of Business Research, 66, 1279-1284.
Exploratory Factor Analysis and Confirmatory Factor Analysis are used in scale development studies. Rasch Analysis method can also be used in scale development. There are some researchers who consider the Rasch Analysis as up-to-date analysis. Frankly, I don't think so, but is there a feature that makes EFA, CFA or Rasch superior to each other in Likert type scale development?
Hello, I am new to R so a little stuck. I have 164 items as part of a scale development project (n=1271), and I want to set a cutoff of .40 for factor analysis. I tried using logical operators and used this script
loload <- tempdf1 %>% filter(Q1.0<.40) which set up the new datafile for 'loload' but didn't put any data in there. I then tried using this script with all 164 items separated by a comma, which returned an error message.
I'm quite stuck; numerous google searches don't offer a lot unless I want to do things to one specific variable.
Any help is appreciated.
Looking for the questionnaire on EMOTIONAL STABILITY SCALE developed by CHATURVEDI (2010) with description.
Dear Sir/Ma'am,
Greetings!
I developed a psychometric tool consisting of 95 items. I had shared with 5 experts for content validation. 51 items are removed after the content validation process. According to Polit & beck (2006) and Polit et al., (2007) the acceptable CVI values should be 1. The 51 items did not get the required value of CVI for acceptance thus, the 44-items are retained.
My question is 'Should I circulate the questionnaire (with 44-items) to students for criterian-releated validity or 'Should I go with 95 original items?'
If I go with 95 items then what is the use of doing content validation or computing CVI. Am I following the right path if I decide to go with 44 retained items? please say
Big Thank you,
Narottam Kumar
I am a researcher student. I noticed that many scholars used scales developed in western context. However, may constructs manifest differently in different cultures. Thus, I have to develop scales by myself. After reading some papers related to scale development, I found the "item generation" stage is the most difficult. I invite researchers who have experience in scale development to share their tips in "item generations" and other stages. For example, is interview the only approach to generate items? Are there any tips for item generation? Are there any other suggestions for scale development?
Thanks !!!
Hello,
Using factor analysis, I recently created a social questionnaire with four factors, each containing four items. My next step is validating the questionnaire.
I want to show that this questionnaire can differentiate between 3 product categories that have different social characteristics.
- What statistical method should I use to prove the differences?
- Should I ask the same respondents to evaluate all 3 products? or it is ok to have separate respondents for each product category.
Thank you
This 2022 article provides extensive details of the steps involved in a robust scale (measurement) study -
Syed Mahmudur Rahman, Jamie Carlson, Siegfried P. Gudergan, Martin Wetzels, Dhruv Grewal. (2022). Perceived Omnichannel Customer Experience (OCX): Concept, measurement, and impact. Journal of Retailing.
Open access; freely available at https://doi.org/10.1016/j.jretai.2022.03.003
See Table 1 for an overview of the steps.
Hello, it will be a great help, if anyone would help me find the scale and the scoring of the levels of emotional awareness scale developed by Richard D Lane in 1990. Thank you in anticipation.
Hi,
I have turned up a questionnaire prepared in English. I want to convert it into Turkish than practice it. But I cannot be sure whether I should analyze its validity and reliability because it is not a scale. I hope I can have a chance to pick your brains. Thank you in advance.
I developed the scale using qualitative approach ( projective technique and depth interview). can i complement scale development procedure on the basis of recommendation given by Carpenter,2018 for developing the scale? Please suggest .
Also throw the light on scale development procedure prescribed by Carpenter,2018.
Dear researchers
I developed a scale with 6 items and I want to compare it with a widely used scale with 4 items. These two scales have similar validity, reliability and in the EFA they are loading on the same factor. In addition, I have done a CFA between these two scales (or better said between these two latent variables) and the correlation among them is 1. Finally, another CFA between two other latent variables and these scales, (CFA was done separately that is one at a a time) and the correlations between the latent variables and the scales were similar - not identical though.
Are there any suggestions on how I should work ?
Thank you
Hello,
I am currently working on the discriminant validity on my scale development.
I have a second-order model consists of four first-order latent variable and one global second order variable (as you can see on the figure). I also have other related constructs that is needed for the discriminant validity.
When I calculated the average variances extracted, do I need to use factor loading and variance of a model that includes the related constructs or without them?
looking forward your answer!
Hello,
I was wondering if anyone would be willing to share any literature/knowledge on determining item order in scale development.
For context, we have developed and tested a scale. Question order was randomized in Qualtrics and the scale has since been validated. I have read several papers on scale development and validation, but they don't discuss item order within individual questionnaires at depth or at all. As the order of the items in a questionnaire should be consistent when used in the future, I was wondering at what stage this order is decided and how it can be tested (e.g., in terms of validity and reliability).
Thank you kindly in advance!
Kat Schneider
Greetings researchers. I am a Ph.D. scholar and I want to go through a Ph.D. research thesis which could guide me regarding the correct methodology of scale or tool development to measure an anxiety. I will be grateful to you for your cooperation in this respect. Thanks.
can I use a combination of a vignette (for one variable i.e. dependent variable) with a self-report survey questionnaire (for all other variables IVs, Mediators, and moderators)? if I can what types of analysis and software for that analysis I may use? if I can't what should I do? (scale development is not a good solution, neither scale for survey research is used nor available in previous research for that Dependent variable). I mean can I use a vignette for one variable with a self-report scale for all other variables in combination (it is somehow a mix of experimental and self-report methodology).
Dear RG Community, Can you please share how to do Item Analysis in scale development? I have developed a scale, now need to do the item analysis before try out.
I am trying to make a risk assessment type scale that measures the likelihood/risk of someone experiencing a psychological disorder in a setting. There are about 10 factors (e.g., gender, sleep quality, self-esteem, resilience, loneliness) that increase the likelihood and I would like to include these in the new scale. There are existing validated scales that examine each of these factors in the literature, but these are relatively long and including all these scales would make the new scale too long.
My plan at the moment is to pick 4-5 best questions (for confirmatory factor analysis later on) from each of these existing scales and adapt them to suit the setting I am interested in, and run the new scale with their full-length counterparts. As 11 scales (10 existing and 1 new one) would be difficult to complete in one sitting, I was thinking of breaking it down to 2 or 3 different testing sessions, preferably with the same participants, but if not possible, with different participants. Is this a reasonable approach to take?
I am aware that there are many books/articles about scale developments (e.g., Scale development: theory and applications by DeVellis), but I feel that these either target single construct (unidimensional) or multidimensional construct in which the subscales sort of “form” the multidimensional construct (e.g., personality with openness, neuroticism, agreeableness, conscientiousness and extraversion), and it is difficult to find the information on the type of scale I am trying to make. If anybody could recommend a good reference for the type of scale I am trying make, it would be most appreciated.
In a scale development, we know that Cronbach's α coefficient will be increased as the number of item increase, generally, when Cronbach's α coefficient above 0.7, the reliability is acceptable. I want to know when the item in a sub-scale is only 3, how much the Cronbach's α coefficient should be reached, the reliability can be accepted?
What should be done if items did not load in EFA?
should they be discarded? Or that means that they are unique and do not have partners?
What if these items are important or represent our dependent variables ? Can we include them as single items?
What are the programs which can be used for structural modeling?
I have used an AMOS previously but the trial period had run off, so are there any other available programs for the use which are free of charge?
thanks
Hello everyone,
My dependent and independent variables are loading on the same component when performing EFA, Also CFA does not show a good fit when they are combined in one model but separated into different components.
I think this is due to the high correlation between them. So should I separate the questionnaire into different scales and perform EFA/CFA for each scale independently. Then combine them together in a path analysis later?
How can I validate a questionnaire for a small sample of hospitals' senior executive managers?
Hello everyone
-I performed a systematic review for the strategic KPIs that are most used and important worldwide.
-Then, I developed a questionnaire in which I asked the senior managers at 15 hospitals to rate these items based on their importance and their performance at that hospital on a scale of 0-10 (Quantitative data).
-The sample size is 30 because the population is small (however, it is an important one to my research).
-How can I perform construct validation for the items which are 46 items, especially that EFA and CFA will not be suitable for such a small sample.
-These 45 items can be classified into 6 components based on literature (such as the financial, the managerial, the customer, etc..)
-Bootstrapping in validation was not recommended.
-I found a good article with a close idea but they only performed face and content validity:
Ravaghi H, Heidarpour P, Mohseni M, Rafiei S. Senior managers’ viewpoints toward challenges of implementing clinical governance: a national study in Iran. International Journal of Health Policy and Management 2013; 1: 295–299.
-Do you recommend using EFA for each component separately which will contain around 5- 9 items to consider each as a separate scale and to define its sub-components (i tried this option and it gave good results and sample adequacy), but am not sure if this is acceptable to do. If you can think of other options I will be thankful if you can enlighten me.
How can i validate a questionnaire for hospitals' senior managers?
Hello everyone
-I performed a systematic review for the strategic KPIs that are most used and important worldwide.
-Then, I developed a questionnaire in which I asked the senior managers at 15 hospitals to rate these items based on their importance and their performance at that hospital on a scale of 0-10 (Quantitative data).
-The sample size is 30 because the population is small (however, it is an important one to my research).
-How can I perform construct validation for the items which are 46 items, especially that EFA and CFA will not be suitable for such a small sample.
-These 45 items can be classified into 6 components based on literature (such as the financial, the managerial, the customer, etc..)
-Bootstrapping in validation was not recommended.
-I found a good article with a close idea but they only performed face and content validity:
Ravaghi H, Heidarpour P, Mohseni M, Rafiei S. Senior managers’ viewpoints toward challenges of implementing clinical governance: a national study in Iran. International Journal of Health Policy and Management 2013; 1: 295–299.
-Do you recommend using EFA for each component separately which will contain around 5- 9 items to consider each as a separate scale and to define its sub-components (i tried this option and it gave good results and sample adequacy), but am not sure if this is acceptable to do. If you can think of other options I will be thankful if you can enlighten me.
ALTS is an altruism scale developed by S.N Rai and S.Singh.
executive functioning = working memory, attention shifting and inhibition if task.
Please suggest me scale who asses all three of them or even i am ready to go with different scales. kindly mention authors also.
I want to use these tests for my m.phil thesis which is non funded and i am unable to bear expenses of publishers.
I am trying to validate a translated version of PCFS scale developed by Klok et al., 2020. in my country. Can you suggest how should I proceed? Such as:
- Confirmatory Factor Analysis
- Comparing with the results with other scale evaluating the same domain, i.e. EQ-5D etc.

I want a beginner friendly book about the process of adapting and validating a scale or health measurement for psychology. If I plan to do a scale validation as a dissertation project, what should I start reading about? Thanks!
I am in the midst of a scale development project for a psychological construct and wondering if a scale can still be used as a composite score in other analyses if there are four factors and there is no higher order factor. If anyone has insights on implications and/or examples of this type of construct/use, that would be great! Thank you.
I am currently undergoing research. I want to find the academic stress of undergraduate students. Anyone with original version of Academic stress scale developed by Kohn and Frazer 1986? Most of the one found was adapted
We can say that these factorial analysis approach are generally used for two main purposes:
1) a more purely psychometric approach in which the objectives tend to verify the plausibility of a specific measurement model; and,
2) with a more ambitious aim in speculative terms, applied to represent the functioning of a psychological construct or domain, which is supposed to be reflected in the measurement model.
What do you think of these generals' use?
The opinion given for Wes Bonifay and colleagues can be useful for the present discussion:
Yes/no response scales have a worst result in factorial analysis than Likert scales? Should we avoid yes/no response scales?
I am completing a study that intends to use a cognitive load scale for different formats of instruction. The simplest, most common scales are “How difficult was the lesson you just studied?” with a 7- point subjective rating scale developed by Kalyuga, Chandler, and Sweller (1999), ranging from (1) extremely easy to (7) extremely difficult, and “What level of effort did you put into learning the lesson?” with a 7-point subjective rating scale developed by Paas and Van Merri¨enboer (1993), ranging from (1) extremely low to (7) extremely high.
What is the benefit of having the two questions instead of just the perceived effort question? Is there a benefit to providing the questionnaire just once rather than after each trial? Finally, I'm having a hard time finding the reported reliability/validity information for the instruments. Much appreciated.
I have developed a scale based on literature review. There are many scales already present on these variables (knowledge management and ICT) and many studies in literature have given various factors as these are multidimensional scales. I have conduct CFA to confirm these factors. Please provide me strong citations based on area of social sciences for this process where we don't need EFA for scale development.
Academic Motivation Scale - College Version (AMS-C 28) score interpretation
Currently i have a psychological scale and is ready to gather data for pre-analyses like EFA and corelation analyses.
my quesitons is:how can i assure that the sample is representative for the total .
Hello all,
This is my first time doing CFA AMOS.
Initially, I developed a scale for a specific industry 17 items 5 factor scale based on theory of other industries. This proposed scale was tested with two ) datasets first with n=91 year 1 and second n=119 year 2 from a single institution. EFA identified 3 underlying factors in both the datasets, no items were deleted.
During year 3, a sample of n=690 consisting of participants all over the nation was used to do CFA using SPSS AMOS. Following is the output:
1. Based on EFA (3 factors, 17 items)
a) Chisquare = 1101.449 and df= 116 [χ2/DF = 9.495]
b) GFI = 0.805
c) NFI = 0.898
d) IFI = 0.908
e) TLI = 0.892
f) CFI = 0.908
g) RMSEA =0.111 (PClose 0.000)
h) Variance
Estimate S.E. C.R. P Label
F1 .573 .056 10.223 ***
F2 .668 .043 15.453 ***
F3 .627 .040 15.620 ***
i) Covariance
Estimate S.E. C.R. P Label
F1 <--> F2 .446 .036 12.502 ***
F1 <--> F3 .365 .032 11.428 ***
2) Based on theory (5 factors, 17 items)
a) Chisquare = 440.594 and df= 109 [χ2/DF = 4.042]
b) GFI = 0.926
c) NFI = 0.959
d) IFI = 0.969
e) TLI = 0.961
f) CFI = 0.969
g) RMSEA =0.066 (PClose 0.000)
h) Variance
Estimate S.E. C.R. P Label
F1 .677 .047 14.334 ***
F2 .670 .043 15.493 ***
F3 .648 .054 12.100 ***
F4 .741 .061 12.103 ***
F5 .627 .040 15.620 ***
i) Covariance
Estimate S.E. C.R. P Label
F1 <--> F2 .503 .036 14.057 ***
F1 <--> F3 .581 .041 14.262 ***
F1 <--> F4 .546 .041 13.388 ***
F1 <--> F5 .398 .032 12.321 ***
F2 <--> F3 .457 .036 12.848 ***
F2 <--> F4 .403 .035 11.405 ***
F2 <--> F5 .458 .033 13.899 ***
F3 <--> F4 .553 .042 13.036 ***
F3 <--> F5 .360 .032 11.275 ***
F4 <--> F5 .358 .033 10.754 ***
My questions:
1. Do I have to normalize the data before CFA analysis? (I am finding conflicting information since my scale is a likert scale and extreme values are not really outliers ?)
2. Can I report that theory based model is a better fit compared to EFA model? Would doing so be appropriate?
3. Is there anything else I need to do ?
Any guidance will be greatly appreciated.
Thank you,
Sivarchana
We often use scales developed by others in our researches. Sometimes the scale may have low validity and reliability values in our sample. In such a situation, I always encounter problems in analysis. So, in this kind of case, is there a problem if items that do not function well are removed in order to obtain acceptable values and if the necessary exploratory and confirmatory factor analyzes are made again?
I would be very happy if you add a source with your feedback. Thank you.
I am reviewing studies that assessed stress. The most commonly used tool is the perceived stress scale. I am attempting to interpret the data which is reported as a mean score. I haven't found any studies that have cut off scores related to levels/intensity of stress. I have read internet articles where scores are interpreted in terms of intensity e.g. but they don't give the research behind the cut off score, see below
I would be grateful if somebody could guide me to research that validates cut off scores for each level/intensity of stress.
Thank
Lloyd
Hi RG colleagues,
Is anyone familiar with a good climate-related vulnerability scale? I have not landed on a valid climate-related psychological vulnerability scale so far. Now I am wondering if someone can make a recommendation. Please advise if adapting The Psychological Vulnerability Scale would be a good idea? The psychological vulnerability scale is a 6 items scale (PVS; Sinclair & Wallston, 1999). I have used it in the past with high reliability in the covid related PVS.
Best wishes,
Gulnaz
I'm proceeding to adapt a scale in English into my own language, and also wanna make some small changes not related to cultural problems. The original scale is designed to measure a general concept, now I want to change the items a little bit so it can measure that concept in a context-specific way.
So I wanna ask what is the right procedure to do it? Should I run a translation-adaptation process first (translate, make some cultural changes if needed and run a pilot study to check the reliability & validity of the adapted scale) then make the context-specificity changes that I want for the research topic;
or translate and make the changes at the same time, then run the pilot study to check the reliability & validity of the adapted scale?
Thank you in advance.
I am a student who needs to do some research for my master thesis about the expected product care of a phone in the future. To be precise: I’m doing research on the impact that a Country of Origin label has on the usage of a mobile phone. In this research, I’ll give different information to different respondent groups (the “made-in-label” will differ). Now I want to develop or use an existing scale that measures how well the respondent thinks he’ll take care of the product. For instance: “I will use this product longer than a year”, “I will take care of this product” etc.
Does anyone know a reliable existing scale or how to develop a new one?
I'm in the process of scale development and have as many as 7 to 11 items cross-loading (>.32) on two or more factors/sub-scales. How acceptable and common is the practice of including an item under two or more of them?
Need the scale for a dissertation. Would be better if I could get it for free since my work is not funded. I need the English version of this scale.
Hello,
I want to use a motivation scale with my students (secondary school) for my research, but the scale is in a 4-point Likert style with no neutral option (Certainly true - True - Not True - Certainly not true). It is a valid and reliable scale. The question is, would it still be valid and reliable if I added a neutral option in the middle of the answers? More importantly, would it be better that way?
A second question is, as seen above, the answers of the scale goes from positive to negative. Because I use the other way around version (negative to positive) in the rest of my research instruments, I'd like to do the same with this motivation scale too. Would that be OK?
Thanks a lot for your time in advance.
Yusuf Polat
I am developing a scale to investigate self-understanding in relation to cultural diversity among university students. I will be using this scale to investigate criterion validity for the scale I developing.
Hi,
I'm conducting an EFA for my scale development article. I was wondering whether anyone has an article or book I can refer to regarding the process of rerunning the EFA after item deletion?
Thanks!
The MMPI-2 and it's much improved variant the MMPI-2-RF produces clinical scales that are derived from participants responding to several hundred true and false items. Test developers reference deploying factor analysis in the scale development, but I thought factor analysis couldn't be computed with dichotomous data? What am I missing?
The scale Development is a sequential, systematic and scientific exploring and confirming the variables under a construct....
I am using three scales in my study, two of them are five Likert Scale, while one of them, Self-perceived communication competence (SPCC) developed by McCroskey & McCroskey, 1988) is a percentage scale measuring the percentage of perception of people about their own communication competence from 1 to 100.
Can I convert the scale using, 1-20=Strongly disagree, 21-40= Disagree, 41-60= Neutral, 61-80=Agree and 81-100= Strongly Agree.
Is there any way to convert scales?
Thank you in Advance
PS: Scale is attached.
I am using AMOS-21 for the estimation of the model fit for a questionnaire. I have found the following values. Are these values enough and appropriate? Which of the following are important to present while reporting?
CMIN/df= 2.426
RMR= .05
GFI=.921
AGFI= .890
NFI= .951
TLI=.964
CFI=.970
RMSEA=.064
Hello
I want to build a questioner about emotional and cognitive Strategies . my question is:
how should I Formulate sentences or items?
they should began mostly with " I thing " or "I feel" to represent the emotional and cognitive side of the subject ?
or
they should began with a verb to present a behavior or an act "I do" to represent the Strategies?
Tank you