ArticlePDF Available

Scale Validity In Exploratory Stages Of Research

Authors:

Abstract

Scale development assumes that certain steps are to be taken in order to obtain a valid measurement instrument. Most of the researchers jump to the confirmatory stage and avoid exploratory measures. However, exploratory methods that are used in the first stages of scale development are recommended so as to avoid further problems regarding the validity of the scale. Before conducting reliability analysis and factorial analysis, exploratory methods can be applied. The main purpose of this paper is to draw the attention on alternative methods for scale validation that should be used in the exploratory phase. The role of these methods is to improve validity of results of the further confirmatory phases of research. The Lawshe (1975) content validity ratio and the Q-sorting procedure for testing construct validity are applied in the process of developing a scale for perceived risk.
The following ad supports maintaining our C.E.E.O.L. service
SCALEVALIDITYINEXPLORATORYSTAGESOFRESEARCH
«SCALEVALIDITYINEXPLORATORYSTAGESOFRESEARCH»
byAdrianaZait;ElenaBERTEA
Source:
Management&MarketingCraiova(Management&MarketingCraiova),issue:1/2013,pages:3846,
onwww.ceeol.com.
SCALE VALIDITY IN EXPLORATORY STAGES OF
RESEARCH
PhD Patricea Elena BERTEA
Romanian Academy Iaşi, Romania
Email: patricia.bertea@yahoo.com
Professor PhD Adriana ZAIŢ
University “A. I. Cuza”, Iaşi, Romania
Email: azait@uaic.ro
Abstract:
Scale development assumes that certain steps are to be taken in order to
obtain a valid measurement instrument. Most of the researchers jump to the
confirmatory stage and avoid exploratory measures. However, exploratory
methods that are used in the first stages of scale development are
recommended so as to avoid further problems regarding the validity of the
scale. Before conducting reliability analysis and factorial analysis, exploratory
methods can be applied. The main purpose of this paper is to draw the
attention on alternative methods for scale validation that should be used in
the exploratory phase. The role of these methods is to improve validity of
results of the further confirmatory phases of research. The Lawshe (1975)
content validity ratio and the Q-sorting procedure for testing construct validity
are applied in the process of developing a scale for perceived risk.
Keywords: scale development, content validity, q-sorting, perceived risk
Introduction
The main purpose of this paper is
to draw the attention on alternative
methods for scale validation that should
be used in the exploratory phase. The
role of these methods is to improve
validity of results of the further
confirmatory phases of research. The
methods are exemplified on a scale that
aims to measure perceived risk in e-
commerce.
Scale development has become
an important research area since
several seminal works (Cronbach, 1951;
Nunnaly, 1967; Churchill, 1979). The
use of scales in Management and
Marketing research has become
common since both fields deal with
studies on latent variables. Thus, the
methodology from Psychology is now
successfully employed by researchers
from the previously mentioned areas.
An important aspect in scale
development is assessing validity.
Validity refers to the ability of a
construct to measure what it was
supposed to measure (Goodwin, 2009).
When assessing the validity of a scale
we are actually looking how accurate
the scale is (Groth-Marnat, 2009).
Establishing the validity of a scale is
rather difficult, especially when we are
dealing with psychological variables.
The main issue is that such variables
are not observable and the researcher
has to identify the underlying latent
variables by constructing measurement
instruments. Validation of measurement
instruments assumes that the
inferences and conclusions that are
drawn in a research are actually valid
(Schultz & Whitney, 2004).
Another issue when talking about
validity is that it should not be
confounded with reliability. Reliability,
which is usually measured using the
Management&Marketing, volume XI, issue 1/2013
39
Cronbach alpha coefficient, refers to the
consistency of the measurement. A
more clarifying perspective is given by
Campbell and Fiske (1959), who explain
that reliability is the agreement of two
attempts to measure the same
underlying construct through similar
methods, while validity refers to the
same issue, but the methods used are
totally different. Cronbach alpha
measures a certain type of reliability
which is defined as internal consistency
and offers information on how items that
form a scale correlate with each other.
An accepted level of internal
consistency has to be at least of 0,7, but
not higher than 0,9 (Cronbach, 1951),
which indicates that some items might
be redundant inside the scale. Alwin
(2007) considers that alpha Cronbach
should be used more as an internal
consistency measure that shows how “a
set items hangs together to form a
scale” and that other approaches should
be employed in assessing reliability.
Among these, Alwin (2007) talks about
using multi-trait multi-method/
confirmatory factor analysis to measure
reliability. As far as validity is
concerned, Alwin (2007) explains that “a
reliable measure is not necessarily a
valid one”.
Types of validity
There are different types of validity
that researchers should look into when
developing a scale. Specialists talk
about three types of validity: criterion
validity, content validity and construct
validity.
Criterion validity
Criterion validity stands for how
well an instrument measures a variable
in comparison with another instrument
or a predictor. There are two types of
criterion validity: concurrent and
predictive validity.
Concurrent validity assumes there
is another construct that measures the
same variable, a construct considered
to be a benchmark in the research
domain. To have concurrent validity for
a construct it is compulsory that there is
a high correlation with the benchmark
construct. Researchers can also
choose the benchmark as being a
totally opposed variable and in this case
low correlation is expected in order to
have good concurrent validity. Usually,
to test for concurrent validity
researchers apply two different
instruments measuring the same
variable on the same sample, just that
one of the instruments must be a
standard in the domain, with previously
tested psychometric characteristics.
Predictive validity refers to the
ability of a measurement instrument to
predict future attitudes or behaviors.
Establishing predictive validity means
that data is collected twice at different
moments in time, so as to check if the
scale predicted or not a certain event. In
this case there is also need to do a
correlation between the variable we are
trying to measure and another variable
that is used as a criterion.
Content validity
Content validity refers to a correct
definition of the domain of the latent
variable that one intends to measure.
Another important aspect is the
identification of possible facets of the
construct. Thus, when we want to
measure a latent variable is important to
introduce in the construct all possible
items which could capture the essence
of the variable (Haynes, et al., 1995).
For instance, if we include items that
have no connection with the variable
that we generate measurement errors,
while if we exclude items that we will
have exclusion errors (Straub, et al.,
2004).
Content validity assumes two
stages (Lynn, 1986): the development
stage and the judgement-quantification
stage. The first stage implies the use of
qualitative methods such as interviews,
focus groups and, of course, an
intensive review of literature. The
second stage, which is intended to
Access via CEEOL NL Germany
Management&Marketing, volume XI, issue 1/2013
40
quantify the validity of a scale, requires
that a panel of experts evaluate the
scale’s items accordingly to the
Although methods have been
developed for the second stage, most
researchers appeal to literature review
and other qualitative methods to assure
content validity of the scale. This
qualitative type of validation is more or
less prone to subjective influences
coming from the researchers. Yet, this
approach is intensively used and there
are few who reach for alternative
quantitative methods. Nevertheless,
using a more empirical method with a
quantitative foundation adds more
scientific value to our research and
prevents validation problems to further
affect results.
Content validity measures
There are several ways to test
content validity using a quantitative
approach.
Lawshe (1975) developed a
quantitative measure for assessing
content validity called the content
validity ratio (CVR). The content validity
ratio offers information about item-level
validity. The procedure consists in
using a panel of experts to rate items
according to the relevance for the
domain of the scale. Each item of a
scale is rated on 3-point rating system
(1- item is irrelevant, 2 – item is
important, but not essential, 3 – item is
essential). For each item a CVR is
computed, that is basically the
proportion of experts that considered
the items important or essential for the
content of the scale. There is also the
possibility of having an overall measure
for the content validity of the scale. This
is called an index and it is computed as
a mean of items’ CVR values.
Another quantitative measure was
proposed by Waltz & Bausell (1983)
and it is called the Content Validity
Index (CVI). The difference between
this measure and the previous (Lawshe,
1975) is that experts rate items on a 4-
points rating scale with slightly different
anchors (1 – not relevant, 2 – somewhat
relevant, 3 – quite relevant and 4 – very
relevant). The index computation is
actually a percentage given by the
number of experts that rate quite
relevant or very relevant an item. A total
index per scale can also be computed.
According to Waltz et. al (2010) the CVI
per scale is recommended when there
are only two experts involved in the
judgment stage. When more than two
judges are involved, Waltz et. al (2010)
recommend to use alpha coefficient,
that quantifies the extent to which there
is agreement between experts.
Construct validity
Construct validity refers more to
the measurement of the variable. The
issue is that the items chosen to build
up a construct interact in such manner
that allows the researcher to capture the
essence of the latent variable that has
to be measured. Content validity must
be assessed priori to construct validity.
Construct validity implies the use of
more quantitatively oriented analyses.
It is important to make the
distinction between internal validity and
construct validity. The first one refers to
assuring a methodology that enables
the research to rule out alternative
explanations for the dependent
variables, while construct validity is
more concerned with the choice of the
instrument and its ability to capture the
latent variable. Internal validity becomes
a problem in experimental studies,
where each experimental group has to
follow the same methodology in order to
be able to correctly isolate the effect.
Construct validity has three
components: convergent, discriminant
and nomological validity. Convergent
validity and discriminant validity refers
to the way the construct relates to other
constructs. Convergent validity tests if
the items of a scale correlate higher
among them and have significant higher
loadings. Convergent validity can also
be assessed buy checking the
correlation between the instrument and
Management&Marketing, volume XI, issue 1/2013
41
other instruments that mean to measure
the same latent variable. Discriminant
validity assumes that items should
correlate higher among them than they
correlate with other items from other
constructs that are theoretically
supposed not to correlate. Nomological
validity tests if the construct has the
same relationships with other variables
that have been previously tested and
confirmed in other studies.
Construct validity can be tested
during early stages of research using
the Q-sorting procedure. The main idea
of the analysis is to separate items in
construct according to their specific
domain. The procedure is more close to
measuring discriminant validity. There
are two ways that it can be done
(Storey, et al., 1997):
Exploratory, when respondents
are given the items and asked to group
and identify category labels for each
group of items.
Confirmatory, when the
categories are already labeled and
respondents are asked to classify each
item in one category.
Q-sorting is applied on experts and
other persons of interest for the
research. It helps eliminate items that
do not discriminate well between
categories.
Research methodology
The present study presents to
alternative methods for assessing scale
validaty: the content validity ratio and
the Q-sorting procedure. Both
procedures were applied on a scale that
measures perceived risk in e-
commerce.
For building up the construct for
perceived risk in e-commerce we
followed the methodology used by
Jacoby and Kaplan (1972). They
divided perceived risk into six
dimensions: financial, performance,
time, social, psychological and physical.
We did not use the same dimensions as
listed above, since Jacoby and Kaplan
(1972) did research on products.
We aimed to study perceived risk
of Internet as an alternative shopping
channel. As a consequence, there was
need to restate the dimensions. In order
to do, that we investigated the work of
Featherman and Pavlou (2003) together
with Crespo, et al. (2009). In the end
we defined six dimensions of perceived
risk in e-commerce: financial,
security/privacy, psychological, social,
time/delivery and product risk. Each
dimension was identified through a
number of items ranging from 3 to 8,
which were extracted from the literature
review and in-depth interviews (table 1).
Table 1
Dimensions of perceived risk in e-commerce
Type of risk Items
I believe that online shopping is risky because I cannot
examine the product.
If I choose to buy online I do not have the certainty that the
product will be of good quality.
I believe online shopping is risky because I cannot touch the
product before buy it.
I cannot be sure that a product bought online has the
characteristics advertised on the website.
I believe that a product bought online will not perform as well
as one bought from a bricks and mortar store.
Product risk
If I buy a product online I risk not to be given the guaranty.
Management&Marketing, volume XI, issue 1/2013
42
I do not trust online payment.
When I pay online there is an increased probability to lose
the money on my credit card.
Using online payment there is a chance I pay more due to
hidden fees.
There is a low probability to lose money for a product ordered
on the Internet if I pay on delivery.
I believe that paying by credit card is a secure payment
method.
There are high chances of losing money when paying online
for a product.
Online shopping means potential money loss due to possible
Internet frauds.
Financial risk
The risk of losing money when buying online is the same
whether I pay by credit card or on delivery.
If I buy online there is a high risk that my personal data would
be used without my consent.
There is high chance that hackers take over my personal
account from a e-shop.
Security/
privacy
risk
If I decide to buy products online I risk losing control over my
personal data.
If I do my shopping online, there is a high risk that I receive a
different product that the one I ordered.
When I buy online I am sure that I will receive exactly the
product I ordered.
If I buy online there are low chances that my product would
have a delivery delay.
Time/
delivery
risk
When I buy online, I not sure that the e-shop will respect the
promised deadline.
There is small chance that my friends will change their
opinion about me because of me using Internet to do
shopping.
If I buy online I am taking the risk that my friends will change
their opinion about me.
Online shopping is positively seen by my family.
Social Risk
My friends do not approve online shopping.
Online shopping does not fit my self-image.
Online shopping is not compatible to my self-image.
Online shopping gives me a state of stress because it does
not fit with my self-image.
Psychological
risk
Online shopping fits me well.
In order to apply the two methods
we had to do two separate studies for
which we developed two
questionnaires.
Methodology for the content
validity ratio
For the content validity ratio we
followed the methodology explained by
Lawshe (1975). We introduced all the
items grouped for each type of risk. We
interviewed six experts that were asked
to answer if each item was “1=
Irrelevant, 2=Important, but not
essential and 3=Essential” for
measuring a certain type of perceived
risk.
Management&Marketing, volume XI, issue 1/2013
43
Table 2
CVR questionnaire example
Product risk item Irrelevant
Important,
but not
essential
Essential
I believe that a product bought
online will not perform as well as one
bought from a bricks and mortar store.
Methodology for the Q-sorting
procedure
For the Q-sorting study we
developed a questionnaire were we
included all items measuring perceived
risk without showing which item belongs
to which type of perceived risk.
Respondents had to classify items into
6 categories: social, psychological,
financial, security, product and delivery
risk (table 3).
Table 3
Q-sorting questionnaire example
Risk Item Risk Type
Social
Financial
Psychological
Security
Delivery
Online shopping gives me a state of stress because it
does not fit with my self-image.
Product
As a quantitative indicator of the
Q-sorting procedure we used the
correct classification percent, which
describes the percent of respondents
that have correctly classified an item
(Straub, et al., 2004).
Results
Content Validity Ratio
To calculate the content validity
ratio we used the methodology
described by Lawshe (1975), which
indicates that all items should be
analyzed by a group of experts, each
expert having the possibility to describe
the item as: 1= Irrelevant, 2=Important,
but not essential and 3=Essential. The
formula to calculate the ratio is:
N
In
CVR
=,
Where n – number of experts who
considered the item to be “Essential” or
“Important, but not essential”;
I – number of experts who
considered the item “Irrelevant”;
N – total number of experts;
The logic behind the formula is that
the more experts are in favor of one
item as being important or essential, the
more we can consider that item as
being part of the construct. Thus, we
can attain content validity of the
construct. As one can easily see, the
formula gives a negative result when
less than 50% of the experts rate the
item as essential or important but not
essential or a null result when 50% rate
it as irrelevant.
A panel formed by six experts
rated the items according to Lawshe
(1975) specifications. After analyzing
the data, we identified 7 items which
presented serious problems, CVR value
being negative, which suggests that
more than 50% of experts found the
items to be irrelevant (table 4).
Management&Marketing, volume XI, issue 1/2013
44
Table 4
CVR values
Item CVR values
Product risk – I believe that a product bought online will not
perform as well as one bought from a bricks and mortar store. -0.67
Social risk – There is small chance that my friends will change
their opinion about me because of me using Internet to do
shopping.
-0.67
Social risk – If I buy online I am taking the risk that my friends
will change their opinion about me. -0.67
Psychological risk – Online shopping does not fit my self-
image. -0.67
Psychological risk – Online shopping gives me a state of stress
because it does not fit with my self-image. -0.33
Psychological risk – Online shopping does not fit with my self-
image. -0.67
Psychological risk – Online shopping suits my self-image. -0.33
These results suggest that the 7
items should be removed from the
construct before advancing the
research.
Q-sorting
In order to calculate the percent of
correct classification, we identified the
frequency of respondents that checked
the correct category for each item. We
had items that obtained a 100% correct
classification – 3 items, items that had
percents higher than 70% – 22 items,
but also items with lower percents -4
items. We considered items with a low
classification percent those who were
below 60% (table 5).
Taking into account that more than
80% of all 26 items were correctly
classified, we can consider that the
scale has a good level of discriminant
validity. However, it is important to
further analyze those items that were
not correctly recognized as belonging to
a certain category of risk.
Table 5
Q-sorting results (items with low classification)
Risk type Item Percent
Psychological Online shopping does not fit my self-image. 0.52
Security/
privacy
There is high chance that hackers take over
my personal account from a e-shop. 0.59
If I do my shopping online, there is a high risk
that I receive a different product that the one I
ordered.
0.52
Time/delivery
When I buy online I am sure that I will receive
exactly the product I ordered. 0.22
Management&Marketing, volume XI, issue 1/2013
45
Conclusions
There is only one item that
presented common problems in both
procedures, the one belonging to
psychological risk. However, the
objective of the research was no to see
whether there are items with problems
in both cases, but to identify items that
do not match validity. So, CVR was
measured to test for content validity,
while Q-sorting was applied to test for
construct validity, more specifically
discriminant validity of items.
Both alternative methods revealed
items with significant problems, items
that should be removed in next stages
of the study or should be refined in
order to express more clearly a certain
type of risk.
The major implications of this
research rest in the importance of
correctly developing a measurement
instrument for a latent variable. There is
need for applying alternative methods to
test scale validity especially when we
develop a whole new construct and we
use qualitative methods such as in-
depth interviews or focus groups, but
also when we want to use a scale that
was previously developed, but never
used on a certain sample. The concern
for applying these types of methods
should exist whenever the aim is to
raise the quality of a research. That
would show a profound investigation of
all possible issues which may affect
scale validity.
Further research should
concentrate on establishing how these
methods can improve convergent
validity, discriminant validity and
nomological validity. Moreover, it could
be useful to examine who are the most
appropriate respondents for each
method. If we have to use only experts
or we could also use non-experts, just
consumers. An interesting approach
would be to compare results coming
from two different samples and to see
whether respondents’ type is an issue.
The problem is, however, that the
experts sample will always be smaller
than the consumers’ one and it is
difficult to obtain representativity.
The value of this research stands in
the revival of rather isolated methods of
scale validation that can prove high
utility in exploratory phases of research.
Content validity ratio and Q-sorting are
less employed, so we wanted to
introduce them and raise researchers’
interest for these alternative methods.
REFERENCES
Alwin, D. (2007), 'Margins of error: A study of reliability in survey measurement',
547.
Campbell, D. & Fiske, D. (1998), 'Convergent and discriminant validation by the
multitrait-multimethods matrix', Personality 56, 162.
Churchill Jr., G. A. (1979), 'A Paradigm for Developing Better Measures of
Marketing Constructs.', Journal of Marketing Research (JMR) 16(1), 64 - 73.
Crespo, Á.. H., del Bosque, I. R. & de los Salmones Sánchez, M. M. G. (2009),
„The Influence Of Perceived Risk On Internet Shopping Behavior: A
Multidimensional Perspective”, Journal of Risk Research, 12(2), 259–277.
Cronbach, L. (1951), „Coefficient alpha and the internal structure of tests”,
Psychometrika 16(3), 297-334.
Management&Marketing, volume XI, issue 1/2013
46
Featherman, M. S. & Pavlou, P. A. (2003), „Predicting e-services adoption: a
perceived risk facets perspective”, International Journal of Human-Computer
Studies, 59(4), 451 - 474.
Goodwin, C. (2009), Research in psychology: Methods and design, Wiley.
Groth-Marnat, G. (2009), Handbook of psychological assessment, Wiley.
Gwet, K. (2001), Handbook of inter-rater reliability.
Haynes, S.; Richard, D. & Kubany, E. (1995), 'Content validity in psychological
assessment: A functional approach to concepts and methods', Psychological
Assessment 7(3), 238--247.
Jacoby, J. & Kaplan, L. B. (1972), „The Components Of Perceived Risk”, in M.
Venkatesan, ed., Proceedings, Third Annual Conference, College Park, ED,
Association for Consumer Research, 382-393.
Lawshe, C. (1975), „A quantative approach to content validity”, Personnel
Psychology 28(4), 563-575.
Lynn, M. (1986), 'Determination and quantification of content validity.', Nursing
research.
Mitchell, V.-W. (1999), 'Consumer Perceived Risk: Conceptualisations And Models',
European Journal of Marketing 33, 163-195(33).
Nunnally, J. (1967), Psychometric theory, Tata McGraw-Hill.
Storey, V., Straub, D., Stewart, K. & Welke, R. (2000), „A conceptual investigation
of the e-commerce industry”, Communications of the ACM 43(7), 117-123.
Straub, D.; Boudreau, M. & Gefen, D. (2004), 'Validation guidelines for IS positivist
research', Communications of the Association for Information Systems 13(24),
380--427.
Waltz, C. & Bausell, R. (1981), Nursing research: Design, statistics, and computer
analysis, FA Davis Company.
Waltz, C.; Strickland, O. & Lenz, E. (2010), Measurement in nursing and health
research, Springer Publishing Company.
This paper is supported by the Sectoral Operational Programme Human Resources Development (SOP
HRD), financed from the European Social Fund and by the Romanian Government under the contract
number POSDRU/89/1.5/S/56815.
... To ensure the reporting of credible research findings, consideration was given to the rigorous development of the survey 12 . It is well understood and documented that drawing accurate and valid inferences from research findings relies upon the instrument's validity, in this case, the survey and that in the absence of rigorous development, the credibility of reported results may be disputed 12,13,14,15,16 . ...
... The use of mixed methods research, in which qualitative data informs item generation, for developing an appropriate survey has been documented by authors such as Onwuegbuzie and colleagues 17 . Using qualitative findings to inform item generation enhances the survey's content validity 11,13,[16][17][18][19] . Using qualitative findings enhances the value of generated items as they are obtained from the population of interest and reflect their opinions, perspectives, perceptions and experiences 11,16,18 . ...
... The evaluation of generated items by individuals who are part of the target population and who are considered knowledgeable in the field is documented as a further means to enhance the overall validity of a survey 11,13,18,[20][21][22] . Qualitative feedback on the acceptability of a survey is a means to ensure face validity 23 , which has been described as the extent to which the items appear relevant, important, and interesting to the respondent 17 (p.60) . ...
Article
Full-text available
BACKGROUND: Wound management is internationally recognised as part of hand therapy practice. However, the role of occupational therapists in this area of upper limb rehabilitation in South Africa is unclear AIM: To develop and validate a survey to examine occupational therapists' wound management practices in hand therapy within the South African context METHODS: A sequential exploratory mixed methods study design was utilised to develop a survey. Thereafter, a panel of 11 occupational therapists with relevant experience in the field rated the relevance of survey items to produce item-content validity indices (I-CVIs RESULTS: An initial survey containing 27 questions containing 214 items was presented for review. Experts rated 171 items (69%) relevant with an I-CVI of 0.90-1.00. A further 32 items (13%) were rated relevant with an ICVI of >0.80 <0.90. A total of 45 items (18%) were irrelevant with I-CVIs <0.80. The final survey consisted of 19 questions containing 139 items, excluding those that obtained relevant demographic data CONCLUSIONS: The survey demonstrated good content and face validity but is limited to use with occupational therapists in South Africa. Validation for use with different populations in different settings is recommended Implications for practice This article offers researchers within the field of occupational therapy methodology for the development and validation of a survey. Suggestions for improving the overall validity of the survey are given. The survey may be used in future studies with occupational therapists providing hand injury care in South Africa. Use of the survey with different populations requires validation using the intended target population.
... Scale development and validation is a rigorous process and a crucial part of the social, behavioral, and health sciences [64]. Scales are developed to capture a latent construct, and several steps are required to ensure that an instrument is both valid and reliable and to measure the intended complex phenomena [64,65]. Researchers are not uncommon to bypass the exploratory process and direct their efforts toward the confirmatory stage [64]. ...
Article
Full-text available
Background: Mirroring is a specific phenomenon that occurs in groups. People observe others enacting different roles in given situations, whereby concretizing the role of the observing aspect of their self as if looking in a mirror. Individuals could acquire emotional, cognitive, imaginative, behavioral, and interpersonal experiences through this process of mirroring; that is, "mirror effects. " Given that there are very few comprehensive scales capable of evaluating mirror effects, the Mirror Effects Inventory (MEI) was developed to measure experiences after participating in counseling education. This study describes the development of the MEI based on an interpersonal perspective and aimed to validate this scale. Methods: The development and validation of the MEI involved a seven-stage process that utilized both qualitative and quantitative methods. Former students who studied counseling (N = 24) were surveyed via semi-structured telephone interviews. A subjective perceived scale (the MEI) was then constructed. The MEI was administered to part-time university students (N = 221) studying counseling courses. Psychometric tests were performed to examine its validity and reliability. Validation of the MEI included examining correlations between three types of mirror effects and age differences and educational attainment. Results: The PCA revealed three types of mirror effects: general, positive, and negative. These components accounted for 53.82% of the total variance. Reliability estimates measured using Cronbach's alpha of .88 indicated that the items were internally consistent. The findings showed that as the general mirror effect increased, there was a corresponding increase in the positive mirror effect and a decrease in the negative mirror effect. However, an increase in positive mirror effects was associated with an increase in negative mirror effects. Conclusions: The development of a reliable and valid measure of mirror effects has benefits and implications for counseling education and clinical practice. Further empirical research could contribute to resolving the controversies concerning the basic conceptual understanding of MEI.
... If the CVI>0.78 the item will be appropriate (Bertea & ZAIŢ, 2013). The number of experts that reviewed the questionnaire was 10 experienced researchers in the field of naturally occurring radioactive materials. ...
Article
Full-text available
The level of knowledge and awareness of naturally occurring radioactive materials and indoor radon among underground goldmine workers were measured using a well-structured closed ended five-point likert scale questionnaire using a one sample T test for the analysis. There was a significant relationship between the awareness and knowledge of NORM and radon gas among the underground workers apart from the awareness and knowledge of the increased risk of cancer due to exposure of radon for workers who smoke. The level of awareness and knowledge is positively correlated with the level of education of respondents and negatively correlated to the years of service and age of respondents, but the ANOVA table indicates a statistical significance between the demographic data and the awareness and knowledge of radon gas and NORM in their working environment with an adjusted R 2 of 46.1%.
... To ensure the validity and reliability of measurement items in the questionnaire, an initial version was sent to a panel of five transport experts in Feb 2022 for pretesting. The experts were requested to assess the relevance of each measurement item to its respective construct using a 4-point rating scale (1 -not relevant, 2 -somewhat relevant, 3 -quite relevant, 4 -very relevant) (Bertea and Zaiţ, 2013). The majority of measurement items received agreement from the experts, with some comments provided regarding duplicated items, wording, grammar, and sentence structure. ...
Article
Motorcycle riding is the principal mode of travel in low-and middle-income countries (LMICs). Motorcycle-based transport is regarded as highly problematic today due to reliance on fossil-based fuels, which contributes to climate change. Additionally, emissions from fuel combustion release air pollutants in urban areas, resulting in public health issues. Unfortunately, in LMICs, functioning and widespread public transport infrastructure will not be feasible in the near future due to economic and social tensions. As such, encouraging a mode shift from conventional motorcycles to electric motorcycles (EMs), a more sustainable transport mode, is a reasonable strategy for many jurisdictions. However, to develop effective programs to increase the adoption of EMs, it is necessary to have a better understanding of motives and barriers that determine the switching intention to EMs. To cover this gap, this investigation aims to develop a conceptual research model based on the Behavioural Reasoning Theory to explore the antecedents of the switching intention. Data from 751 conventional motorcycle users in Vietnam were used to test the model using partial least squares structural equation modeling (PLS-SEM). The moderating roles of trust and knowledge were also examined in this study. The results showed that 'reasons for' have no direct effect on the switching intention; however, this variable affects the intention indirectly via global motives. In contrast, 'reasons against' show a negative direct impact on the intention, and no indirect effect is found. The findings provide interesting insights for governments and the electric vehicle industry.
... lack of foreign market knowledge and information (LFMKI); lack of international business experiences (LIBE); declining internationalisation of women entrepreneurship (DIWE) adapted from past studies related to women entrepreneurs and international trade context. We utilised the Q-sorting procedure for assessing convergent and discriminant validity between items and construct due to the exploratory nature of the study (Bertea and Zait, 2013). Cohen's Kappa (Cohen, 1960) and "Hit Ratio" (Moore and Benbasat, 1991) were used to assess convergent and discriminant validity. ...
Article
Full-text available
Internationalisation of women-owned firms is considered a new strategy for unlocking the full economic potential of a country. However, there has been a desultory work on gender and trade, particularly the epistemological and methodological approaches to explore this phenomenon, which has received inadequate attention. Drawing on the institution-based- theory and incorporating the feminist perspective, this study identifies barriers to socio-economic freedom as a multidimensional concept that influences the internationalisation of women entrepreneurship. Using mixed methods and based on findings from focus group discussion (study-1), in-depth interviews (study-2) and a survey (study-3) of SME women entrepreneurs, the study develops and validates a gender-specific model of barriers to socio-economic freedom that women entrepreneurs face in the internationalisation of their firms. The findings confirm that the additional barriers that women entrepreneurs face emerge from three dimensions (i.e. socio-interactional patterns, socio-psychological attachment and socio-cultural embeddedness) that influence the internationalisation of women-owned SMEs through the lack of foreign market knowledge and information and lack of international business experience. This study extends internationalisation and women entrepreneurship research by identifying novel dimensions of the barriers to socio-economic freedom landscape. Implications of our findings for theory, methods and practice with limitations and future research directions are discussed critically.
... To evaluate the content validity of the measurement scales, we invited three academics with expertise in T&H, and five industry experts who are human resource managers in resorts and travel agencies to review and evaluate the instrument. These experts were provided with the conceptualizations of constructs in the research and asked to assess each item's relevance based on a four-point rating scale with 1 anchored as not relevant and 4 anchored as very relevant (Bertea and Zaiţ, 2013). This review indicated that most academic and industry experts agreed on the scales to measure all six constructs with some items being amended. ...
Article
By applying the protection motivation theory, the study aims to investigate factors influencing tourism and hospitality (T&H) workers' career resilience when faced with health-related risks at work during a pandemic. Data were collected from 495 part-time and full-time employees in the Vietnamese tourism and hospitality sector. The study found that workers' perceived vulnerability and perceived severity of the pandemic were positively associated with career resilience. Perceived severity was positively related to self-efficacy and response efficacy, while perceived vulnerability was positively associated with self-efficacy only. Both self-efficacy and response efficacy positively influenced career resilience through the mediating role of health risk preventative behavior. Theoretically, the study advances the stream of research in resilience in general and career resilience in particular among T&H workers when faced with a health-related crisis. Practical implications are provided with recommendations on how to facilitate career resilience among T&H employees working in the current high-risk environment.
Article
Purpose: The reliability of test scores is the extent to which they are consistent across different occasions of testing, different editions of the test, or different raters scoring the test taker’s responses. The purpose of this study is to assess the various approaches in determining the reliability of research instruments in management sciences research. Design/methodology/approach: The study used an exploratory research technique and relied on information from previous studies and publications, including journals, textbooks, periodicals, and the internet. Findings: Consequence upon several articles reviewed on the subject matter by different researchers on reliability of research instrument, it was observed that some scholars were able to test and measure data credibility through different modes such as internal consistency, inter rater, alternate form and reliability coefficient. Practical implications: The paper explored all of the pertinent concerns surrounding quantitative research instrument reliability and reviewed test reliability which include but not limited to: “alternate-forms reliability,” “inter-rater reliability,” “internal consistency,” “reliability coefficient,” “classification consistency,” with illustrations. Originality/value: Popular and commonly used reliability assessment approaches in Nigeria and in the field of management are the use of Cronbach alpha and Test-retest reliability tests for instrument reliability. Despite these, there are different types of reliabilities which are less reported in the field of management in the Nigeria academia. Based on this, the study reviewed various approaches and types of reliability test commonly utilised in Management sciences. Keywords: Research Instrument, Reliability, Alternate-forms Reliability, Inter-rater Reliability, Internal Consistency. JEL CODE: M30.
Book
Full-text available
Article
Full-text available
Reviews the literature on consumer-perceived risk over the past 30 years. The review begins by establishing perceived risk’s relationship with related marketing constructs such as involvement and trust. It then tackles some debates within the literature, concerning subjective and objective risk and differences between the concepts of risk and uncertainty. It describes how different models have been devised and operationalised to measure risk and how these have developed over the years. Aims to identify and report the theoretical and model developments over the past 30 years and to propose criteria which researchers can use in deciding the most useful model for their own research. The criteria are: understanding, prediction, suitability for reliability and validity assessment, practicality and usability. It is suggested that the basic two-component model is still the most generally useful for researchers and practitioners alike.
Article
Full-text available
The present study analyzes the influence that perceived risk in online shopping has on the process of e-commerce adoption by end consumers. With this aim, the Technology Acceptance Model is taken as a reference framework, proposing an Extended E-Commerce Acceptance Model that includes the diverse constructs of perceived risk: financial, performance, social, time, psychological and privacy. Empirical evidence is obtained from two samples, one is composed by Internet users with no experience in web shopping and the other is formed by online buyers. The results obtained confirm that the intention to shop through the Internet is positively influenced by general attitude toward the system and negatively influenced by the risk associated with the Web. Regarding the importance of the risk dimensions considered in the study, the economic and performance facets are the ones that have a greater influence on e-commerce adoption, while social and time dimensions are the less relevant.
Article
Full-text available
This article examines the definition, importance, conceptual basis, and functional nature of content validity, with an emphasis on psychological assessment in clinical situations. The conditional and dynamic nature of content validity is discussed, and multiple elements of content validity along with quantitative and qualitative methods of content validation are reviewed. Finally, several recommendations for reporting and interpreting content validation evidence are offered. (PsycINFO Database Record (c) 2012 APA, all rights reserved)
Article
Full-text available
The issue of whether IS positivist researchers were validating their instruments sufficiently was initially raised fifteen years ago. Rigor in IS research is still one of the critical scientific issues facing the field. Without solid validation of the instruments that are used to gather data on which findings and interpretations are based, the very scientific basis of the profession is threatened. This study builds on four prior retrospectives of IS research that conclude that IS positivist researchers continue to face major barriers in instrument, statistical, and other forms of validation. It goes beyond these studies by offering analyses of the state-of-the-art of research validities and deriving specific heuristics for research practice in the validities. Some of these heuristics will, no doubt, be controversial. But we believe that it is time for the IS academic profession to bring such issues into the open for community debate. This article is a first step in that direction. Based on our interpretation of the importance of a long list of validities, this paper suggests heuristics for reinvigorating the quest for validation in IS research via content/construct validity, reliability, manipulation validity, and statistical conclusion validity. New guidelines for validation and new research directions are offered.
Article
Enhance the quality of survey results by recognizing and reducing measurement errors. Margins of Error: A Study of Reliability in Survey Measurement demonstrates how and hwy identifying the presence and extent of measurement errors in survey data is essential for improving the overall collection and analysis of the data. The author outlines the consequences of ignoring survey measurement errors and also discusses ways to detect and estimate the impact of these errors. This book also provides recommendations of improving the quality of survey data. Logically organized and clearly written, this book: • Deconstructs the data gathering process into six main elements of the response process: question adequacy, comprehension, accessibility, retrieval, motivation, and communication • Provides an exhaustive review of valuable reliability estimation techniques that can be applied to survey data • Identifies the types of questions and interviewer practices that are essential to the collection of reliable data • Addresses hypotheses regarding which survey questions, sources of information, and questionnaire formats produce the most reliable data. In conjunction with research data gathered on nearly 500 survey measures and the application of an empirical approach grounded in classical measurement theory, this book discusses the sources of measurement error and provides the tools necessary for improving survey data collection methods. Margins of Error enables statisticians and researchers in the fields of public opinion and survey research to design studies that can detect, estimate, and reduce measurement errors that may have previously gone undetected. This book also serves as a supplemental textbook for both undergraduate and graduate survey methodology courses.
Article
A critical element in the evolution of a fundamental body of knowledge in marketing, as well as for improved marketing practice, is the development of better measures of the variables with which marketers work. In this article an approach is outlined by which this goal can be achieved and portions of the approach are illustrated in terms of a job satisfaction measure.
Article
CIVIL rights legislation, the attendant actions of compliance agencies, and a few landmark court cases have provided the impetus for the extension of the application of content validity from academic achieve- ment testing to personnel testing in business and industry. Pressed by the legal requirement to demonstrate validity, and constrained by the limited applicability of traditional criterion-related methodologies, practitioners are more and more turning to content validity in search of solutions. Over time, criterion-related validity principles and strate- gies have evolved so that the term, "commonly accepted professional practice" has meaning. Such is not the case with content validity. The relative newness of the field, the proprietary nature of work done by professionals practicing in industry, to say nothing of the ever present legal overtones, have predictably militated against publication in the journals and formal discussion at professional meetings. There is a paucity of literature on content validity in employment testing, and much of what exists has eminated from civil service commissions. The selectipn of civil servants, with its eligibility lists and "pass-fail" con- cepts, has always been something of a special case with limited trans- ferability to industry. Given the current lack of consensus in profes- sional practice, practitioners will more and more face each other in adversary roles as expert witnesses for plaintiff and defendant. Until professionals reach some degree of concurrence regarding what con- stitutes acceptable evidence of content validity, there is a serious risk that the courts and the enforcement agencies will play the major determining role. Hopefully, this paper will modestly contribute to the improvement of this state of affairs (1) by helping sharpen the content ' A paper presented at Content Validity (1, a conference held at Bowling Green
Article
A general formula (α) of which a special case is the Kuder-Richardson coefficient of equivalence is shown to be the mean of all split-half coefficients resulting from different splittings of a test. α is therefore an estimate of the correlation between two random samples of items from a universe of items like those in the test. α is found to be an appropriate index of equivalence and, except for very short tests, of the first-factor concentration in the test. Tests divisible into distinct subtests should be so divided before using the formula. The index [`(r)]ij\bar r_{ij} , derived from α, is shown to be an index of inter-item homogeneity. Comparison is made to the Guttman and Loevinger approaches. Parallel split coefficients are shown to be unnecessary for tests of common types. In designing tests, maximum interpretability of scores is obtained by increasing the first-factor concentration in any separately-scored subtest and avoiding substantial group-factor clusters within a subtest. Scalability is not a requisite.