ArticlePDF Available

Likert Scale: Explored and Explained

Authors:

Abstract and Figures

Likert scale is applied as one of the most fundamental and frequently used psychometric tools in educational and social sciences research. Simultaneously, it is also subjected to a lot of debates and controversies in regards with the analysis and inclusion of points on the scale. With this context, through reviewing the available literature and then clubbing the received information with coherent scientific thinking, this paper attempts to gradually build a construct around Likert scale. This analytical review begins with the necessity of psychometric tools like Likert scale andits variants and focuses on some convoluted issues like validity, reliability and analysis of the scale.
Content may be subject to copyright.
_____________________________________________________________________________________________________
*Corresponding author: E-mail: drankurjoshi7@gmail.com;
British Journal of Applied Science & Technology
7(4): 396-403, 2015, Article no.BJAST.2015.157
ISSN: 2231-0843
SCIENCEDOMAIN international
www.sciencedomain.org
Likert Scale: Explored and Explained
Ankur Joshi
1*
, Saket Kale
2
, Satish Chandel
3
and D. K. Pal
1
1
Department of Community Medicine, Gandhi Medical College, Bhopal, 462001,
India.
2
Technical Support Unit, Madhya Pradesh State AIDS Control Society, India.
3
Department of Pharmacology, All India Institute of Medical Sciences, Bhopal,
India.
Authors’ contributions
This work was carried out in collaboration between all authors. Author AJ initiated the idea, wrote the
first draft and contributed in further refinement with critical inputs to literature review. Author SK
conceptualized the variation of Likert scale and provided critical input to the several drafts of
manuscript and literature review. Author SC contributed for common understanding of psychometrics
with critical inputs. Author DKP reviewed and facilitated the final shape of paper. All authors read and
approved the final manuscript.
Article Information
DOI: 10.9734/BJAST/2015/14975
Editor(s):
(1)
Meng Ma, Anhui University, Hefei, Anhui, China and Icahn Institute for Genomics and Multiscale Biology, Icahn School of
Medicine at Mount Sinai, New York, USA.
Reviewers:
(1)
Anonymous, USA.
(2)
Adalberto Campo-Arias, Faculty of Researches and Publications, Human Behavioral Research Institute, Bogota, Colombia.
(3)
David Magis, Department of Education, University of Liège, Belgium.
(4)
Anonymous, Croatia.
(5)
Anonymous, Canada.
Complete Peer review History:
http://www.sciencedomain.org/review-history.php?iid=773&id=5&aid=8206
Received 30
th
October 2014
Accepted 27
th
January 2015
Published 20
th
February 2015
ABSTRACT
Likert scale is applied as one of the most fundamental and frequently used psychometric tools in
educational and social sciences research. Simultaneously, it is also subjected to a lot of debates
and controversies in regards with the analysis and inclusion of points on the scale. With this
context, through reviewing the available literature and then clubbing the received information with
coherent scientific thinking, this paper attempts to gradually build a construct around Likert scale.
This analytical review begins with the necessity of psychometric tools like Likert scale andits
variants and focuses on some convoluted issues like validity, reliability and analysis of the scale.
Keywords: Psychometrics; Likert scale; points on scale; analysis; education.
Opinion Article
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
397
1. INTRODUCTION
Nothing is more than a fear you cannot name.
― Cornelia Funke, Inkheart
Since the inception of human race there is an
inclination to capture the ethereal attributes of
human behaviour and performance.
Simultaneously, it has been a challenge from the
same time to quantify the thing which cannot be
measured through conventional measurement
techniques. The perceived need of this
quantification lies in the necessity to transform an
individual's subjectivity into an objective reality.
Attitude, perceptions and opinions are such
qualitative attributes amenable for quantitative
transformation due to above mention reason.
Qualitative research techniques do try to
compensate, by depicting the complexity of
human thoughts, feelings and outlooks through
several social science techniques, still the
quantification of these traits remains a
requirement and that’s how psychometric
techniques come into picture.
2. PSYCHOMETRICS AND LIKERT SCALE
Psychometrics techniques are being developed,
instituted and refined in order to meet the
quantification of traits like ability, perceptions,
qualities and outlooks- the requirement of social
sciences and educational researches [1,2].
Psychometrics operates through two ways; the
first is to formulate approaches (theoretical
construct) for measurements, followed by
development of measuring instruments and their
validation. Stanford Binet test (measures human
intelligence) and Minnesota Multiphasic
Personality Inventory (measures human
personality) are the example for the same. The
content in such instruments are rather ‘pre-fixed’
[3,4,5]. The another path is same up to
formulation of theoretical construct for the
measurement. This conceptualization is followed
by operational assembly of abstract
ideas/experiences/issues under investigation into
some statements (items) largely guided by the
aim of the study. This permits the contents
(items) in such scales/models to be rather
flexible and need based. Rasch measurement
model (use for estimation of ability), Likert scale
(measures human attitude) are the examples of
such scales in Psychometrics used widely in the
social science & educational research [3,4,5].
Likert scale was devised in order to measure
‘attitude’ in a scientifically accepted and validated
manner in 1932 [6,7]. An attitude can be defined
as preferential ways of behaving/reacting in a
specific circumstance rooted in relatively
enduring organization of belief and ideas (around
an object, a subject or a concept) acquired
through social interactions [8]. This is clear from
this discourse mentioned above that thinking
(cognition), feeling (affective) and action
(psychomotor) all together in various
combination/permutation constitute delivery of
attitude in a specified condition. The issue is how
to quantify these subjective preferential thinking,
feeling and action in a validated and reliable
manner: a help is offered by Likert scale [9,10].
The original Likert scale is a set of statements
(items) offered for a real or hypothetical situation
under study. Participants are asked to show their
level of agreement (from strongly disagree to
strongly agree) with the given statement (items)
on a metric scale. Here all the statements in
combination reveal the specific dimension of the
attitude towards the issue, hence, necessarily
inter-linked with each other [11].
With this context, this exploratory article attempts
to describe two confusing issues related with
Likert scale- (would be) preferable numbers of
points on a scale and analysis of the scale.
During one of the contributing authors’
participation in a web based conversational
learning forum on medical education. These two
issues emerged as thrust area amenable for
further exploration and lucid explanation for the
educational researchers. An initial literature
searched by authors led to aggregation of mutual
conflicting evidences which compelled us to re-
explore and further construct arguments based
upon accumulated knowledge.
3. LIKERT SCALE AND ITS VARIATION
Before proceeding further, let’s have a brief look
on several constructional diversities of a Likert
scale as the analytical treatment and
interpretation with Likert scale largely depends
upon these diversities.-Symmetric versus
asymmetric Likert scale- If the position of
neutrality (neutral/don't know) lies exactly in
between two extremes of strongly disagree (SD)
to strongly agree (SA), it provides independence
to a participant to choose any response in a
balanced and symmetric way in either directions.
This construction is known as symmetric scale.
On the other hand, asymmetric Likert scale offer
less choices on one side of neutrality (average)
as compared to other side. Asymmetric scale in
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
398
some cases also indicatesipsative (forced)
choices where there is no perceived value of
indifference/neutrality of the researcher
[12, 13,14].
Seven /ten point scale - They are the variation of
5 point scale in which adjacent options are less
radically different(or more gradually different)
from each other as compare to a 5 point scale.
This larger (step by step) spectrum of choices
offers more independence to a participant to
pick the exact’ one (which he prefers most)
rather than to pick some ‘nearby’ or ‘close’ option
[15]. These variations are discussed in more
details (in reference with validity and reliability)
further in this paper.
Likert and Likert type scale- The construction of
Likert (or Likert type) scale is rooted into the aim
of the research Sometimes the purpose of the
research is to understand about the
opinions/perceptions of participants related with
single ‘latent’ variable (phenomenon of interest)
.This ‘latent’ variable is expressed by several
‘manifested’ items in the questionnaire. These
constructed items in a mutually exclusive manner
address a specific dimension of phenomenon
under inquiry and in cohesion measure the whole
phenomena. Here during analysis, the scores of
the all items of the questionnaire are combined
(sum) to generate a composite score, which
logically in totality measures anuni-dimensional
trait. This instrument is known as Likert scale.
Sometimes the primary interest of the researcher
is not to synthesize the stance of the participants
per se but to capture feelings, actions and
pragmatic opinion of the participants about
mutually exclusive issues around phenomenon/s
under study. This fact demands the individual
analysis of item to ascertain the participants’
collective degree of agreement around that
issue. The scale used so can be labeled as Likert
type and not Likert scale [16]. A word of caution;
this ‘direction of enquiry’ must be decided during
the planning phase and at least during the
designing of questionnaire and not at the time of
analysis.
4. IS 7 POINT LIKERT SCALE BETTER
THAN 5 POINT LIKERT SCALE? - A
PERSPECTIVE CONTROVERSY OR
ESTABLISHED WITH A CONSENSUS?
Since the advent of Likert scale in 1932, there
have been debates among the users about its
best possible usability in term of reliability and
validity of number of points on the scale [17-20].
Likert (1932,7) in his original paper, discussed
about the infinite number of definable attitudes
existing in a given person with possibility of
grouping them into “clusters” of responses. He
further conversed about the assumption of his
“survey of opinions” on which he provided his
results and psychological interpretations [21].
The key assumptions of his survey being firstly,
the presentation of item on scale are such that,
so as to allow the participants to choose clearly
opposed alternatives. Secondly, the conflicting
issues chosen were empirically important issues
thus, results themselves constituting an empirical
check on the degree of success.
Thus, it is argued in particular context of
clustering of attitudes. Considering reliability of
the responses from participants in a survey,
chances are that the 7 point scale may perform
better compared to 5 point scale owing to the
choice of items on scale defined by the construct
of survey. The 7 point scale provides more
varieties of options which in turn increase the
probability of meeting the objective reality of
people. As a 7-point scale reveals more
description about the motif and thus appeals
practically to the “faculty of reason” of the
participants [19,20].
A respondents’ absolute agreement with the
motif of topic may lie between the two descriptive
options provided on a 5 point scale. On repeated
administration, he/she may differ in choosing one
of the options, e.g. 3 instead of 4 when the
person thinks in between the two of the response
options on 5 point scale. A 7 point scale may
eliminate this problem up to an extent, by eliciting
retrieval beyond the utmost level of agreement
provided by a 5 point scale, the dilemma of
choosing between the two undesirable points on
5 point. Hence this dilemma of forced choosing
between two equally undesirable point imposed
by the 5-point Likert scale may be addressed up
to a extent by offering more choices (in between)
by a 7-point scale [22-24]. The provision of
number of scale points, 5 point or 7 point, would
be more engaging to the minds of respondents
when the items on the scale carry the statement
of ideas near the truth of the universe for both
the participants and the surveyor. It may create
the ‘curves of reliability’ around the ‘zenith of
validity’. The dilemma of choice and explicit
greater extent of measurement by 7 point scale
is very much in the territory of the reason of
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
399
response without which consideration of
reliability is of no weight [19].
Validity of Likert scale is driven by the
applicability of the topic concerned; in context of
respondents’ understanding and judged by
creator of the response item. We can appreciate
it by an example: “How efficacious is a
therapeutic modality in treating a particular
disease?” This question when asked to a group
of individuals, indifferent with the disease or the
modality, the response pattern may remain
similar, independent of the number of point on
the scale. The responses may cluster around
center or to the extreme ends. On the contrary,
when the topic concerned is relevant to the
respondents’ context provision of more option,
may add to the content & construct validity of the
scale. Providing options more close to the
original view of the respondent reduce the role of
ambiguity in the responses [23,12]. Furthermore,
comprehension of all items and points on a scale
needs a judgment time and a memory span
different for different means and also depends on
communication mode. While listening to the
responses of a long scale may discern the
various options on the scale with lesser time to
judge compared to a written scale. Written scale
thus will add to validity even with more points on
the Likert scale. Also research concerning span
of immediate memory support this notion of
accuracy of response categories around seven,
as human mind has span of absolute judgment
that can distinguish 7 categories at a time [25].
5. ANALYSIS OF THE ITEM RESPONSE
Before we proceed to the method of analysis
available to Likert scale, a very fundamental but
equally controversial question should be
addressed- which type of scale Likert is?
There are two schools of thoughts - One school
considers Likert scale as ordinal and other treats
it as Interval scale. This conflict is primarily
rooted into the question: whether points on a
items are equivalent and equidistant? Points on
scale are not close enough to consider them
equal (in other words strongly agree is definitely
away from agree and agree is away from
neutral), they should be considered as non-
equivalent entity. There is an agreement in both
schools for the above fact. The conflict arises on
asking another question: if the points on scale
are non –equivalent, are they equi-distant (in
other words is ‘neutral of same distance from
‘agree’ as ‘agree’ from ‘strongly ‘agree’)? This
question is important as by answering of this
question only, one can decide whether Likert
scale can be treated as Interval scale?
The first school of researchers and statisticians
consider Likert scale as ordinal scale. They
argue that choices or responses are arranged in
some ranking order. However, as this scale
doesn’t show the relative magnitude and
distance between two responses quantitatively, it
can’t be treated as interval scale. The other
school interprets this dilemma from a different
perspective, stating that when the aim of the
researcher is to ‘combine’ all the items in order to
generate a ‘composite’ score for an individual
rather than separate analysis of single item
responded by all individuals, then this
individualistic summative score (for all the items)
of a participant shows a sensible realistic
distance from the individual summative score of
another individual; hence, can be labeled as
‘interval estimates’ [26,16].
To understand this concept, let’s assume a
scenario in which the aim of the researcher is to
measure the attitude towards classroom lectures
and to make out relative preferences (library
reading and small group teaching) compared
with lecture. (Fig. 1) He designs the following
survey instrument on a 5 point Likert scale for the
stated aim-
The first question of importance is: ‘Can these
items be clubbed (see together) in order to
generate a composite index for measuring the
attitude?’ In order to evaluate their
appropriateness for transformation into a single
composite index, following points can be
considered-
1. Whether the items are arranged in logical
sequence?
2. Whether the items are closely interrelated
but provide some independent information
as well?
3. Whether there is some element of
‘coherence/expectedness’ between
responses (whether next response can be
predicted up to some extent based upon
previous one)?
4. Whether each item measures a distinct
element of the issue?
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
400
Fig. 1. Survey instrument for measuring attitude towards classroom lectures
Fig. 2. Choice of Analysis of Likert Items: Aim and Construct of Research
If answer to all the above questions is affirmative
for all the items of a set, they may be combined
to construct a composite index which measures
the collective stance of the participant towards
phenomenon under study. In the above example
as item 1, 2 and 3 fulfill all four criteria for each
other, they may be combined and can be treated
further in unison.
On the other hand, item-4 and item-5, offer
separate and sovereign (mutually exclusive)
preferences regarding two different teaching-
learning methods: self-directed reading and small
group teaching. Hence, they can’t be combined
and further they should be analyzed
independently from item 1, 2 and 3 and even
from each other.
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
401
After this assertion of eligibility for combination,
the next question arises- On what scale can item
1, 2 and 3 be treated and what is the appropriate
measurement scale for item4 and item-5?
The answer of the above question lies in another
question asked by Stevens in his famous paper:
‘what are the rules (if any) under which numerals
are assigned?’ Here we see (a) the minimum
score one can secure for first three items is 3
(and not an absolute zero). The reason for this
apparently dislike for zero lies in the fact that in
psychometrics, attitude is preferably measured in
positive degree and being the ‘strongly disagree
‘cannot be equated with ‘absolute disagreement’;
there is always something below than strongly
disagree. Zero also gives the notion of neutrality
rather disagreement (the attitude is zero; means
one is apathetic to issue) (b) Each numeral
conveys the same meaning in all three items (i.e.
3 denotes the neutral in all three items) (c) As
mentioned above, all three items can be clubbed
while satisfying the content and criterion validity.
This sentence needs a little more explanation.
The idea or concept behind framing item 1, 2 and
3 is to capture the opinion of participants about
the lecture. This theoretical construct how well
can be transformed into operating reality, can be
ascertained by looking at relevant content
domains (content validity/reflection of construct),
ability to distinguish opinion on lecture from other
teaching modality (concurrent validity) and
similarities among items 1 to 3 and dissimilarities
from item 4 and 5 (convergent and discriminant
validity). Concurrent, convergent and
discriminant validities are the domains of criterion
validity. Before deciding any statistical treatment
to items, all the items must be scrutinized for
validity issues.
If we look into point (a), (b) and (c) in cohesion
for the set of item 1, 2 and 3, that composite
score for the item-1, 2 and 3 can be compared
with another composite score for another
individual on an interval scale. A ‘rank-order’
among the composite scores can be presumed
as well as equality of interval among related
composite scores can also be postulated. The
specific point on a particular item is conveying
the same meaning for all individuals (for item -2
point 3 on Likert scale denotes ‘neutral’ among
all individuals.) Moreover a specific point (say 2
for disagree) is conveying the same meaning
(same extent of disagreement) in all the items
and there is no absolute zero in scale (minimum
achievable score is 3). From the discourse, this
can besafely assumed (after going through all
these mathematical characteristics with due
consideration of validity related issues) that the
obtained composite data for item 1, 2 and 3 for
all the participants can be treated on an interval
scale.
The truth has different dimension in case of item
4 and item 5. Item -4 and 5 being a mutually
exclusive observation from each other (opinion
on self-directed reading/ small group teaching)
and from item 1, 2 and 3 should be treated
differently. They may not be combined (validity
restriction) for an individual as they are nowhere
providing complementary observation.
Still item 4 and 5 can be treated on a certain
measurement scale. The arguments for this
assumption are –first, a specific point (say point -
4) for a particular item ( (say for item-4) conveys
the same meaning (agree) for all individuals
treated on that item and second, response
variables obtained for a single item from all the
individuals can be arranged in any order
preserving transformation (like square,
multiplication, square root etc.) to the response
variable(the rank order remains unaffected) ....
so an ordinal scale’s assumptions and treatment
is applicable on this subset of items (4 and 5).
Once it is clear that under which rules the items
are categorized and what the direction of inquiry
is, it becomes obvious that the further statistical
treatment as per their assignment into ordinal or
interval scale.
6. CONCLUSION
The crux that can be extracted from the above
inductive arguments and logical interpretation is
that the methods adopted for Likert scale
analysis largely depends on the item response
variable assignment into ordinal or interval scale
which in turn depends on the construct of the
research instrument. This construct of research
instrument can be derived from objectives of
study and objectives are the operational form of
theoretical construct of phenomenon under
inquiry. In other words, designing of instruments
based upon objectives and frameworks of study
decides further statistical treatment.
Hence if one wishes to combine the items in
order to generate a composite score (Likert
scale) of a set of items for different participants,
then the assigned scale will be an interval scale
(Fig. 2 above). The measures for central
tendency and dispersion for an interval scale are
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
402
mean and standard deviation. Further this data
set can be statistically treated with Pearsons’
correlation coefficient (r), Analysis of Variance
(ANOVA) and regression analysis.
As opposed to, if researcher wishes to analyze
separate item (no composite score; Likert type
scale), the assigned scale for such data set will
be ordinal (Fig. 2 above). Needless to say, the
recommended measure of central tendency and
dispersion for the ordinal data are the median (or
the mode) & frequency (or range). An ordinal
data set can further be statistically tested by non-
parametric techniques such as Chi-square test,
Kendall Tau B or C test.
Before wrapping up, it is imperative to transform
an abstract issue into figurative shape in order to
measure it up to best possible extent.
Simultaneously, this is an integrate process
reason being influenced by perspective and
subjectivity of researcher. Still all attempts should
be directed for quantification of such qualitative
attributes as -‘what get measured, get managed.’
(Peter Druker).
COMPETING
INTERESTS
Authors have declared that no competing
interests exist.
REFERENCES
1. Cook DA, Beckman TJ. Current concepts
in validity and reliability for psychometric
instruments: Theory and application. The
American Journal of Medicine. 2006;119
(2). DOI:10.1016/j.amjmed.2005.10.036.
2. Shea JA, Fortna GS. Psychometric
Methods. In: Norman GR, Van der Vleuten
CPM, Newble DI, editors. International
Handbook of Research in Medical
Education. Dordrecht: Kluwer Academic
Publisher. 2002;97-126.
3. Youngstrom EA, Glutiing JJ, Watkins MW.
Stanford-Binet intelligence scale: Fourth
Edition (SB4): Evaluating the empirical
bases for interpretations. In: Reynolds CR,
Kamphaus RW, editors. Handbook of
psychological and educational
assessment: Intelligence, aptitude, and
achievement. New York: Guilford press.
2003;217-242.
4. Wu M, Adams R. Applying the Rasch
model to psycho-social measurement:
A practical approach. Melbourne:
Educational Measurement Solutions; 2007.
5. Croasmun JT, Ostrom L. Using Likert-type
scales in the social sciences. Journal of
Adult Education. 2011;40:19-22.
6. Edmondson DR. Likert scales: A history.
Proceedings of the 12
th
Conference on
Historical Analysis and Research in
Marketing (CHARM). April 28-May1,
California, USA; 2005.
7. McLeod S. Likert Scale; 2014.
Simplypsychology.org.
Retrieved:http://www.simplypsychology.org
/Likert-scale.html/pdf
8. Park K. Social sciences and health. In Park
K, editor. Park Text Book of Preventive
and Social Medicine. Jabalpur: Bhanot
Publishers. 2013;54x-54x.
9. Culbertson JM. What is an attitude?
Journal of Cooperative Extension. 1963;
79-84.
10. Schwarz N, Bohner G. The construction of
attitudes. In: Tesser A & Schwarz N,
Editors. Blackwell Handbook of Social
Psychology. Oxford: Blackwell. 2001;436-
457.
11. Singh YK. Fundamental of Research
Methodology and Statistics. New Delhi:
Newage International (P) Ltd. Publisher;
2006.
12. Lobsy J, Wetmore A. CDC Coffee Break:
Using Likert Scales in Evaluation Survey
Work [Internet]; 2014. Cdc.gov.
Retrieved:www.cdc.gov/dhdsp/pubs/docs/
CB_February_14_2012.pdf
13. Tsang KK. The use of midpoint on Likert
scale: The implications for educational
research. Hong Kong Teachers Centre
Journal. 2012;11:121-130.
14. Malhotra NK. Questionnaire Design and
Scale Development. In: Grover R, Vriens
M, editors. The Handbook of Marketing
Research. California: Sage Publications,
Inc; 2006.
15. Dawes J. Do data characteristics change
according to the number of scale points
used? An experiment using 5-point, 7-point
and10-point scales. International Journal of
Market Research. 2008;50(1):61-77.
16. Boone HN, Boone DA. Analyzing Likert
Data. Journal of Extension. 2012;50(2).
Retrieved:http://www.joe.org/joe/2012april/t
t2.php
17. Colman AM, Norris CE, Preston CC.
Comparing rating scales of different
lengths: Equivalence of scores from 5-
point and 7-point scales. Psychological
Reports. 1997;80:355-362.
Joshi et al.; BJAST, 7(4): 396-403, 2015; Article no.BJAST.2015.157
403
18. Preston CC, Colman AM. Optimal number
of response categories in rating scales:
reliability, validity, discriminating power,
and respondent Preferences. Acta
Psychologica. 2000;104;1-15.
19. Chang L. A psychometric evaluation of 4-
point and 6-point Likert-type scale in
relation to reliability and validity. Applied
Psychological Measurement. 1994;18:205-
215.
20. Cox EP. The optimal number of response
alternate for a scale: A review. Journal of
Marketing Research. 1980;17(4):407-422.
21. Likert R. A technique for the
measurements of attitudes. Archives of
psychology. 1932;140(22):5-55.
22. Pearse N. Deciding on the scale
granularity of response categories of Likert
type scales: The Case of a 21-Point Scale.
Electronic Journal of Business Research
Methods. 2011;9(2):2011.
23. Finstad K. Response interpolation and
scale sensitivity: Evidence against 5-point
scales. Journal of Usability Studies.
2010;5(3):104-110.
24. Komorita SS and Graham WK. Number of
scale points and the reliability of scales.
Educational and Psychological
Measurement. 1965;25(4):987-995.
25. Miller GA. The magical number seven, plus
or minus two: Some limits on our capacity
for processing information. The
Psychological Review. 1956;63:81-97.
26. Carifio J, Perla RJ. Ten common
misunderstandings, misconceptions,
persistent myths and urban legends about
Likert scales and Likert response formats
and their antidotes. Journal of Social
Sciences. 2007;3(3):106-116.
_________________________________________________________________________________
© 2015 Joshi et al.; This is an Open Access article distributed under the terms of the Creative Commons Attribution License
(http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium,
provided the original work is properly cited.
Peer-review history:
The peer review history for this paper can be accessed here:
http://www.sciencedomain.org/review-history.php?iid=773&id=5&aid=8206
... To evaluate the usability of Build_PC, a survey based on the IBM-CSUQ tool was used, which employs a 7-point Likert scale, where 1 indicates strong disagreement or very unlikely, and 7 indicates strong agreement or very likely. [61,62]. This questionnaire consists of 19 questions designed to measure user satisfaction with the developed application [61]. ...
... [61,62]. This questionnaire consists of 19 questions designed to measure user satisfaction with the developed application [61]. The survey aims to gather data on various aspects, such as the ease of use of the system (SYSUSE), the quality of the information provided (INFOQUIAL), the quality of the interfaces (INTERQUIAL), and an overall evaluation of the application and its ease of use (OVERALL). ...
... The IBM-CSUQ tool is widely used and recognized in the field of usability, it also guarantees in this research, effectiveness, ease of use, wide coverage, standardization, and adaptability [61]. ...
Article
Full-text available
The proliferation of mobile applications for educational purposes has highlighted the need to evaluate their usability, especially in diverse international contexts. This study addresses the problem of insufficient engagement and effectiveness of educational tools related to PC hardware training, a problem exacerbated by cultural and contextual differences between regions. Understanding the importance of this issue is crucial, as effective educational tools can improve learning outcomes on a global scale. Previous research has explored various educational technologies but often failed to comprehensively address usability across different cultural contexts, limiting the generalization and impact of the results. This gap underscores the need for robust evaluation of educational applications in diverse populations. In this context, our research proposes the analysis of Build_PC, a mobile augmented reality (MAR) application designed to teach PC hardware, using the IBM Computer System Usability Questionnaire (CSUQ) to assess user satisfaction. This study was conducted in three universities from three countries—Ecuador, Indonesia, and Lebanon—covering a variety of cultural and educational settings. The results indicate remarkably high levels of user satisfaction with the augmented reality (AR) application across the three participating universities. Positive feedback suggests that the application effectively engages students and improves their understanding of PC hardware training, regardless of regional differences. The implications of these findings are significant, as they suggest that augmented reality applications may be a viable solution for overcoming educational barriers related to PC hardware training on an international scale. This study highlights the potential of such technology to enhance educational outcomes and provides a framework for future research in the global deployment of educational technologies. Doi: 10.28991/ESJ-2025-09-02-024 Full Text: PDF
... The Likert scale was first mentioned by Rensis Likert in 1932 on his research article "A Technique for the Measurement of Attitudes" in the journal "Archive of Psychology" [25]. The Likert scale was designed to measure attitudes in a scientifically acceptable and validated way [26]. Likert in his research discussed the possibility of classifying an infinite number of attitudes of a person into groups of responses [26]. ...
... The Likert scale was designed to measure attitudes in a scientifically acceptable and validated way [26]. Likert in his research discussed the possibility of classifying an infinite number of attitudes of a person into groups of responses [26]. In applying the Likert scale, participants were asked to indicate their level of conformity with the statements given on the metric scale [26]. ...
... Likert in his research discussed the possibility of classifying an infinite number of attitudes of a person into groups of responses [26]. In applying the Likert scale, participants were asked to indicate their level of conformity with the statements given on the metric scale [26]. Likert stated that two polar choices on the scale needed to be assigned an exact value, while the choices between them were left without an exact value [27]. ...
Article
Sentiment analysis is the application of natural language processing which aims to identify the sentiment of texts. To carry out sentiment analysis, data which has been labeled sentiment is needed to be included in the training model. Crowdsourcing is considered as the most optimal method to label data because it has a high level of accuracy at a relatively low cost. However, the use of crowdsourcing platforms has its own challenge, which is to increase user interest and motivation. A solution which can be applied is to design and build a crowdsourcing platform or application using the gamification method. The definition of gamification is an effort to increase one's intrinsic motivation for an activity by applying game elements to it. Therefore, a cross-platform mobile based crowdsourcing application for sentiment labeling using gamification method was carried out. The gamification design process was done based on the 6D framework and the application was developed using the Ionic-React framework. Application was examined through black box testing and the result showed that the application was functioning properly and according to the design requirements. There was also an evaluation carried out by distributing Intrinsic Motivation Inventory questionnaires to users who had used the application for 2 weeks. From a total of 40 respondents, the result showed that the level of user motivation and interest in using the application was high with a percentage of 83.10%.
... To analyze the questionnaire data, we calculated the mean and standard deviation for each question using the Likert scale, where "strongly agree" corresponds to 5 and "strongly disagree" corresponds to 1 [42]. A mean score above 3 indicates general agreement or a positive response, while a mean below 3 reflects disagreement or negative feedback. ...
Preprint
Transforming educational technologies through the integration of large language models (LLMs) and virtual reality (VR) offers the potential for immersive and interactive learning experiences. However, the effects of LLMs on user engagement and attention in educational environments remain open questions. In this study, we utilized a fully LLM-driven virtual learning environment, where peers and teachers were LLM-driven, to examine how students behaved in such settings. Specifically, we investigate how peer question-asking behaviors influenced student engagement, attention, cognitive load, and learning outcomes and found that, in conditions where LLM-driven peer learners asked questions, students exhibited more targeted visual scanpaths, with their attention directed toward the learning content, particularly in complex subjects. Our results suggest that peer questions did not introduce extraneous cognitive load directly, as the cognitive load is strongly correlated with increased attention to the learning material. Considering these findings, we provide design recommendations for optimizing VR learning spaces.
... In this study, a five-point Likert scale survey was adopted and designed with statements related to various aspects of FlexiPavD as a teaching and learning tool. The five-point scale provides a compromise between detail and simplicity, by offering a neutral midpoint, giving respondents a balanced option between two opposing perspectives (Joshi et al., 2015). Moreover, it gives researchers enough data to conduct statistical analyses without overwhelming respondents (Revilla et al., 2014). ...
Article
Full-text available
This study presents the findings from a survey conducted to gather student feedback on the efficacy and user experience of a teaching and learning tool for flexible pavement design, namely FlexiPavD. In most Malaysian universities, the Highway and Traffic Engineering course syllabus includes the design of flexible pavements. The student needs to go through a manual (more than 30 pages) to assist them in designing a flexible pavement. This process may result in human error in addition to wasting money, time, and paper. With this in mind, we develop a tool using Microsoft Excel to help assist students in carrying out the design process anytime and anywhere from hp, tablet, laptop, or pc. However, no matter how well the tool is developed, it will serve no purpose if the student does not benefit from it. Therefore, the survey aimed to evaluate the tool's impact on students' learning process and overall satisfaction. Data were gathered from a diverse cohort of 50 undergraduate students through an online survey platform to ensure ease of access and involvement. The results indicate a generally positive reception of the tool, with 90% of respondents being satisfied with the design tool, and 10% feeling neutral. It is concluded that the tool is pertinent to the teaching and learning process based on the fact that 92% of students concur that FlexiPavD is helpful for their learning process in flexible pavement design.
... For each requirement, they were asked to rate four key items: Item 1: The requirement clearly references a component or system. It is important to note that the evaluation metrics presented here were developed based on an exhaustive search of papers related to requirements evaluation [144]. ...
Thesis
Full-text available
Defining system requirements in engineering design has always been challenging and complex. This research explores the potential for Large Language Models to support and enhance requirements development. A mixed-methods approach is employed to explore this potential, combining quantitative surveys and qualitative interviews with industry professionals who manage requirements. The original data set consisted of human-created system requirements, which were compared to AI-generated requirements that were assessed for completeness using four criteria: specificity, functionality, target values, and verifiable. The interviews provided valuable insights into current workflows and the common challenges faced in requirement definition, and also the potential benefits and limitations of AI solutions. The results indicated that AI-generated requirements can help make the process more manageable and act as a collaborative partner for human engineers. Although AI may miss some important details, there is still significant potential for its improvement to create models capable of accurately defining requirements.
... The use of the Likert scale enabled the quantification of subjective opinions and attitudes into measurable data, thereby enhancing the reliability and validity of the findings. Likert scales are widely recognized for their simplicity, ease of administration, and ability to generate interval-level data suitable for advanced statistical analysis (Joshi, Kale, Chandel, & Pal, 2015). ...
Article
Full-text available
Microfinance institutions (MFIs) have emerged as crucial agents in fostering financial inclusion, particularly in developing economies. This paper explores the role of MFIs in promoting financial inclusion in Kerala, India. Through a structured survey using a 5-point Likert scale-based questionnaire administered to a randomly selected sample of 300 beneficiaries, the study examines the extent of financial inclusion facilitated by MFIs. Data analysis was carried out using EDUSTAT software, employing descriptive statistics and hypothesis testing. Results indicate that MFIs significantly contribute to enhancing access to financial services, particularly among marginalized communities. The findings underline the need for strengthening the microfinance framework to ensure sustainable inclusion. This paper highlights the importance of strategic policy initiatives and recommends avenues for future development in the sector.
Article
The Controller Area Network (CAN) facilitates efficient communication among vehicle components. While it ensures fast and reliable data transmission, its lightweight design makes it susceptible to data manipulation in the absence of security layers. To address these vulnerabilities, machine learning (ML)-based intrusion detection systems (IDS) have been developed and shown to be effective in identifying anomalous CAN traffic. However, these models often function as black boxes, offering limited transparency into their decision-making processes, which hinders trust in safety-critical environments. To overcome these limitations, this paper proposes a novel method that combines Local Interpretable Model-agnostic Explanations (LIME) with a vision-language model (VLM) to generate detailed textual interpretations of an ML-based CAN IDS. This integration mitigates the challenges of visual-only explanations in traditional XAI and enhances the intuitiveness of IDS outputs. By leveraging the multimodal reasoning capabilities of VLMs, the proposed method bridges the gap between visual and textual interpretability. The method supports both global and local explanations by analyzing feature importance with LIME and translating results into human-readable narratives via VLM. Experiments using a publicly available CAN intrusion detection dataset demonstrate that the proposed method provides coherent, text-based explanations, thereby improving interpretability and end-user trust.
Article
Full-text available
This study examined how using Likert-type scales with either 5-point, 7-point or 10-point format affects the resultant data in terms of mean scores, and measures of dispersion and shape. Three groups of respondents were administered a series of eight questions (group n's = 300, 250, 185). Respondents were randomly selected members of the general public. A different scale format was administered to each group. The 5-and 7-point scales were rescaled to a comparable mean score out of ten. The study found that the 5-and 7-point scales produced the same mean score as each other, once they were rescaled. However, the 10-point format tended to produce slightly lower relative means than either the 5-or 7-point scales (after the latter were rescaled). The overall mean score of the eight questions was 0.3 scale points lower for the 10-point format compared to the rescaled 5-and 7-point formats. This difference was statistically significant at p = 0.04. In terms of the other data characteristics, there was very little difference among the scale formats in terms of variation about the mean, skewness or kurtosis. This study is 'good news' for research departments or agencies who ponder whether changing scale format will destroy the comparability of historical data. 5-and 7-point scales can easily be rescaled with the resultant data being quite comparable. In the case of comparing 5-or 7-point data to 10-point data, a straightforward rescaling and arithmetic adjustment easily facilitates the comparison. The study suggests that indicators of customer sentiment – such as satisfaction surveys – may be partially dependent on the choice of scale format. A 5-or 7-point scale is likely to produce slightly higher mean scores relative to the highest possible attainable score, compared to that produced from a 10-point scale.
Article
Full-text available
Using a self-administered questionnaire, 227 respondents rated service elements associated with a restaurant, retail store, or public transport company on several 5-point and 7-point rating scales. Least-squares regression showed that linear equations for estimating 7-point from 5-point and 5-point from 7-point ratings explained over 85% of the variance and fitted the data almost as well as higher-order polynomials and power functions. In a cross-validation on a new data set, the proportion of variance explained fell to about 76%. Functionally inverse versions of the derived linear equations were calculated for the convenience of researchers and psychometricians.
Article
A conceptual framework employing the distinction between stimulus-centered and subject-centered scales is presented as a basis for reviewing 80 years of literature on the optimal number of response alternatives for a scale. Concepts and research from information theory and the absolute judgment paradigm of psychophysics are used. The author reviews the major factors influencing the quality of scaled information, points out areas in particular need of additional research, and makes some recommendations for the applied researcher.
Article
This article provides information for Extension professionals on the correct analysis of Likert data. The analyses of Likert-type and Likert scale data require unique data analysis procedures, and as a result, misuses and/or mistakes often occur. This article discusses the differences between Likert-type and Likert scale data and provides recommendations for descriptive statistics to be used during the analysis. Once a researcher understands the difference between Likert-type and Likert scale data, the decision on appropriate statistical procedures will be apparent.
Article
Reliability and validity of 4-point and 6-point scales were assessed using a new model-based ap proach to fit empirical data. Different measurement models were fit by confirmatory factor analyses of a multitrait-multimethod covariance matrix. 165 gradu ate students responded to nine items measuring three quantitative attitudes. Separation of method from trait variance led to greater reduction of reliability and heterotrait-monomethod coefficients for the 6-point scale than for the 4-point scale. Criterion-related valid ity was not affected by the number of scale points. The issue of selecting 4- versus 6-point scales may not be generally resolvable, but may rather depend on the empirical setting. Response conditions theorized to in fluence the use of scale options are discussed to pro vide directions for further research. Index terms: Likert-type scales, multitrait-multimethod matrix, reli ability, scale options, validity.
Chapter
From the Introduction: A growing body of literature suggests that attitudes may be much less enduring and stable than has traditionally been assumed. ... self-reports of attitudes are highly context dependent and can be profoundly influenced by minor changes in question wording, question format, or question order. For some researchers, this malleability simply reflects measurement error ... For other researchers, the same findings indicate that all we assess in attitude measurement are evaluative judgments that respondents construct ... based on whatever information happens to be accessible (e.g. Schwarz & Strack, 1991). From this perspective, the traditional attitude concept may not be particularly useful and we may learn more about human cognition and behavior from a detailed analysis of the underlying judgmental processes. Other researchers have taken intermediate positions ... For example, Lord & Lepper (in press) and Tourangeau and his colleagues (e.g. Tourangeau, 1992) equate attitudes with relatively stable memory structures, but assume that individuals sample from these structures when they answer attitude questions. Hence, a stable attitude can result in variable attitude reports, depending on which aspect of the knowledge structure (attitude) is accessed. Others (e.g., Wilson, 1998) suggested that individuals may hold multiple attitudes about an object, accessing different ones at different points in time. As we illustrate below, it is surprisingly difficult to design conclusive empirical tests to evaluate the relative merit of these proposals ... Yet, a scientific concept like “attitude” is to be evaluated on the basis of its explanatory power – and without taking judgmental processes into account, there is little that the attitude concept explains. In fact, the contemporary definition of attitudes as “likes and dislikes” (Bem, 1970, p. 14) equates attitudes with evaluative judgments. Hence, the first section of this chapter highlights judgmental processes and the second section applies these process assumptions to some findings that are typically considered evidence for the enduring nature of attitudes. In response to the malleability of attitude reports, social psychologists have repeatedly tried to replace or supplement verbal self-report measures with other, presumably more direct, ways to assess individuals’ evaluative responses to attitude objects. These attempts range from the “bogus pipeline” (Jones & Sigall, 1971) of the 1970s to the recent development of sophisticated “implicit” measures of attitudes (e.g. Dovidio & Fazio, 1992). Recent findings suggest that such measures may be just as context dependent as verbal reports, although the relevant contextual variables may differ. The third section addresses these developments, which are discussed in more detail by Banaji and colleagues (Chapter 7, this volume) and Bassili (Chapter 4, this volume). Much as the enduring nature of attitudes has been called into question, another body of research suggested that attitudes may not be closely related to behavior either (see Wicker, 1969, for an influential early review). Instead, we may expect a close relationship between attitudes and behavior only under some specific, and relatively narrow, conditions (see Chapter 19, this volume). These conditions can be fruitfully conceptualized within a judgment perspective, as we review in the final section.