ArticlePDF Available

Abstract

At a first glance, the emergence of a new scientific journal in the ever-growing market of scientific journals may raise eyebrows. Would a new journal rather aid a solution or contribute to the “replicability crisis” recently acknowledged in some social scientific disciplines? Are there not already more scientific outlets than anyone can ever follow? And do not we all have enough items on our reading lists already? At a second and closer look, it becomes clear that if, in the past, effort had been devoted to developing high-quality measurement instruments, it often went unnoticed. Documenting such effort is not a typical goal of journals focusing on substantive research, because measurement is just perceived to be a vehicle to answer substantive questions. In addition, even if an instrument was published, it was most often only recognized and reused in its discipline of origin. A transfer of knowledge among disciplines, even though they aimed to study the same constructs, only seldom happens. A slight imbalance then needs to be addressed; substantive questions can only be meaningfully answered if measurement as such is sound. We reason that a new interdisciplinary journal specifically devoted to disseminating open measurement instruments freely, across cultural and language barriers, while supporting open access, open data, and open methodology can aid researchers around the globe and across many disciplines.
E D I T O R I A L Open Access
Measurement Instruments for the Social
Sciences
Beatrice Rammstedt
*
and Matthias Bluemke
Editorial
At a first glance, the emergence of a new scientific
journal in the ever-growing market of scientific jour-
nals may raise eyebrows. Would a new journal rather
aid a solution or contribute to the replicability crisis
recently acknowledged in some social scientific disci-
plines? Are there not already more scientific outlets
than anyone can ever follow? And do not we all have
enough items on our reading lists already? At a
second and closer look, it becomes clear that if, in
the past, effort had been devoted to developing
high-quality measurement instruments, it often went
unnoticed. Documenting such effort is not a typical
goal of journals focusing on substantive research, be-
cause measurement is just perceived to be a vehicle
to answer substantive questions. In addition, even if
an instrument was published, it was most often only
recognized and reused in its discipline of origin. A
transfer of knowledge among disciplines, even though they
aimed to study the same constructs, only seldom happens.
A slight imbalance then needs to be addressed; sub-
stantive questions can only be meaningfully answered
if measurement as such is sound. We reason that a
new interdisciplinary journal specifically devoted to
disseminating open measurement instruments freely,
across cultural and language barriers, while support-
ing open access, open data, and open methodology
can aid researchers around the globe and across many
disciplines.
Relevance of sound measurement
Measurement instruments are the central tools to ac-
quire sound and scientifically based knowledge and to
make theoretical progress in a variety of social scientific
disciplines. Across these disciplines, empirical re-
searchers rely on information that has to be collected in
an objective manner, in many cases, to be quantified and
statistically analyzed afterwards. Irrespective of different
underpinnings in test-theoretical terms, this requires at
least a standardized approach of collecting information
and integrating (often numerical) survey responses or
other participant data, before making inferences at
the construct level and quantifying individual differ-
ences. This process applies to various disciplines, re-
gardless whether the goal is to describe political
systems and societies, implement societal change, pre-
pare individual interventions, or predict economic
decision-making in the lab or in the field, to name a
few examples.
Regardless whether political surveys, personality as-
sessment, educational testing, changes in societal values,
or health screenings are concerned, the adequacy of in-
terpretations of measurement outcomes as well as the
correct tracking of trends in societies, the tools we use
are fundamental for obtaining meaningful measures.
They need to be objective andideallytransparent for
independent inspection and evaluation. Different re-
search purposes may prolong the parallel existence of
instruments measuring the same construct, yet after
some time, so one might hope, scientific evolution will
have helped some of the bestmeasures to shine,
whereas others that appear to be weaker will have paled
in comparison.
Current challenges
According to our reading, the current state of measure-
ment in the social sciences is suboptimal in some
regards: (a) there is multiplicity of instruments in a
world full of diverse disciplines; (b) in some cases, the
instruments themselves, or existing translations, are not
available to the research community; and (c) despite the
need for brief measures or cross-cultural adaptations,
there is a lack of recognition for this kind of work (e.g.,
resulting in desk rejections due to the lack of originality),
leading to lower effort in this regard.
To elaborate, scientific disciplines exist next to each
other and continue to diversify, so a number of measure-
ment approaches compete and the sameor highly
relatedconstructs are being assessed by economists,
* Correspondence: beatrice.rammstedt@gesis.org
GESIS - Leibniz-Institute for the Social Sciences, Cologne, Mannheim, Germany
© The Author(s). 2019 Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and
reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to
the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver
(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
Rammstedt and Bluemke Measurement Instruments for the Social Sciences
(2019) 2:4
https://doi.org/10.1186/s42409-018-0003-3
psychologists, or sociologists, albeit for different pur-
poses. Consequently, multiple measures for the same
construct exist, rendering comparisonsfor instance,
across disciplinesdifficult.
Furthermore, there is an unknown number of meas-
urement instruments hidden in file drawers, which is
probably true for each discipline. This refers to newly
developed instruments as well as to abbreviated mea-
sures (Rammstedt & Beierlein, 2014). Whereas long item
batteries are often not feasible in large-scale assessments
or general population surveys, selecting items in an ad
hoc manner is not wise either, yet some existing short
scales are hardly visible (Ziegler, Kemper, & Kruyen,
2014). Also, questionnaires that have been translated
to a foreign language, or adapted for a different
population, context, and culture, rarely find the atten-
tion and respect they deserve. The efforts in provid-
ing and testing them do not only contribute to
replicable science but to the only way that science
can shed light on human universals.
Unless a journal is devoted to assessment or measure-
ment instruments, the typical description of measure-
ment instruments is rather succinct. Often crucial
psychometric evidence on an instrument is missing (if
only for space constraints). Sometimes, even a sound
definition of the construct is lacking, not to mention the
intended uses, target population, and users of a meas-
urement tool (Ziegler, 2014). Sometimes relevant aspects
of a questionnaire are not disclosed (e.g., full item set,
item translations/adaptations, instructions, response
options, aggregation rules, measurement model). Not
surprisingly then, data and/or crucial statistical code
used for analyses might be missing, too. This current
practice can undermine a thorough investigation of the
quality of instruments. It definitely hampers attempts at
replicating the original research and conducting further
research on the instrument.
Yet even if all the information one seeks were
available, the comparability of many tools across cul-
tures and languages, or across subgroups in the same
population, might appear rather doubtful. We would
like to rest assured that bias does not distort the
comparability of measurement outcomes across
groups. Some researchers might devote attention to
developing cross-culturally valid items from the be-
ginning (e.g., by cognitive pretesting); others might
test (or not) for measurement equivalence at least in
retrospect (Thalmeyer & Saucier, 2014). Sadly, the
current practice is that measurement equivalence is
not considered often enough, even in cross-cultural
research where this topic should be ranking high on
the agenda (Boer, Hanke, & He, 2018). Accordingly,
researchersconclusions do not always live up to
scrutiny.
Journal development
When we at GESIS-Leibniz-Institute for the Social
Sciencespondered the launch of a new open access
journalwe discovered that the scientific community
was open to the idea of a new journal specifically de-
voted to Measurement Instruments in the Social Sciences
(MISS). We were encouraged in our view that there is a
fundamental need in the scientific community address-
ing the aforementioned challenges and that GESIS is
ideally suited for setting up an interdisciplinary journal,
for gathering an initial editorial board, and for helping
the fields to develop an integrated view on measure-
ment. It goes without saying that our newest addendum
to the journal list is not the only outlet to further the
cause of measurement in science. Each discipline has
established its own flagship.
Nonetheless, we have tried to achieve a combination
of features that uniquely characterize the new journal.
The open access journal functions as a platform to be
shared by various disciplines (e.g., sociology, psychology,
education, political science, economics), and the journal
disseminates public measurement instruments intended
for scientific use across multiple languages and settings.
Though focusing predominantly on social surveys for
the general population, the instruments may also be
relevant for the study of individual differences and useful
for specific groups or in specific diagnostic contexts.
Submitting authors have to subscribe to high scientific,
open-science, and ethical standards (more information
can be found at the journal homepage: https://measure
mentinstrumentssocialscience.biomedcentral.com).
All submissions undergo double-blind peer review.
Six types of articles are acceptable for publication,
which may further evolve in the future (for details,
see the journal submission guidelines: https://
measurementinstrumentssocialscience.biomedcentral.com/
submission-guidelines):
1) New measurement instruments present an
instrument that may either be completely new or
offer a better version of existing measurement
approaches.
2) International adaptations of measurement
instruments present a tool (e.g., a questionnaire) in
different languages to foster the international
harmonization of measurement instruments.
3) Validation of measurement instruments presents
validation studies on established questionnaires or
tests, enhancing what has been known and tested
about their psychometric properties.
4) Advances in methodology portray best practice in
social-scientific measurement or describe recom-
mended changes in testing and analytical proce-
dures on the basis of scientific evidence.
Rammstedt and Bluemke Measurement Instruments for the Social Sciences (2019) 2:4 Page 2 of 3
5) Test reviews focus on a standardized evaluation of a
measurement instruments characteristics and its
materials as well as current evidence of
psychometric quality in line with established test
criteria and guidelines (e.g., international test
commissions).
6) Systematic reviews with or without meta-analysis
may occasionally be published. These critical assess-
ments of literature and data sources focus on a
topic that concerns the use and analysis of measure-
ment instruments in the social sciences. They
should emphasize quality indicators such as object-
ivity of tools, reliability estimates, construct validity,
factorial validity/measurement model, predictive
validity, or fairness/comparability/equivalence across
groups.
7) Meeting reports summarize the major themes of a
meeting, symposium, or workshop, focusing on the
key developments as well as new discoveries
relevant to measurement instruments, changes in
best practices regarding measurement and use of
measurement instruments, and the application of
recently gained knowledge in the social sciences.
Recommendations for authors
On the basis of the requests that our interdisciplinary
and international editorial board members have received
so far, the first three article types are about to become
the most frequent publication types appearing in Meas-
urement Instruments for the Social Sciences for a while.
If you feel that your approach does not fit one of the
recommended article types and structures, please be in
touch with the editors.
It should be evident that any progress in
measurement-related issues across the social sciences
will always be gradual and incremental. We appreciate
you contributing to this worthwhile endeavor by submit-
ting a manuscript or helping out with peer review. Start-
ing from 2018 through 2020, GESIS is sponsoring a
number of articles through a full waiver (further details
at https://measurementinstrumentssocialscience.biomed-
central.com/submission-guidelines/fees-and-funding).
Most of all, we hope to stimulate discussion across
fields. Next to submitting your work as a paper to MISS,
you can actively contribute to this enterprise by suggest-
ing future special issues or thematic series (also see
journal homepage).
Funding
Not applicable
Availability of data and materials
Not applicable
Authorscontributions
Both authors read and approved the final manuscript.
Competing interests
The authors declare that they have no competing interests.
PublishersNote
Springer Nature remains neutral with regard to jurisdictional claims in
published maps and institutional affiliations.
Received: 2 December 2018 Accepted: 12 December 2018
References
Boer, D., Hanke, K., & He, J. (2018). On detecting systematic measurement error in
cross-cultural research: a review and critical reflection on equivalence and
invariance tests. Journal of Cross-Cultural Psychology, 49, 713734.
Rammstedt, B., & Beierlein, C. (2014). Cant we make it any shorter? The limits of
personality assessment and ways to overcome them. Journal of Individual
Differences, 35, 212220.
Thalmayer, A. G., & Saucier, G. (2014). The questionnaire big six (QB6) in 26
nations: developing cross-culturally applicable big six, big five, and big two
inventories. European Journal of Personality, 28, 482496.
Ziegler, M. (2014). Stop and state your intentions! Lets not forget the ABC of test
construction. European Journal of Psychological Assessment, 30, 239242.
Ziegler, M., Kemper, C. J., & Kruyen, P. (2014). Short scales: five misunderstandings
and ways to overcome them. Journal of Individual Differences, 35, 185189.
Rammstedt and Bluemke Measurement Instruments for the Social Sciences (2019) 2:4 Page 3 of 3
... En conjunto, la validez ha dado lugar a una reflexión profunda sobre el rol que desempeña la medición dentro de la ciencia, en especial en ciencias sociales (Flake et al., 2017;Rammstedt & Bluemke, 2019), lo que conlleva que la validez de un instrumento deba ser vista de manera crítica e integral (Markus, 2016;Norman, 2016), en especial porque ha surgido una discusión sobre lo que es la validez (Stone, 2019) y su utilidad (Larroulet, 2021), donde se promueve que lo estadístico no prevalezca sobre lo práctico y teórico (Guyon, 2018;Mercado-Salgado et al., 2016). ...
Article
Full-text available
Objetivo: Identificar la estabilidad entre los resultados de la validez de constructo, convergente y discriminante de un instrumento de capital intelectual. Método cuantitativo: Análisis estadístico de tres tipos de validez, previa traducción inversa (inglés-españolinglés) del cuestionario; muestra no probabilística de 226 empresas de México y Bolivia; cálculo del análisis factorial exploratorio, análisis factorial confirmatorio, varianza promedio extraída y criterio Fornell-Larcker. Resultados: Se confirma la conveniencia de aplicar pruebas diversas para obtener la validez de un instrumento de capital intelectual. Originalidad: Privilegiar lo teórico sobre lo estadístico en el análisis de la validez de instrumentos; confianza de replicar este cuestionario en contextos similares. Limitaciones: Tipo y tamaño de muestra; inclusión marginal de la validez de contenido.
Article
This study developed a brief student-reported Perceived Teacher Support Scale (PTSS) to assess how well students felt supported by their teachers. Using a theory-driven approach, the PTSS with instrumental, emotional, informational, and appraisal support subscales was developed based on the framework of social support. A total of 1,138 middle school students in grades seven to ten from middle schools in mainland China participated in this survey. The psychometric features of the PTSS were studied using factor and Rasch analyses. Exploratory factor analysis revealed a three-factor model while confirmatory factor analysis supported three- and four-factor solutions. The Rasch analysis further demonstrated the psychometric quality of the four subscales: scale dimensionality, rating scale functioning, and item fit. Measurement invariance across gender was confirmed. The final PTSS had 25 items in four subscales evaluating students’ perceived teacher support: instrumental, emotional, informational, and appraisal support. The correlation between the PTSS and student engagement supported concurrent validity. Finally, the study’s limitations and implications are discussed. In general, the PTSS scale is a more effective tool for measuring students’ perceived teacher support. It can be used to understand the situation of teacher support in different dimensions, and can also be used to conduct relevant cross-sectional and longitudinal research experiments.
Article
Full-text available
Psychological constructs are becoming increasingly important in social surveys. Scales for the assessment of these constructs are usually developed primarily for individual assessment and decision-making. Hence, in order to guarantee high levels of reliability, measurement precision, and validity, these scales are in most cases much too long to be applied in surveys. Such settings call for extremely short measures validated for the population as a whole. However, despite the unquestionable demand, appropriate measures are still lacking. There are several reasons for this. In particular, short scales have often been criticized for their potential psychometric shortcomings with regard to reliability and validity. In this article, the authors discuss the advantages of short scales as alternative measures in large-scale surveys. Possible reasons for the assumed limited psychometric qualities of short scales will be highlighted. The authors show that commonly used reliability estimators are not always appropriate for judging the quality of scales with a minimal number of items, and they offer recommendations for alternative estimation methods and suggestions for the construction of a thorough short scale.
Article
Full-text available
What do we know about short psychological scales? The idea for this special issue was born out of a common interest in short scales, defined as scales measuring a specific psychological construct with fewer than, say, ten items. All editors of this special issue were recently either active in short scale construction themselves (Kemper, Beierlein, Kovaleva, & Rammstedt, 2013; Ziegler, Kemper, & Rammstedt, 2013) or evaluated the current status of short scales in the field (Kemper, Brähler, & Zenger, 2013; Kruyen, Emons, & Sijtsma, 2012, 2013a, 2013b, 2014). Based on discussions with other scientists about the challenges and caveats of short scales, they felt a strong need to share contemporary issues regarding short scales with a broader scientific audience. This special issue brings together a variety of researchers who have approached the topic of short scales from different angles. It is our ambition to resolve some misunderstandings and open questions regarding the construction, psychometric quality, and use of short scales by bringing together methodological, statistical and construction-oriented perspectives . We aim at contributing to an ongoing debate.
Article
Full-text available
The Big Five is a useful model of attributes now commonly used in cross-cultural research, but without the support of strong measurement invariance (MI) evidence. The Big Six has been proposed as a cross-culturally informed update, and the broader Big Two (Social Self-Regulation and Dynamism) draws on even more cross-cultural evidence. However, neither has been rigorously tested for cross-cultural MI. Here a Big Six inventory (36QB6) and measures of the Big Five and Big Two derived from it were tested and refined for cross-cultural usability in samples from 26 nations, divided into three subsets. Confirmatory factor analysis of the models in the first subset of nations demonstrated fit as strong in translation as typical personality measures achieve in their nation of origin (although poor per standard benchmarks). Items that performed inconsistently across cultures were removed, and alternates considered in a second subset of nations. Fit and invariance were improved for refined 30-item QB6, 25-item Big Five and 14-item Big Two measures in the third subset of nations. For all models, decrease in comparative fit index between MI levels was larger than .01, indicating lack of support for higher levels. Configural and factorial invariance were relatively stronger, compared to scalar and full. Copyright © 2014 European Association of Personality Psychology
Article
One major threat to revealing cultural influences on psychological states or processes is the presence of bias (i.e., systematic measurement error). When quantitative measures are not targeting the same construct or they differ in metric across cultures, the validity of inferences about cultural variability (and universality) is in doubt. The objectives of this article are to review what can be done about it and what is being done about it. To date, a multitude of useful techniques and methods to reduce or assess bias in cross-cultural research have been developed. We explore the limits of invariance/equivalence testing and suggest more flexible means of dealing with bias. First, we review currently available established and novel methods that reveal bias in cross-cultural research. Second, we analyze current practices in a systematic content analysis. The content analysis of more than 500 culture-comparative quantitative studies (published from 2008 to 2015 in three outlets in cross-cultural, social, and developmental psychology) aims to gauge current practices and approaches in the assessment of measurement equivalence/invariance. Surprisingly, the analysis revealed a rather low penetration of invariance testing in cross-cultural research. Although a multitude of classical and novel approaches for invariance testing is available, these are employed infrequent rather than habitual. We discuss reasons for this hesitation, and we derive suggestions for creatively assessing and handling biases across different research paradigms and designs.
  • Bluemke Rammstedt
Rammstedt and Bluemke Measurement Instruments for the Social Sciences (2019) 2:4