ArticlePDF Available

The dilemmas of performance indicators of individual researchers—An urgent debate in bibliometrics

Authors:
ISSI NEWSLETTER VOL. . NR. .
© International Society for Scientometrics and Informetrics
SHORT COMMUNICATIONS, ARTICLES
48
INTRODUCTION
In the last quarter of the 20th century, bib-
liometrics evolved from a sub-discipline of
library and information science to an in-
strument for evaluation and benchmark-
ing (Glänzel, InScit, 2006; Wouters 2013).
As a consequence of this shift in perspec-
tive, new fields of applications and chal-
lenges opened to bibliometrics, although
many tools were still designed for use in
the context of scientific information, in-
formation retrieval and libraries. In other
words, these became used in a context for
which they were not designed (e.g., the
Journal Impact Factor (JIF)).
THE DILEMMAS
OF PERFORMANCE
INDICATORS OF
INDIVIDUAL RESEARCHERS
AN URGENT DEBATE IN
BIBLIOMETRICS
This development has been joined by an
intensified interest in the evaluation of in-
dividual researchers. The publication of the
Hirsch Index in 2005 (Hirsch, 2005) and its
popularisation by the journal Nature (Ball,
2005) has given this a strong stimulus. Accord-
ing to Hirsch, his index seemed the perfect
indicator to assess the scientific performance
of an individual author because “it is trans-
parent, unbiased and very hard to rig”. The h-
index combines publication activity with ci-
tation impact. For example, an author with a
h-index of 14 has created 14 publications that
each have been cited at least 14 times each. So
neither authors with a long list of mediocre
publications, nor an author with a one won-
PAUL
WOUTERS
WOLFGANG
GLÄNZEL
JOCHEN
GLÄSER
ISMAEL
RAFOLS
ISSI NEWSLETTER VOL. . NR. .
© International Society for Scientometrics and Informetrics
SHORT COMMUNICATIONS, ARTICLES
49
der hit are rewarded by this indicator. Never-
theless, the h-index turned out to have several
severe disadvantages to be wearing the crown
of “the perfect indicator” (cf. Glänzel, 2006).
As Hirsch acknowledged himself, it cannot
be used for cross-disciplinary comparison. A
field in which many citations are exchanged
among authors will produce a much higher
average Hirsch index than a field with much
less citations and references per publication.
Moreover, the older one gets, the higher ones
h-index will be. Furthermore, confidence in-
tervals of empirical h-indexes are huge (Glän-
zel, 2010) such that this indicator is not suited
for ranking individuals or research units and
rankings based on the h-index may also be
influenced in rather counter-intuitive ways
(Waltman & Eck, 2012). Although many vari-
ants of the h-index have been published, none
of them has turned out to be the perfect indi-
cator. In fact, we cannot expect any indicator
to be the perfect one. Nevertheless, there is
ample evidence that the use of the h-index and
other bibliometric indicators (such as the JIF)
has become pervasive in the estimation of the
scientific and even scholarly impact of a body
of work by an individual scientific author. For
example, many biomedical researchers men-
tion the value of their h-index on their CV.
In publications lists, one can regularly see the
value of the JIF mentioned after the journal’s
name. In some countries, for example Turkey
and China, ones salary can be determined by
the value of either the h-index or the journal’s
Impact Factor one has published in.
This situation is clearly not desirable. If
researchers are being evaluated, they should
be aware of the criteria used and these crite-
ria should be justified for the purpose at hand.
This requires that users of performance indi-
cators should have clear guidelines. It seems
rather obvious that the bibliometric commu-
nity has an important responsibility to inform
and provide such guidelines. However, at the
moment, there is no consensus yet about such
guidelines. Individual bibliometric centres do
indeed inform their clients about the use and
limitations of their indicators. Moreover, all
bibliometric centres have the habit of pub-
lishing their work in the scientific literature,
often including technical details of their in-
dicators. However, this published work is not
easily accessible to non-expert users such as
deans of faculties and research directors. The
literature is too technical and distributed over
too many journals and books. It needs syn-
thesizing and translation into plain language
which is easily understandable.
So how should the community of sciento-
metricians relate to this development? What
should the responsibility be of scientometric
and bibliometric experts in the process of re-
search evaluation? Should science and tech-
nology indicators be used at this level? If so,
how should their limitations be interpreted?
In what sense are we in need of a heigthened
ethical awareness in the field of scientomet-
rics, informetrics and bibliometrics? We are
fully aware that these questions are not new.
In fact, they have been raised several times at
scientometric and bibliometric conferences,
almost from the very start of the field. But so-
cial relationships are always dynamic and this
certainly holds for the scientific and schol-
arly system. The increased role of indicators
in general and of scientometric performance
indicators in particular makes it necessary
to address these questions again and in the
context of the evolving practices of research
evaluations and assessments. This was our
motivation to propose two debates at the sub-
sequent scientometric conferences this year.
At the 14th ISSI Conference 15-19 July in Vi-
enna, a special plenary session was organized
with a joint presentation by Wolfgang Glän-
zel and Paul Wouters, followed by responses
by Henk Moed and Gunnar Sivertsen. At the
STI2013 conference, “Translational twists and
turns: science as a socio-economic endeav-
our” 4-6 September in Berlin, a full plenary
was devoted to bibliometrics of individual
researchers, chaired by Ben Martin (SPRU),
with presentations by Wolfgang Glänzel, Paul
Wouters, Marc Luwel, and Jochen Gläser. In
this short report, we wish to give an impres-
sion of this discussion with the aim to further
stimulate this exchange of ideas, experiences
and, of course, technical knowledge.
ISSI NEWSLETTER VOL. . NR. .
© International Society for Scientometrics and Informetrics
SHORT COMMUNICATIONS, ARTICLES
50
THE ISSI CONFERENCE IN
VIENNA
To initiate a process of a more professional
guidance for the application of bibliomet-
ric indicators in the evaluation of individual
researchers, we asked the organizers of the
ISSI conference to devote a plenary to this
problem, which they kindly agreed to. At the
plenary, Wolfgang Glänzel and Paul Wouters
presented “The dos and don’ts in individual
level bibliometrics”. We do not think this is a
final list, more a good start with ten dos and
don’ts. A start for reflection, experiments
and the exchange of experiences. In the fol-
lowing, we sketch our proposals for applying
bibliometrics on individual researchers as
well as the ensuing debate at the conference.
TEN THINGS YOU MUST NOT DO:
1. Don’t reduce individual performance
to a single number
2. Don’t use journal impact factors as
measures of quality of individual
researchers
3. Don’t apply hidden “bibliometric
filters” for selection
4. Don’t apply arbitrary weights to
co-authorships
5. Don’t rank scientists according to one
indicator
6. Don’t merge incommensurable
measures
7. Don’t use flawed statistics
8. Dont blindly trust one-hit wonders
9. Don’t compare apples and oranges
10. Don’t allow deadlines and workload
to compel you to drop good practices.
TEN THINGS YOU CAN DO:
1. Basic measures such as numbers of
publications and citations are still
relevant statistical measures
2. Analyze collaboration patterns of
researchers
3. Always combine quantitative and
qualitative methods
4. Use citation context analysis
5. Analyze subject profiles of individual
researchers
6. Make an explicit choice between
the analysis of the full oeuvre or
comparative analysis using a citation
window
7. Combine bibliometrics with career
analysis
8. Clean bibliometric data carefully and
use external sources
9. Don’t take this list of dos and don’ts
too absolutely: even some don’ts can
be used given the right context
10. Help users to interpret and use your
results.
Of course, the complex business of research
assessments cannot be reduced to a simple
list of “commandments”. In other words, we
do not want to initiate a bibliometric police
with absolute rules. The context of the eval-
uation should always determine which indi-
cators and methods to use. Therefore, some
don’ts in our list may sometimes be perfectly
useable, such as the application of biblio-
metric indicators to make a first selection
among a large number of candidates. Also,
in informed peer review of large institutions
it may be inevitable to use bibliometric filters
to zoom in on the most relevant work for
closer inspection. In all those cases, these fil-
ters should however be made explicit. After
all, the researchers who are subjected to as-
sessment should be able to provide counter-
arguments if they think the filters have been
used inappropriately.
Our presentation was commented on by
Henk Moed (Elsevier) with a presentation
on “Author Level Bibliometrics” and by Gun-
nar Sivertsen (NIFU, Oslo University) with
comments on the basis of his extensive ex-
periences in research evaluation. Henk Moed
built on the concept of the multi-dimensional
research matrix which was published by the
European Expert Group on the Assessment of
University Based Research in 2010, of which
he was a member (Assessing Europes Univer-
sity-Based Research - Expert Group on Assess-
ISSI NEWSLETTER VOL. . NR. .
© International Society for Scientometrics and Informetrics
SHORT COMMUNICATIONS, ARTICLES
51
ment of University-Based Research, 2010). This
matrix aims to give global guidance to the use
of indicators at various levels of the university
organization. The first row of the matrix dis-
cusses goals, output dimensions and biblio-
metric and other indicators at the individual
level. The matrix as a whole does not focus
on the problem of how to evaluate individual
researchers. Still, the matrix is surely a valua-
ble contribution to the development of more
professional standards in the application of
performance indicators. In his presentation,
Moed discussed a number of case studies
which clearly showed that no absolute rules
can be expected. It all depends on the goal of
the assessment as well as on the state of af-
fairs in the research area involved. Moed ar-
gued that the data should be verified by the
researchers themselves (already a standing
practice in most if not all of the main biblio-
metric centres). A key problem he identified
is the attribution of scientific performance to
an individual when in reality most research is
based on collaborative work within and be-
tween teams. One of the three cases he pre-
sented involved a country in which science
policy suspects that their researchers are not
oriented enough toward international net-
works. In this case, a policy measure could be
to stimulate and reward publication in top in-
ternational journals. For this, the number of
publications in those type of journals could be
an appropriate bibliometric indicator.
Gunnar Sivertsen strongly agreed with
the main thrust of the discussion. Moreover,
he made clear that the discussion should not
be restricted to the bibliometric community
itself. On the contrary, the main audience of
bibliometric guidelines and standards should
be the researchers themselves and admin-
strators in universities and funding agencies.
The ensuing debate led to a large number
of suggestions. A few speakers emphasized
that this debate is indeed not new and was
already addressed at the ISSI conference in
Chicago in 1995. A key point was the issue
of responsibility: it is clear that researchers
themselves and the evaluating bodies should
carry the main responsibility for the use
of performance indicators. However, they
should be able to rely on clear guidance from
the technical experts. How must this balance
be struck? Should bibliometricians refuse to
deliver indicators when they think their ap-
plication would be unjustified? Should the as-
sociation of scientometricians publicly com-
ment on misapplications? Or should this be
left to the judgment of the universities them-
selves? Several calls were made to publicly
criticize applications of bibliometric meas-
ures that are deemed harmful. At the same
time, it is not yet clear in whose name these
statements should be made. The plenary did
not solve these issues yet. However, a con-
sensus seemed to be emerging that more ex-
plicit guidance by bibliometricians is required
(building on the work from the past) and that
researchers should have a clear address to
which they can turn to with questions about
the application of performance indicators ei-
ther by themselves or by their evaluators.
THE STI2013 CONFERENCE
The plenary at the STI2013 conference start-
ed with an introduction by Ben Martin. He
reported on a dramatic case of the rise and
fall of a young professor in economics in
Germany. This researcher had conducted
one research project for his PhD and was
able to generate a rather impressive number
of publications on the basis of this dataset.
Because he was so productive, he was able to
attract more external research funding. Sub-
sequently, he became an attractive target for
headhunting by universities looking for high
profile researchers who might help them in-
crease their ranking position in various league
tables and in getting grants for the Excellence
Initiative program in Germany. And indeed,
a university was quite eager to hire him for
precisely this purpose. It was only then that
a few members of the selection committee
decided to actually read the publications by
this young economics researcher. It struck
them that these publications were quite like
each other. Not very surprising given that
ISSI NEWSLETTER VOL. . NR. .
© International Society for Scientometrics and Informetrics
SHORT COMMUNICATIONS, ARTICLES
52
the research was all based on a single dataset
from his PhD project. It turned out that he
had published a large number of variations of
the same article in different journals without
anyone noticing these duplications. This dis-
covery was the beginning of the end of the
career of this formerly promising economist.
A number of journals began retracting these
publications, although not with the coop-
eration of the researcher. This process is still
ongoing. A sobering tale, according to Mar-
tin, and one which shows that the abuse of
bibliometrics is now a serious concern for all
parties involved in the management and de-
velopment of scientific research.
Wolfgang Glänzel then informed the au-
dience about the discussion at the ISSI con-
ference. He emphasized especially the need
for more information and guidelines among
researchers, managers and policy makers.
The bibliometrics community should play
a role in providing at least the core of these
guidelines. He also emphasized that the ISSI
conference had made clear that we need
some organization to which questions about
the proper use of bibliometric indicators can
be addressed. ISSI as an organization may
not be in the best position to play this role,
given its scientific rather than professional
role. This may therefore be a role that the
main bibliometric centres should take up, in
a somewhat coordinated way.
In his presentation, Paul Wouters argued
in favour of a portfolio approach and showed
how CWTS has been developing bibliomet-
ric profiles at the level of the individual re-
searcher. He also presented the philosophy
behind the 7th Framework project ACUMEN
which aims to enable individual researchers
to enrich their Curriculum Vitae with prop-
erly calculated and relevant bibliometric in-
dicators as well as qualitative evidence. The
portfolio approach has also been proposed by
other researchers such as Bornmann (2013).
Marc Luwel focused on the tensions in
the concrete practice of science policy mak-
ing. There is an increasing need for justifica-
tion of the budgets devoted to research. This
needs to be based on verifiable empirical
evidence, hence the need for performance
indicators among which also bibliometric
indicators. Nevertheless, Luwel stressed,
quantitative indicators cannot be used as the
sole basis. Luwel: “Beware of the lone librar-
ian cooking a toxic cocktail of publication
and citation data!”
The presentation by Jochen Gläser
linked to Luwel’s talk by taking one more
step in the area of applied ethics and the
dilemmas in the application and use of bib-
liometric indicators. Gläser made clear that
he does not see himself as a bibliometrician
pur sang, but more as a sociologist who
is interested in combining bibliometric
methods with other social science meth-
ods, such as surveys and interviews. More-
over, he is not himself involved in applying
bibliometrics for research assessments and
can therefore take a step back and present
reflections that may be useful to the com-
munity. He presented a couple of scenarios
in which bibliometric reports are carefully
crafted but basically ignored by the deci-
sion makers or only cherry-picked in order
to justify decisions that were going to be
taken anyway. And indeed, often it is not
clear to what extent and how the biblio-
metric reports that we produce are actu-
ally useful for quality decision making. The
reports may very well start to live a life of
their own. Gläser discussed to what extent
the current available literature on applied
ethics is useful for the bibliometric com-
munity (not very much) and in what sense
we can learn from the communities that
have been producing guidelines and stand-
ards for their profession (more promising
perhaps). His presentation made clear that
ethics is indeed inevitable and highly rel-
evant, albeit not in all respects complete-
ly new. Some of the bibliometric centres
have been following implicit and explicit
ethical rules without always formalizing
them. Gläser also mentioned the proposal
by Wouters in a parallel session at STI2013
on bibliometric standards to initiate a bib-
liometric “Ombudsman Oce” to which
researchers that have been evaluated can
ISSI NEWSLETTER VOL. . NR. .
© International Society for Scientometrics and Informetrics
SHORT COMMUNICATIONS, ARTICLES
53
turn in case they feel misjudged or harmed
by the performance analysis and these con-
flicting interests cannot be solved with the
bibliometric experts involved. He saw this
proposal as a long-term goal. In the shorter
term, he proposed that the next few STI
conferences should all devote more atten-
tion to these ethical and political issues, for
example by organizing special session de-
voted to them.
As authors of this short report, we think
this last proposal is an excellent idea. It
might for example be possible to not only
discuss formal research papers in these ses-
sions, but we might want to focus more on
exchanging experiences. This could take
the form of submitting problems or case
studies rather than the normal papers. The
discussion could also be organized in more
engaging and discussion oriented way.
The ensuing discussion at the STI2013
conference again made clear that these is-
sues have become quite urgent for many
practitioners in the field of science & tech-
nology indicators. The issue of responsi-
bility rose to the top again of most urgent
issues, although it is also clear that it will
also remain a rather complex web of prob-
lems. This complexity was illustrated by a
number of participants who addressed the
limitation of the present discussion to the
European and Anglo-Saxon context. The
role of indicators in China, Turkey, Iran,
and South East Asia is clearly different and
perhaps even more dominating. As a result,
many researchers have had to set explicit
indicator based targets in their career de-
velopment. This means that the current
debate should be linked to these practices.
In the discussion an additional funda-
mental problem was raised: to what ex-
tent is the ethical problem in reality caused
by a fundamental problem in the state of
our knowledge in bibliometrics? Often,
we actually do not know what exactly is
represented by the patterns we see by the
indicators. Therefore, it was argued, we
should first of all try to generate more ro-
bust knowledge, using also a much larger
variety of databases in addition to the tra-
ditional citation indexes.
Perhaps this questions might be a good
agenda for the next series of STI and ISSI
conferences? In parallel, we will be organ-
izing a workshop on these issues targeted
at the users of bibliometric indicators, in-
cluding the scientific communities in the
first half of 2014. Stay tuned!
REFERENCES
Assessing Europe’s University-Based Research - Expert
Group on Assessment of University-Based
Research. (2010). Research Policy. European
Commission. doi:10.2777/80193
Ball, P. (2005). Index aims for fair ranking of
scientists. Nature, 436(7053), 900. Retrieved
from http://dx.doi.org/10.1038/436900a
Bornmann, L. (2013). A better alternative to the
h index. Journal of Informetrics, 7(1), 100.
doi:10.1016/j.joi.2012.09.004
Glänzel, W., On the opportunities and limitations
of the h-index (in Chinese). Science Focus, 1
(1), 2006, 10–11. English version accessible
at: http://eprints.rclis.org/9378/
Glänzel, W., On Reliability and Robustness
of Scientometrics Indicators Based on
Stochastic Models. An evidence-based
opinion paper. Journal of Informetrics, 2010,
4 (3), 313-319
Hirsch, J. E. (2005). An index to quantify an
individual’s scientific research output.
Proceedings of the National Academy
of Sciences of the United States of
America, 102(46), 16569–72. doi:10.1073/
pnas.0507655102
Waltman, L., & Eck, N. J. Van. (2012). The
Inconsistency of the h-index. Journal
of the American Society for Information
Science and Technology, 63(2007), 406–415.
doi:10.1002/asi
... Barré's despair about the evolution of STI indicators is shared. Other key pioneers in the adoption and use of scientometrics for science policy such as Ben Martin (Martin, Nightingale and Rà fols 2014;Martin 2016), Diana Hicks (Hicks et al. 2015), or Wolfgang Glä nzel (Glä nzel and Schoepflin 1994;Wouters et al. 2013) have also expressed concerns. 3 Barré's critique is consistent with many analyses of the construction and effects of indicators under managerialism, or what has often been called the 'neoliberal academy'meaning the application of New Public Management (NPM) frameworks in higher education and R&D management (Weingart 2005;Glä ser and Laudel 2007;Burrows 2012;Halffman and Radder 2015;Fochler and De Rijcke 2017). ...
... First, traces such as publications, keywords, and citations appear with stochastic patterns (i.e. with probability distributions). Therefore, at low levels of aggregation, particularly at the level of individual researchers, one cannot robustly differentiate meaningfully the 'performance' of different researchers (Wouters et al. 2013). Even at much higher levels of aggregation, the Leiden University ranking shows that dozens of universities are crammed at the tails of the citation distribution within the same range of 'interval of stability'. ...
Article
The use of indicators in research policy and evaluation is widely perceived as problematic. Responding to demands for explicit normative framings in STI governance, I propose an agenda for transforming the place and role of indicators in policy. Given that expert advice should not separate knowledge formation from decision-making under conditions of uncertainty and lack of value consensus, I argue that current scientometrics is too focused on technical issues, too reductionist, and too isolated from the contexts and values of its use. Using Callon's analytical framework of 'secluded research' vs. 'research in the wild', I propose three moves for improving design and use of science, technology, and innovation (STI) indicators. First, to continue ongoing trends towards pluralizing the data sources, processing and visualization techniques, and expand the research communities involved in scientometrics. Second, to develop forms of quantitative evidence that can be contextualized with the participation of a more diverse set of stakeholders. Third, to open up the policy framings implicit in measurement, and use quantitative analyses to reveal more balanced perspectives of existing and alternative STI options. I conclude by arguing that these shifts are necessary to preserve epistemic diversity and pluralism in the face of ongoing managerial push for standardization via 'platforms' run by commercial oligopolies. © The Author(s) 2018. Published by Oxford University Press. All rights reserved.
... Barré's despair about the evolution of STI indicators is shared. Other key pioneers in the adoption and use of scientometrics for science policy such as Ben Martin (Martin, Nightingale and Rà fols 2014;Martin 2016), Diana Hicks (Hicks et al. 2015), or Wolfgang Glä nzel (Glä nzel and Schoepflin 1994;Wouters et al. 2013) have also expressed concerns. 3 Barré's critique is consistent with many analyses of the construction and effects of indicators under managerialism, or what has often been called the 'neoliberal academy'meaning the application of New Public Management (NPM) frameworks in higher education and R&D management (Weingart 2005;Glä ser and Laudel 2007;Burrows 2012;Halffman and Radder 2015;Fochler and De Rijcke 2017). ...
... First, traces such as publications, keywords, and citations appear with stochastic patterns (i.e. with probability distributions). Therefore, at low levels of aggregation, particularly at the level of individual researchers, one cannot robustly differentiate meaningfully the 'performance' of different researchers (Wouters et al. 2013). Even at much higher levels of aggregation, the Leiden University ranking shows that dozens of universities are crammed at the tails of the citation distribution within the same range of 'interval of stability'. ...
... In general, the scientometric community had developed a consensus about the need to use bibliometrics as complement, rather than replacement, of peer review, which is summarized in the concept of informed peer review. With the rise of the web and the wider availability of both traditional and novel metrics, the scientometric professionals lost their monopoly and what was variously called amateur scientometrics or citizen scientometrics started to take off (Glänzel, 1996;Leydesdorff, Wouters, & Bornmann, 2016;Thelwall, 2009;Wouters, Glänzel, Gläser, & Rafols, 2013). This required a new approach and a more explicit non-technical development of guidelines, for which the groundwork was laid at a series of conferences in the years 2013 -2016 and in the context of the debates about the role of metrics in national research assessments, especially in Northwestern Europe. ...
... At the 2013 International Scientometric and Informetric Society (July 2013, Vienna) and at the 2013 Science and Technology Indicator / ENID Conference (September 2013, Berlin) another set of recommendations was discussed, specifically aimed at the use of indicators to assess the contribution of individual researchers (Wouters et al., 2013). A year later, the EU funded project ACUMEN resulted in a more detailed evaluation guideline for both researchers and evaluators . ...
Preprint
This chapter approaches, both from a theoretical and practical perspective, the most important principles and conceptual frameworks that can be considered in the application of social media metrics for scientific evaluation. We propose conceptually valid uses for social media metrics in research evaluation. The chapter discusses frameworks and uses of these metrics as well as principles and recommendations for the consideration and application of current (and potentially new) metrics in research evaluation.
... Below we provide an open-ended list of the virtues that enable the constitution of good research practice and constitute something that should be taken into account by those making comparative judgments about the relative value of different research practices in a complementary way with respect to currently used individual bibliometric indicators. For a critical discussion of performance indicators of individual researchers, see Wouters et al. (2013). For an extensive discussion on these individual bibliometric indicators, see Schubert and Schubert (2019) and Wildgaard (2019). ...
Article
Full-text available
In this paper we answer the question of how evaluation should be by proposing a good evaluation of research practices. A good evaluation of research practices, intended as social practices à la MacIntyre, should take into account the stable motivations and the traits of the characters (i.e. the virtues ) of researchers. We also show that a good evaluation is also just , beyond the sense of fairness , as working on good research practices implies keep into account a broader sense of justice . After that, we propose the development of a knowledge base for the assessment of “good” evaluations of research practices to implement a questionnaire for the assessment of researchers’ virtues. Although the latter is a challenging task, the use of ontologies and taxonomic knowledge, and the reasoning algorithms that can draw inferences on the basis of such knowledge represents a way for testing the consistency of the information reported in the questionnaire and to analyse correctly and coherently how the data is gathered through it. Finally, we describe the potential application usefulness of our proposal for the reform of current research assessment systems.
... Research assessment is an arena in which growing practice of 'governance by indicators' has been especially diverse and intense (Burrows, 2012). From university rankings (van Raan, 2005), to performance-based funding systems (Hicks, 2012), or to individual level assessments (Wouters et al., 2013), the use of 'metrics' has become pervasive in research evaluation. ...
... Research assessment is an arena in which growing practice of 'governance by indicators' has been especially diverse and intense (Burrows, 2012). From university rankings (van Raan, 2005), to performance-based funding systems (Hicks, 2012), or to individual level assessments (Wouters et al., 2013), the use of 'metrics' has become pervasive in research evaluation. ...
Preprint
Full-text available
The use of indicators is generally associated with a reduction of perspectival diversity in evaluation that often facilitates making decisions along dominant framings – effectively closing down debate. In this chapter we will argue that while this is indeed often the case, indicators can also be used to help support more plural evaluation and foster more productively critical debate. In order to achieve this shift, it is necessary equally to change understandings, forms and uses of indicators in decision making. These shifts involve, first, broadening out the range of ‘inputs’ taken into account; and second, opening up the ‘outputs', in the sense of developing methodologies for indicator-based analyses to help in considering plural perspectives. In practice, this means a move towards more situated and participatory use of quantitative evidence in evaluation, a shift from universal indicators to contextualised indicating.
... In research evaluation there has been increased discussion recently on the use of bibliometric data for decisions on a research proposal, fellowship, or an academic position to complement or even to replace the peer review procedure (e.g., Abramo, Cicero, & D'Angelo, 2013;Bornmann & Marx, 2013, 2014aBornmann & Marx, 2014b;Mutz, 2016;Sahel, 2011;Wouters, 2013;Wouters, Glänzel, Gläser, & Rafols, 2013). Similar to assessment of institutions or countries, such as the Leiden Ranking, bibliometric indicators may be used also at the level of the individual researcher. ...
Article
Full-text available
Regarding evaluation of individual researchers, the bibliometric indicators approach has been increasingly discussed recently, but there are some problems involved with it: construct definition, measurement errors, level of scale, dimensionality, normalization. Based on a psychometric model, the Rasch model, we developed a measuring scale for the theoretical construct 'researcher's performance capacity,' defined as the competency of a researcher to write influential papers. The aim was a scale that is one-dimensional and continuous, is applicable to bibliometric count variables, and takes measurement errors into account. In this paper we present the psychometric model (Bayesian Poisson Rasch model, BPR) and its assumptions and examine the behavior of the model under various sampling conditions. For a sample of N = 254 researchers in a quantitative methodology section of an undisclosed German academic society for social sciences, using the BPR model we developed a scale that we named 'Bibliometric Quotient' (BQ, M = 100, SD = 15) (following the term 'intelligence quotient'). The scale fulfills most of the test-theoretical requirements (e.g., high reliability αt=.96, no differential item functioning except for academic age and German states) and in addition allows researchers to be ranked. Women's BQ scores were 8.3 points lower on the scale than men's BQ scores.
Article
Full-text available
In recent years, the research performance evaluation of members of the academic community conducted by government or institutions has been applied with multiple indicators and peer review, however, there are many controversies about the design and application of research evaluation indicators. This article aims to introduce the development process of San Francisco Declaration on Research Assessment (DORA) and Leiden Manifesto and summarize their contents of guidelines and attempts to compare the differences between these two documents. It hopes that this article can arise the attention and reflection of research evaluation indicators and relevant issues from Taiwan academic community to reach consensus of utilization of research evaluation indicators. It will be beneficial to develop the version of declaration with local characteristics in the future.
Article
Full-text available
A new indicator for the assessment of the research performance of individual scientists was suggested by Jorge E. Hirsch in 2005. Several opportunities and limitations of this new measure are discussed.
Article
In the present paper we give an overview over the opportunities of probabilistic models in scientometrics. Four examples from different topics are used to shed light on some important aspects of reliability and robustness of indicators based on stochastic models. Limitations and future tasks are discussed as well.
Article
The h-index is a popular bibliometric indicator for assessing individual scientists. We criticize the h-index from a theoretical point of view. We argue that for the purpose of measuring the overall scientific impact of a scientist (or some other unit of analysis) the h-index behaves in a counterintuitive way. In certain cases, the mechanism used by the h-index to aggregate publication and citation statistics into a single number leads to inconsistencies in the way in which scientists are ranked. Our conclusion is that the h-index cannot be considered an appropriate indicator of a scientist's overall scientific impact. Based on recent theoretical insights, we discuss what kind of indicators can be used as an alternative to the h-index. We pay special attention to the highly cited publications indicator. This indicator has a lot in common with the h-index, but unlike the h-index it does not produce inconsistent rankings.
Article
'H-index' sums up publication record.
Article
I propose the index h, defined as the number of papers with citation number ≥h, as a useful index to characterize the scientific output of a researcher. • citations • impact • unbiased
The Inconsistency of the h-index
  • L Waltman
  • N J Eck
  • Van
Waltman, L., & Eck, N. J. Van. (2012). The Inconsistency of the h-index. Journal of the American Society for Information Science and Technology, 63(2007), 406-415. doi:10.1002/asi