ArticlePDF Available

Measure for measures

Authors:

Abstract

Are some ways of measuring scientific quality better than others? Sune Lehmann, Andrew D. Jackson and Benny E. Lautrup analyse the reliability of commonly used methods for comparing citation records.
Measures for measures
Are some ways of measuring scientific quality better than others? Sune Lehmann, Andrew D. Jackson and
Benny E. Lautrup analyse the reliability of commonly used methods for comparing citation records.
Although quantifying the quality of individual
scientists is difficult, the general view is that
it is better to publish more than less and that
the citation count of a paper (relative to cita-
tion habits in its field) is a useful measure of
its quality. How citation counts are weighed
and analysed in practice becomes important
as publication records are increasingly used
in funding, appointment and promotion deci-
sions. Typically, a scientist’s full citation record
is summarized by simpler
measures, such as average cita-
tions per paper, or the recently
proposed Hirsch index
1
, which
is ever more being used as an
indicator of scientific quality
2
.
Despite their growing impor-
tance, there have been few
attempts to discover which
of the popular citation meas-
ures is best and whether any
such measure is statistically
reliable.
Measures of citation quality are of value only
if they can be assigned to individual authors
with high confidence. Previous bibliometric
studies
2
have compared different measures of
scientific quality, but just because two meas-
ures agree does not mean that either one is
accurate or reliable. We will argue that some
citation-based measures can provide useful
information given data of sufficient quality,
but others fail to meet minimum acceptable
standards. This should concern every work-
ing scientist.
Unfair discrimination
Because citation practice differs markedly
between disciplines and subfields, a homo-
geneous set of authors is essential for any
statistical analysis of citations. Here we use
data from the theory section of the SPIRES
database in high-energy physics, which has
the requisite homogeneity
3
. Within this data-
base, the probability that a paper will receive
k citations falls slowly with increasing k and
is described by a power-law distribution,
a/k
γ
with γ ≈ 2.8, for large k. This long-tailed
distribution has a number of consequences.
About 50% of all papers have two or fewer
citations; the average number of citations is
12.6. The top 4.3% of papers produces 50% of
all citations whereas the bottom 50% of papers
yields just 2.1% of all citations. Measuring an
author’s mean or median citation count per
paper probe different aspects of their full cita-
tion record: which is better? Fortunately, this
question can be posed in a way that yields a
statistically compelling answer.
The purpose of comparing citation records is
to discriminate between scientists. An author’s
citation record is a list of the number of cita-
tions of each of the author’s publications. Until
reduced to a single number, this list cannot
provide a means of ranking
scientists. But whatever the
intrinsic merits of the chosen
number, it will be of no practi-
cal use unless the uncertainty
in assigning it to individual
scientists is small. From
this perspective, the ‘best
measure will be that which
minimizes uncertainty in the
values assigned and hence
maximizes discrimination
between individuals. We analyse three meas-
ures of author quality: mean number of cita-
tions per paper, number of papers published
per year, and the Hirsch index. A scientist is
said to have Hirsch index h if h of their total,
N, papers have at least h citations each, and
the remaining (N-h) papers have fewer than
h citations
1
. For this study, we adopt Hirschs
assumption that h divided by Nshould pro-
vide a useful yardstick. To calibrate our results,
we also consider an obviously mean-
ingless measure;
we rank authors
alphabetically by
name.
We st ar t w ith
one of the three
measures we had chosen and sort the SPIRES
authors into decile bins. We use the full cita-
tion records for all authors in a given bin, n,
to calculate the conditional probability that a
paper written by an author in bin n will have k
citations. From these conditional probabilities,
we use Bayes’ theorem to determine the average
probability that an author initially assigned to
bin n should instead be assigned to bin m. (To
do this, we calculate the probability that the full
publication record of each author in bin n was
drawn, at random, on the conditional prob-
ability appropriate for bin m; see Supplemen-
tary Information.) Because the m assignment
is based on an author’s full citation record, it
is more reliable than the n assignment. This
process is repeated for each decile bin and for
each measure considered.
Quality testing
A perfect measure of author quality would
place all weight in the diagonal entries of a plot
of m versus n (Fig. 1, overleaf). The better the
measure, the more weight will be found in the
diagonal boxes. Figure 1 reveals that both accu-
racy and certainty are sensitive to the choice
of indicator.
An alphabetical ranking of authors contains
no information regarding scientific quality,
and so every author is assigned to every decile
with equal probability (Fig. 1a). The resulting
root-mean-square (rms) uncertainty in author
assignment thus has the maximum value of
±29 percentile points. One of the most widely
used measures of scientific quality is the
average number of
papers published
by an author per
“There have been
few attempts to
discover which of
the popular citation
measures is best
and whether any are
statistically reliable.
A. MACDONALD
1003
COMMENTARY
Vol 444|21/28 December 2006
year
1,4
. This measure has a similar rms varia-
tion to alphabetization (Fig. 1b). Publication
frequency would be more useful if all papers
were cited equally but, as noted above, this is
not the case. The best that can be said of pub-
lication frequency is that it measures industry
rather than ability.
Impact factors are widely used to intro-
duce a citation measure into calculations of
publication frequency. The impact factor for
each journal, as defined by Thomson Scien-
tific/ISI
5
, is the average number of citations
acquired during the past two years for papers
published over the same period. But weighting
each paper by the journal’s current impact fac-
tor is unlikely to improve the situation, espe-
cially when estimating scientific quality across
an authors entire career. The impact factor for
reputable journals is determined by a small
fraction of highly cited papers, so the citation
rate for individual papers is largely uncor-
related to the impact factor of the journal in
which it was published
6
. The widespread use
of publication frequency — with or without
an impact factor — is disturbing and requires
further study.
Word of caution
Hirschs h-index attempts to strike a balance
between productivity and quality and to avoid
the heavy weight that power-law distributions
place on a relatively small number of highly
cited papers. Hirschs measure is obtained by
ranking papers in order of decreasing citations
with paper i having C(i) citations and solving
the equation h = C(h). This is the simplest ver-
sion of h = AC(h)
Κ
. Hirschs choice of A = Κ = 1
is unsupported by any data. Nevertheless,
Fig. 1 indicates that this measure does better
than publication frequency, because the h-
index depends on the entire citation record.
Hirschs measure overestimates the initial
n-assignments by some 8 percentile points
as indicated by higher densities above the
diagonal (Fig. 1c). Moreover, the rms uncer-
tainty in the assignment of h is ±16 percentile
points, which is only a factor of two better than
alphabetization. Although capturing certain
aspects of quality, Hirschs index cannot make
decile assignments with confidence.
Compared with the h-index, the mean
number of citations per paper is a superior
indicator of scientific quality, in terms of both
accuracy and precision. The average assign-
ment of each n-bin is in error by 1.8 percentile
points with an associated rms uncertainty of
±9. Similar calculations based
on authors’ median citation
give an accuracy of 1.5 and
an uncertainty of only ±7 per-
centile points, suggesting that
the median copes better with
long-tailed distributions.
Simple scaling arguments
4
show that the rms uncertainty
for any measure decreases
rapidly (exponentially) as
the total number of papers
increases. Thus, for example,
no more than 50 papers are
required to assign a typical
author to deciles 2–3 or 8–9
with 90% confidence when using the mean
citation rate as a measure. Fewer papers suf-
fice for deciles 1 and 10. Any attempt to assess
the quality of authors using substantially fewer
publications must be treated with caution.
Data access
The methods used here are not specific to
high-energy physics. Given suitably homo-
geneous data sets, they can be applied to
any scientific field and permit a meaningful
(probabilistic) comparison of scientists work-
ing in different fields by assuming the equal-
ity of scientists in the same percentile of their
respective peer groups. Similarly, probabilities
can be combined to make meaningful quality
assignments to authors with publications in
several disjointed subfields.
There are strong indications that an authors
initial publications are drawn on the same
probability distribution as their remaining
papers
7
. Therefore, with sufficient numbers of
publications to draw meaningful conclusions
(50 or more) the mean or median citation
counts can be a useful factor in the academic
appointment process.
Unfortunately, the potential benefits of
careful citation analyses are overshadowed
by their harmful misuse. Institutions have a
misguided sense of the fairness of decisions
reached by algorithm, and unable to measure
what they want to maximize (quality), insti-
tutions will maximize what they can meas-
ure. Decisions will continue
to be made using measures
of quality that either ignore
citation data entirely (such
as frequency of publication)
or rely on data sets of insuffi-
cient quality.
Access to the full citation
distribution for an entire sub-
field is essential to our analy-
sis. Existing databases such as
the ISI can therefore actively
help to improve the situation
by compiling field-specific
homogeneous data sets simi-
lar to what we have generated
for SPIRES. This would allow institutions and
scientists alike to evaluate the quality of any
citation record using all available informa-
tion. For their part, scientists should insist that
their institutions disclose their uses of citation
data, making both data and the methods used
for data analysis available for scrutiny. In the
meantime, we shall have to continue to do
things the old-fashioned way and actually read
the papers.
Sune Lehmann is at the Department of Informatics
and Mathematical Modeling, Technical University
of Denmark, DK-2800, Lyngby, Denmark.
Andrew D. Jackson and Benny E. Lautrup are at
The Niels Bohr Institute, Blegdamsvej 17, DK-
2100, Copenhagen, Denmark.
1. Hirsch, J. E. Proc. Natl Acad. Sci. USA 102, 16569 (2005).
2. van Raan, A. F. J. Scientometrics 67, 491 (2006).
3. Lehmann, S., Lautrup, B. E. & Jackson, A. D. Phys. Rev. E 68,
026113 (2003).
4. van Raan, A. F. J. J. Am. Soc. Inf. Sci. 57, 408 (2005).
5. Thomson Scientific/ISI http://www.isinet.com/
6. Seglen, P. O. J. Am. Soc. Inf. Sci. 45, 1 (1994).
7. Lehmann, S. Thesis, The Niels Bohr Institute (2003).
Supplementary information accompanies this
Commentary on Natures website.
Alphabetized Papers/year
1
2
3
4
5
6
7
8
9
10
n
m
1
2
3
4
5
6
7
8
9
10
n
m
h-index Mean
1
2
3
4
5
6
7
8
9
10
n
m
1
2
3
4
5
6
7
8
9
10
n
m
acd
b
12345678910123456789101234567891012345678910
Figure 1 | The probabilities for four different measures. ad, Each horizontal row, indexed by n, shows the average probabilities that authors initially
assigned to a given decile bin n are predicted to lie in a different decile bin m. The probabilities are proportional to the areas of the corresponding squares.
“Institutions have
a misguided sense
of the fairness of
decisions reached by
algorithm; unable to
measure what they
want to maximize
(quality), they will
maximize what they
can measure.
1004
COMMENTARY
NATURE|Vol 444|21/28 December 2006
Article
Full-text available
In academia, scholarly impact is a prototypical performance criterion used for a variety of purposes, most noticeably, but not exclusively, promotion and tenure decisions. In making these decisions, one may ask, “What constitutes ‘good’ scholarly impact for an industrial-organizational psychologist?” Acknowledging some critiques raised against them, bibliometric indicators of scholarly impact have nevertheless become widely accepted. However, the availability of normative data would potentially further enhance their utility. So, using data based on all 2021 paid-up members of the Society for Industrial and Organizational Psychology who listed academia as their primary work sector, restricted to only those in research-focused academic positions (N = 1075), the study’s objectives were to (1) present normative scholarly impact data for industrial-organizational psychologists in academia, (2) examine the relationships between various scholarly impact metrics in a test of their construct validity, (3) investigate how normative impact varies as a function of various factors, and (4) as a further test of their construct validity, explore the convergence between individual-level rankings based on textbook citations and impact metrics. In doing so, this study contributes to the broader discussion of “how we might define, generate, make sense of, measure, reward, and shape” (Haley et al., 2022, p. 347) scholarly impact. Finally, we provide an interactive dashboard < https://hindex.shinyapps.io/shinyweb/> that presents normative data which can be searched and sorted based on a variety of user-specified variables. The approach outlined and data presented here serve as a template for researchers seeking to generate similar normative data for other fields.
Article
Women have historically encountered numerous barriers and biases that hinder their complete participation and acknowledgement in scientific research. In this study, we scrutinise the gender makeup of scientific teams publishing in top business journals based on a cross-sectional sample of 46,708 publications. Scientific impact is based on the citations, and novelty by using the NLP co-occurrence matrix to compute the cosine similarity between pairs of knowledge entities. Our findings reveal an inverse U-shaped relationship between the proportion of women scientists and scientific impact and a positive U-shaped association with novelty, significantly moderated by team size. These outcomes persist even after many controls and potentially relevant characteristics are taken into account. By delving into the divergent effects of female participation within teams, we not only reveal the nonlinear relationship between team gender composition and scientific discovery but also provide relevant advice for cross-gender collaborative researchers. Our evidence underscores the importance of a more detailed scholarly discussion of the role of women scientists in team performance, at least in business science research.
Article
Full-text available
No one has yet investigated the information reach and influence of thermal analysis. Even so the heat and entropy are bound closely. The very problem of the term entropy is that it was defined specifically within the classical, equilibrium Thermodynamics like Clausius entropy, while its meaning is more general. It is usable in all areas of physics, in the communication theory and technology, and also, in social sciences, just as an universal quantitative model. The bridge to this wider world is in Boltzmann’s statistical point of view used in the Shannon’s information point of view. The difference is that the Shannon entropy is defined more generally, both for the uniform and for the not uniform probability distributions. But with renormalization and with the Boltzmann constant we can speak, using the Clausius entropy, about a certain thermodynamic system considered as a heat model of the area of our wider interests. So we will to deal with the heat and its entropy embodied to a wider consideration. In addition to the standard physical meaning, another thermodynamic realization in the form of a message is derived from Shannon’s information theory, where the sample carries a measure of Shannon’s information given by its structure or internal organization. This is also a contribution to improving the approach to thinking about the importance of observation and measurement, reflecting at the same time the costs and influence of the measurement itself and its organization on the measured object or, on its environment or, even on the environment of the measuring itself generally. Statistical sorting also carries with it the internal movement of individual elementary particles, the value of which indicates the resulting state of heat transfer and dissipation to and from the system providing a state of called equilibrium. It is statistically described by a weighted quantity called temperature, the changes of which via momentum alteration satisfy Newton’s laws of inertia allowing enumeration of the related heat information processing. The terminology used is still non-institutional.
Article
Measuring the influence of scientists and their activities on science and society is important and indeed essential for many studies. Despite the substantial efforts devoted to exploring the influence’s measures and patterns of an individual scientific enterprise, it remains unclear how to quantify the mutual impact of multiple scientific activities. This work quantifies the relationship between the scientists’ interactive activities and their influences with different patterns in the AMiner dataset. Specifically, inflation treatment and field normalization are introduced to process the big data of paper citations as the scientist’s influence, and then the evolution of the influence is investigated for scientific activities in the citation and cooperation patterns through the Hawkes process. The results show that elite scientists have higher individual and interaction influences than ordinary scientists in all patterns found in the study, with permutation tests verifying the significance of the new findings. Moreover, the study compares the patterns found in two largest disciplines, i.e., STEM and Humanities , revealing the higher value of individual influence in STEM than in Humanities . Furthermore, it is found that the opposite trend of STEM and Humanities in the cooperation pattern suggests different cooperation habits of scientists in different disciplines. Overall, this investigation provides a feasible approach to addressing the scientific influence issue and deepening the quantitative understanding of the mutual influence of multiple scientific activities in science and society.
Article
Full-text available
Background Mesenchymal stem cell-derived exosomes (MSC-exosomes) have been found to effectively improve the systemic inflammatory response caused by acute lung injury and acute respiratory distress syndrome (ALI/ARDS), regulate systemic immune disorders, and help injured cells repair. The purpose of this study was to take a holistic view of the current status and trends of MSC-exosomes research in ALI/ARDS. Methods Bibliometrix, Citespace and VOSviewer software were used for bibliometric analysis of the data. We analysed the world trends, country distribution, institution contribution, most relevant journals and authors, research hotspots, and research hotspots related to Coronavirus Disease 2019 (COVID-19) based on the data collected. Results China possessed the largest number of publications, while the USA had the highest H-index and the number of citations. Both China and the USA had a high influence in this research field. The largest number of publications in the field of MSC-exosomes and ALI/ARDS were mainly from the University of California system. Stem Cell Research & Therapy published the largest number of papers in this scope. The author with the greatest contribution was LEE JW, and ZHU YG published an article in Stem Cell with the highest local citation score. The most frequent keyword and the latest research hotspot were “NF-κB” and “Coronavirus Disease 2019”. Furthermore, our bibliometric analysis results demonstrated that MSC-exosomes intervention and treatment can effectively alleviate the inflammatory response caused by ALI/ARDS. Conclusion Our bibliometric study suggested the USA and China have a strong influence in this field. COVID-19-induced ALI/ARDS had become a hot topic of research.
Article
Full-text available
There is a significant gender gap in accounting academia that places women at a disadvantage in terms of recruitment, hiring, promotion, tenure, status, high-level areas or positions (both research and administrative), burden distribution of work, and remuneration. Women are disproportionately represented in part-time or non-tenure tracks, such as lecturers, instructors, and assistant professors. They experience a slower rate of advancement and have lower pay and prestige. Given that various authors attribute this situation to the level of research and production of papers in top-tier scientific journals, this article aims to describe women's participation as authors in cost and management accounting to contribute to clarifying possible causes of gender disparity in the accounting case.
Article
Evaluation of papers’ academic influence is a hot issue in the field of scientific research management. Academic big data provides a data treasure with the coexistence of different types of academic entities, which can be used to evaluate academic influence from a more macro and comprehensive perspective. Based on academic big data, a heterogeneous academic network composed of links within and between three types of academic entities (authors, papers and venues) is constructed. In addition, a new academic influence ranking algorithm, AIRank, is proposed to evaluate papers’ academic influence. Different from the existing academic influence ranking algorithms, AIRank has made innovations in the following two aspects. (1) AIRank distinguishes the influence transmission intensity between different node pairs. Different from the strategy of evenly distributing influence among different node pairs, AIRank quantifies the intensity of influence transmission between node pairs based on investigating the citation emotional attribute, semantic similarity and academic quality differences between node pairs. Based on the intensity characteristics, AIRank realises the distribution and transmission of influence among different node pairs. (2) AIRank incorporates the influence transmission from heterogeneous neighbours in evaluating papers’ influence. According to the academic influence of author nodes and venue nodes, AIRank fine-tunes the iteration formula of paper influence to obtain the ranking of papers under the joint influence of homogeneous and heterogeneous neighbours. Experimental results show that, compared with the ranking results based on citation frequency and PageRank algorithm, AIRank algorithm can produce more differentiated and reasonable academic influence ranking results.
Article
We seek a unified and distinctive citation description of both journals and individuals. The journal impact factor has a restrictive definition that constrains its extension to individuals, whereas the h-index for individuals can easily be applied to journals. Going beyond any single parameter, the shape of each negative slope Hirsch curve of citations vs. rank index is distinctive. This shape can be described through five minimal parameters or ‘flags’: the h-index itself on the curve; the average citation of each segment on either side of h; and the two axis endpoints. We obtain the five flags from real data for two journals and 10 individual faculty, showing they provide unique citation fingerprints, enabling detailed comparative assessments. A computer code is provided to calculate five flags as the output, from citation data as the input. Since papers (citations) can form nodes (links) of a network, Hirsch curves and five flags could carry over to describe local degree sequences of general networks.
Article
A well-designed method for evaluating scientists is vital for the scientific community. It can be used to rank scientists in various practical tasks, such as hiring, funding application and promotion. However, a large number of evaluation methods are designed based on citation counts which can merely evaluate scientists’ scientific impact but can not evaluate their innovation ability which actually is a crucial characteristic for scientists. In addition, when evaluating scientists, it has become increasingly common to only focus on their representative works rather than all their papers. Accordingly, we here propose a hybrid method by combining scientific impact with innovation under representative works framework to evaluate scientists. Our results are validated on the American Physical Society journals dataset and the prestigious laureates datasets. The results suggest that the correlation between citation and disruption is weak, which enables us to incorporate them. In addition, the analysis shows that using representative works framework to evaluate scientists is advantageous and our hybrid method can effectively identify the Nobel Prize laureates and several other prestigious prizes laureates with higher precision and better mean ranking. The evaluation performance of the hybrid method is shown to be the best compared with the mainstream methods. This study provides policy makers an effective way to evaluate scientists from more comprehensive dimensions.
Article
Full-text available
Scientific literature, as the major medium that carries knowledge between scientists, exhibits explosive growth in the last century. Despite the frequent use of many tangible measures, to quantify the influence of literature from different perspectives, it remains unclear how knowledge is embodied and measured among tremendous scientific productivity, as knowledge underlying scientific literature is abstract and difficult to concretize. In this regard, there has laid a vacancy in the theoretical embodiment of knowledge for their evaluation and excavation. Here, for the first time, we quantify the knowledge from the perspective of information structurization and define a new measure of knowledge quantification index (KQI) that leverages the extent of disorder difference caused by hierarchical structure in the citation network to represent knowledge production in the literature. Built upon 214 million articles, published from 1800 to 2021, KQI is demonstrated for mining influential classics and laureates that are omitted by traditional metrics, thanks to in-depth utilization of structure. Due to the additivity of entropy and the interconnectivity of the network, KQI assembles numerous scientific impact metrics into one and gains interpretability and resistance to manipulation. In addition, KQI explores a new perspective regarding knowledge measurement through entropy and structure, utilizing structure rather than semantics to avoid ambiguity and attain applicability.
Article
Full-text available
In this paper we present characteristics of the statistical correlation between the Hirsch (h-) index and several standard bibliometric indicators, as well as with the results of peer review judgment. We use the results of a large evaluation study of 147 university chemistry research groups in the Netherlands covering the work of about 700 senior researchers during the period 1991-2000. Thus, we deal with research groups rather than individual scientists, as we consider the research group as the most important work floor unit in research, particularly in the natural sciences. Furthermore, we restrict the citation period to a three-year window instead of life time counts in order to focus on the impact of recent work and thus on current research performance. Results show that the h-index and our bibliometric crown indicator both relate in a quite comparable way with peer judgments. But for smaller groups in fields with less heavy citation traffic the crown indicator appears to be a more appropriate measure of research performance.
Article
The relationship between article citedness and journal impact was investigated on the basis of complete publication lists provided by 16 senior scientists from a major Norwegian biomedical research institute. The citedness of each individual journal article was measured as the mean annual number of citations during the second to fourth year after publication, and compared with the impact factor (mean article citedness) of each corresponding journal, recorded during the first two years after publication. The distribution of article citedness was very skewed (corresponding to a topnormal, extreme-property distribution yielding linearity in a semilogarithmic plot), even for individual authors and within defined journal impact cohorts. The large variability was not due to random citation of the individual article, but rather to a fundamental heterogeneity among the articles. A similar skewness was found in the distribution of journal impact values, whereas the author impact distribution was less heterogeneous. The skewed distributions resulted in poor correlations between article citedness and journal impact, for the whole article population (correlation coefficient, c=0.41) as well as for individual authors (mean c=0.32; range 0.05–0.66) and journal pairs (mean c=0.22). Pooling of articles into defined journal impact cohorts dramatically improved the correlation (c=0.999 for the study group as a whole), indicating that the authors tended to submit their most cited work to journals of higher impact (although several notable exceptions were observed among individual authors). However, very large numbers of articles (50–100) had to be pooled in order to obtain good correlations (c=0.8–0.9). Furthermore, the mean citedness of the study group (3.34 ± 0.16 citations/article/year; mean ± SE of 907 articles) differed significantly from the corresponding mean journal impact (2.61 ± 0.07); for individual authors even larger discrepancies were observed. Use of journal impact as an evaluation parameter may therefore yield highly misleading results, unless the evaluated unit (author, research group, institution, or country) happens to be equal to the world average. By dividing the authors into a highly cited group (the eight most cited authors) and a less cited group (the eight least cited authors) it was possible to observe a twofold ratio in citedness between the two groups throughout the journal impact range. This difference could not be accounted for by journal choice, and did not diminish with increasing journal impact. The citedness of journal articles thus does not seem to be detectably influenced by the status of the journal in which they are published. © 1994 John Wiley & Sons, Inc.
Article
In this article we present an empirical approach to the study of the statistical properties of bibliometric indicators on a very relevant but not simply “available” aggregation level: the research group. We focus on the distribution functions of a coherent set of indicators that are used frequently in the analysis of research performance. In this sense, the coherent set of indicators acts as a measuring instrument. Better insight into the statistical properties of a measuring instrument is necessary to enable assessment of the instrument itself. The most basic distribution in bibliometric analysis is the distribution of citations over publications, and this distribution is very skewed. Nevertheless, we clearly observe the working of the central limit theorem and find that at the level of research groups the distribution functions of the main indicators, particularly the journal- normalized and the field-normalized indicators, approach normal distributions. The results of our study underline the importance of the idea of “group oeuvre,” that is, the role of sets of related publications as a unit of analysis.
Article
The citation network constituted by the SPIRES database is investigated empirically. The probability that a given paper in the SPIRES database has k citations is well described by simple power laws, P(k) proportional to k(-alpha), with alpha approximately 1.2 for k less than 50 citations and alpha approximately 2.3 for 50 or more citations. A consideration of citation distribution by subfield shows that the citation patterns of high energy physics form a remarkably homogeneous network. Further, we utilize the knowledge of the citation distributions to demonstrate the extreme improbability that the citation records of selected individuals and institutions have been obtained by a random draw on the resulting distribution.
Article
I propose the index h, defined as the number of papers with citation number ≥h, as a useful index to characterize the scientific output of a researcher. • citations • impact • unbiased
  • S Lehmann
  • B E Lautrup
  • A D Jackson
Lehmann, S., Lautrup, B. E. & Jackson, A. D. Phys. Rev. E 68, 026113 (2003).
  • A F J Van Raan
van Raan, A. F. J. Scientometrics 67, 491 (2006).
The Niels Bohr Institute
  • S Lehmann
  • Thesis
Lehmann, S. Thesis, The Niels Bohr Institute (2003).
  • P O J Seglen
Seglen, P. O. J. Am. Soc. Inf. Sci. 45, 1 (1994).