Research Paper

Equalities between h-type Indices and Definitions of Rational h-type Indicators

  • Leo Egghe 1 ,
  • Yves Fassin 2 ,
  • Ronald Rousseau , 3, 4†
Expand
  • 1University of Hasselt, Belgium
  • 2Department of Marketing, Innovation and Organisation, Faculty of Economics and Business Administration, Ghent University, Tweekerkenstraat, 2, 9000 Gent, Belgium
  • 3University of Antwerp, Faculty of Social Sciences, Middelheimlaan 1, 2020, Antwerpen, Belgium
  • 4KU Leuven, Facultair Onderzoekscentrum ECOOM, Naamsestraat 61, 3000 Leuven, Belgium
Corresponding author: Ronald Rousseau (E-mail: ).

Received date: 2018-11-24

  Request revised date: 2018-12-08

  Accepted date: 2018-12-11

  Online published: 2019-01-31

Copyright

Open Access

Abstract

Purpose: To show for which publication-citation arrays h-type indices are equal and to reconsider rational h-type indices. Results for these research questions fill some gaps in existing basic knowledge about h-type indices.

Design/methodology/approach: The results and introduction of new indicators are based on well-known definitions.

Findings: The research purpose has been reached: answers to the first questions are obtained and new indicators are defined.

Research limitations: h-type indices do not meet the Bouyssou-Marchant independence requirement.

Practical implications: On the one hand, more insight has been obtained for well-known indices such as the h- and the g-index and on the other hand, simple extensions of existing indicators have been added to the bibliometric toolbox. Relative rational h-type indices are more useful for individuals than the existing absolute ones.

Originality/value: Answers to basic questions such as “when are the values of two h-type indices equal” are provided. A new rational h-index is introduced.

Cite this article

Leo Egghe , Yves Fassin , Ronald Rousseau . Equalities between h-type Indices and Definitions of Rational h-type Indicators[J]. Journal of Data and Information Science, 2019 , 4(1) : 22 -31 . DOI: 10.2478/jdis-2019-0002

1 Introduction

Definition: The classical h-index
Consider a set S of publications, ranked decreasingly according to the number of citations each of these publications has received. Publications with the same number of citations are given different rankings. Then the h-index of set S is h if the first h publications received each at least h citations, while the publication ranked h+1 received strictly less than h+1 citations. Stated otherwise: the h-index of set S is the largest natural number h such that the first h publications received at least h citations (Hirsch, 2005).
When applied to the publication list of a researcher the previous definition favors more prolific, e.g. older, scientists above those with less publications, e.g. younger ones. For this reason one may use a publication window in calculating an h-index. Also the citation window can be adapted to make a difference between short-term and long - term influence. As databases differ in content an h-index may also differ according to the used database. Besides these adaptations of the original definition it is also possible to calculate h-indices for other types of citations, e.g. of patents and for fractionally counted items.
Definition: the g-index
Additional citations to publications among the first h play no role at all. For this reason another indicator has been introduced. This is the g-index, proposed by Egghe (2006a). It is defined as follows: articles are ranked in decreasing order of received citations (as for the h-index). Then the g-index of this set of articles is defined as the highest rank g such that these g articles together received at least g2 citations. If necessary, fictitious articles with zero citations are added to the publication list.
Definition: Kosmulski’s index and its generalizations
Another variation on the h-index was introduced by Kosmulski (2006). He proposed the h(2)-index as follows. Again one ranks the set of articles for which one wants to determine the h(2)-index in decreasing order of received citations. Now this set (authors, journals, etc.) has an h(2)-index equal to h2 if r = h2 is the highest rank such that the first h2 articles each received at least (h2)2 citations.
As a next step, colleagues observed that one may define in a similar manner an h(k)-index (k = 1, 2, 3, ….). This has been done e.g. in (Deineko & Woeginger, 2009), who proposed an axiomatic characterization of an even more general family of indices and in (Egghe, 2011), who studied this index in a Lotkaian framework.
Concretely the h(3) index is defined as follows. Consider a list of articles ranked decreasingly according to the number of citations each of these articles has received. Articles with the same number of citations are given different rankings. Then the h(3)-index of this set S is h3 if the first h3 articles received each at least (h3)3 citations, while the article ranked h3+1 received strictly less than (h3+1)3 citations. Stated otherwise: the h(3)-index of a set S is the largest natural number h3 such that the first h3 publications each received at least (h3)3 citations (Fassin & Rousseau, 2018).
In this contribution we represent the units of attention (authors, journals, research groups, etc. ) as a finite array such as A = (10, 7, 7, 2, 0). This symbol shows that author A has five publications with respective (ranked) citations equal to 10, 7, 7, 2 and 0. Clearly author A has an h-index equal to 3 and a g-index equal to 5. The h(2)-index is equal to 2 and the h(3)-index is equal to 1. The number of items with a non-zero number of citations is called the length of the array. This array has length 4. For simplicity we will always assume that values in array A are natural numbers (including the value zero). In this contribution we restrict our attention to the g, h, h(2) and the h(3) index, to which we refer as h-type indices.

2 When do we have equality?

It follows from their definitions that always g ≥ h ≥ h(2) ≥ h(3). In this section we tackle the question: for which arrays are two different h-type indices equal?
A. When is h = h(2)?
We recall the two conditions: a set of articles has h-index h if the first h articles received at least h citations and the article ranked h+1 received strictly less than h+1 citations. Similarly: a set of articles has h(2)-index h2 if the first h2 articles received at least (h2)2 citations each and the article ranked h2+1 received strictly less than (h2+1)2 citations. If the two conditions must be satisfied at the same time, then the first h articles must have received at least h2 citations each and the article ranked h+1 must have strictly less than h+1 citations. Obviously, h = h(2) can only occur for an array of length at least equal to h.
The following array A = (100, 30, 9, 3) is an example for which h = h(2) = 3.
The least number of citations for the case h = h(2) = 3, occurs for the array (9, 9, 9, 0). We added a non-essential zero at the end to make it clear that the length of this array is three. Generally, the least number of citations for the case h = h(2) is \((\begin{matrix} \underbrace{ h^2+h^2,...,h^2},0 \\ h \ times\end{matrix})\). Of course, there is no upper limit to the corresponding number of citations.
B. When is h = h(3) or equivalently, when is h = h(2) = h(3) ?
As, by definition h(2) is always situated between h and h(3), it suffices to solve the problem: when is h = h(3)?
Again we recall the two conditions: a set of articles has h-index h if the first h articles received at least h citations and the article ranked h+1 received strictly less than h+1 citations. Similarly: a set of articles has h(3)-index h3 if the first h3 (here equal to h) articles received at least (h3)3 citations each and the article ranked h3+1 received strictly less than (h3+1)3 citations. If the two conditions must be satisfied at the same time, then the first h articles must have received at least h3 citations each and the article ranked h+1 must have strictly less than h+1 citations.
The following array A = (100, 30, 27, 3) is an example for which h = h(3) = 3.
The least number of citations for the case h = h(3) = 3, occurs for the array (27, 27, 27, 0). Generally, the least number of citations for the case h = h(3) is \((\begin{matrix} \underbrace{ h^3+h^3,...,h^3},0 \\ h \ times\end{matrix})\). Of course, even when h = h(2) = h(3) there is no upper limit to the corresponding number of citations.
C. When is h(2) = h(3)?
Recall that a set of articles has h(2)-index h2 if the first h2 articles received at least (h2)2 citations each and the article ranked h2+1 received strictly less than (h2+1)2 citations; and a set of articles has h(3)-index h3 if the first h3 articles received at least (h3)3 citations each and the article ranked h3+1 received strictly less than (h3+1)3 citations. If the two conditions must be satisfied at the same time then the first h2articles must have received at least (h2)3 citations each and the article ranked h2+1 must have strictly less than (h2+1)2 citations.
The following array A = (100, 30, 27, 15) is an example for which h(2) = h(3) = 3. Note that this array has an h-index equal to 4.
The least number of citations for the case h(2) = h(3) = 3, occurs for the array (27, 27, 27, 0). Generally, the least number of citations for the case h(2) = h(3) is \((\begin{matrix} \underbrace{ h^3+h^3,...,h^3},0 \\ h \ times\end{matrix})\).
D. When is h = g?
A set of articles has g-index h if the sum of the citations of the first h articles is at least h2 and the sum of the first h+1 articles is strictly less than (h+1)2. If X = (x1, x2, …xj,…) then we see that if g(X) = h, then \(\sum\limits_{i=1}^{h}{x_i}≥{h}^{2}\). This inequality always holds if the h-index of X is equal to h. Now from \(\sum\limits_{i=1}^{h+1}{x_i}<{(h+1)}^{2}\) and the fact that \(\sum\limits_{i=1}^{h}{x_i}≥{h}^{2}\), we see that if xh+1 = h then \({h}^{2}<\sum\limits_{i=1}^{h}{x_i}<{h}^{2}+h+1\).
For h = g = 3, xh+1 = x4 = 3 and for the largest possible number of citations for x1 we have: (6, 3, 3, 3) as an example. If xh+1 = x4 = 0 we have (9, 3, 3, 0) again for the largest possible value of x1. An example, still for g=h=3, of an intermediate case is (5, 4, 3, 2). In general, again trying to give the first item the largest possible value, we have for the largest possible integer value, namely h, for item h+1 an array of the form \((\begin{matrix} \underbrace{ 2h+h+h,...,h},h \\ h-1 \ times\end{matrix})\). If item h+1 has value zero, then we have: \((\begin{matrix} \underbrace{ 3h+h+h,...,h},0 \\ h \ times\end{matrix})\)
E. When is g = h = h(2) = h(3)?
If h = 3 then the condition h = h(2) = h(3) leads to an array of the form (27, 27, 27, 0), or with higher values. As 27+27+27+0 = 81 = 92 we observe that the g-index is at least 9. Hence the equality g = h = h(2) = h(3) is not possible for h=3, and certainly not for higher values.
If h = 2, then h = h(2) = h(3) leads to an array of the form (8, 8, 0), or with higher values. As 8+8+0=16=42 the g-index is at least 4. Hence, also for h=2 it is impossible to have equality.
Finally for h=1 it is easy to find examples for which g = h(3) such as (2, 1). As the sum of the first two citations must be at most equal to 3, this example is an extreme. Similarly (3, 0) is an extreme. Among publication-citation arrays of length one (3), (2) and (1) are the only three cases; among publication arrays of length two we have (2, 1) and (1, 1). From this we conclude that equality among the four indices can only occur for h=1 and, even then, occurs in just a few cases.
Note that there are no conditions on the tail so that there is no condition on the total number of citations. The array (2, 1, 1, 1,...) has h-index = g-index = h(3)-index = 1, but there is no upper limit on the total number of received citations. If the number of articles, N, is given, then the upper limit for the number of received citations is N+1; the lower limit is 1.

3 Rational indices

Rational h-type indices can be used to make a distinction between cases with the same h-type value. The rational variant of the h-index, denoted as hrat, was introduced by Ruane and Tol (2008) in the context of publications and citations. It is defined as follows.
Definition: Consider a researcher with h-index h. Let n be the smallest possible number of citations necessary to reach an h-index equal to h+1, then the rational h-index, denoted hrat, is defined as:
\[h_{rat}=h+1-\frac {n}{2h+1}\ \ (1)\]
We next explain this formula. If a researcher has h-index h, then one may ask about the minimum number of citations necessary to reach an h-index equal to h + 1. This number is denoted here as n. The next question is now: if you only know that this scientist’s h-index is h what is then the largest number of citations that this researcher needs to reach an h-index equal to h+1. The answer is 2h+1, corresponding with the “worst case scenario” that there are h publications with h citations each and the publication at rank h + 1 has 0 citations. This explains the occurrence of the factor 2h+1 in the formula for the rational h-index (Rousseau et al., 2018). In a similar way a rational g-index was introduced in (Guns & Rousseau, 2009). Next we define the rational h(2) and h(3) indices.
Similar to the case of the h-index we note that the worst case for a set of articles with h(2) index equal to h2 happens when the first h2 articles received (h2)2 citations and the article ranked h2+1 has no citations. Such an article needs h2 times (h2+1)2-(h2)2 = h2(2h2+1) extra citations plus (h2+1)2 new citations, leading to a total of 3(h2)2+3h2+1 citations. Consequently:
\[{(h_{2})}_{rat}=h_2+1-\frac {n_2}{3h^{2}_{2}+3h_{2}+1}\ \ (2)\]
where n2 is the minimum number of citations necessary to reach an h(2)-index equal to h(2) + 1.
Finally, for the h(3) index we note that the worst case for a set of articles with h(3) index equal to h3 happens when the first h3 articles received (h3)3 citations and the article ranked h3+1 has no citations. Such an article needs h3 times (h3+1)3-(h3)3 = 3(h3)2 + 3(h3) +1 extra citations plus (h3+1)3 new citations, leading to a total of 4(h3)3 + 6(h3)2 + 4h3 + 1 citations. Consequently:
\[{(h^{(3)})}_{rat}=h^{(3)}+1-\frac {n_3}{4h^{3}_{3}+6h^{2}_{3}+4h_{3}+1}\ \ (3)\]
where n3 is the minimum number of citations necessary to reach an h(3)-index equal to h(3) + 1.
An example. Array A = (100, 30, 27, 3) has a rational h(3)-index of \(3+1-\frac {0+34+37+61}{4*27+6*9+4*3+1}=4-\frac {132}{175}≈3.246 \).

4 The relative rational h-index

When researchers reach an h-index of h, it will rarely occur that they really need 2h+1 new citations to reach the value h+1. Usually some of these citations may already have been received. In the extreme case they will only need two new citations, namely when their publication-citation array is \((\begin{matrix} \underbrace{ h+1,h+1,...,h+1},h,h \\ h-1 \ times\end{matrix})\) or with more citations. If, at the moment a researcher reaches an h-index of h, they need m new citations to reach an h-index of h+1, then at a later moment their relative (or individual) rational h-index, denoted hr,rat, is
\[h_{r,rat}=h+1-\frac {n}{m}\ \ (4)\]
where n has the same meaning as before, namely: the minimum number of citations still necessary to reach an h-index equal to h + 1. As m ≤ 2h+1, hr,rat ≤ hrat. For an individual researcher this relative rational h-index is clearly more meaningful than the absolute one. An example: if A0 = (6, 5, 3, 1) when an h-index of 3 was reached and if this researcher’s publication-citation array is now A = (9, 6, 4, 2), then their relative rational h-index is 4 - 2/4 = 3.5; the absolute one would be 4-2/7 ≈ 3.71. Similarly, one may define relative rational g, h(2) and h(3) indices and apply them not only to persons, but also to journals or other units of interest.

5 Equality between h and g in a Lotkaian framework

In this section we use a continuous framework. This has no direct application in research evaluation, but it is part of a context in which researchers use a continuous version of h-type indices for modelling purposes (Egghe, 2005). We first recall the definition of the h- and the g-index in this framework. If f(r) is a given rank-frequency function (Zipf-type) then the h-index is the solution of the equality (in r):
\[f(r)=r (5)\]
while the g-index is the solution, g, of the equality
\[\int ^{g}_{0}f(r)dr=g^2 (6)\]
We recall that always (in a continuous as well as a discrete framework) g ≥ h. It has been shown (Egghe & Rousseau, 2006) that in a Lotkaian framework
\[h=T^{(1-α)} \ for \ \ α>1 \\ (7)\]
where α is the exponent of the underlying Lotka (power) function and T is the total number of sources. Similarly, it has been shown (Egghe, 2006b) that
\[g=(\frac {α-1}{α-2})^{{(α-1)/α} \ } \ {{T}^{1/α}} \ for \ \ α>2 \\ (8)\]
Now we prove that, for α > 2, and for fixed T, g-h is decreasing in α.
Proof. We consider the derivative of g-h with respect to α and prove that this derivative is always negative.
\[\frac{d}{dα}{(g(α)-h(α))}=\frac{d}{dα}(((\frac {α-1}{α-2})^{(α-1)/α}-1){{T}^{1/α}})=((\frac {α-1}{α})(\frac {α-1}{α-2})^{{\frac {α-1}{α}}-1} \ {\frac {α-2-(α-1)}{{(α-2)}^2}}+ \\ ln(\frac {α-1}{α})(\frac {α-1}{α-2})^{{\frac {α-1}{α}}} \ {\frac {α-2-(α-1)}{{α}^2}} \}{T^{1/α}} \\ +(((\frac {α-1}{α-2})^{{({α-1})/α}}-1\} \ 1n(T^{1/α})(-\frac{1}{α^{2}}))\ (9)\]
The first factor is positive; the first term of the second factor is clearly negative, being a product of a positive and a negative factor. Now the second term of the second factor is a positive number multiplied by negative one (shown below), so that the derivative of g-h with respect to α is negative. This proves that, for fixed T, the difference between g and h decreases with α.
Now we consider the factor $(\frac {-1}{α(α-2)})+ln(\frac {α-1}{α(α-2)})\frac {1}{α^{2}}$ and show that it is negative. This holds if . in $(\frac {α-1}{α-2})<\frac {α}{α-2}$ . Taking exponentials we have to show that $\frac {α-1}{α-2}<e^{α/(α-2)}=1+\frac {α}{α-2}+\frac {1}{2} (\frac {α}{α-2})^{2}$. Now $ e^{α/(α-2)}=1+\frac {α}{α-2}+\frac {1}{2} (\frac {α}{α-2})^{2}+…>1+\frac {α}{α-2}=\frac {2α-2}{α-2}$. As α > 2, this boils down to the inequality 2α-2 > α-1, which is clearly true. This proves the required inequality.
Now $ \lim_{{α \to \infty} \\ _{T \ CONSTAN}} \ (g-h)= lim_{{α \to \infty} \\ _{T \ CONSTAN}} (((\frac {α-1}{α-2})^{(α-1)/α}-1){{T}^{1/α}})=0$. This shows that for fixed T, g tends to h. This is easy to understand: indeed, if the Lotka-coefficient α tends to infinity, the Zipf-coefficient β (the coefficient of the Zipf distribution equivalent of the Lotka distribution with coefficient α) tends to zero (recall that $β=\frac {1}{α-1}$). Now a Zipf-coefficient equal to zero corresponds to a ranking in which all elements are equal, which means that g = h.

6 Discussion and conclusion

We derived conditions under which h-type indices, and in particular the h and the g-index, are equal. Next we introduced the rational h(2) and h(3)-index. We moreover proposed a relative or individual rational h-index. Finally we studied the limiting behavior of the difference g-h in a continuous Lotkaian framework.
Although this article is explicitly meant to be a contribution in theoretical informetrics, it might have some practical use. This holds, in particular, for the introduction of the relative rational h-index.
We recognize that all h-type indicators do not always behave in a logical way (Bouyssou & Marchant, 2011; Waltman & van Eck, 2012). Like many other indicators the h, h(2), h(3) and g indicator are only PAC (Probably Approximately Correct) (Rousseau, 2016). However, this practical observation has no direct relation with the mathematical properties studied in this contribution. We further note that these h-type indices may play a role in heuristic approaches to support informed peer review (Bornmann et al., 2018).

Author Contributions

All the authors, Leo Egghe (leo.egghe@uhasselt.be), Yves Fassin (fassin@skynet.be, Yves.Fassin@Ugent.be), and Ronald Rousseau (ronald.rousseau@uantwerpen.be, ronald.rousseau@kuleuven.be) conceived and designed the analysis, contributed to the development of original idea, and worte the paper.

The authors have declared that no competing interests exist.

[1]
Bornmann L., Hug S., & Marewski J.N. (2018). Bibliometrics-based heuristics: What is their definition and how can they be studied? arXiv:1810.13005Abstract: Paradoxically, bibliometric indicators (i.e., publications and citation counts) are both widely used and widely criticized in research evaluation. At the same time, a common methodological and theoretical framework for conceptually understanding, empirically investigating, and effectively training end-users of bibliometrics (e.g., science managers, scientists) is lacking. In this paper, we outline such a framework - the fast-and-frugal heuristics research framework developed by Gigerenzer et al. [1] - and discuss its application to evaluative bibliometrics. Heuristics are decision strategies that use part of the available information (and ignore the rest). In so doing, they can aid to make accurate, fast, effortless, and cost-efficient decisions without that trade-offs are incurred (e.g., effort versus accuracy). Because of their simple structure, heuristics are easy to understand and communicate and can enhance the transparency of decision-making processes. We introduce three bibliometrics-based heuristics and discuss how these heuristics can be employed in the evaluative practice (using the evaluation of applicants for funding programs as example).

[2]
Bouyssou D., &Marchant T. (2011). Ranking scientists and departments in a consistent manner. Journal of the American Society for Information Science and Technology, 62(9), 1761-1769.The standard data that we use when computing bibliometric rankings of scientists are their publication/ citation records, i.e., so many papers with 0 citation, so many with 1 citation, so many with 2 citations, etc. The standard data for bibliometric rankings of departments have the same structure. It is therefore tempting (and many authors gave in to temptation) to use the same method for computing rankings of scientists and rankings of departments. Depending on the method, this can yield quite surprising and unpleasant results. Indeed, with some methods, it may happen that the “best” department contains the “worst” scientists, and only them. This problem will not occur if the rankings satisfy a property called consistency, recently introduced in the literature. In this article, we explore the consequences of consistency and we characterize two families of consistent rankings.

DOI

[3]
Deineko V.G., &Woeginger G.J. (2009). A new family of scientific impact measures: The generalized Kosmulski-indices . Scientometrics, 80(3), 821-828.This article introduces the generalized Kosmulski-indices as a new family of scientific impact measures for ranking the output of scientific researchers. As special cases, this family contains the well-known Hirsch-index h and the Kosmulski-index h (2) . The main contribution is an axiomatic characterization that characterizes every generalized Kosmulski-index in terms of three axioms.

DOI

[4]
Egghe L. (2005). Power laws in the Information Production Process: Lotkaian Informetrics. Amsterdam: Elsevier.This book describe informetric results from the point of view of Lotkaian size-frequency functions, i.e. functions that are decreasing power laws. Explanations and examples of this model are given showing that it is the most important regularity amongst other possible models. This theory is then developed in the framework of IPPs (Information Production Processes) herby also indicating its relation with e.g. the law of Zipf. Applications are given in the following fields: three-dimensional informetrics (positive reinforcement and Type/Token-Taken informetrics), concentration theory (including the description of Lorenz curves and concentration measures in Lotkaian informetrics), fractal complexity theory(Lotkaian informetrics as self-similar fractals), Lotkaian informetrics in which items can have multiple sources(where frational size-frequency functions are constructed), the theory of first-citation distributions and the N-fold Cartesian product of IPPs (describing frequency functions for N-grams and N-word phrases. In the Appendix, methods are given to determine the parameters in the law of Lotka, based on a set of discrete data. The book explains numerous informetric regularities, only based on a decreasing power law as size-frequecy function, i.e. Lotka's law. It revives the historical formulation of Alfred Lotka of 1926 and shows the power of this power law, both in classical aspects of informetrics(libraries, bibliographies) as well as in 'new' applications such as social networks (citation or collaboration networks and the internet).

[5]
Egghe L. (2006a). An improvement of the h-index: The g-index. ISSI Newsletter, 2(1), 8-9.

[6]
Egghe L. (2006b). Theory and practice of the g-index. Scientometrics, 69(1), 131-152.

DOI

[7]
Egghe L. (2011). Characterizations of the generalized Wu- and Kosmulski-indices in Lotkaian systems. Journal of Informetrics, 5(3), 439-445.We define the generalized Wu- and Kosmulski-indices, allowing for general parameters of multiplication or exponentiation. We then present formulae for these generalized indices in a Lotkaian framework.Next we characterise these indices in terms of their dependence on the quotient of the average number of items per source in the m-core divided by the overall average (m is any generalized Wu- or Kosmulski-index).As a consequence of these results we show that the fraction of used items (used in the definition of m) in the m-core is independent of the parameter and equals one divided by the overall average.

DOI

[8]
Egghe L. , & Rousseau, R. (2006). An informetric model for the h-index. Scientometrics, 69(1), 121-129.The h -index (or Hirsch-index) was defined by Hirsch in 2005 as the number h such that, for a general group of papers, h papers received at least h citations while the other papers received no more than h citations. This definition is extended here to the general framework of Information Production Processes (IPPs), using a source-item terminology. It is further shown that in each practical situation an IPP always has a unique h -index. In Lotkaian systems h = T 1 / a , where T is the total number of sources and 伪 is the Lotka exponent. The relation between h and the total number of items is highlighted.

DOI

[9]
Fassin .Y. , &Rousseau ,R. (2018). The h(³) - index for academic journals. Preprint.

[10]
Guns R.,& Rousseau , R. (2009). Real and rational variants of the h-index and the g-index. Journal of Informetrics, 3(1), 64-71.The definitions of the rational and real-valued variants of the -index and -index are reviewed. It is shown how they can be obtained both graphically and by calculation. Formulae are derived expressing the exact relations between the -variants and between the -variants. Subsequently these relations are examined. In a citation context the real -index is often, but not always, smaller than the rational -index. It is also shown that the relation between the real and the rational -index depends on the number of citations of the article ranked . Maximum differences between , r and rat on the one hand and between , r and rat on the other are determined.

DOI

[11]
Hirsch J.E. (2005). An index to quantify an individual’s scientific research output. Proceedings of the National Academy of Sciences USA 102(46), 16569-16572.

[12]
Kosmulski M. (2006). A new Hirsch-type index saves time and works equally well as the original h-index. ISSI Newsletter, 2(3), 4-6.

[13]
Rousseau R.(2016). Citation data as a proxy for quality or scientific influence are at best PAC (Probably Approximately Correct). Journal of the Association for Information Science and Technology, 67(12), 3092-3094.Abstract In this communication I give a brief introduction to Valiant's probably approximately correct (PAC) theory, provide an extension that goes beyond Valiant's ideas (and beyond the domain for which this theory was meant), and come to an interpretation in terms of research evaluation. As such, PAC provides a framework for a theory of research evaluation.

DOI

[14]
Rousseau R., Egghe L., & Guns R. (2018). Becoming metric-wise. A bibliometric guide for researchers. Kidlington (UK): Chandos-Elsevier.

[15]
Ruane F., &Tol ,R.S.J. (2008). Rational (successive) h-indices: An application to economics in the Republic of Ireland, Scientometrics, 75(2), 395-405.We rank economics departments in the Republic of Ireland according to the number of publications, number of citations, and successive h -index of research-active staff. We increase the discriminatory power of the h 1 -index by introducing three generalizations, each of which is a rational number. The first ( h 1 + ) measures the excess over the actual h -index, while the other two ( h 1 *, h 1 Δ ) measures the distance to the next h -index. At the individual level, h * and h Δ coincide while h + is undefined.

DOI

[16]
Waltman L., & van Eck ,N.J. (2012). The inconsistency of the h-index. Journal of the American Society for Information Science and Technology, 63(2), 406-415.The h-index is a popular bibliometric indicator for assessing individual scientists. We criticize the h-index from a theoretical point of view. We argue that for the purpose of measuring the overall scientific impact of a scientist (or some other unit of analysis), the h-index behaves in a counterintuitive way. In certain cases, the mechanism used by the h-index to aggregate publication and citation statistics into a single number leads to inconsistencies in the way in which scientists are ranked. Our conclusion is that the h-index cannot be considered an appropriate indicator of a scientist's overall scientific impact. Based on recent theoretical insights, we discuss what kind of indicators can be used as an alternative to the h-index. We pay special attention to the highly cited publications indicator. This indicator has a lot in common with the h-index, but unlike the h-index it does not produce inconsistent rankings.

DOI

Outlines

/

京ICP备05002861号-43

Copyright © 2023 All rights reserved Journal of Data and Information Science

E-mail: jdis@mail.las.ac.cn Add:No.33, Beisihuan Xilu, Haidian District, Beijing 100190, China

Support by Beijing Magtech Co.ltd E-mail: support@magtech.com.cn