| Literature DB >> 17445275 |
Abstract
Quantifying the impact of scientific research is almost always controversial, and there is a need for a uniform method that can be applied across all fields. Increasingly, however, the quantification has been summed up in the impact factor of the journal in which the work is published, which is known to show differences between fields. Here the h-index, a way to summarize an individual's highly cited work, was calculated for journals over a twenty year time span and compared to the size of the journal in four fields, Agriculture, Condensed Matter Physics, Genetics and Heredity and Mathematical Physics. There is a linear log-log relationship between the h-index and the size of the journal: the larger the journal, the more likely it is to have a high h-index. The four fields cannot be separated from each other suggesting that this relationship applies to all fields. A strike rate index (SRI) based on the log relationship of the h-index and the size of the journal shows a similar distribution in the four fields, with similar thresholds for quality, allowing journals across diverse fields to be compared to each other. The SRI explains more than four times the variation in citation counts compared to the impact factor.Entities:
Year: 2007 PMID: 17445275 PMCID: PMC1868756 DOI: 10.1186/1742-5581-4-3
Source DB: PubMed Journal: Biomed Digit Libr ISSN: 1742-5581
Figure 1A double logarithm plot of the h-index and the number of citable items in a journal since 1986. The line of best fit has a slope of 0.57.
Figure 2The ranking of the strike rate index across four fields.
Figure 3The relationship between strike rate index and impact factor across four fields.
Figure 4The relationship of citations for cattle QTL articles published in 2003–4 against (a) the impact factor and (b) the strike rate index of the journals in which they were published.
| Term | Explanation |
| Cook's Distance | measures the influence of a particular data point on all the other data points in a linear regression, it indicates how important a particular data point is for the method [ |
| F | ratio of the variance or mean square between groups to the variance within groups |
| Linear Regression | where one variable is expressed as a function of another variable in a statistical analysis using simple least squares methods |
| log-log plot | double logarithm plot, if y = cxa, where x is the independent variable, c is a constant and a is an exponent, then logy = alogx + logc and the slope of the resulting line is the exponent a. An exponent of 2 would imply a square or quadratic relationship while an exponent of 0.5 would imply a square root relationship between the variables |
| median | half the values in a distribution are higher and half the values are lower than the median value |
| P value | with a null hypothesis of no difference between two or more samples, the P value is the probability that the null hypothesis is true, and that the observed difference is due to a chance event |
| Quantiles of the standard normal | QQplot. Plot of data against the corresponding quantiles of a standard normal distribution, one with a mean of zero and a variance of one. If the plot is fairly linear, the data are reasonably Gaussian or normal [ |
| R2 | the square of the correlation coefficient. It is an estimate of the variance explained by a particular statistical model |
| Robust Regression | the robust fit is minimally influenced by outliers in the data, minimizing bias in the estimates of the coefficients [ |
| s.e. | standard error, which is an estimate of the accuracy of a mean (s.e.m.) or other coefficient given the variability found in a particular set of data; it is fundamental to understanding whether two means are likely to be from the same or from different distributions |
| t | calculated from the difference between means divided by the standard error of the difference between two means (Student's t-test) and in ordinary least-squares regression analyses to determine whether a slope is significantly different from zero by comparing the slope to its standard error |