Measuring scientific impact beyond academia: An assessment of existing impact metrics and proposed improvements

How does scientific research affect the world around us? Being able to answer this question is of great importance in order to appropriately channel efforts and resources in science. The impact by scientists in academia is currently measured by citation based metrics such as h-index, i-index and citation counts. These academic metrics aim to represent the dissemination of knowledge among scientists rather than the impact of the research on the wider world. In this work we are interested in measuring scientific impact beyond academia, on the economy, society, health and legislation (comprehensive impact). Indeed scientists are asked to demonstrate evidence of such comprehensive impact by authoring case studies in the context of the Research Excellence Framework (REF). We first investigate the extent to which existing citation based metrics can be indicative of comprehensive impact. We have collected all recent REF impact case studies from 2014 and we have linked these to papers in citation networks that we constructed and derived from CiteSeerX, arXiv and PubMed Central using a number of text processing and information retrieval techniques. We have demonstrated that existing citation-based metrics for impact measurement do not correlate well with REF impact results. We also consider metrics of online attention surrounding scientific works, such as those provided by the Altmetric API. We argue that in order to be able to evaluate wider non-academic impact we need to mine information from a much wider set of resources, including social media posts, press releases, news articles and political debates stemming from academic work. We also provide our data as a free and reusable collection for further analysis, including the PubMed citation network and the correspondence between REF case studies, grant applications and the academic literature.

[1]  James G. Corrigan,et al.  Programmatic evaluation and comparison based on standardized citation scores , 1983, IEEE Transactions on Engineering Management.

[2]  Blaise Cronin,et al.  The citation process: The role and significance of citations in scientific communication , 1984 .

[3]  Emma Hill,et al.  Show me the data. , 1998 .

[4]  J. E. Hirsch,et al.  An index to quantify an individual's scientific research output , 2005, Proc. Natl. Acad. Sci. USA.

[5]  E. Garfield The history and meaning of the journal impact factor. , 2006, JAMA.

[6]  Lokman I. Meho The Rise and Rise of Citation Analysis , 2006, ArXiv.

[7]  Lee A Vucovich,et al.  Analyzing the impact of an author's publications. , 2008, Journal of the Medical Library Association : JMLA.

[8]  Lutz Bornmann,et al.  What do citation counts measure? A review of studies on citing behavior , 2008, J. Documentation.

[9]  Christoph Bartneck,et al.  Detecting h-index manipulation through self-citation analysis , 2010, Scientometrics.

[10]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[11]  Julia Lane,et al.  Star Metrics and the Science of Science Policy , 2012 .

[12]  Laura Paglione,et al.  ORCID: a system to uniquely identify researchers , 2012, Learn. Publ..

[13]  Cassidy R. Sugimoto,et al.  Do Altmetrics Work? Twitter and Ten Other Social Web Services , 2013, PloS one.

[14]  Euan A. Adie,et al.  Altmetric: enriching scholarly content with article‐level discussion and metrics , 2013, Learn. Publ..

[15]  Heather A. Piwowar,et al.  Altmetrics: Value all research products , 2013, Nature.

[16]  Lutz Bornmann,et al.  Do altmetrics point to the broader impact of research? An overview of benefits and disadvantages of altmetrics , 2014, J. Informetrics.

[17]  Jane Grimson,et al.  Measuring research impact: not everything that can be counted counts, and not everything that counts can be counted , 2014 .

[18]  Wenyi Huang,et al.  RefSeer: A citation recommendation system , 2014, IEEE/ACM Joint Conference on Digital Libraries.

[19]  Oren Etzioni,et al.  Identifying Meaningful Citations , 2015, AAAI Workshop: Scholarly Big Data.

[20]  S. Rijcke,et al.  Bibliometrics: The Leiden Manifesto for research metrics. , 2015, Nature.

[21]  S. Rijcke,et al.  Bibliometrics: The Leiden Manifesto for research metrics , 2015, Nature.

[22]  Rodrigo Costas,et al.  Do “altmetrics” correlate with citations? Extensive comparison of altmetric indicators with citations from a multidisciplinary perspective , 2014, J. Assoc. Inf. Sci. Technol..

[23]  Lutz Bornmann,et al.  How many scientific papers are mentioned in policy-related documents? An empirical investigation using Web of Science and Altmetric data , 2016, Scientometrics.

[24]  Luis Gravano,et al.  Predicting the impact of scientific concepts using full‐text features , 2016, J. Assoc. Inf. Sci. Technol..

[25]  Ludo Waltman,et al.  A review of the literature on citation impact indicators , 2015, J. Informetrics.

[26]  Vincent Larivière,et al.  Tweets as impact indicators: Examining the implications of automated “bot” accounts on Twitter , 2014, J. Assoc. Inf. Sci. Technol..

[27]  Lutz Bornmann,et al.  To what extent does the Leiden manifesto also apply to altmetrics? A discussion of the manifesto against the background of research into altmetrics , 2016, Online Inf. Rev..

[28]  M. Edwards,et al.  Academic Research in the 21st Century: Maintaining Scientific Integrity in a Climate of Perverse Incentives and Hypercompetition , 2017, Environmental engineering science.