Accpeted for publication in the Journal of Korean Medical Science (JKMS)

Similar documents
Which percentile-based approach should be preferred. for calculating normalized citation impact values? An empirical comparison of five approaches

Quality assessments permeate the

Methods for the generation of normalized citation impact scores. in bibliometrics: Which method best reflects the judgements of experts?

UNDERSTANDING JOURNAL METRICS

Citation analysis: Web of science, scopus. Masoud Mohammadi Golestan University of Medical Sciences Information Management and Research Network

Publication Output and Citation Impact

1.1 What is CiteScore? Why don t you include articles-in-press in CiteScore? Why don t you include abstracts in CiteScore?

INTRODUCTION TO SCIENTOMETRICS. Farzaneh Aminpour, PhD. Ministry of Health and Medical Education

THE USE OF THOMSON REUTERS RESEARCH ANALYTIC RESOURCES IN ACADEMIC PERFORMANCE EVALUATION DR. EVANGELIA A.E.C. LIPITAKIS SEPTEMBER 2014

Journal Citation Reports Your gateway to find the most relevant and impactful journals. Subhasree A. Nag, PhD Solution consultant

On the causes of subject-specific citation rates in Web of Science.

INTRODUCTION TO SCIENTOMETRICS. Farzaneh Aminpour, PhD. Ministry of Health and Medical Education

Bibliometrics & Research Impact Measures

Citation analysis: State of the art, good practices, and future developments

Promoting your journal for maximum impact

DISCOVERING JOURNALS Journal Selection & Evaluation

STRATEGY TOWARDS HIGH IMPACT JOURNAL

WHO S CITING YOU? TRACKING THE IMPACT OF YOUR RESEARCH PRACTICAL PROFESSOR WORKSHOPS MISSISSIPPI STATE UNIVERSITY LIBRARIES

SCOPUS : BEST PRACTICES. Presented by Ozge Sertdemir

InCites Indicators Handbook

Embedding Librarians into the STEM Publication Process. Scientists and librarians both recognize the importance of peer-reviewed scholarly

Demystifying Citation Metrics. Michael Ladisch Pacific Libraries

The Impact Factor and other bibliometric indicators Key indicators of journal citation impact

Standards for the application of bibliometrics. in the evaluation of individual researchers. working in the natural sciences

A Correlation Analysis of Normalized Indicators of Citation

and social sciences: an exploratory study using normalized Google Scholar data for the publications of a research institute

Web of Science Unlock the full potential of research discovery

Measuring Academic Impact

Tracing the origin of a scientific legend by Reference Publication Year Spectroscopy (RPYS): the legend of the Darwin finches

USING THE UNISA LIBRARY S RESOURCES FOR E- visibility and NRF RATING. Mr. A. Tshikotshi Unisa Library

What are Bibliometrics?

Cited Publications 1 (ISI Indexed) (6 Apr 2012)

MEASURING EMERGING SCIENTIFIC IMPACT AND CURRENT RESEARCH TRENDS: A COMPARISON OF ALTMETRIC AND HOT PAPERS INDICATORS

What is bibliometrics?

A brief visual history of research metrics. Rights / License: Creative Commons Attribution-NonCommercial-NoDerivatives 4.

Research Evaluation Metrics. Gali Halevi, MLS, PhD Chief Director Mount Sinai Health System Libraries Assistant Professor Department of Medicine

Focus on bibliometrics and altmetrics

JOURNAL IMPACT FACTOR. 3-year calculation window (2015, 2016, and 2017)

Appropriate and Inappropriate Uses of Journal Bibliometric Indicators (Why do we need more than one?)

Percentile Rank and Author Superiority Indexes for Evaluating Individual Journal Articles and the Author's Overall Citation Performance

Integrated Impact Indicators (I3) compared with Impact Factors (IFs): An alternative research design with policy implications

Professional and Citizen Bibliometrics: Complementarities and ambivalences. in the development and use of indicators. A state-of-the-art report.

An Introduction to Bibliometrics Ciarán Quinn

The largest abstract and citation database

Developing library services to support Research and Development (R&D): The journey to developing relationships.

CITATION CLASSES 1 : A NOVEL INDICATOR BASE TO CLASSIFY SCIENTIFIC OUTPUT

ISSN: ISO 9001:2008 Certified International Journal of Engineering Science and Innovative Technology (IJESIT) Volume 3, Issue 2, March 2014

SCIENTOMETRICS AND RELEVANT BIBLIOGRAPHIC DATABASES IN THE FIELD OF AQUACULTURE

DON T SPECULATE. VALIDATE. A new standard of journal citation impact.

Russian Index of Science Citation: Overview and Review

Where to present your results. V4 Seminars for Young Scientists on Publishing Techniques in the Field of Engineering Science

Using Bibliometric Analyses for Evaluating Leading Journals and Top Researchers in SoTL

and social sciences: an exploratory study using normalized Google Scholar data for the publications of a research institute

Scopus. Advanced research tips and tricks. Massimiliano Bearzot Customer Consultant Elsevier

THE TRB TRANSPORTATION RESEARCH RECORD IMPACT FACTOR -Annual Update- October 2015

CITATION DATABASES: SCOPUS, WEB OF SCIENCE CITESCORE SJR SNIP H INDEX IF ISSUES

The largest abstract and citation database

PUBLIKASI JURNAL INTERNASIONAL

University of Liverpool Library. Introduction to Journal Bibliometrics and Research Impact. Contents

Coverage of highly-cited documents in Google Scholar, Web of Science, and Scopus: a multidisciplinary comparison

Scopus. Dénes Kocsis PhD Elsevier freelance trainer

Your research footprint:

A MATTER OF CITATIONS how scientific research is evaluated: from Impact Factor to CiteScore

Mendeley readership as a filtering tool to identify highly cited publications 1

Edited Volumes, Monographs, and Book Chapters in the Book Citation Index. (BCI) and Science Citation Index (SCI, SoSCI, A&HCI)

Individual Bibliometric University of Vienna: From Numbers to Multidimensional Profiles

The Journal Impact Factor: A brief history, critique, and discussion of adverse effects

On the relationship between interdisciplinarity and scientific impact

UvA-DARE (Digital Academic Repository)

Special Article. Prior Publication Productivity, Grant Percentile Ranking, and Topic-Normalized Citation Impact of NHLBI Cardiovascular R01 Grants

The use of bibliometrics in the Italian Research Evaluation exercises

Normalization of citation impact in economics

Bibliometrics and the Research Excellence Framework (REF)

Predicting the Importance of Current Papers

BIBLIOMETRIC INDICATORS AND CORE JOURNALS IN PHYSICAL AND REHABILITATION MEDICINE

SEARCH about SCIENCE: databases, personal ID and evaluation

Bibliometric practices and activities at the University of Vienna

STI 2018 Conference Proceedings

Citation Metrics. BJKines-NJBAS Volume-6, Dec

Coverage of highly-cited documents in Google Scholar, Web of Science, and Scopus: a multidisciplinary comparison

Visualizing the context of citations. referencing papers published by Eugene Garfield: A new type of keyword co-occurrence analysis

RPYS i/o: A web-based tool for the historiography and visualization of. citation classics, sleeping beauties, and research fronts

2018 Journal Citation Reports Every journal has a story to tell

Scientometric and Webometric Methods

What is Web of Science Core Collection? Thomson Reuters Journal Selection Process for Web of Science

Keywords: Publications, Citation Impact, Scholarly Productivity, Scopus, Web of Science, Iran.

Citation Impact on Authorship Pattern

Evaluating Research and Patenting Performance Using Elites: A Preliminary Classification Scheme

Bibliometric Indicators for Evaluating the Quality of Scientific Publications

AN INTRODUCTION TO BIBLIOMETRICS

Introduction to Citation Metrics

Bibliometric measures for research evaluation

Mapping the Research Productivity of Three Medical Sciences Journals Published in Saudi Arabia: A Comparative Bibliometric Study

Measuring the reach of your publications using Scopus

Bibliometric Rankings of Journals Based on the Thomson Reuters Citations Database

International Journal of Library and Information Studies ISSN: Vol.3 (3) Jul-Sep, 2013

MURDOCH RESEARCH REPOSITORY

BIBLIOMETRIC REPORT. Bibliometric analysis of Mälardalen University. Final Report - updated. April 28 th, 2014

WOUTER GERRITSMA, VU UNIVERSITY

Normalizing Google Scholar data for use in research evaluation

Transcription:

The Journal Impact Factor Should Not Be Discarded Running title: JIF Should Not Be Discarded Lutz Bornmann, 1 Alexander I. Pudovkin 2 1 Division for Science and Innovation Studies, Administrative Headquarters of the Max Planck Society, Munich, Germany 2 A. V. Zhirmunsky Institute of Marine Biology, Far Eastern Branch of Russian Academy of Sciences, Vladivostok, Russian Federation Address for Correspondence: Lutz Bornmann, PhD Division for Science and Innovation Studies, Administrative Headquarters of the Max Planck Society, Hofgartenstraße 8, 80539 München, Germany E-mail: bornmann@gv.mpg.de Received: 28 November 2016 Accepted: 28 November 2016

Journal editors and experts in scientometrics are increasingly concerned with the reliability of the Journal Impact Factor (JIF, Clarivate Analytics, formerly the IP & Science business of Thomson Reuters) as a tool for assessing the influence of scholarly journals. A paper bylarivière et al. (1), which was reposited on bioarxiv portal and commented on in Nature (2), reminded all stakeholders of science communication that the citability of most papers in an indexed journal deviates significantly from its JIF. These authors recommend to display journal citation distribution instead of the JIF, and the proposal is widely discussed on social networking platforms (3,4). The overall impression is that the discussion over the JIF is endless. The JIF along with the h-index is the simplest and most studied indicator in scientometrics (5,6). However, the commentary in Nature (2) and subsequent debates over the citation distribution revived interest of the scientific community toward empirical analyses of the JIF and its uses and misuses in research evaluation. After all the endless discussions, research evaluators should have realized that the JIF should not be used to measure the impact of single papers. But there are still some experts, who argue that the use of the JIFs at the level of single papers cannot be simply distinguished from its use at the journal level (4). In some circumstances, the JIFs may help authors and readers to pick, read, and cite certain papers. Papers from high-impact journals are more likely to be picked and cited than similar ones from low-impact periodicals. The JIF should not be demonized. It still can be employed for research evaluation purposes by carefully considering the context and academic environment. Elsevier provider of the Scopus database rates the JIF as so important that the company introduced the near-doppelgänger CiteScore, recently (see https://journalmetrics.scopus.com/). The JIF measures the average impact of papers, which are published in a journal, with a citation window of only one year. The JIFs are calculated and published annually in the Journal Citation Reports (JCR, Clarivate Analytics). Papers counted in the denominator of the JIF formula are published within 2 years prior to this citation metric calculation. In contrast to the JIF, the new CiteScore metric considers the papers from 3 years (instead of 2 years). As such, the JIF (and also the CiteScore) covers rather short term of interest toward papers (i.e., interest at the research front) and overlooks long-term implications of publication activity (the socalled sticky knowledge) (7). Focus on the short-term attention of the field-specific community makes

sense since the JIF was initially designed to guide librarians purchase the most used modern periodicals for their libraries. Accordingly, the JIF cannot and should not be employed for evaluating the average impact of papers in a journal in the long and distant run. The JIF formula aims at calculating average numbers that reveal the central tendency of a journal's impact. As such, one or few highly-cited papers, which are published within the 2 years, may boost the JIF. That is particularly the case with Nature, Science, and other influential journals (1). The skewed citation distribution implies that the JIF values do not reflect the real impact of most papers published in the index journal. The absolute number of citations received by a single paper is the correct measure of its impact. Currently, the Web of Science and Scopus databases can provide an outlook at citations for evaluating the impact of single papers. Importantly, the JIF is the best predictor of single papers' citability (8). Studies examining the predictive value of the JIF along with number of authors and pages proved that notion (9). One can expect more citations to single papers, which are published in higher-impact journals, compared to those in lower-impact ones. Another important point is the field-dependency of citations contributing to the JIFs. There are differing citation rates across different disciplines and subject categories, regardless of the scientific quality of the papers, and confounded by field-specific authorship rules, publication activity, and referencing patterns (10). Such differences justified the development of field-normalized indicators, which are employed for evaluating individual researchers, research groups, and institutions (11,12). Since the JIF is not a field-normalized indicator, it can only be used for evaluations within a single subject category. The SCImago Journal Rank SJR indicator a variant of the JIF were employed for institutional excellence mapping at www.excellencemapping.net (13,14). For institutions worldwide, this site maps the results of 2 indicators. First, the best paper rate measures the long-term impact of papers in a size-independent way, using percentiles as a field-normalized indicator. Second, the best journal rate, which is based on the citation impact of the journals publishing the institutions' papers. That indicator is the proportion of papers published in journals belonging to the 25% of the best journals in their subject categories in terms of citation impact. Through the consideration of journal sets, the indicator

is a field-normalized metric at the journal level. The indicator demonstrates how successful are academic institutions in terms of publishing their papers in high-impact journals (13,14). Thus, the socalled success at www.excellencemapping.net is measured by the ability of publishing in high-impact target journals and by receiving the long-term attention of the scientific community. The JIF can be used to measure the ability of individual researchers and institutions to successfully publish their research. However, the JIF should not be used as a proxy for measuring the impact of single papers. In this regard, more appropriate indicators should be considered (e.g., data from the Field Baselines tables in the Essential Science Indicators [ESI] by Clarivate Analytics). The baselines can be used to assess whether a specific paper received an impact which is far above or below the worldwide average performance in a field. For example, the 2006 baseline for chemistry is approximately 23 (November 14, 2016). If a chemistry paper from 2006, which is published by an evaluated entity, attracts 50 citations, the impact of that paper is far above the baseline, whereas with 10 citations the impact would be far below the baseline. There is only one scenario when the use of the JIFs is justifiable for the assessment of individual scientists (15). It is when recently published papers are considered for research evaluation, which is routinely practised for intramural monitoring of staff productivity, academic promotion, or recruitment. The evaluators pay particular attention to the most recent publications. But for these items, the citation window is too short for quantifying their citation impact reliably (16), and in that case reputation of the publishing journals along with their JIFs can be conditionally employed as the proxies of single papers' impact (9). InCites (Clarivate Analytics) has already implemented the calculation of specialty-specific percentile-transformed JIFs (17), which reflect field-normalized journal impact values and can be used for assessing recently published papers.

DISCLOSURE The authors have no potential conflicts of interest to disclose. AUTHOR CONTRIBUTION Study design: Bornmann L, Pudovkin AI. Writing: Bornmann L, Pudovkin AI. Revision and final approval of manuscript: all authors. ORCID Lutz Bornmann http://orcid.org/0000-0003-0810-7091

REFERENCES 1. Larivière V, Kiermer V, MacCallum CJ, McNutt M, Patterson M, Pulverer B, Swaminathan S, Taylor S, Curry S. A simple proposal for the publication of journal citation distributions. biorxivorg 2016: 062109. 2. Callaway E. Beat it, impact factor! Publishing elite turns against controversial metric. Nature 2016; 535: 210-1. 3. de Rijcke S. Let's move beyond too simplistic notions of misuse and unintended effects in debates on the JIF [Internet]. Available at https://www.cwts.nl/blog?article=n-q2x234 [accessed on 16 November 2016]. 4. Waltman L. The importance of taking a clear position in the impact factor debate [Internet]. Available at https://www.cwts.nl/blog?article=n-q2w2c4&title=the-importance-of-taking-a-clearposition-in-the-impact-factor-debate [accessed on 16 November 2016]. 5. Hönekopp J, Kleber J. Sometimes the impact factor outshines the H index. Retrovirology 2008; 5: 88. 6. Bornmann L, Marx W, Gasparyan AY, Kitas GD. Diversity, value and limitations of the journal impact factor and alternative metrics. Rheumatol Int 2012; 32: 1861-7. 7. Baumgartner SE, Leydesdorff L. Group-based trajectory modeling (GBTM) of citations in scholarly literature: dynamic qualities of transient and sticky knowledge claims. J Assoc Inf Sci Technol 2014; 65: 797-811. 8. Bornmann L, Leydesdorff L. Does quality and content matter for citedness? A comparison with para-textual factors and over time. J Informetr 2015; 9: 419-29. 9. Onodera N, Yoshikane F. Factors affecting citation rates of research articles. J Assoc Inf Sci Technol 2015; 66: 739-64. 10. Gasparyan AY, Yessirkepov M, Voronov AA, Gerasimov AN, Kostyukova EI, Kitas GD. Preserving the integrity of citations and references by all stakeholders of science communication. J Korean Med Sci 2015; 30: 1545-52. 11. Hicks D, Wouters P, Waltman L, de Rijcke S, Rafols I. Bibliometrics: the Leiden Manifesto for research metrics. Nature 2015; 520: 429-31.

12. Waltman L. A review of the literature on citation impact indicators. J Informetr 2016; 10: 365-91. 13. Bornmann L, Stefaner M, de Moya Anegón F, Mutz R. Ranking and mapping of universities and research-focused institutions worldwide based on highly-cited papers: a visualisation of results from multi-level models. Online Inf Rev 2014; 38: 43-58. 14. Bornmann L, Stefaner M, de Moya Anegón F, Mutz R. What is the effect of country-specific characteristics on the research performance of scientific institutions? Using multi-level statistical models to rank and map universities and research-focused institutions worldwide. J Informetr 2014; 8: 581-93. 15. Wouters P, Thelwall M, Kousha K, Waltman L, de Rijcke S, Rushforth A, Franssen T. The Metric Tide: Literature Review (Supplementary Report I to the Independent Review of the Role of Metrics in Research Assessment and Management). London: Higher Education Funding Council for England (HEFCE), 2015. 16. Wang J. Citation time window choice for research impact evaluation. Scientometrics 2013; 94: 851-72. 17. Pudovkin AI, Garfield E. Rank-normalized impact factor: a way to compare journal performance across subject categories. Proc Assoc Inf Sci Technol 2004; 41: 507-15.