Citation analysis may severely underestimate the impact of clinical research as compared to basic research

Similar documents
A systematic empirical comparison of different approaches for normalizing citation impact indicators

Source normalized indicators of citation impact: An overview of different approaches and an empirical comparison

Discussing some basic critique on Journal Impact Factors: revision of earlier comments

Citation analysis: State of the art, good practices, and future developments

CITATION CLASSES 1 : A NOVEL INDICATOR BASE TO CLASSIFY SCIENTIFIC OUTPUT

Results of the bibliometric study on the Faculty of Veterinary Medicine of the Utrecht University

A Taxonomy of Bibliometric Performance Indicators Based on the Property of Consistency

F1000 recommendations as a new data source for research evaluation: A comparison with citations

CitNetExplorer: A new software tool for analyzing and visualizing citation networks

BIBLIOMETRIC REPORT. Bibliometric analysis of Mälardalen University. Final Report - updated. April 28 th, 2014

Bibliometric report

A Correlation Analysis of Normalized Indicators of Citation

Constructing bibliometric networks: A comparison between full and fractional counting

PBL Netherlands Environmental Assessment Agency (PBL): Research performance analysis ( )

Scientometrics & Altmetrics

AN INTRODUCTION TO BIBLIOMETRICS

Predicting the Importance of Current Papers

Comparing Bibliometric Statistics Obtained from the Web of Science and Scopus

Developing library services to support Research and Development (R&D): The journey to developing relationships.

Scientometric Measures in Scientometric, Technometric, Bibliometrics, Informetric, Webometric Research Publications

THE KISS OF DEATH? THE EFFECT OF BEING CITED IN A REVIEW ON

2nd International Conference on Advances in Social Science, Humanities, and Management (ASSHM 2014)

The Operationalization of Fields as WoS Subject Categories (WCs) in. Evaluative Bibliometrics: The cases of Library and Information Science and

The Impact Factor and other bibliometric indicators Key indicators of journal citation impact

Citation Proximity Analysis (CPA) A new approach for identifying related work based on Co-Citation Analysis

Focus on bibliometrics and altmetrics

On the relationship between interdisciplinarity and scientific impact

The real deal! Applying bibliometrics in research assessment and management...

Edited Volumes, Monographs, and Book Chapters in the Book Citation Index. (BCI) and Science Citation Index (SCI, SoSCI, A&HCI)

The journal relative impact: an indicator for journal assessment

Methods for the generation of normalized citation impact scores. in bibliometrics: Which method best reflects the judgements of experts?

Identifying Related Documents For Research Paper Recommender By CPA and COA

The use of citation speed to understand the effects of a multi-institutional science center

Which percentile-based approach should be preferred. for calculating normalized citation impact values? An empirical comparison of five approaches

Counting the Number of Highly Cited Papers

1.1 What is CiteScore? Why don t you include articles-in-press in CiteScore? Why don t you include abstracts in CiteScore?

Kent Academic Repository

Publication Output and Citation Impact

Should author self- citations be excluded from citation- based research evaluation? Perspective from in- text citation functions

Normalizing Google Scholar data for use in research evaluation

Professor Birger Hjørland and associate professor Jeppe Nicolaisen hereby endorse the proposal by

HIGHLY CITED PAPERS IN SLOVENIA

Self-citations at the meso and individual levels: effects of different calculation methods

Peter Ingwersen and Howard D. White win the 2005 Derek John de Solla Price Medal

Edited volumes, monographs and book chapters in the Book Citation Index (BKCI) and Science Citation Index (SCI, SoSCI, A&HCI)

Citation Analysis in Research Evaluation

Bibliometrics and the Research Excellence Framework (REF)

hprints , version 1-1 Oct 2008

Automatic selection of references for the creation of a biomedical literature review using citation mapping

Impact Factors: Scientific Assessment by Numbers

Universiteit Leiden. Date: 25/08/2014

The problems of field-normalization of bibliometric data and comparison among research institutions: Recent Developments

THE USE OF THOMSON REUTERS RESEARCH ANALYTIC RESOURCES IN ACADEMIC PERFORMANCE EVALUATION DR. EVANGELIA A.E.C. LIPITAKIS SEPTEMBER 2014

Bibliometric Rankings of Journals Based on the Thomson Reuters Citations Database

What is Web of Science Core Collection? Thomson Reuters Journal Selection Process for Web of Science

Swedish Research Council. SE Stockholm

BIBLIOMETRIC REPORT. Netherlands Bureau for Economic Policy Analysis (CPB) research performance analysis ( ) October 6 th, 2015

On the causes of subject-specific citation rates in Web of Science.

For Your Citations Only? Hot Topics in Bibliometric Analysis

Keywords: Publications, Citation Impact, Scholarly Productivity, Scopus, Web of Science, Iran.

Scientometric and Webometric Methods

The 2016 Altmetrics Workshop (Bucharest, 27 September, 2016) Moving beyond counts: integrating context

Getting started with CitNetExplorer version 1.0.0

Coverage of highly-cited documents in Google Scholar, Web of Science, and Scopus: a multidisciplinary comparison

Looking for the impact of peer review: does count of funding acknowledgements really predict research impact?

SCOPUS : BEST PRACTICES. Presented by Ozge Sertdemir

Is Scientific Literature Subject to a Sell-By-Date? A General Methodology to Analyze the Durability of Scientific Documents

Normalization of citation impact in economics

Citation Analysis with Microsoft Academic

In basic science the percentage of authoritative references decreases as bibliographies become shorter

Elsevier Databases Training

A bibliometric analysis of publications by staff from Mid Yorkshire Hospitals NHS Trust,

INTRODUCTION TO SCIENTOMETRICS. Farzaneh Aminpour, PhD. Ministry of Health and Medical Education

Complementary bibliometric analysis of the Health and Welfare (HV) research specialisation

Valeria Aman Does the Scopus author ID suffice to track scientific international mobility? A case study based on Leibniz laureates (abstract IS10)

INTRODUCTION TO SCIENTOMETRICS. Farzaneh Aminpour, PhD. Ministry of Health and Medical Education

Research Ideas for the Journal of Informatics and Data Mining: Opinion*

Corso di dottorato in Scienze Farmacologiche Information Literacy in Pharmacological Sciences 2018 WEB OF SCIENCE SCOPUS AUTHOR INDENTIFIERS

STI 2018 Conference Proceedings

Scientometric Profile of Presbyopia in Medline Database

Embedding Librarians into the STEM Publication Process. Scientists and librarians both recognize the importance of peer-reviewed scholarly

Using Bibliometric Analyses for Evaluating Leading Journals and Top Researchers in SoTL

Lokman I. Meho and Kiduk Yang School of Library and Information Science Indiana University Bloomington, Indiana, USA

Journal Citation Reports Your gateway to find the most relevant and impactful journals. Subhasree A. Nag, PhD Solution consultant

Standards for the application of bibliometrics. in the evaluation of individual researchers. working in the natural sciences

Scopus. Advanced research tips and tricks. Massimiliano Bearzot Customer Consultant Elsevier

KTH RAE BIBLIOMETRIC REPORT

Alphabetical co-authorship in the social sciences and humanities: evidence from a comprehensive local database 1

A tutorial for vosviewer. Clément Levallois. Version 1.6.5,

Visualizing the context of citations. referencing papers published by Eugene Garfield: A new type of keyword co-occurrence analysis

ISSN: ISO 9001:2008 Certified International Journal of Engineering Science and Innovative Technology (IJESIT) Volume 3, Issue 2, March 2014

Publication boost in Web of Science journals and its effect on citation distributions

Classic papers: déjà vu, a step further in the bibliometric exploitation of Google Scholar

Mendeley readership as a filtering tool to identify highly cited publications 1

Identifying Related Work and Plagiarism by Citation Analysis

Web of Science Unlock the full potential of research discovery

Global Journal of Engineering Science and Research Management

2013 Environmental Monitoring, Evaluation, and Protection (EMEP) Citation Analysis

Promoting your journal for maximum impact

A Scientometric Study of Digital Literacy in Online Library Information Science and Technology Abstracts (LISTA)

Transcription:

Citation analysis may severely underestimate the impact of clinical research as compared to basic research Nees Jan van Eck 1, Ludo Waltman 1, Anthony F.J. van Raan 1, Robert J.M. Klautz 2, and Wilco C. Peul 3 1 Leiden University, Centre for Science and Technology Studies, Leiden, The Netherlands {ecknjpvan, waltmanlr, vanraan}@cwts.leidenuniv.nl 2 Leiden University Medical Center, Department of Cardiothoracic Surgery, Leiden, The Netherlands r.j.m.klautz@lumc.nl 3 Leiden University Medical Center, Department of Neurosurgery, Leiden, The Netherlands w.c.peul@lumc.nl Background Citation analysis has become an important tool for research performance assessment in the medical sciences. However, different areas of medical research may have considerably different citation practices, even within the same medical field. Because of this, it is unclear to what extent citationbased bibliometric indicators allow for valid comparisons between research units active in different areas of medical research. Methodology A visualization methodology is introduced that reveals differences in citation practices between medical research areas. The methodology extracts terms from the titles and abstracts of a large collection of publications and uses these terms to visualize the structure of a medical field and to indicate how research areas within this field differ from each other in their average citation impact. Results Visualizations are provided for 32 medical fields, defined based on journal subject categories in the Web of Science database. The analysis focuses on three fields: Cardiac & cardiovascular systems, Clinical neurology, and Surgery. In each of these fields, there turn out to be large differences in citation practices between research areas. Low-impact research areas tend to focus on clinical intervention research, while high-impact research areas are often more oriented on basic and diagnostic research. 1

Conclusions Popular bibliometric indicators, such as the h-index and the impact factor, do not correct for differences in citation practices between medical fields. These indicators therefore cannot be used to make accurate between-field comparisons. More sophisticated bibliometric indicators do correct for field differences but still fail to take into account within-field heterogeneity in citation practices. As a consequence, the citation impact of clinical intervention research may be substantially underestimated in comparison with basic and diagnostic research. 1. Introduction Citation analysis is widely used in the assessment of research performance in the medical sciences (Patel et al., 2011). Especially the h-index (Hirsch, 2005) and the impact factor (Chew, Villanueva, & Van der Weyden, 2007; Garfield, 1996, 2006) are extremely popular bibliometric indicators. However, the use of these indicators for performance assessment has important limitations. In particular, both the h-index and the impact factor fail to take into account the enormous differences in citation practices between fields of science (e.g., Radicchi, Fortunato, & Castellano, 2008). For instance, the average length of the reference list of a publication is much larger in molecular biology than in mathematics. As a consequence, publications in molecular biology on average are cited much more frequently than publications in mathematics. This difference can be more than an order of magnitude (Waltman, Van Eck, Van Leeuwen, Visser, & Van Raan, 2011a). More sophisticated bibliometric indicators used by professional bibliometric centers perform a normalization to correct for differences in citation practices between fields of science (e.g., Glänzel, Thijs, Schubert, & Debackere, 2009; Waltman, Van Eck, Van Leeuwen, Visser, & Van Raan, 2011b). These fieldnormalized indicators typically rely on a field classification system in which the boundaries of fields are explicitly defined (e.g., the journal subject categories in the Web of Science database). Unfortunately, however, practical applications of fieldnormalized indicators often suggest the existence of differences in citation practices not only between but also within fields of science. As shown in this paper, this phenomenon can be observed especially clearly in medical fields, in which the citation impact of clinical intervention research may be substantially underestimated in comparison with basic and diagnostic research. Within-field heterogeneity in citation practices is not corrected for by field-normalized bibliometric indicators and therefore poses a serious threat to the accuracy of these indicators. 2

This paper presents an empirical analysis of the above problem, with a focus on the medical sciences. An advanced visualization methodology is used to show how citation practices differ between research areas within a medical field. In particular, substantial differences are revealed between basic and diagnostic research areas on the one hand and clinical intervention research areas on the other hand. Implications of the analysis for the use of bibliometric indicators in the medical sciences are discussed. 2. Methodology The analysis reported in this paper starts from the idea that drawing explicit boundaries between research areas, for instance between basic and clinical areas, is difficult and would require many arbitrary decisions, for instance regarding the treatment of multidisciplinary topics that are in between multiple areas. To avoid the difficulty of drawing explicit boundaries between research areas, the methodology adopted in this paper relies strongly on the use of visualization. The methodology uses so-called term maps (e.g., Van Eck & Waltman, 2011; Waaijer, Van Bochove, & Van Eck, 2010, 2011) to visualize scientific fields. A term map is a two-dimensional representation of a field in which strongly related terms are located close to each other and less strongly related terms are located further away from each other. A term map provides an overview of the structure of a field. Different areas in a map correspond with different subfields or research areas. In the term maps presented in this paper, colors are used to indicate differences in citation practices between research areas. For each term in a map, the color of the term is determined by the average citation impact of the publications in which the term occurs. We note that the use of visualization to analyze the structure and development of scientific fields has a long history (e.g., Börner, 2010), but visualization approaches have not been used before to study differences in citation practices between research areas. The use of term maps, also referred to as co-word maps, has a 30-year history, with early contributions dating back to the 1980s and the beginning of the 1990s (e.g., Peters & Van Raan, 1993; Rip & Courtial, 1984; Tijssen & Van Raan, 1989). The first methodological step is the definition of scientific fields. This study uses data from the Web of Science (WoS) bibliographic database. This database has a good coverage of the medical literature (Moed, 2005) and is the most popular data source for professional bibliometric analyses. Because of their frequent use in field- 3

normalized bibliometric indicators, the journal subject categories in the WoS database are employed to define fields. There are about 250 subject categories in the WoS database, covering disciplines in the sciences, the social sciences, and the arts and humanities. The analyses reported in this paper are based on all publications in a particular subject category that are classified as article or review and that were published between 2006 and 2010. For each publication, citations are counted until the end of 2011. Using natural language processing techniques, the titles and abstracts of the publications in a field are parsed. This yields a list of all noun phrases (i.e., sequences of nouns and adjectives) that occur in these publications. An additional algorithm (Van Eck & Waltman, 2011) selects the 2000 noun phrases that can be regarded as the most characteristic terms of the field. This algorithm aims to filter out general noun phrases, like for instance result, study, patient, and clinical evidence. Filtering out these noun phrases is crucial. Due to their general meaning, these noun phrases do not relate specifically to one topic, and they therefore tend to distort the structure of a term map. Apart from excluding general noun phrases, noun phrases that occur only in a small number of publications are excluded as well. This is done in order to obtain sufficiently robust results. The minimum number of publications in which a noun phrase must occur depends on the total number of publications in a field. For the three fields discussed in the next section, thresholds between 70 and 135 publications were used. Given a selection of 2000 terms that together characterize a field, the next step is to determine the number of publications in which each pair of terms co-occurs. Two terms are said to co-occur in a publication if they both occur at least once in the title or abstract of the publication. The larger the number of publications in which two terms co-occur, the stronger the terms are considered to be related to each other. In neuroscience, for instance, Alzheimer and short-term memory may be expected to cooccur a lot, indicating a strong relation between these two terms. The matrix of term co-occurrence frequencies serves as input for the VOS mapping technique (Van Eck, Waltman, Dekker, & Van den Berg, 2010). This technique determines for each term a location in a two-dimensional space. Strongly related terms tend to be located close to each other in the two-dimensional space, while terms that do not have a strong relation are located further away from each other. The VOS mapping technique is closely related to the technique of multidimensional scaling (e.g., Borg & Groenen, 4

2005), but for the purpose of creating term maps the VOS mapping technique has been shown to yield more satisfactory results, as discussed by Van Eck et al. (2010). It is important to note that in the interpretation of a term map only the distances between terms are relevant. A map can be freely rotated, because this does not affect the inter-term distances. This also implies that the horizontal and vertical axes have no special meaning. In the final step, the color of each term is determined. First, in order to correct for the age of a publication, each publication s number of citations is divided by the average number of citations of all publications that appeared in the same year. This yields a publication s normalized citation score. A score of 1 means that the number of citations of a publication equals the average of all publications that appeared in the same field and in the same year. Next, for each of the 2000 terms, the normalized citation scores of all publications in which the term occurs (in the title or abstract) are averaged. The color of a term is determined based on the resulting average score. Colors range from blue (average score of 0) to green (average score of 1) to red (average score of 2 or higher). Hence, a blue term indicates that the publications in which a term occurs have a low average citation impact, while a red term indicates that the underlying publications have a high average citation impact. The VOSviewer software (Van Eck & Waltman, 2010) is used to visualize the term maps resulting from the above steps. 1 3. Results Figures 1, 2, and 3 show the term maps obtained for the WoS fields Cardiac & cardiovascular systems, Clinical neurology, and Surgery. These fields were selected because they match well with our areas of expertise. The maps are based on, respectively, 75,314, 105,405, and 141,155 publications from the period 2006 2010. Only a limited level of detail is offered in Figures 1, 2, and 3. To explore the term maps in full detail, the reader is invited to use the interactive versions of the maps that are available on a webpage. 2 The webpage also provides maps of 29 other medical fields as well as of all 32 medical fields taken together. 1 The VOSviewer software is freely available at www.vosviewer.com. 2 The interactive maps can be found at www.neesjanvaneck.nl/basic_vs_clinical/. 5

The term maps shown in Figures 1, 2, and 3 all indicate a clear distinction between different research areas. Clinical research areas tends to be located mainly in the left part of a map and basic research areas mainly in the right part, although making a perfect distinction between basic and clinical research areas is definitely not possible. The basic-clinical distinction is best visible in the Cardiac & cardiovascular systems and Clinical neurology maps (Figures 1 and 2), in which the left part consists of clinical intervention research areas (e.g., cardiac surgery and neurosurgery) while the right part includes important basic and diagnostic research areas (e.g., cardiology and neurology). The Surgery map (Figure 3) gives a somewhat different picture, probably because of the more clinical focus of surgical research. In this map, clinical research areas (e.g., orthopedic surgery, oncological surgery, and cardiac surgery) are concentrated in the left, middle, and upper parts, while research areas with a more basic focus can be found in the lower-right part. Connections between basic research areas on the one hand and clinical research areas on the other hand are also visible in the term maps. The maps display bridges that seem to represent translational research, that is, research aimed at translating basic research results into clinical practice. In the Cardiac & cardiovascular systems map (Figure 1), for instance, two bridges are visible, one in the upper part of the map and one in the lower part. In the upper part, the topic of atherosclerosis can be found, starting in the upper-right part of the map with basic research on vascular damage, continuing in the middle part with research on cholesterol and cholesterol lowering drugs, and extending in the upper-left part with interventional therapies such as coronary bypass surgery and percutaneous interventions (PCI) and its modifications (BMS and DES). In the lower part of the map, the topic of arrhythmias can be identified. It starts in the lower-right part of the map with basic research on electrophysiological phenomena, it continues in the middle part with diagnostic tools, and it ends in the lower-left part with the clinical application of ablation therapy for arrhythmias. Looking at Figures 1, 2, and 3, a crucial observation is that the distinction between different research areas is visible not only in the structure of the maps but also in the colors of the terms. In general, in the right part of each map, in which the more basic and diagnostic research areas are located, there are many yellow, orange, and red terms, which clearly indicates an above-average citation impact. (As indicated by the color bar in the lower right in Figures 1, 2, and 3, yellow and orange correspond with 6

a citation impact that is, respectively, about 25% and about 50% above the average of the field. Red corresponds with a citation impact that is 100% or more above average.) On the other hand, in the left part of each map, research areas can be found with mainly blue and green terms, implying a below-average citation impact. This pattern is most strongly visible in the Clinical neurology map (Figure 2) and can also be observed in the Surgery map (Figure 3). In the Cardiac & cardiovascular systems map (Figure 1), a clear distinction between high- and low-impact research areas is visible as well, but it coincides only partially with the left-right distinction. We further note that within an area in a map terms are usually colored in a quite consistent way. In other words, terms tend to be surrounded mainly by other terms with a similar color. This is an important indication of the robustness of the maps. The general picture emerging from Figures 1, 2, and 3, and supported by term maps for other medical fields provided online, is that within medical fields there is often a considerable heterogeneity in citation impact, with some research areas on average receiving two or even three times more citations per publication than other research areas. In general, low-impact research areas tend to focus on clinical research, in particular on surgical interventions. Research areas that are more oriented on basic and diagnostic research usually have an above average citation impact. 4. Discussion and conclusion The citation impact of a publication can be influenced by many factors. In the medical sciences, previous studies have for instance analyzed the effect of study design (e.g., case report, randomized controlled trial, or meta-analysis; Patsopoulos, Analatos, & Ioannidis, 2005), article type (i.e., brief report or full-size article; Mavros, Bardakas, Rafailidis, Sardi, Demetriou, & Falagas, 2013), and article length (Falagas, Zarkali, Karageorgopoulos, Bardakas, & Mavros, 2013). In this paper, the effect of differences in citation practices between medical research areas has been investigated. Different fields of science have different citation practices. In some fields, publications have much longer reference lists than in others. Also, in some fields researchers mainly refer to recent work, while in other fields it is more common to cite older work. Because of such differences between fields, publications in one field 7

may on average receive many more citations than publications in another field. 3 Popular bibliometric indicators, such as the h-index and the impact factor, do not correct for this. The use of these indicators to make comparisons between fields may therefore easily lead to invalid conclusions. 4 The results obtained using the visualization methodology introduced in this paper go one step further and show that even within a single field of science there can be large differences in citation practices. Similar findings have been reported in earlier studies (Neuhaus & Daniel, 2009; Smolinsky & Lercher, 2012; Van Leeuwen & Calero Medina, 2012), but based on smaller analyses and not within the medical domain. The present results suggest that in medical fields low-impact research areas tend to be clinically oriented, focusing mostly on surgical interventions. Basic and diagnostic research areas usually have a citation impact above the field average, although not all high-impact research areas need to have a basic focus. The coloring of the term maps indicates that two- or even threefold impact differences between research areas within a single medical field are not uncommon. Although differences in citation impact between basic and clinical research have been mentioned in earlier studies (e.g., Seglen, 1997), only a limited amount of empirical evidence of such differences has been collected. We are aware of only a few earlier studies in which differences in citation impact between basic and clinical research have been analyzed (Lewison & Dawson, 1998; Lewison & Devey, 1999; Lewison & Paraje, 2004; Opthof, 2011). These studies are based on much smaller amounts of data than the present analysis. Contrary to the present results, Opthof (2011) concludes that clinical research is cited more frequently than basic research. However, the study is limited in scope. It is restricted to a single medical field, and it 3 It could be suggested that differences in citation impact between research areas may also be caused by the size of an area. In a larger research area, there are more researchers and more publications than in a smaller research area, and therefore there are also more citations. However, one should be careful with this argument. In a large research area, there are many publications, each of them giving citations to earlier work, but at the same time there are also many publications that can be cited. Given this balance between citing and citable publications, one may expect that in general the average number of citations of the publications in a research area is not affected by the size of the area. 4 This is by no means the only objection one may have against these indicators. An important objection against the impact factor could be that the impact of a journal as a whole may not be representative of the impact of individual publications in the journal (Seglen, 1997). An objection against the h-index could be that it suffers from inconsistencies in its definition (Waltman & Van Eck, 2012a). 8

considers publications from only a small set of journals. 5 In another relatively small study, reported by Lewison and Paraje (2004), no difference in citation impact between basic and clinical research is detected. This study has the limitation of being restricted to publications from only two journals. Two earlier studies (Lewison & Dawson, 1998; Lewison & Devey, 1999) provide some evidence for a citation advantage for basic publications over clinical ones. A number of limitations of the methodology of the present study need to be mentioned. First of all, because the visualization methodology does not draw explicit boundaries between research areas, no exact figures can be provided on citation impact differences between, for instance, basic and clinical research. On the other hand, by not drawing explicit boundaries, many arbitrary choices are avoided and more fine-grained analyses can be performed. Another methodological limitation is the ambiguity in the meaning and use of terms. Some terms may for instance be used both in basic and in clinical research. Although a term selection algorithm was employed to filter out the most ambiguous terms, some degree of ambiguity cannot be avoided when working with terms. Other limitations relate to the bibliographic database that was used. The WoS database has a good coverage of the medical literature, but to some extent the analysis might have been affected by gaps in the coverage of the literature. Also, the analysis depends strongly on the field definitions offered by the WoS database. The results reported in this paper lead to the conclusion that one should be rather careful with citation-based comparisons between medical research areas, even if in a bibliographic database such as WoS the areas are considered to be part of the same field. Field-normalized bibliometric indicators, which are typically used by professional bibliometric centers, correct for differences in citation practices between 5 Replicating the two analyses reported by Opthof (2011) confirmed their results. The first analysis reported by Opthof is based on six cardiovascular journals, three basic ones and three clinical ones. The difference between the outcomes of this analysis and the analysis reported in the present paper appears to be related to the particular characteristics of the selected journals. The publications in these journals turn out not to be fully representative for basic and clinical publications in all cardiovascular journals. The second analysis reported by Opthof is based on the distinction between basic and clinical publications within a single cardiovascular journal (Circulation). In this case, the difference with the outcomes of the analysis reported in the present paper seems to indicate that the selected journal differs from the cardiovascular field as a whole in terms of the characteristics of its basic and clinical publications. 9

fields, but at present they fail to correct for within-field differences. The use of bibliometric indicators, either the h-index and the impact factor or more sophisticated field-normalized indicators, may therefore lead to an underestimation of the impact of certain types of research compared with others. In particular, the impact of clinical intervention research may be underestimated, while the impact of basic and diagnostic research may be overestimated. There is an urgent need for more accurately normalized bibliometric indicators. These indicators should correct not only for differences in citation practices between fields of science, but also for differences between research areas within the same field. Research areas could for instance be defined algorithmically based on citation patterns (e.g., Klavans & Boyack, 2010; Waltman & Van Eck, 2012b). Alternatively, a normalization could be performed at the side of the citing publications by giving a lower weight to citations from publications with long reference lists and a higher weight to citations from publications that cite only a few references. A number of steps towards such citing-side normalization procedures have already been taken (e.g., Glänzel, Schubert, Thijs, & Debackere, 2011; Leydesdorff & Opthof, 2010; Moed, 2010; Rafols, Leydesdorff, O Hare, Nightingale, & Stirling, 2012; Waltman & Van Eck, in press; Waltman, Van Eck, Van Leeuwen, & Visser, 2013; Zitt & Small, 2008), but more research in this direction is needed. Using the presently available bibliometric indicators, one should be aware of biases caused by differences in citation practices between areas of medical research, especially between basic and clinical areas. Acknowledgment We would like to thank Cathelijn Waaijer for helpful suggestions in the interpretation of the term maps. References Borg, I., & Groenen, P.J.F. (2005). Modern multidimensional scaling (2nd ed.). Springer. Börner, K. (2010). Atlas of science: Visualizing what we know. MIT Press. Chew, M., Villanueva, E.V., & Van der Weyden, M.B. (2007). Life and times of the impact factor: Retrospective analysis of trends for seven medical journals (1994 10

2005) and their editors views. Journal of the Royal Society of Medicine, 100(3), 142 150. Falagas, M.E., Zarkali, A., Karageorgopoulos, D.E., Bardakas, V., & Mavros, M.N. (2013). The impact of article length on the number of future citations: A bibliometric analysis of general medicine journals. PLoS ONE, 8(2), e49476. Garfield, E. (1996). How can impact factors be improved? British Medical Journal, 313(7054), 411 413. Garfield, E. (2006). The history and meaning of the journal impact factor. JAMA, 295(1), 90 93. Glänzel, W., Schubert, A., Thijs, B., & Debackere, K. (2011). A priori vs. a posteriori normalisation of citation indicators. The case of journal ranking. Scientometrics, 87(2), 415 424. Glänzel, W., Thijs, B., Schubert, A., & Debackere, K. (2009). Subfield-specific normalized relative indicators and a new generation of relational charts: Methodological foundations illustrated on the assessment of institutional research performance. Scientometrics, 78(1), 165 188. Hirsch, J.E. (2005). An index to quantify an individual s scientific research output. Proceedings of the National Academy of Sciences, 102(46), 16569 16572. Klavans, R., & Boyack, K.W. (2010). Toward an objective, reliable and accurate method for measuring research leadership. Scientometrics, 82(3), 539 553. Lewison, G., & Dawson, G. (1998). The effect of funding on the outputs of biomedical research. Scientometrics, 41(1 2), 17 27. Lewison, G., & Devey, M.E. (1999). Bibliometric methods for the evaluation of arthritis research. Rheumatology, 38(1), 13 20. Lewison, G., & Paraje, G. (2004). The classification of biomedical journals by research level. Scientometrics, 60(2), 145 157. Leydesdorff, L., & Opthof, T. (2010). Scopus s source normalized impact per paper (SNIP) versus a journal impact factor based on fractional counting of citations. Journal of the American Society for Information Science and Technology, 61(11), 2365 2369. Mavros, M.N., Bardakas, V., Rafailidis, P.I., Sardi, T.A., Demetriou, E., & Falagas, M.E. (2013). Comparison of number of citations to full original articles versus brief reports. Scientometrics, 94(1), 203 206. Moed, H.F. (2005). Citation analysis in research evaluation. Springer. 11

Moed, H.F. (2010). Measuring contextual citation impact of scientific journals. Journal of Informetrics, 4(3), 265 277. Neuhaus, C., & Daniel, H.-D. (2009). A new reference standard for citation analysis in chemistry and related fields based on the sections of Chemical Abstracts. Scientometrics, 78(2), 219 229. Opthof, T. (2011). Differences in citation frequency of clinical and basic science papers in cardiovascular research. Medical and Biological Engineering and Computing, 49(6), 613 621. Patel, V.M., Ashrafian H., Ahmed, K., Arora, S., Jiwan, S., Nicholson, J.K., Darzi, A., & Athanasiou, T. (2011). How has healthcare research performance been assessed? A systematic review. Journal of the Royal Society of Medicine, 104(6), 251 261. Patsopoulos, N.A., Analatos, A.A., & Ioannidis, J.P.A. (2005). Relative citation impact of various study designs in the health sciences. JAMA, 293(19), 2362 2366. Peters, H.P.F., & Van Raan, A.F.J. (1993). Co-word-based science maps of chemical engineering. Part I: Representations by direct multidimensional scaling. Research Policy, 22(1), 23 45. Radicchi, F., Fortunato, S., & Castellano, C. (2008). Universality of citation distributions: Toward an objective measure of scientific impact. Proceedings of the National Academy of Sciences, 105(45), 17268 17272. Rafols, I., Leydesdorff, L., O Hare, A., Nightingale, P., & Stirling, A. (2012). How journal rankings can suppress interdisciplinary research: A comparison between Innovation Studies and Business & Management. Research Policy, 41(7), 1262 1282. Rip, A., & Courtial, J.P. (1984). Co-word maps of biotechnology: An example of cognitive scientometrics. Scientometrics, 6(6), 381 400. Seglen, P.O. (1997). Why the impact factor of journals should not be used for evaluating research. British Medical Journal, 314(7079), 498 502. Smolinsky, L., & Lercher, A. (2012). Citation rates in mathematics: A study of variation by subdiscipline. Scientometrics, 91(3), 911 924. Tijssen, R.J.W., & Van Raan, A.F.J. (1989). Mapping co-word structures: A comparison of multidimensional scaling and LEXIMAPPE. Scientometrics, 15(3 4), 283 295. 12

Van Eck, N.J., & Waltman, L. (2010). Software survey: VOSviewer, a computer program for bibliometric mapping. Scientometrics, 84(2), 523 538. Van Eck, N.J., & Waltman, L. (2011). Text mining and visualization using VOSviewer. ISSI Newsletter, 7(3), 50 54. Van Eck, N.J., Waltman, L., Dekker, R., & Van den Berg, J. (2010). A comparison of two techniques for bibliometric mapping: Multidimensional scaling and VOS. Journal of the American Society for Information Science and Technology, 61(12), 2405 2416. Van Leeuwen, T.N., & Calero Medina, C. (2012). Redefining the field of economics: Improving field normalization for the application of bibliometric techniques in the field of economics. Research Evaluation, 21(1), 61 70. Waaijer, C.J.F., Van Bochove, C.A., & Van Eck, N.J. (2010). Journal editorials give indication of driving science issues. Nature, 463, 157. Waaijer, C.J.F., Van Bochove, C.A., & Van Eck, N.J. (2011). On the map: Nature and Science editorials. Scientometrics, 86(1), 99 112. Waltman, L., & Van Eck, N.J. (2012a). The inconsistency of the h-index. Journal of the American Society for Information Science and Technology, 63(2), 406 415. Waltman, L., & Van Eck, N.J. (2012b). A new methodology for constructing a publication-level classification system of science. Journal of the American Society for Information Science and Technology, 63(12), 2378 2392. Waltman, L., & Van Eck, N.J. (in press). Source normalized indicators of citation impact: An overview of different approaches and an empirical comparison. Scientometrics. Waltman, L., Van Eck, N.J., Van Leeuwen, T.N., & Visser, M.S. (2013). Some modifications to the SNIP journal impact indicator. Journal of Informetrics, 7(2), 272 285. Waltman, L., Van Eck, N.J., Van Leeuwen, T.N., Visser, M.S., & Van Raan, A.F.J. (2011a). Towards a new crown indicator: An empirical analysis. Scientometrics, 87(3), 467 481. Waltman, L., Van Eck, N.J., Van Leeuwen, T.N., Visser, M.S., & Van Raan, A.F.J. (2011b). Towards a new crown indicator: Some theoretical considerations. Journal of Informetrics, 5(1), 37 47. 13

Zitt, M., & Small, H. (2008). Modifying the journal impact factor by fractional citation weighting: The audience factor. Journal of the American Society for Information Science and Technology, 59(11), 1856 1860. 14

Figure 1. Term map of the Cardiac & cardiovascular systems field. The map shows 2000 terms extracted from titles and abstracts of publications in the WoS field Cardiac & cardiovascular systems. In general, the closer two terms are located to each other, the stronger their relation. The size and the color of a term indicate, respectively, the number of publications in which the term occurs and the average citation impact of these publications (where blue represents a low citation impact, green a normal citation impact, and red a high citation impact). Each term occurs in at least 70 publications. 15

Figure 2. Term map of the Clinical neurology field. The map shows 2000 terms extracted from titles and abstracts of publications in the WoS field Clinical neurology. In general, the closer two terms are located to each other, the stronger their relation. The size and the color of a term indicate, respectively, the number of publications in which the term occurs and the average citation impact of these publications (where blue represents a low citation impact, green a normal citation impact, and red a high citation impact). Each term occurs in at least 100 publications. 16

Figure 3. Term map of the Surgery field. The map shows 2000 terms extracted from titles and abstracts of publications in the WoS field Surgery. In general, the closer two terms are located to each other, the stronger their relation. The size and the color of a term indicate, respectively, the number of publications in which the term occurs and the average citation impact of these publications (where blue represents a low citation impact, green a normal citation impact, and red a high citation impact). Each term occurs in at least 135 publications. 17