How Scholarly Is Google Scholar? A Comparison of Google Scholar to Library Databases

Similar documents
ENSC 105W: PROCESS, FORM, AND CONVENTION IN PROFESSIONAL GENRES

A Case Study of Web-based Citation Management Tools with Japanese Materials and Japanese Databases

1.1 What is CiteScore? Why don t you include articles-in-press in CiteScore? Why don t you include abstracts in CiteScore?

INFS 321 Information Sources

An Introduction to Bibliometrics Ciarán Quinn

Citation-Based Indices of Scholarly Impact: Databases and Norms

Ebook Collection Analysis: Subject and Publisher Trends

King's College STUDY GUIDE # 4 D. Leonard Corgan Library Wilkes-Barre, PA 18711

Citation analysis: Web of science, scopus. Masoud Mohammadi Golestan University of Medical Sciences Information Management and Research Network

Web of Science Unlock the full potential of research discovery

EndNote X8 Workbook. Getting started with EndNote for desktop. More information available at :

The Financial Counseling and Planning Indexing Project: Establishing a Correlation Between Indexing, Total Citations, and Library Holdings

Battle of the giants: a comparison of Web of Science, Scopus & Google Scholar

University of Liverpool Library. Introduction to Journal Bibliometrics and Research Impact. Contents

Absolute Relevance? Ranking in the Scholarly Domain. Tamar Sadeh, PhD CNI, Baltimore, MD April 2012

INTRODUCTION TO ENDNOTE

USING ENDNOTE X2 SOFTWARE

Selected Databases and EndNote

EVALUATING THE IMPACT FACTOR: A CITATION STUDY FOR INFORMATION TECHNOLOGY JOURNALS

Echocardiography Research

Should the Journal of East Asian Libraries Be a Peer- Reviewed Journal? A Report of the Investigation and Decision

Code Number: 174-E 142 Health and Biosciences Libraries

What Can I Trust? Identifying Trustworthy Information. Health 205- September 2014 Michele Chittenden Queen s University Library

Citation & Journal Impact Analysis

Exporting Citations from Databases into EndNote

Library and IT Services Manual EndNote import filters Tilburg University

Google vs. the Library (Part III): Assessing the Quality of Sources Found by Undergraduates

Collection Development Policy, Film

CITATION INDEX AND ANALYSIS DATABASES

Library Guide to EndNote Online

Critical Evaluation of Information Sources True or false: Authority Objectivity Quality Coverage Currency

KEAN UNIVERSITY LIBRARY GUIDE Graduate Research Resources

GPLL234 - Choosing the right journal for your research: predatory publishers & open access. March 29, 2017

How to Choose the Right Journal? Navigating today s Scientific Publishing Environment

Research Resources for Graduate Bilingual Education

DIGITAL MEASURES BIBTEX AND PUBMED IMPORT- MANUAL

BME100 Library Resources Session

The Literature Resource Center User Guide

Finding Periodical Articles

Practice with PoP: How to use Publish or Perish effectively? Professor Anne-Wil Harzing Middlesex University

The largest abstract and citation database

Direct export allows you to mark items in a database or catalogue, and then export them directly into your EndNote library.

and Beyond How to become an expert at finding, evaluating, and organising essential readings for your course Tim Eggington and Lindsey Askin

Information Literacy for German Language and Literature at the Graduate Level: New Approaches and Models

Cracking the PubMed Linkout System

Horizon 2020 Policy Support Facility

Microsoft Academic is one year old: the Phoenix is ready to leave the nest

A Citation Analysis of Articles Published in the Top-Ranking Tourism Journals ( )

THE INFORMATION MATRIX

RESEARCH MATERIALS AND STRATEGIES FOR COMM 498E Alan Mattlage, Communication Librarian

Academic honesty. Bibliography. Citations

Contract Cataloging: A Pilot Project for Outsourcing Slavic Books

Collection Development Policy, Modern Languages

The Library Reference Collection: What Kinds of Materials will you find in the Reference Collection?

Finding & Using Different Article Types

Journal Citation Reports on the Web. Don Sechler Customer Education Science and Scholarly Research

Helping your students identify academic sources

Journal of Undergraduate Research at Minnesota State University, Mankato

of Nebraska - Lincoln

Library resources & guides APA style Your research questions Primary & secondary sources Searching library e-resources for articles

with a librarian at encyclopedia on your subject such as Mosby s Paramedic

EndNote X4 An Introduction. EndNote X4 - introduction Page 1

EndNote Basics Fall 2010, Room 14N-132 Peter Cohn, x8-5596


EDITORIAL POLICY. Open Access and Copyright Policy

UCSB LIBRARY COLLECTION SPACE PLANNING INITIATIVE: REPORT ON THE UCSB LIBRARY COLLECTIONS SURVEY OUTCOMES AND PLANNING STRATEGIES

The Online Search Process and Building an EndNote Library

Citation Indexes: The Paradox of Quality

Using Bibliometric Analyses for Evaluating Leading Journals and Top Researchers in SoTL

College of Southern Nevada

LIBRARY TUTORIAL 4: ERIC (EDUCATION RESOURCE INFORMATION CENTER)

Follow this and additional works at: Part of the Library and Information Science Commons

A BEGINNER'S GUIDE TO ENDNOTE ONLINE

Research Smarter NOT Harder. College of Science & Mathematics

Arjumand Warsy

Google Labs, for products in development:

EndNote basics. Francesca Frati, MLIS Jacynthe Touchette, MSI JGH Health Sciences Library

Google Tools for Scholars

Scopus. Advanced research tips and tricks. Massimiliano Bearzot Customer Consultant Elsevier

LIBRARY RESOURCES FOR SUSTAINABILITY 100

The College Student s Research Companion:

NURS 300 BS Nursing Program Orientation. Wenli Gao Instructional Services Librarian MSC Butcher Library

ENDNOTE X6 FOR HEALTH

Indexing in Databases. Roya Daneshmand Kowsar Medical Institute

Our E-journal Journey: Where to Next?

Semi-automating the manual literature search for systematic reviews increases efficiency

Your main aim should be to capture references electronically, avoid typing in reference information by hand. This is a last resort.

Coverage analysis of publications of University of Mysore in Scopus

You can listen to or view the contents of this tutorial on the left menu.

Introduction to Bell Library Resources

DISCOVERING JOURNALS Journal Selection & Evaluation

GETTING STARTED WITH ENDNOTE

Dr Paper. Online Researcher. 5Plus. Contents... Copyright Oberon and The Write Direction Columbus Ohio All rights reserved.

DOCTORAL DISSERTATIONS OF MAHATMA GANDHI UNIVERSITY A STUDY OF THE REFERENCES CITED

All academic librarians, Is Accuracy Everything? A Study of Two Serials Directories. Feature. Marybeth Grimes and

Simple Steps to Effective Library Research :

Using EndNote 6 to create bibliographies

Homework Help for Parents Academic resources from the Library and the Web

Introduction to EndNote Desktop

Lifelong Learning Research tips

Transcription:

Brigham Young University BYU ScholarsArchive All Faculty Publications 2008-07-08 How Scholarly Is Google Scholar? A Comparison of Google Scholar to Library Databases Jared L. Howland jared_howland@byu.edu Thomas C. Wright tom_wright@byu.edu See next page for additional authors Follow this and additional works at: https://scholarsarchive.byu.edu/facpub Part of the Library and Information Science Commons Original Publication Citation Jared L. Howland, Thomas C. Wright, Rebecca A. Boughan and Brian C. Roberts. "How Scholarly Is Google Scholar? A Comparison to Library Databases". College & Research Libraries vol. 70 no. 3 227-234, May 2009. doi: 10.5860/crl.70.3.227 BYU ScholarsArchive Citation Howland, Jared L.; Wright, Thomas C.; and Boughan, Rebecca A., "How Scholarly Is Google Scholar? A Comparison of Google Scholar to Library Databases" (2008). All Faculty Publications. 1263. https://scholarsarchive.byu.edu/facpub/1263 This Presentation is brought to you for free and open access by BYU ScholarsArchive. It has been accepted for inclusion in All Faculty Publications by an authorized administrator of BYU ScholarsArchive. For more information, please contact scholarsarchive@byu.edu, ellen_amatangelo@byu.edu.

Authors Jared L. Howland, Thomas C. Wright, and Rebecca A. Boughan This presentation is available at BYU ScholarsArchive: https://scholarsarchive.byu.edu/facpub/1263

How Scholarly is? A Comparison of Google Scholar to Library Databases Jared L. Howland Thomas C. Wright Rebecca A. Boughan Brian C. Roberts Brigham Young Univeristy 1

Introduction Literature Review 2

3

Methodology Subject specialists 7 Subject specialists from 3 disciplines: 3 from sciences 2 from humanities 2 from social sciences Blind to purpose of study Asked them to give us 3 things... 4

How does the acquisition and use of a second language in children affect their general cognitive development? A question they would typically receive from a student (humanities) 5

(bilingual* OR L2) AND (child* OR toddler) AND cognitive development A structured query they would use to search a database 6

Linguistic and Language Behavior Abstracts The database they would use to search for that question 7

Academic Discipline Database Query Library Database Science (ACL or anterior cruciate ligament* ) and injur* and (athlet* or sport or sports) and (therap* or treat* or rehab*) SportDiscus Science Science lung cancer and (etiol* or caus*) and (cigarette* or smok* or nicotine*) dark matter and evidence Medline Applied Science and Technology Abstracts Social Science ( fast food or mcdonald s or wendy s or burger king or restaurant) and franchis* and (knowledge n3 transfer or knowledge management or train*) Business Source Premier Social Science ( standardized test* or high stakes test* ) and ( learning disabilit* or Dyslexia or learning problem ) and accommodat* PsycINFO Humanities Humanities (bilingual* or L2) and (child* or toddler) and cognitive development (memor* or remembrance or memoir*) and (holocaust) and (Spiegelman or Maus) Linguistics and Language Behavior Abstracts JSTOR This is what things looked like after we got all the information back from the librarians 8

Methodology Search using query Then we took that information and used it in 2 ways. 9

Native database results The first was to actually run the search query in the suggested database. 10 We put the first 30 citations into a bibliographic citation manager and saved all of the actual full text We chose 30 because usability studies (Jakob Neilsen) tell us that less than 1% of all users ever go beyond the 3rd page of results and very few people ever change the defaults (ie, once they run a search they stick with it, success or failure). Most of our DBs present 10 results per page so 30 results should represent a large enough sample to represent the actual set of results the majority of our users is ever going to see after performing a search.

Google Scholar results We ran the same query in Google Scholar and saved the results again in a bibliographic Manager. 11 We used Zotero to quickly export all of the results. We also saved the full text of each citation for later use in our study.

Methodology Search using citations So, the first searches we ran using the native DBs and GS was for the query given to us by the librarian 12 The second set of searches we ran was to see if the citations we found in the DB were available in GS and vice versa

Is this citation available in Google Scholar? Here is the same screenshot we saw just a minute ago. 13 We took the bibliographic information for each citation and searched for the citation within Google Scholar.

Yes, it is available We then did the same thing in reverse. We took the 30 results from GS and searched for each citation within the database 14

in GS in both in DB Exclusivity This allowed us to later calculate something we called exclusivity 15 We put the citations into 1 of 3 possible exclusivity categories Shows proportion of citations within our study that overlap. As you can see, within our study we found that, on average, GS had a larger result set overall as well as more exclusively than the databases.

Methodology Citation grading 16 So now that we have the citations from the database and the citations from Google Scholar. We used the bibliographic manager to generate a list of references that we input into an Excel spreadsheet. Then, using a random number table, we completely randomized the order of the citations for each subject specialist.

Finally, to deliver the content to the librarians in a way in which it would be easiest for them to evaluate, we saved the full-text of each citation according to its randomly assigned citation number. Then we used Excel to create hyperlinks to the full-text of each citation and delivered this list along with the full-text on a CD to the subject librarians. We asked them to evaluate each citation using a rubric which we provided in hard copy form. As you can see, the subject librarians were only able to see the citation number and the bibliographic information. By clicking on the hyperlinked citation number, the full-text of that citation would appear and the subject librarians could easily rate the citation on the rubric. 17 Have full text appear on this page after click to simulate linking from provided document.

This resulted in a total possible score of 18 for each citation - we called this a scholarliness score Accuracy: Authority: Objectivity: Currency: Coverage: Relevancy: reliability, fact checkers/editors, peer review author s qualifications, reputable publisher minimum bias, extent to which persuasion is the goal information up to date, date of publication indicated depth of coverage related to research topic Rubric and Full Text This screen shows the rubric that we used. It is based on a rubric that has popularly been used to evaluate print resources (Alexander, 1999) 18 Alexander, J. E. (1999). Web wisdom: How to evaluate and create information quality on the Web. We asked each subject librarian to assign a score of between 1 and 3 within 6 different categories to each of the citations (1 was below average, 2 was average and 3 was above average). These six categories were: Accuracy which looks at Authority specifically the Objectivity looking for Currency is the information up to date? How deep is the Coverage And finally Relevancy how well does the citation relate to the research question

Methodology total scholarliness score = μ + Ei + Lj + ELij + εijkl where μ = Average total score E = Effect due to exclusivity (i = 1, 2, 3) L = Effect due to librarian (j = 1, 2,... 7) EL = Interaction between exclusivity and librarian ε = Error term We used this statistical model to evaluate the data. Essentially this formula says 2 important things about the way we used the data: 19 1. We controlled for the differences between the way librarians grade 2. We controlled for the differences in how exclusively the citation was available This allowed us to pinpoint and measure any differences there may have been between disciplines in our data as well as any differences that can be attributed to the source of the citations

Results Google Scholar was 17.6% more scholarly Citations found only in GS had, on average, a 17.6% higher scholarliness score than citations found only in the DB 20

Results Highest scholarliness score when found in both Citations found in both GS and the DB were even higher than citations found only in GS 21

Results No difference between disciplines We found no statistically significant difference in the scholarliness scores between disciplines (ie, humanities citations in GS are just as scholarly as science citations found in GS) 22

18 Average Scholarliness Score 14.2 13.5 14.6 15.6 14.3 14.3 13.9 11.5 9 14.0 16.1 13.8 12.0 13.5 11.6 12.8 14.4 0 11.9 AVERAGE 11.7 Physics 13.2 Linguistics Biology 10.0 Medicine 10.0 Literature 11.7 Business 16.5 Education Only in database Only in GS In both 23

100% 100% 97% 93% 76% 77% 83% 82% 50% 0% 47% AVERAGE Physics 43% Linguistics 97% Biology 80% Medicine 28% Literature 47% Business 35% Education database citations in GS GS citations in database 24

Future Studies Generally applicable results This study can only be extrapolated statistically to the specific topics and subject specialists used in this study 25 A more robust statistical methodology would need to be employed to make these results generally applicable We are encouraged by the results we received and feel that they would probably hold up but cannot say so until another study is done

Future Studies Improved rubric If we had to do it over again, we would have increased the Likert scale on our rubric from 1-3 to 1-7 or 1-10 26 This would have allowed for a more nuanced statistical analysis and made it easier to spot significant differences, if any, between GS and databases

Future Studies Scholarliness calculation Our scholarliness calculation, ultimately, was based on the subjective opinions of librarians with subject expertise. 27 There are lots of ways to create a scholarliness score (citation counts, impact factors, etc). Which is best is still debatable

Future Studies Comparison to federated searching Our study compared GS to individual library databases. A more appropriate comparison may be GS to federated search tools. 28

Questions? jared_howland@byu.edu tom_wright@byu.edu rebecca_boughan@byu.edu http://dspace.byu.edu/handle/1877/634 29