Bibliometrics in higher education: the impact on libraries Keith Webster

Transcription

Bibliometrics in higher education: the impact on libraries Keith Webster
Bibliometrics in higher education:
the impact on libraries
Keith Webster
University of Queensland, Australia
Dr Berenika M. Webster
CIBER, Univ. College London
Outline
• Bibliometrics and university libraries
– Structure of disciplines
– Collection management
– Research evaluation
• Bibliometric techniques
• Sample projects
Structure of disciplines
• Domain analysis
• Patterns of scholarship
Collection management
• Patterns of use
• Core journals in discipline
• International comparison
Research evaluation
• Approaches to research assessment: experiences from
UK, NZ and Australia
• Peer assessment and metrics
• Bibliometrics
– Assessment tools
– Challenges of bibliometric approach
– How to increase the quality and impact of published outputs
What do governments want for
their money?
• Economic outcomes
– increase wealth creation & prosperity
– improve nation’s health, environment &
quality of life
• Innovation
• R&D from private sector
• Improved competitiveness
• Less “curiosity-driven” activity
Research Assessment Exercise (UK)
• Started in mid-1980s to determine size of research funding from HEFCEs to
universities. Early 1990s brought shift from quantity to quality
• 60 subject panels examine, for quality, selected outputs of UK academics.
Also, previous funding, evidence of esteem, PhD completions and res.
environment are taken into consideration
• Grades for submission units; correlates with bibliometrics measures (Charles
Oppenheim’s research)
• It will be scraped in 2008. New systems will be a hybrid of a “basket of
metrics” and peer-review (???)
• Changed behaviour of scholars
– Increased volume of publications (salami slicing?)
– Increased quality of outputs (numbers of 5 and 5* departments increase)
Performance-Based Research Fund
(PBRF) New Zealand
• Introduced in early 2003
• 12 panels grades individuals in 41 subject categories
• Portfolio has three components
– Publications (4 best and the rest)
– Peer esteem
– Contribution to research environment
• 2003 round produced lower scores than expected and great
variability between subjects (from 4.7 in philosophy to 0.3 in
nursing). 40% of submitted staff were judged research inactive
(grade R)
• Scores for institutions are calculated from these and used together
with PhD completion and external funding metrics to calculate
awards to institutions
• 2006 round incorporates “breaks” for new researchers
Research Quality Framework (RQF)
- Australia
• To be introduced in 2007 (?)
• Quality includes the intrinsic merit of original research and academic
impact. This relates to recognition of the originality of research by peers
and its impact on the development of the same or related discipline areas
within community of peers.
Assessment: peer review and metrics
• Impact or use of the original research outside of the peer community that
will typically not be reported in traditional peer review literature. Relates to
the recognition by qualified end-users that quality research has been
successfully applied to achieve social, cultural, economic and/or
environmental outcomes.
Assessment: expert opinion and narratives or metrics?
RQF evidence portfolio
• Four “best” outputs (with rationale for selection)
• List of all published outputs in the last six years
• Statement of “early impact” as assessed by qualified
end-users (description of outcome; identification of
beneficiaries; metrics illustrating benefits; linkage
between claimant and beneficiary)
• Context statement for the research group
Evaluation: peer vs. metrics
• Evaluate/reward past performance
• Measures short to medium term impacts
• Established use for the evaluation of academic
impact (within the group of peers)
• No established/accepted procedures for
evaluation of impact outside academe
Peer review
• “Traditional” assessment is an outputs-based peer-review
–
–
–
–
–
–
–
–
–
–
a panel of experts “reads” submissions or outputs
not always transparent (selection; evaluation criteria)
subjective (concept of quality is very difficult to objectivise)
relies on proxies (e.g. ranking/prestige of publishing journal; inst.
affiliation of authors, etc.)
composition of panels will affect their judgements
lacks comparability between panels
“punishes” innovative or multidisciplinary research
“old boys” club disadvantaging young researchers
assesses what already has been assessed by journals’ peer-review
processes
time and resource intensive
Metrics
• Range of unobtrusive and transparent
measures can be:
– Input-based
• Income
• Success rates
– Output-based
• Long-term impact (patents, IP, etc.)
• Bibliometric measures (volume and impact/quality)
Problems with metrics (1)
• Input-based
– Retrospective
– Creates “Matthew effect”
– Disadvantages younger researchers (without
previous track record)
– Undervalues research done “on the string”
Problems with metrics (2)
• Output based
–
–
–
–
–
–
–
–
Citations as a measure of intellectual influence?
Comparing like with like (benchmarking)
Inadequate tools
Social science and humanities (in search for new metrics)
Research on the periphery (local citation indices)
New models of communication (capturing it)
Impacts on wider community (“tangible” outcomes)
Lack of expertise in analysis and interpretation (education)
Indicators of quality as measured
using published outputs
•
Number of publications
•
Citation counts to these publications (adjusted for self-citations)
•
Citations per publication
•
Percentage of uncited papers
•
Impact factors (of publishing journals)
•
Diffusion factor (of citing journals) – profile of users of research (who, where, when and what)
•
“Impact factor” of a scholar - Hirsh index (h index)
-what “window” should be used? 4, 5, 10 years?
–
–
(numbers of papers with this number of citations).
Your h index =75 if you wrote at least 75 papers with 75 citations each.
Note: These should not be seen as “absolute” numbers but always seen in the context of the
discipline, research type, institution profile, seniority of a researcher, etc.
Compare like with like!
• Applied research attracts fewer citations than basic research.
• Differences in citation behaviour between disciplines (e.g. papers in
organisational behaviour attract 5 times as many citations as papers
in accounting).
• Highest IF journal in immunology is Ann Rev Immun (IF 47.3) Mean
for cat. 4.02; and in health care and services category is Milbank Q.
(IF of 3.8). Mean for cat. 1.09.
• Matthew effect.
Benchmarking must be done using comparable variables!
Tools available
• Publication and citation data
–
–
–
–
Web of Science
SCOPUS
Google Scholar
Local/national databases
• Other bibliometric indicators
– Journal Citation Reports (JCR)
– Other indicators databases (national, essential, university,
institutional)
– ISIHighlyCited.com
WoS and Scopus: subject coverage
(% of total records)
WoS
Social
Sciences, 14
SCOPUS
Arts &
Humanities,
9
Social
Sciences, 2
Physical
Sciences, 25
Science, 77
Biological &
Environmental
Sciences, 13
Health & Life
Sciences, 60
Google Scholar ?
Jacso, 2005
Web of Science
• Covers around 9,000 journal titles and 200 book
•
•
•
•
•
•
series divided between SCI, SSCI and A&HCI.
Electronic back files available to 1900 for SCI
and mid- 50s for SSCI and mid-70s for A&HCI.
Very good coverage of sciences; patchy on
“softer” sciences, social sciences and arts and
humanities.
US and English-language biased.
Full coverage of citations.
Name disambiguation tool.
Limited downloading options.
Scopus
• Positioning itself as an alternative to ISI
• More journals from smaller publishers and open access (13,000
•
•
•
•
•
•
•
•
journal titles; 600 open access journals; 750 conf proceedings; 600
trade publications, 2.5 mil “quality” web pages)
Source data back to 1960.
Excellent for physical and biological sciences; poor for social
sciences; does not cover humanities or arts.
Better international coverage (60% of titles are non-US)
Not much of a back file (e.g. citation data for the last decade only)
Not “cover to cover” and not up to date
Easy to use in searching for source publications; clumsy in searching
cited publications.
Citation tracker works up to 1000 records only.
Limited downloading options.
Google Scholar
• Coverage and scope?
• Inclusion criteria?
• Very limited search options
• No separate cited author search
• Free!
Taiwan in the three sources,1996-2005
70000
Scopus (1.23%)
60000
WoS (0.98%)
Google Scholar
50000
40000
30000
20000
10000
0
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
Social sciences
• Communication in social sciences
–
–
–
–
Books
National journals
International journals
Non-scholarly outputs (culture-creating role)
–
–
–
–
–
Chinese Social Sciences Citation Index
Polish Sociology Citation Index
Serbian Social Sciences Index
Polish Citation Index for Humanities
European Humanities Citation index
• “International” tools inadequate
• Creation of “national” and “regional” tools
Science on the periphery
• Poor coverage in international databases
• Intl. benchmarking difficult
• Impossible for local evaluations
• Development of national sources of data
• While volume is growing, citation rates lag
behind
Science on the periphery
Figure 1: Polish scientific papers in SCI (in Polish and foreign
journals), three-year running means and Polish SCI journals, 19801999
30
Polish SCI
journals (right
scale)
7000
Papers per year
6000
25
20
5000
4000
15
3000
10
Papers in
foreign journals
2000
5
1000
Papers in Polish journals
0
journals
1981
1983
1985
0
1987
1989
1991
1993
1995
1997
Number of Polish journals in SCI
8000
Capturing new forms of
communication
• Use of online journals (deep log analysis)
• Traffic on websites and downloads
• In-links to websites
Impact of research: example of
biomedicine
• Informing policies (citations on guidelines, govt.
policy, development of medicines)
• Building capacity (training; development)
• Relationship between research and health
outcomes and cost savings
• Healthier workforce
Citations on clinical guidelines
• 47 Clinical Guidelines and 64 Health Technology
Appraisals prepared by the UK’s National
Institute for Clinical Excellence (NICE) were
analysed for citations
– cited 25% of UK papers (2.5 more than expected)
– Majority were clinical papers
– Median lag time between publication of res. paper
and its citation was 3.3 years
Relative commitment to research and
burden of disease
0.8
Percent of biomed research
Prostate cancer
SE
NL
0.6
US
y = 0.1112x + 0.1645
R 2 = 0.087
0.4
IT
CA
UK
DE
ES FR
JP
0.2
AU
CH
0.0
0.0
0.5
1.0
1.5
Percent of all deaths
2.0
2.5
3.0
How to increase “quality” of your
publications?
•
Publish in the right journals (prestige; importance to the discipline; impact factor vs. diffusion
factor)
•
Publish in English
•
Write review articles
•
Engage in basic research
•
Become a journal editor (Lange, 1997)
•
Acquire a co-author (preferably from US or UK)
•
Get external funding (from different sources)
•
Make your outputs available in open access (own website, institutional and subject repositories)
(Antelman, 2004; Harnard various)
•
“Advertise” your publications on listservs and discussion groups
•
Make and maintain professional/social contacts with others in your research area (Rowlands,
2000)