W. Glänzel and K. Debackere

Download Report

Transcript W. Glänzel and K. Debackere

MEASURING COMMUNICATION
IN SCIENCE
OPPORTUNITIES AND LIMITATIONS OF
BIBLIOMETRIC METHODS
Wolfgang Glänzel and Koenraad Debackere
SooS, Leuven, Belgium
STRUCTURE OF THE PRESENTATION
Introduction
1. Structure of Bibliometrics
2. Data sources of bibliometric research and technology
3. Elements, units and measures of bibliometric
research
4. Citations and Self-citations
5. “Against absolute methods”
6. Journal Impact Factor
7. Distorted behaviour based on policy use and misuse
of bibliometric data
Conclusions
W. Glänzel and K. Debackere
Measuring communication in science
2
INTRODUCTION
W. Glänzel and K. Debackere
Measuring communication in science
3
Introduction
What is bibliometrics?
The terms bibliometrics and scientometrics were almost
simultaneously introduced by Pritchard and by Nalimov &
Mulchenko in 1969.
According to Pritchard bibliometrics is
“the application of mathematical and statistical
methods to books and other media of
communication”.
W. Glänzel and K. Debackere
Measuring communication in science
4
Introduction
Nalimov and Mulchenko defined scientometrics as
“the application of those quantitative methods which
are dealing with the analysis of science viewed as an
information process”.
The two terms have become almost synonyms;
nowadays, the field informetrics (Gorkova, 1988)
stands for a more general subfield of information
science dealing with mathematical-statistical analysis
of communication processes in science.
W. Glänzel and K. Debackere
Measuring communication in science
5
Introduction
Present-day use of bibliometrics
• Bibliometrics has evolved to a standard tool of
science policy and research management.
• A vast array of indicators to measure and to map
research activity and its progress is available.
• Science indicators relying on comprehensive
publication and citation statistics and other, more
sophisticated bibliometric techniques, are used in
science policy and research management.
• A growing, often controversial, policy interest to use
bibliometric techniques in measurements of
research productivity and efficiency.
W. Glänzel and K. Debackere
Measuring communication in science
6
Introduction
Common misbeliefs on bibliometrics
• Main task of bibliometrics should be the expeditious
issuing of “prompt” and “comprehensible” indicators
for science policy and research management.
• Research on bibliometric methodology is
unnecessary; instead bibliometricians should
elaborate guidelines explaining the use of their
indicators.
• Bibliometrics might be reduced to simple counting
activities in order to replace/supplement qualitative
assessment by quantitative indicators and to set
publication output off against funding.
W. Glänzel and K. Debackere
Measuring communication in science
7
Introduction
Facts about bibliometrics
• Bibliometrics is a powerful, multifaceted endeavour
encompassing subareas such as structural, dynamic
and evaluative scientometrics.
- Structural scientometrics came up with results
like the re-mapping of the epistemological
structure of science.
- Dynamic scientometrics constructed
sophisticated models of scientific growth,
obsolescence, citation processes, etc.
- Evaluative scientometrics developed arrays of
indicators to be used to characterise research
performance at different levels of aggregation.
W. Glänzel and K. Debackere
Measuring communication in science
8
Introduction
What is bibliometrics dealing with and what can
bibliometrics not be responsible for?
• Bibliometrics can be used to develop and provide
tools to be applied to research evaluation, but is not
designed to evaluate research results.
• Bibliometrics does not aim at replacing qualitative
methods by quantitative approaches.
• Consequently, bibliometrics is not designed to correct
or even substitute peer reviews or evaluation by
experts but qualitative and quantitative methods in
science studies should complement each other.
W. Glänzel and K. Debackere
Measuring communication in science
9
1. STRUCTURE OF BIBLIOMETRICS
W. Glänzel and K. Debackere
Measuring communication in science
10
1.
Structure of Bibliometrics
The three “components” of present-day bibliometrics
according to its three main target-groups
Bibliometrics for bibliometricians (Methodology)
This is the domain of bibliometric “basic research”.
Bibliometrics for scientific disciplines (Scientific information)
A large but also the most diverse interest-group in bibliometrics.
Due to the scientists’ primary scientific orientation, their interests are
strongly related to their speciality. Here we also find joint borderland
with quantitative aspects of information retrieval.
Bibliometrics for science policy and management (Science policy)
At present the most important topic in the field. Here the national,
regional, and institutional structures of science and their comparative
presentation are in the foreground.
W. Glänzel and K. Debackere
Measuring communication in science
11
1.
Structure of Bibliometrics
Links of bibliometrics with related research fields and application services
Science policy
Research management
Scientific information
Librarianship
Services for
Research in
Economics
Sociology of science
History of science
applied
Library and
Information Science
Scientometrics
basic
Life sciences
Informetrics
Mathematics/Physics
Webometrics
W. Glänzel and K. Debackere
Measuring communication in science
12
2. DATA SOURCES OF BIBLIOMETRIC
RESEARCH AND TECHNOLOGY
W. Glänzel and K. Debackere
Measuring communication in science
13
2.
Sources of Bibliometrics
Data sources of bibliometric research and technology
Data sources of bibliometrics are bibliographies and
bibliographic databases. Large scale analyses can only be
based on bibliographic databases.
Prominent specialised databases are, e.g., Medline, Chemical
Abstracts, INSPEC and Mathematical Reviews in the sciences
and, e.g., Econlit, Sociological Abstracts and Humanities
Abstracts in the social sciences and humanities.
 Disadvantage: Lack of reference literature, incomplete
address recording
The databases of the Institute for Scientific Information
(Thomson - ISI), above all, the Science Citation Index
(Expanded) have become the most generally accepted
source of bibliometrics.
W. Glänzel and K. Debackere
Measuring communication in science
14
2.
Sources of Bibliometrics
Although, there are several objections against the journal
coverage and the data processing policy of the ISI in
preparing the SCI, its unique features are basic
requirements of bibliometric technology. Among these
features we have
•
•
•
Multidisciplinarity
Selectiveness
Completeness of addresses
•
•
Full coverage
Bibliographical references
 Disadvantage: no individual subject classification for
papers available.
W. Glänzel and K. Debackere
Measuring communication in science
15
2.
Sources of Bibliometrics
ProceedingsSM is available in two editions (Science &
Technology and Social Sciences & Humanities) covering
about 2,000,000 papers from over 60,000 conferences
since 1990.
Non-serial literature (except for proceedings) such as
monographs and books is not indexed in these ISI
databases.
Since non-serial literature is an important conveyor of
information in the social sciences and humanities, journal
based data-sources are accepted by scientists only with
certain reservations.
W. Glänzel and K. Debackere
Measuring communication in science
16
3. ELEMENTS, UNITS AND MEASURES OF
BIBLIOMETRIC RESEARCH
W. Glänzel and K. Debackere
Measuring communication in science
17
3.
Elements of Bibliometrics
Elements, units and measures of bibliometric research
Basic units in bibliometrics are usually not further subdivided.
These form the elements of bibliometric analyses. Elements are,
e.g., publications, (co-)authors, references and citations.
Publications can be assigned to the journals in which they
appeared, through the corporate addresses of their authors to
institutions or countries, references and citations to subject
categories, and so on.
Units are specific sets of elements, e.g., journals, subject
categories, institutions, regions and countries to which elements
can – not necessarily uniquely – be assigned. The clear definition
of the assignment – or in mathematical parlance – of mappings
between elements and units allows the application of
mathematical models.
W. Glänzel and K. Debackere
Measuring communication in science
18
3.
Elements of Bibliometrics
Publication activity and authorship
Publication activity is influenced by several factors. At the micro
level, we can distinguish the following four factors.
1.
the subject matter
2.
the author’s age
3.
the author’s social status
4.
the observation period
The publication activity in theoretical fields (e.g., mathematics)
and in engineering is lower than in experimental fields or in the
life sciences.
Cross-field comparison – without appropriate normalisation –
would not be valid. This applies above all to comparative
analyses at the meso level (universities and departments).
W. Glänzel and K. Debackere
Measuring communication in science
19
3.
Elements of Bibliometrics
Can scientific collaboration be measured through
co-authorship?
•Laudel (2001) (micro study): A large share of persons involved in
the preparation of a scientific paper does thus not appear either
as co-author or as a sub-author. Katz & Martin (1997) argue that
co-authorship is no more than a partial indicator of collaboration.
•Intensifying collaboration, however, goes with growing coauthorship (Patel, 1973). There is a positive correlation between
collaboration and co-authorship at the level of individual actors,
too.
•The phenomenon described by Laudel and Katz & Martin rather
applies to intramural collaboration. Extramural collaboration,
above all international collaboration, is usually well
acknowledged.
W. Glänzel and K. Debackere
Measuring communication in science
20
4. CITATIONS AND SELF-CITATIONS
W. Glänzel and K. Debackere
Measuring communication in science
21
4.
Citations and Self-citations
The notion of citations in information science and
bibliometrics
Citations became a widely used measure of the impact of
scientific publications.
Cozzens: “Citation is only secondarily a reward system.
Primarily, it is rhetorical-part of persuasively arguing for the
knowledge claims of the citing document.”
L. C. Smith: "citations are signposts left behind after
information has been utilized".
Cronin: Citations are "frozen footprints in the landscape of
scholarly achievement … which bear witness to the passage
of ideas“.
W. Glänzel and K. Debackere
Measuring communication in science
22
4.
Citations and Self-citations
Glänzel and Schoepflin: Citations are “one important form of use
of scientific information within the framework of documented
science communication,” Although citations cannot describe
the totality of the reception process, they give, “a formalised
account of the information use and can be taken as a strong
indicator of reception at this level.”
Westney: “Despite its flaws, citation analysis has demonstrated
its reliability and usefulness as a tool for ranking and evaluating
scholars and their publications. No other methodology permits
such precise identification of the individuals who have
influenced thought, theory, and practice in world science and
technology.”
Garfield and Weinstock have listed 15 different reasons for giving
citations to others’ work.
W. Glänzel and K. Debackere
Measuring communication in science
23
4.
Citations and self-citations
The process of re-interpreting the notion of citation and its
consequences
interpretation
citation
Bibliometrics/
Information science
Signpost of information use
repercussion
re-interpretation
(possible distortion of
citation behaviour)
Research evaluation/
Science policy
W. Glänzel and K. Debackere
uncitedness: unused information
frequent cite: good reception
self-cite: part of scient. communication
Rewarding system/
Quality measure
uncitedness: low quality
frequent cite: high quality
self-cite: manipulation of impact
Measuring communication in science
24
5. “AGAINST ABSOLUTE METHODS”
W. Glänzel and K. Debackere
Measuring communication in science
25
“Against absolute methods”
5.
Factors influencing citation impact
Citation impact is mainly influenced by the following five factors
that are analogously to the case of publication activity at higher
levels of aggregation practically quite inseparable.
1. the subject matter and within a subject, the “level of
abstraction”
2. the paper’s age
3. the paper’s “social status” (through authors and journal)
4. the document type
5. the observation period (”citation window”)
W. Glänzel and K. Debackere
Measuring communication in science
26
“Against absolute methods”
5.
Complexity of influences and biases in calculating
citation impact measures
Mean citation rate of two journals in time as a function of time
(source year: 1980)
Citation
window
1980-80
1980-81
1980-82
1980-85
1980-89
W. Glänzel and K. Debackere
Mean citation rate
ASR
LANCET
0.2
0.6
1.8
2.4
4.3
4.5
12.1
9.7
20.9
14.0
Measuring communication in science
27
“Against absolute methods”
5.
Impact of different document types
”3-year impact measure” for selected journals by document types
(source: 1995/96)
# Journal
1 SCIENCE
Total
32.86
2 NATURE
3 LANCET
Impact Factor
Articles
Reviews
Letters
42.30
145.35
0.41
32.88
49.73
96.07
3.93
5.25
17.55
14.68
1.99
4 CELL
75.68
74.82
78.63
75.64
5 ANGEW CHEM INT ED
11.01
9.37
32.03
19.00
6 J ACQ IMMUN DEFIC SYND HUM R
4.05
4.64
39.00
1.04
7 INT J RAD ONCOL BIOL PHY
3.52
4.15
35.00
0.37
8 J PHYS CONDENS MATTER
2.72
2.47
9.57
3.99
W. Glänzel and K. Debackere
Measuring communication in science
28
“Against absolute methods”
5.
Influence of subject characteristics
Mean citation rate of subfields
(source: 1996, citation window: 1996-1998)
Mechanical, civil and other engineering
Mathematics
Analytical chemistry
Solid state physics
Neurosciences
1.12
1.46
3.00
3.06
4.54
Citation measures are thus – without normalisation – not
appropriate for cross-field comparisons.
W. Glänzel and K. Debackere
Measuring communication in science
29
“Against absolute methods”
5.
 The only possible way to compensate for
the subject-specific characteristics is an
appropriate normalisation and the
application of exactly the same
underlying publication period and
citation window to all units under study.
W. Glänzel and K. Debackere
Measuring communication in science
30
“Against absolute methods”
5.
Citation indicators can be normalised using a reference standard
based on journals or subjects in which the papers under study
have been published.
Problem: Subject assignament is not unique.
The Relative Citation Rate (RCR) introduced by Schubert et al.
in 1983 gauges observed citation rates of the papers against the
standards set by the specific journals. It has largely been
applied to comparative macro and meso studies since. A version
of this relative measure, namely, CPP/JCSm, where JCSm
denotes the mean Journal Citation Score is used at CWTS.
The Normalised Mean Citation Rate (NMCR) introduced by
Braun and Glänzel in 1993 normalises observed citation rates
by weighted average of the mean citation rates of subfields.
A similar measure (CPP/FCSm, FCSm being the mean Field
Citation Score) is used at CWTS.
W. Glänzel and K. Debackere
Measuring communication in science
31
6. JOURNAL IMPACT FACTOR
W. Glänzel and K. Debackere
Measuring communication in science
32
6.
Journal Impact Factor
On the role of the Impact Factor
• The Garfield ISI Impact Factor (IF) represents a
paradigm in bibliometric/information science research.
• The IF is used frequently and has obtained a very strong
‘market’ position.
• From the mathematical viewpoint, the IF is the mean
value, i.e., an arithmetic mean of citations in a particular
(citing) year to a particular set of articles published in a
particular journal one or two years earlier.
• The Impact Factor has become perhaps the most
popular bibliometric product used in bibliometrics and
outside the scientific community.
W. Glänzel and K. Debackere
Measuring communication in science
33
6.
Journal Impact Factor
Problems in using the ISI Impact Factors
•
The strengths of the Impact Factor lies first of all in the
comprehensibility, stability and seeming reproducibility, but some
flaws have provoked critical and controversial discussions about
its correctness and use.
•
The above-mentioned popularity involves also dangers. The use
of impact factors ranges from well-documented and methodically
sound applications to rather ‘grey’ applications as background
information for scientific journalism or in the context of refereeing
procedures. Impact factors are sometimes used even as
substitutes for missing citation data.
•
Although it is difficult to theoretically define the concept of
(journal) impact, there is a wide spread belief that the ISI Impact
Factor is affected or ‘disturbed’ by factors that have nothing to do
with (journal) impact.
W. Glänzel and K. Debackere
Measuring communication in science
34
6.
Journal Impact Factor
•
Being a statistical mean, the IF should be size-independent.
Large journals might, however, often have a higher visibility.
•
The robustness, comprehensibility and methodological
reproducibility of the ISI journal Impact Factor is contrasted by
methodological shortcomings and its technical irreproducibility.
•
It became quite tempting to apply the impact factor as a universal
bibliometric measure. This is certainly one source of possible
uninformed use.
•
Methodological improvements in combination with
complementary measures and an appropriate documentation
may help to overcome limitations described above.
•
The question of reproducibility can thus at least partially be
solved for those who have access to the bibliographic databases
and the technology to produce journal indicators.
W. Glänzel and K. Debackere
Measuring communication in science
35
6.
Journal Impact Factor
Visibility vs. publication targeting vs. citation impact
Publication in a high-IF journal might guarantee excellent visibility, but
not automatically imply high citation rates, too.
In several fields, targeting, i.e., reaching the desired audience is more
important than publishing in high-impact journals (e.g., in clinical
medicine, mathematics).
The latter observation substantiates that research may have other
impact than citations. In order to gain new insight in the utility of
biomedical research Grant Lewison studied citations from clinical
guidelines, textbooks, government policy documents, international or
national regulations and newspaper articles. Moreover, publications in
technical sciences and clinical medicine might find practical application
that cannot be measured through citations.
 Citations and, above all, the IF do not measure all
aspects of impact published research results might have.
W. Glänzel and K. Debackere
Measuring communication in science
36
6.
Journal Impact Factor
The myth of delayed recognition
An often-heard argument on limitations of citation-based
indicators is that important publications are often not cited in the
beginning, and only become recognised in a time that is beyond
the standard citation windows used in most bibliometric studies.
Studies by Glänzel at al. and Glänzel & Garfield in 2004 have
shown that the chance that a paper, uncited for three to five
years after publication, will ever be cited is quite low, even in
slowly aging fields such as mathematics. The citation impact of
papers not cited initially usually remains low even 15 to 20 years
later.
The potential number of delayed recognition papers is extremely
small. A statistically marginal share of 1.3 per 10,000 papers
published in 1980 were "neglected" at first, and then, belatedly,
received relatively high citational recognition.
W. Glänzel and K. Debackere
Measuring communication in science
37
7. DISTORTED BEHAVIOUR BASED ON
POLICY USE AND MISUSE OF
BIBLIOMETRIC DATA
W. Glänzel and K. Debackere
Measuring communication in science
38
7.
Distorted behaviour
Distorted behaviour based on policy use and misuse of
bibliometric data
An additional issue concerns the changes in the publication,
citation and collaboration behaviour of scientists (both
positive and negative) that the consistent policy use of
bibliometric indicators might potentially induce.
Studies on the problem choice behaviour of academic
scientists have revealed that both cognitive and social
influences determine the manner in which scientists go about
choosing the problems they work on (Debackere and Rappa
1994). Hence the issue should be raised to what extent the
policy use of bibliometrics might or could affect this behaviour.
W. Glänzel and K. Debackere
Measuring communication in science
39
7.
Distorted behaviour
The problem of inappropriate use ranges from uninformed use,
over selecting and collecting ‘most advantageous’ indicators to
the obvious and deliberate misuse of data.
Uninformed use and misuse are not always beyond the
responsibility of bibliometricians. Unfortunately, bibliometricians
do not always resist the temptation to follow popular, even
populist, trends in order to meet the expectations of the
customers.
Clearly, any kind of uninformed use or misuse of bibliometric
results involves the danger of bringing bibliometric research
itself into disrepute.
W. Glänzel and K. Debackere
Measuring communication in science
40
7.
Distorted behaviour
Uninformed use
• incorrect presentation, interpretation of bibliometric
indicators or their use in an inappropriate context caused
by insufficient knowledge of methodology, background and
data sources
• generalisation (induction) of special cases or of results
obtained at lower levels of aggregation
Misuse
• intentionally incorrect presentation, interpretation of
bibliometric indicators or their deliberate use in
inappropriate context
• tendentious application of biases
• tendentious choice of (incompatible) indicators
W. Glänzel and K. Debackere
Measuring communication in science
41
7.
Distorted behaviour
But even correct use might have undesired consequences.
Example: Re-interpreting underlying contexts such as the notion
of citation (cf. Section 4) shows author self-citations in an
unfavourable light. Authors might thus be urged avoiding selfcitations – a clear intervention into the mechanism of scientific
communication.
Less obvious repercussions might be observed when bibliometric
tools are used in decision-making in science policy and research
management and the scientific community recognises the
feedback in terms of their funding.
Butler (2004) has shown on the example of Australia what might
happen when funding is linked to publication counts. She found
that the publications component of the Composite Index has
stimulated an increased publication activity in the lower impact
journals.
W. Glänzel and K. Debackere
Measuring communication in science
42
7.
Distorted behaviour
Schematic visualisation of the feedback of policy use of
bibliometrics on the scientific community
Bibliometrics
Scientific community
Publication & citation
behaviour
Science policy &
Research management
Funding &
Promotion
W. Glänzel and K. Debackere
Measuring communication in science
43
7.
Distorted behaviour
Possible positive effects
Scientists might recognise that scientific collaboration and
publishing in high-impact or even top journals pays. Also their
publication activity might be stimulated.
Possible negative effects
Exaggerated collaboration, even trends towards hyperauthorship, inflating publication output by splitting up
publications to sequences, inflating citation impact by selfcitations and forming citation cliques, etc. Trend towards
replacing quality and recognition by visibility at any price or
towards preferring journals as publication channels in social
sciences and humanities might be among these effects.
W. Glänzel and K. Debackere
Measuring communication in science
44
CONCLUSIONS
W. Glänzel and K. Debackere
Measuring communication in science
45
Conclusions
• The future will show in how far these negative effects will
become reality. Empirical monitoring and examination of
hypothetical biases will be worthwhile.
• Similar trends could already be observed far before the time of
bibliometrics:
Striving after visibility and reputation is part of human nature.
Most negative effects will probably be hindered or prevented
through the natural competition and peer review among
researchers.
• The only negative feedback from policy use and misuse of
bibliometric data might on the long run results in general
‘inflationary values’ described, e.g., by Cronin (2001) and
Persson et al. (2003). Bibliometricians have the tools to
normalise and standardise indicators under such conditions,
and are thus able to cope with this problem, too.
W. Glänzel and K. Debackere
Measuring communication in science
46