- Open Access
The meaning of the 'impact factor' in the case of an open-access journal
© Alexandrov; licensee BioMed Central Ltd. 2011
Received: 25 November 2010
Accepted: 2 March 2011
Published: 2 March 2011
The dominant model of journal evaluation emerged at the time when there were no open-access journals, and nobody has assessed yet whether this model is able to cope with this modern reality. This commentary attempts to fill the gaps in the common understanding of the role that 'impact factor' should play in evaluation of open-access journals.
Fifty years ago, when computers were a rarity, librarians used index cards to keep records of library holdings. The cards were arranged into an author index and a subject index. To find publications related to the subject of his or her research, a researcher went to the subject index, drew the box with the cards on the subject, look at the long row of cards, and did not know where to start. Since the cards were ordered alphabetically, the seminal works on the subject were lost in the mass of less important publications.
Pointing out that subject indexes were "limited in their attempt to provide an ideal key to the literature", Eugene Garfield proposed, in 1955, a new bibliographic tool -- a "citation index". The underlying idea of this tool may be formulated as follows. A seminal work could be an ideal key to the literature on the subject to which it gives the birth, if all the references to this work were to be listed on its card (see details in Appendix).
In 1960, Garfield founded the Institute for Scientific Information (ISI), and started a bibliographic indexing service-the Scientific Citation Index, which has been continued since 1992 by the Thomson Reuters Corporation.
In 1975, ISI began to publish Journal Citation Reports (JCR), which became later "the recognized authority for evaluating journals" . The logic of journal evaluation was developed by Garfield  at the time when there were no open-access journals, and nobody has assessed yet whether the logic of counting citations is able to cope with this modern reality.
If a library is subscribed for two journals with roughly the same title, and cannot continue subscription to both of them, then the journal impact factor has a certain meaning for the librarian: it gives an argument for choosing the journal that should be removed from the library collection.
If the subject of an article falls within the scope of two journals, then the journal impact factor has a certain meaning for the article's author: it gives an argument for choosing the journal where to submit the article.
If the impact of a journal is higher than those of other journals in the same field of science, then the journal impact factor has a certain meaning for the journal's editor: it gives an argument for convincing a highly-cited scientist to give an article for the journal.
If the impact of a journal is higher than those of other journals in the same field of science, then the journal impact factor has a certain meaning for the journal's publisher: it gives an argument for convincing a librarian to subscribe to the journal.
Should the meaning of the impact factor remain the same in the case of an open-access journal? Obviously, the impact factor of an open-access journal should not have a practical meaning for a librarian, as the libraries need not subscribe to the open-access journals. As to the other examples, the answer is not so obvious.
Open-access journals have a number of features that make them different from the subscription-based journals. The most important of them is that the readership of an open-access journal is not restricted by subscription. This is a huge advantage in sense of dissemination of research results: each article may receive as wide readership as it deserves. It does not matter whether the journal impact factor is high or low: in any case, the articles published in an open-access journal are not less available for a broad readership than the articles published in Science or Nature.
Another feature is that open-access journals are integrated to the information environment that is formed by search engines, social networks and other tools for harnessing internet power for quick "association of ideas". All these tools provide direct access to the articles. In this information environment, we read articles, not journals. To read an article, we need not open a journal and read the contents. We even need not read the journal title. Thus, again, it does not matter whether the journal impact factor is high or low: a search engine will find the article, and its position in the search list will depend on its relevance to the search query.
In other words, the impact factor of an open-access journal should not have a significant meaning for an author, if the author strives for dissemination of research results.
It would be credulous to believe, however, that dissemination of research results is the only purpose of a scientific publication. An article published in a peer-reviewed journal is a sign of scholarly recognition. Moreover, a publication in a high-rank journal is not merely a sign of recognition -- it is a sign of distinction. Hence, the rank of the journal, whatever it means, is important. Having no rank, received from JCR (or other relevant "authority"), an open-access journal would not attract the author who need such a sign of distinction as a publication in a high-rank journal.
Shall open-access journals compete for the authors who seek distinction? If we are all agree that subscription-based journals should not be extinct, it would be prudent to reserve them for distinctive authors. Open-access journals have been introduced under the flag of research dissemination, and may thrive under this flag.
Of course, we need some metrics to assess usability of a journal for disseminating research results. Bibliographic metrics are not very suitable for this purpose, as the number of citations received by an article depends on the quality of the article. It would be better to develop the metrics that would be less sensitive to the bibliographic quality of a journal and more sensitive to its quality as a medium for conveying research results. Since the websites of the journals published by BioMed Central look like web-portals to corresponding fields of science, it might be reasonable to use the metrics that are proposed for evaluation of web-portals .
The open-access business model has been introduced for solving the problem that could not be solved by means of the subscription-based business model -- that is, the problem of how to give every research article as wide readership as it deserves.
The value of an open-access journal as a tool for communicating research results may be more important than the bibliographic value of the journal as a collection of research articles.
The use of the journal impact factor for evaluating an open-access journal may obscure the role of the journal as an information and communication tool.
The journal impact factor should not play a crucial role in evaluation of open-access journals, if we all agree that the open-access journals are not to replace the subscription-based journals.
It is worth noting here that CBM hopes to start reporting its impact factor in the future like many other journals published by BioMed Central. Meanwhile, authors and readers may retrieve bibliographic indicators of CBM impact from Scopus Journal Analyzer , and related bibliometric services [6–8]. They also may find at the journal website the statistics of accesses that gives some impression about the size of "non-citing" readership, which is ignored by bibliographic indicators, but which is of certain importance for a journal that publish articles on policy relevant aspects of global carbon cycle.
Appendix. CBM as a starting point to a literature search
Proposing a new bibliographic tool, "citation index", Garfield argued that any comprehensive index to the literature of science may provide only a better starting point for a literature search than the one provided in the selective indexes. He wrote, "Proponents of classified indexes may suggest that classification is the solution to this problem, but this no means the case" . The new bibliographic tool was "to span the gap between the subject approach of those who create documents -- that is, authors -- and the subject approach of the scientist who seeks information".
In the later work , Garfield specified more clearly the major difference between subject indexes and citation indexes. The user of a citation index must know some work associated with the subject of his or her research. "If the user does not know of a previous work on the subject he must find one through a book, an encyclopedia, or a colleague." That is to say, the user should know key works, instead of keywords.
The publications that cite the same key work form a well defined class. They are all somehow associated with the subject of this key work. Besides, the publications citing these publications also belong to this class. Hence, quite a large bibliography may be created starting from a single key work.
Additional file 1:A demonstration of how CBM can be used as a starting point to a literature search. The file can be viewed by using Windows Media Player. (WMV 11 MB)
Some articles published in CBM received more than 30 first-order citations. They may be used to generate bibliographies consisting of hundreds of entries. However, an extensive bibliography is not the purpose of every literature search. More than often, a literature search is to find a dozen of the most recent articles that somehow correspond to the completed research and hence could be cited to demonstrate the novelty and significance of the research result.
In the example considered above (Figure 1), 9 of 13 references to the articles published in the year of 2010 cite the most cited first-order citation. Hence, if the purpose of a literature search is to find a dozen of the most recent references, then 'citation surfing', that is, tracking only the most cited citations would be more reasonable than a comprehensive citation analysis.
Additional file 2:A demonstration of how the names of editorial board members can be used as starting points for 'citation surfing'. The file can be viewed by using Windows Media Player. (WMV 14 MB)
In the earlier works of Garfield [1, 9], citation analysis was promoted as a tool for communicating information. Now, it is promoted as a tool for evaluating information. Citation analysis has potential for both communication and evaluation. The obsession with the impact factor obscures the value of citation analysis as a tool for communicating information.
- Garfield E: Citation indexes for science. A new dimension in documentation through association of ideas (Reprinted from Science, vol 122, pg 108–11, 1955). International Journal of Epidemiology 2006, 35: 1123–1127. 10.1093/ije/dyl189View ArticleGoogle Scholar
- Journal Citation Reports: The recognized authority for evaluating journals. Thomson Reuters. Ref Type: Electronic Citation 2010. [http://wokinfo.com/media/pdf/jcrwebfs.pdf]Google Scholar
- Garfield E: Citation analysis as a tool in journal evaluation. Science 1972, 178: 471–479. 10.1126/science.178.4060.471View ArticleGoogle Scholar
- Sampson D, Manouselis N: A flexible evaluation framework for web portals based on multi-criteria analysis. In Web portals. Edited by: Tatnall A. London: Idea Group; 2005:185–211.View ArticleGoogle Scholar
- Journal AnalyzerElsevier B.V; 2010. [http://www.info.sciverse.com/scopus/scopus-in-detail/tools/journalanalyzer] Ref Type: Electronic Citation
- SNIP & SJR: a new perspective in journal metrics.Elsevier B.V; 2010. [http://info.scopus.com/journalmetrics/snip.html] Ref Type: Electronic CitationGoogle Scholar
- CWTS: CWTS Journal Indicators.University of Leiden; 2010. [http://www.journalindicators.com/] Ref Type: Electronic CitationGoogle Scholar
- SCImago: SJR -- SCImago Journal & Country Rank.SCImago Research Group; 2010. [http://www.scimagojr.com] Ref Type: Electronic CitationGoogle Scholar
- Garfield E: "Science Citation Index" - A New Dimension in Indexing. Science 1964, 144: 649–654. 10.1126/science.144.3619.649View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.