Five-year impact factor
   HOME

TheInfoList



OR:

The impact factor (IF) or journal impact factor (JIF) of an
academic journal An academic journal or scholarly journal is a periodical publication in which scholarship relating to a particular academic discipline is published. Academic journals serve as permanent and transparent forums for the presentation, scrutiny, and ...
is a
scientometric Scientometrics is the field of study which concerns itself with measuring and analysing scholarly literature. Scientometrics is a sub-field of informetrics. Major research issues include the measurement of the impact of research papers and academi ...
index calculated by
Clarivate Clarivate Plc is a British-American publicly traded analytics company that operates a collection of subscription-based services, in the areas of bibliometrics and scientometrics; business / market intelligence, and competitive profiling for ph ...
that reflects the yearly mean number of citations of articles published in the last two years in a given journal, as indexed by Clarivate's Web of Science. As a journal-level metric, it is frequently used as a
proxy Proxy may refer to: * Proxy or agent (law), a substitute authorized to act for another entity or a document which authorizes the agent so to act * Proxy (climate), a measured variable used to infer the value of a variable of interest in climate ...
for the relative importance of a journal within its field; journals with higher impact factor values are given the status of being more important, or carry more prestige in their respective fields, than those with lower values. While frequently used by universities and funding bodies to decide on promotion and research proposals, it has come under attack for distorting good scientific practices.


History

The impact factor was devised by
Eugene Garfield Eugene Eli Garfield (September 16, 1925 – February 26, 2017) was an American linguist and businessman, one of the founders of bibliometrics and scientometrics. He helped to create ''Current Contents'', ''Science Citation Index'' (SCI), ''Journ ...
, the founder of the
Institute for Scientific Information The Institute for Scientific Information (ISI) was an academic publishing service, founded by Eugene Garfield in Philadelphia in 1956. ISI offered scientometric and bibliographic database services. Its specialty was citation indexing and analysi ...
(ISI) in Philadelphia. Impact factors began to be calculated yearly starting from 1975 for journals listed in the ''
Journal Citation Reports ''Journal Citation Reports'' (''JCR'') is an annual publicationby Clarivate Analytics (previously the intellectual property of Thomson Reuters). It has been integrated with the Web of Science and is accessed from the Web of Science-Core Collec ...
'' (JCR). ISI was acquired by Thomson Scientific & Healthcare in 1992, and became known as Thomson ISI. In 2018,
Thomson-Reuters Thomson Reuters Corporation ( ) is a Canadian multinational corporation, multinational media conglomerate. The company was founded in Toronto, Ontario, Canada, where it is headquartered at the Bay Adelaide Centre. Thomson Reuters was crea ...
spun off and sold ISI to Onex Corporation and
Baring Private Equity Asia BPEA EQT (formerly known as Baring Private Equity Asia and BPEA) is an Asian private equity firm. Previously it was an affiliate of Barings Bank before becoming an independent firm. In 2022 it was acquired by EQT Partners to act as its Asian inve ...
. They founded a new corporation, Clarivate, which is now the publisher of the JCR.


Calculation

In any given year, the two-year journal impact factor is the ratio between the number of citations received in that year for publications in that journal that were published in the two preceding years and the total number of "citable items" published in that journal during the two preceding years: \text_y = \frac. For example, ''Nature'' had an impact factor of 41.577 in 2017: \text_ = \frac = \frac = 41.577. This means that, on average, its papers published in 2015 and 2016 received roughly 42 citations each in 2017. Note that 2017 impact factors are reported in 2018; they cannot be calculated until all of the 2017 publications have been processed by the indexing agency. The value of impact factor depends on how to define "citations" and "publications"; the latter are often referred to as "citable items". In current practice, both "citations" and "publications" are defined exclusively by ISI as follows. "Publications" are items that are classed as "article", "review" or "proceedings paper" in the Web of Science (WoS) database; other items like editorials, corrections, notes, retractions and discussions are excluded. WoS is accessible to all registered users, who can independently verify the number of citable items for a given journal. In contrast, the number of citations is extracted not from the WoS database, but from a dedicated JCR database, which is not accessible to general readers. Hence, the commonly used "JCR Impact Factor" is a proprietary value, which is defined and calculated by ISI and can not be verified by external users. New journals, which are indexed from their first published issue, will receive an impact factor after two years of indexing; in this case, the citations to the year prior to volume 1, and the number of articles published in the year prior to volume 1, are known zero values. Journals that are indexed starting with a volume other than the first volume will not get an impact factor until they have been indexed for three years. Occasionally, ''Journal Citation Reports'' assigns an impact factor to new journals with less than two years of indexing, based on partial citation data. The calculation always uses two complete and known years of item counts, but for new titles one of the known counts is zero. Annuals and other irregular publications sometimes publish no items in a particular year, affecting the count. The impact factor relates to a specific time period; it is possible to calculate it for any desired period. For example, the JCR also includes a five-year impact factor, which is calculated by dividing the number of citations to the journal in a given year by the number of articles published in that journal in the previous five years.


Use

While originally invented as a tool to help university librarians to decide which journals to purchase, the impact factor soon became used as a measure for judging academic success. This use of impact factors was summarised by Hoeffel in 1998:
Impact Factor is not a perfect tool to measure the quality of articles but there is nothing better and it has the advantage of already being in existence and is, therefore, a good technique for scientific evaluation. Experience has shown that in each specialty the best journals are those in which it is most difficult to have an article accepted, and these are the journals that have a high impact factor. Most of these journals existed long before the impact factor was devised. The use of impact factor as a measure of quality is widespread because it fits well with the opinion we have in each field of the best journals in our specialty....In conclusion, prestigious journals publish papers of high level. Therefore, their impact factor is high, and not the contrary.
As impact factors are a journal-level metric, rather than an article- or individual-level metric, this use is controversial. Eugene Garfield, the inventor of the JIF agreed with Hoeffel, but warned about the "misuse in evaluating individuals" because there is "a wide variation f citationsfrom article to article within a single journal". Despite this warning, the use of the JIF has evolved, playing a key role in the process of assessing individual researchers, their job applications and their funding proposals. In 2005, ''The Journal of Cell Biology'' noted that:
Impact factor data ... have a strong influence on the scientific community, affecting decisions on where to publish, whom to promote or hire, the success of grant applications, and even salary bonuses.
More targeted research has begun to provide firm evidence of how deeply the impact factor is embedded within formal and informal research assessment processes. A review in 2019 studied how often the JIF featured in documents related to the review, promotion, and tenure of scientists in US and Canadian universities. It concluded that 40% of universities focussed on academic research specifically mentioned the JIF as part of such review, promotion, and tenure processes. And a 2017 study of how researchers in the life sciences behave concluded that "everyday decision-making practices as highly governed by pressures to publish in high-impact journals". The deeply embedded nature of such indicators not only effect research assessment, but the more fundamental issue of what research is actually undertaken: "Given the current ways of evaluation and valuing research, risky, lengthy, and unorthodox project rarely take center stage."


Criticism

Numerous critiques have been made regarding the use of impact factors, both in terms of its statistical validity and also of its implications for how science is carried out and assessed. A 2007 study noted that the most fundamental flaw is that impact factors present the
mean There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value (magnitude and sign) of a given data set. For a data set, the '' ari ...
of data that are not normally distributed, and suggested that it would be more appropriate to present the median of these data. There is also a more general debate on the validity of the impact factor as a measure of journal importance and the effect of policies that editors may adopt to boost their impact factor (perhaps to the detriment of readers and writers). Other criticism focuses on the effect of the impact factor on behavior of scholars, editors and other stakeholders. Others have made more general criticisms, arguing that emphasis on impact factor results from the negative influence of neoliberal politics on academia. These more politicised arguments demand not just replacement of the impact factor with more sophisticated metrics but also discussion on the social value of research assessment and the growing precariousness of scientific careers in higher education.


Inapplicability of impact factor to individuals and between-discipline differences

It has been stated that impact factors in particular and citation analysis in general are affected by field-dependent factors which invalidate comparisons not only across disciplines but even within different fields of research of one discipline. The percentage of total citations occurring in the first two years after publication also varies highly among disciplines from 1–3% in the mathematical and physical sciences to 5–8% in the biological sciences. Thus impact factors cannot be used to compare journals across disciplines. Impact factors are sometimes used to evaluate not only the journals but the papers therein, thereby devaluing papers in certain subjects. In 2004, the Higher Education Funding Council for England was urged by the
House of Commons The House of Commons is the name for the elected lower house of the bicameral parliaments of the United Kingdom and Canada. In both of these countries, the Commons holds much more legislative power than the nominally upper house of parliament. T ...
Science and Technology Select Committee The Science and Technology Select Committee is a select committee of the House of Commons in the Parliament of the United Kingdom. The original Science and Technology Committee was abolished upon the creation of the Innovation, Universities, Sci ...
to remind
Research Assessment Exercise The Research Assessment Exercise (RAE) was an exercise undertaken approximately every five years on behalf of the four UK higher education funding councils (HEFCE, SHEFC, HEFCW, DELNI) to evaluate the quality of research undertaken by British hig ...
panels that they are obliged to assess the quality of the content of individual articles, not the reputation of the journal in which they are published. Other studies have repeatedly stated that impact factor is a metric for journals and should not be used to assess individual researchers or institutions.


Questionable editorial policies that affect the impact factor

Because impact factor is commonly accepted as a proxy for research quality, some journals adopt editorial policies and practices, some acceptable and some of dubious purpose, to increase its impact factor. For example, journals may publish a larger percentage of
review article A review article is an article that summarizes the current state of understanding on a topic within a certain discipline. A review article is generally considered a secondary source since it may analyze and discuss the method and conclusions i ...
s which generally are cited more than research reports. Research undertaken in 2020 on dentistry journals concluded that the publication of "systematic reviews have significant effect on the Journal Impact Factor ... while papers publishing clinical trials bear no influence on this factor. Greater yearly average of published papers ... means a higher impact factor." Journals may also attempt to limit the number of "citable items"—i.e., the denominator of the impact factor equation—either by declining to publish articles that are unlikely to be cited (such as case reports in medical journals) or by altering articles (e.g., by not allowing an abstract or bibliography in hopes that Journal Citation Reports will not deem it a "citable item"). As a result of negotiations over whether items are "citable", impact factor variations of more than 300% have been observed. Items considered to be uncitable—and thus are not incorporated in impact factor calculations—can, if cited, still enter into the numerator part of the equation despite the ease with which such citations could be excluded. This effect is hard to evaluate, for the distinction between editorial comment and short original articles is not always obvious. For example, letters to the editor may be part of either class. Another less insidious tactic journals employ is to publish a large portion of its papers, or at least the papers expected to be highly cited, early in the calendar year. This gives those papers more time to gather citations. Several methods, not necessarily with nefarious intent, exist for a journal to cite articles in the same journal which will increase the journal's impact factor. Beyond editorial policies that may skew the impact factor, journals can take overt steps to
game the system Gaming the system (also rigging, abusing, cheating, milking, playing, working, or breaking the system, or gaming or bending the rules) can be defined as using the rules and procedures meant to protect a system to, instead, manipulate the system ...
. For example, in 2007, the specialist journal ''
Folia Phoniatrica et Logopaedica Karger Publishers (also: Karger Medical and Scientific Publishers, S. Karger AG) is an academic publisher of scientific and medical journals and books. The current CEO is Daniel Ebneter. History The company was founded in 1890 in Berlin by Sam ...
'', with an impact factor of 0.66, published an editorial that cited all its articles from 2005 to 2006 in a protest against the "absurd scientific situation in some countries" related to use of the impact factor. The large number of citations meant that the impact factor for that journal increased to 1.44. As a result of the increase, the journal was not included in the 2008 and 2009 ''Journal Citation Reports''.
Coercive citation Coercive citation is an academic publishing practice in which an editor of a scientific or academic journal forces an author to add spurious citations to an article before the journal will agree to publish it. This is done to inflate the journal' ...
is a practice in which an editor forces an author to add extraneous citations to an article before the journal will agree to publish it, in order to inflate the journal's impact factor. A survey published in 2012 indicates that coercive citation has been experienced by one in five researchers working in economics, sociology, psychology, and multiple business disciplines, and it is more common in business and in journals with a lower impact factor. Editors of leading business journals banded together to disavow the practice. However, cases of coercive citation have occasionally been reported for other disciplines.


Assumed correlation between impact factor and quality

The journal impact factor was originally designed by Eugene Garfield as a metric to help librarians make decisions about which journals were worth indexing, as the JIF aggregates the number of citations to articles published in each journal. Since then, the JIF has become associated as a mark of journal "quality", and gained widespread use for evaluation of research and researchers instead, even at the institutional level. It thus has significant impact on steering research practices and behaviours. By 2010, national and international research funding institutions were already starting to point out that numerical indicators such as the JIF should not be considered as a measure of quality. In fact, research was indicating that the JIF is a highly manipulated metric, and the justification for its continued widespread use beyond its original narrow purpose seems due to its simplicity (easily calculable and comparable number), rather than any actual relationship to research quality. Empirical evidence shows that the misuse of the JIF—and journal ranking metrics in general—has a number of negative consequences for the scholarly communication system. These include gaps between the reach of a journal and the quality of its individual papers and insufficient coverage of social sciences and humanities as well as research outputs from across Latin America, Africa, and South-East Asia. Additional drawbacks include the marginalization of research in
vernacular language A vernacular or vernacular language is in contrast with a "standard language". It refers to the language or dialect that is spoken by people that are inhabiting a particular country or region. The vernacular is typically the native language, n ...
s and on locally relevant topics and inducement to unethical authorship and citation practices. More generally, the impact factors fosters a reputation economy, where scientific success is based on publishing in prestigious journals ahead of actual research qualities such as rigorous methods, replicability and social impact. Using journal prestige and the JIF to cultivate a competition regime in academia has been shown to have deleterious effects on research quality. A number of regional and international initiatives are now providing and suggesting alternative research assessment systems, including key documents such as the
Leiden Manifesto The Leiden Manifesto for research metrics is a list of "ten principles to guide research evaluation", published as a comment in Volume 520, Issue 7548 of ''Nature'', on 22 April 2015. It was formulated by public policy professor Diana Hicks, scie ...
and the
San Francisco Declaration on Research Assessment The San Francisco Declaration on Research Assessment (DORA) intends to halt the practice of correlating the journal impact factor to the merits of a specific scientist's contributions. Also according to this statement, this practice creates bias ...
(DORA).
Plan S Plan S is an initiative for open-access science publishing launched in 2018 by "cOAlition S", a consortium of national research agencies and funders from twelve European countries. The plan requires scientists and researchers who benefit from s ...
calls for a broader adoption and implementation of such initiatives alongside fundamental changes in the scholarly communication system. As appropriate measures of quality for authors and research, concepts of research excellence should be remodelled around transparent workflows and accessible research results. JIFs are still regularly used to evaluate research in many countries which is a problem since a number of issues remain around the opacity of the metric and the fact that it is often negotiated by publishers.


Negotiated values

Results of an impact factor can change dramatically depending on which items are considered as "citable" and therefore included in the denominator. One notorious example of this occurred in 1988 when it was decided that meeting abstracts published in ''FASEB Journal'' would no longer be included in the denominator. The journal's impact factor jumped from 0.24 in 1988 to 18.3 in 1989. Publishers routinely discuss with Clarivate how to improve the "accuracy" of their journals' impact factor and therefore get higher scores. Such discussions routinely produce "negotiated values" which result in dramatic changes in the observed scores for dozens of journals, sometimes after unrelated events like the purchase by one of the larger publishers.


Distribution skewness

Because citation counts have highly
skewed distribution In probability theory and statistics, skewness is a measure of the asymmetry of the probability distribution of a real-valued random variable about its mean. The skewness value can be positive, zero, negative, or undefined. For a unimoda ...
s, the mean number of citations is potentially misleading if used to gauge the typical impact of articles in the journal rather than the overall impact of the journal itself. For example, about 90% of ''
Nature Nature, in the broadest sense, is the physical world or universe. "Nature" can refer to the phenomena of the physical world, and also to life in general. The study of nature is a large, if not the only, part of science. Although humans are ...
''s 2004 impact factor was based on only a quarter of its publications. Thus the actual number of citations for a single article in the journal is in most cases much lower than the mean number of citations across articles. Furthermore, the strength of the relationship between impact factors of journals and the citation rates of the papers therein has been steadily decreasing since articles began to be available digitally. The effect of outliers can be seen in the case of the article "A short history of SHELX", which included this sentence: "This paper could serve as a general literature citation when one or more of the open-source SHELX programs (and the Bruker AXS version SHELXTL) are employed in the course of a crystal-structure determination". This article received more than 6,600 citations. As a consequence, the impact factor of the journal ''
Acta Crystallographica ''Acta Crystallographica'' is a series of peer-reviewed scientific journals, with articles centred on crystallography, published by the International Union of Crystallography (IUCr). Originally established in 1948 as a single journal called ''Act ...
Section A'' rose from 2.051 in 2008 to 49.926 in 2009, more than ''
Nature Nature, in the broadest sense, is the physical world or universe. "Nature" can refer to the phenomena of the physical world, and also to life in general. The study of nature is a large, if not the only, part of science. Although humans are ...
'' (at 31.434) and ''
Science Science is a systematic endeavor that Scientific method, builds and organizes knowledge in the form of Testability, testable explanations and predictions about the universe. Science may be as old as the human species, and some of the earli ...
'' (at 28.103). The second-most cited article in ''Acta Crystallographica Section A'' in 2008 only had 28 citations. Critics of the JIF state that use of the arithmetic mean in its calculation is problematic because the pattern of citation distribution is skewed and citation distributions metrics have been proposed as an alternative to impact factors. However, there have also been pleas to take a more nuanced approach to judging the distribution skewness of the impact factor. Waltman and Traag, in their 2021 paper, ran numerous simulations and concluded that "statistical objections against the use of the IF at the level of individual articles are not convincing", and that "the IF may be a more accurate indicator of the value of an article than the number of citations of the article".


Lack of reproducibility

While the underlying mathematical model is publicly known, the dataset which is used to calculate the JIF is not publicly available. This prompted criticism: "Just as scientists would not accept the findings in a scientific paper without seeing the primary data, so should they not rely on Thomson Scientific's impact factor, which is based on hidden data". However, a 2019 article demonstrated that "with access to the data and careful cleaning, the JIF can be reproduced", although this required much labour to achieve. A 2020 research paper went further. It indicated that by querying open access or partly open-access databases, like Google Scholar, ResearchGate, and Scopus, it is possible to calculate approximate impact factors without the need to purchase Web of Science / JCR.


Broader negative impact on science

Just as the impact factor has attracted criticism for various immediate problems associated with its application, so has there also been criticism that its application undermines the broader process of science. Research has indicated that bibliometrics figures, particularly the impact factor, decrease the quality of peer review an article receiving, a reluctance to share data, decreasing quality of articles, and a reduced scope in terms of what they can research. "For many researchers the only research questions and projects that appear viable are those that can meet the demand of scoring well in terms of metric performance indicators - and chiefly the journal impact factor.". Furthermore, the process of publication and science is slowed down - authors automatically try and publish with the journals with the highest impact factor - "as editors and reviewers are tasked with reviewing papers that are not submitted to the most appropriate venues."


Institutional responses to criticism of the impact factor

Given the growing criticism and its widespread usage as a means of research assessment, organisations and institutions have begun to take steps to move away from the journal impact factor. In November 2007 the
European Association of Science Editors The European Association of Science Editors (EASE ) is a non-profit membership organisation for people interested in science communication and editing. Founded in 1982, in France, EASE now has an international membership. Members EASE has nearly ...
(EASE) issued an official statement recommending "that journal impact factors are used only—and cautiously—for measuring and comparing the influence of entire journals, but not for the assessment of single papers, and certainly not for the assessment of researchers or research programmes". In July 2008, the International Council for Science Committee on Freedom and Responsibility in the Conduct of Science issued a "statement on publication practices and indices and the role of
peer review Peer review is the evaluation of work by one or more people with similar competencies as the producers of the work ( peers). It functions as a form of self-regulation by qualified members of a profession within the relevant field. Peer revie ...
in research assessment", suggesting many possible solutions—e.g., considering a limit number of publications per year to be taken into consideration for each scientist, or even penalising scientists for an excessive number of publications per year—e.g., more than 20. In February 2010, the
Deutsche Forschungsgemeinschaft The German Research Foundation (german: Deutsche Forschungsgemeinschaft ; DFG ) is a German research funding organization, which functions as a self-governing institution for the promotion of science and research in the Federal Republic of Germ ...
(German Research Foundation) published new guidelines to reduce the number of publications could submit when applying for funding: "The focus has not been on what research someone has done but rather how many papers have been published and where." They noted that for decisions concerning "performance-based funding allocations, postdoctoral qualifications, appointments, or reviewing funding proposals,
here Here is an adverb that means "in, on, or at this place". It may also refer to: Software * Here Technologies, a mapping company * Here WeGo (formerly Here Maps), a mobile app and map website by Here Technologies, Here Television * Here TV (form ...
increasing importance has been given to numerical indicators such as the
h-index The ''h''-index is an author-level metric that measures both the productivity and citation impact of the publications, initially used for an individual scientist or scholar. The ''h''-index correlates with obvious success indicators such as ...
and the impact factor". The UK's
Research Assessment Exercise The Research Assessment Exercise (RAE) was an exercise undertaken approximately every five years on behalf of the four UK higher education funding councils (HEFCE, SHEFC, HEFCW, DELNI) to evaluate the quality of research undertaken by British hig ...
for 2014 also banned the journal impact factor although evidence suggested that this ban was often ignored. In response to growing concerns over the inappropriate use of journal impact factors in evaluating scientific outputs and scientists themselves, the
American Society for Cell Biology The American Society for Cell Biology (ASCB) is a professional society that was founded in 1960.San Francisco Declaration on Research Assessment The San Francisco Declaration on Research Assessment (DORA) intends to halt the practice of correlating the journal impact factor to the merits of a specific scientist's contributions. Also according to this statement, this practice creates bias ...
(DORA). Released in May 2013, DORA has garnered support from thousands of individuals and hundreds of institutions, including in March 2015 the League of European Research Universities (a consortium of 21 of the most renowned research universities in Europe), who have endorsed the document on the DORA website. Publishers, even those with high impact factors, also recognised the flaws. ''Nature'' magazine criticised the over reliance of JIF, pointing not just to its statistical but to negative effects on science: "The resulting pressures and disappointments are nothing but demoralizing, and in badly run labs can encourage sloppy research that, for example, fails to test assumptions thoroughly or to take all the data into account before submitting big claims." Various publishers now use a mixture of metrics on their website; the PLOS series of journals does not display the impact factor. Microsoft Academic took a similar view, stating that h-index, EI/SCI and journal impact factors are not shown because "the research literature has provided abundant evidence that these metrics are at best a rough approximation of research impact and scholarly influence." In 2021,
Utrecht University Utrecht University (UU; nl, Universiteit Utrecht, formerly ''Rijksuniversiteit Utrecht'') is a public research university in Utrecht, Netherlands. Established , it is one of the oldest universities in the Netherlands. In 2018, it had an enrollme ...
promised to abandon all quantitative bibliometrics, including the impact factor. The university stated that "it has become a very sick model that goes beyond what is really relevant for science and putting science forward." This followed a 2018 decision by the main Dutch funding body for research, NWO, to remove all references to journal impact factors and the h-index in all call texts and application forms. Utrecht's decision met with some resistance. An open letter signed by over 150 Dutch academics argued that while imperfect, the JIF is still useful, and that omitting it "will lead to randomness and a compromising of scientific quality."


Closely related indices

Some related metrics, also calculated and published by the same organization, include: * ''Cited half-life'': the median age of the articles that were cited in ''Journal Citation Reports'' each year. For example, if a journal's half-life in 2005 is 5, that means the citations from 2001 to 2005 are half of all the citations from that journal in 2005, and the other half of the citations precede 2001. * ''Aggregate impact factor'' for a subject category: it is calculated taking into account the number of citations to all journals in the subject category and the number of articles from all the journals in the subject category. * ''Immediacy index'': the number of citations the articles in a journal receive in a given year divided by the number of articles published. * '' Journal citation indicator'' (''JCI''): a JIF that adjusts for scientific field; it is similar to Source Normalized Impact per Paper, calculated based on the
Scopus Scopus is Elsevier's abstract and citation database launched in 2004. Scopus covers nearly 36,377 titles (22,794 active titles and 13,583 inactive titles) from approximately 11,678 publishers, of which 34,346 are peer-reviewed journals in top-l ...
database. As with the impact factor, there are some nuances to this: for example, Clarivate excludes certain article types (such as news items, correspondence, and errata) from the denominator.


Other measures of scientific impact

Additional
journal-level metrics Journal ranking is widely used in academic circles in the evaluation of an academic journal's impact and quality. Journal rankings are intended to reflect the place of a journal within its field, the relative difficulty of being published in that ...
are available from other organizations. For example, '' CiteScore'' is a metric for serial titles in
Scopus Scopus is Elsevier's abstract and citation database launched in 2004. Scopus covers nearly 36,377 titles (22,794 active titles and 13,583 inactive titles) from approximately 11,678 publishers, of which 34,346 are peer-reviewed journals in top-l ...
launched in December 2016 by
Elsevier Elsevier () is a Dutch academic publishing company specializing in scientific, technical, and medical content. Its products include journals such as '' The Lancet'', ''Cell'', the ScienceDirect collection of electronic journals, '' Trends'', ...
. While these metrics apply only to journals, there are also
author-level metrics Author-level metrics are citation metrics that measure the bibliometric impact of individual authors, researchers, academics, and scholars. Many metrics have been developed that take into account varying numbers of factors (from only considering t ...
, such as the
h-index The ''h''-index is an author-level metric that measures both the productivity and citation impact of the publications, initially used for an individual scientist or scholar. The ''h''-index correlates with obvious success indicators such as ...
, that apply to individual researchers. In addition,
article-level metrics Article-level metrics are citation metrics which measure the usage and impact of individual scholarly articles. Adoption Traditionally, bibliometrics have been used to evaluate the usage and impact of research, but have usually been focused on j ...
measure impact at an article level instead of journal level. Other more general alternative metrics, or "
altmetrics In scholarly and scientific publishing, altmetrics are non-traditional bibliometrics proposed as an alternative or complement to more traditional citation impact metrics, such as impact factor and ''h''-index. The term altmetrics was proposed i ...
", that include article views, downloads, or mentions in
social media Social media are interactive media technologies that facilitate the creation and sharing of information, ideas, interests, and other forms of expression through virtual communities and networks. While challenges to the definition of ''social medi ...
, offer a different perspective on research impact, concentrating more on immediate social impact in and outside academia.


Counterfeit impact factors

Fake impact factors or bogus impact factors are produced by certain companies or individuals. According to an article published in the ''Electronic Physician'', these include Global Impact Factor, Citefactor, and Universal Impact Factor.
Jeffrey Beall Jeffrey Beall is an American librarian and library scientist, best known for drawing attention to " predatory open access publishing", a term he coined, and for creating what is now widely known as Beall's list, a list of potentially predatory ...
maintained a list of such misleading metrics. Another deceitful practice is reporting "alternative impact factors", calculated as the average number of citations per article using
citation indices A citation index is a kind of bibliographic index, an index of citations between publications, allowing the user to easily establish which later documents cite which earlier documents. A form of citation index is first found in 12th-century Hebre ...
other than JCR, even if based on reputable sources such as
Google Scholar Google Scholar is a freely accessible web search engine that indexes the full text or metadata of scholarly literature across an array of publishing formats and disciplines. Released in beta in November 2004, the Google Scholar index includes ...
(e.g., "Google-based Journal Impact Factor"). False impact factors are often used by predatory publishers. Consulting Journal Citation Reports' master journal list can confirm if a publication is indexed by the ''Journal Citation Reports''. The use of fake impact metrics is considered a red flag.


See also

* Author impact factor *
Citation impact Citation impact is a measure of how many times an academic journal article or book or author is cited by other articles, books or authors. Citation counts are interpreted as measures of the impact or influence of academic work and have given ris ...
* Goodhart's law *
Journalology Journalology (also known as publication science) is the scholarly study of all aspects of the academic publishing process. The field seeks to improve the quality of scholarly research by implementing evidence-based practices in academic publishing. ...


Notes on alternatives


References


Further reading

* * * * * * * {{DEFAULTSORT:Impact factor Journal ranking