Methodology

The Highly Cited Researchers™ list from Clarivate™ identifies scientists and social scientists who have demonstrated significant influence through publication of multiple highly cited papers during the last decade.

Researchers are selected for their exceptional performance in one or more of 21 fields (those used in the Essential Science Indicators™, or ESI) or across several fields.

arrow_forward Read the detailed methodology

2022 Analysis

Experts from the Institute for Scientific Information™ provide exclusive insight into the list of Highly Cited Researchers 2022, including regional, institutional and field of study breakdowns and much more.

Read analysis of this year’s list.

There are approximately 7,200 Highly Cited Researcher designations in 2022 — some 4,000 in specific fields and about 3,200 for cross-field performance.

This is the fifth year that we have identified researchers with cross-field impact. The recognition of researchers with substantial influence in several fields keeps the list of Highly Cited Researchers contemporary and relevant.

All Highly Cited Researcher records are reviewed. Factors such as retractions, misconduct, and extreme self-citation — all of which would detract from true community-wide research influence — may lead to an author being excluded from the list. This year, we expanded the qualitative analysis of the Highly Cited Researchers list to address increasing concerns over potential misconduct (such as plagiarism, image manipulation, fake peer review). With the assistance of Retraction Watch and its unparalleled database of retractions, Clarivate analysts searched for evidence of misconduct in all publications of those on the preliminary list of Highly Cited Researchers. Researchers found to have committed scientific misconduct in formal proceedings conducted by a researcher’s institution, a government agency, a funder or a publisher are excluded from the list of Highly Cited Researchers.

The number of researchers selected in each field is based on the square root of the population of authors listed on the field’s highly cited papers. The number of those with cross-field influence is determined by finding those who have influence equivalent to those identified in the 21 fields.

For the Highly Cited Researchers 2022 analysis, the papers surveyed were the most recent papers available to us — those published and cited during 2011 to 2021 and which then ranked in the top 1% by citations for their ESI field and year (the definition of a highly cited paper).

The threshold number of highly cited papers for selection differs by field, with Clinical Medicine requiring the most and Pharmacology/Toxicology the least.

A second criterion for selection is the citation count to highly cited papers that ranks a researcher in the top 1% by citations in an ESI field for the period surveyed.

To identify researchers with cross-field influence, highly cited paper and citation counts are normalized through fractionating according to the thresholds required for each field (thus, each Clinical Medicine paper has a smaller unit fraction than one in Pharmacology/Toxicology). Citation counts are fractionated in a similar manner. If the sum of the publication counts and the sum of the citation counts for a researcher equals 1.0 or more, the individual exhibits influence equivalent to a researcher selected in one or more ESI defined fields and is therefore selected as a Highly Cited Researcher for exceptional cross-field performance.

There is no unique or universally agreed concept of what constitutes exceptional research performance and elite status in the sciences and social sciences. Consequently, no quantitative indicators will point to a list that satisfies all expectations or requirements. Moreover, a different basis or formula for selection would generate a different — though likely overlapping — list of names. Thus, the absence of a name on our list cannot be interpreted as inferior performance or stature in comparison to those selected this year.

Although this list is updated and refreshed each year, a Highly Cited Researcher is always a Highly Cited Researcher — whether their name was included in 2014 or 2022.

Methodology

Please read the full methodology with care to understand both the meaning and the inevitable limitations of our analytical approach.

Highly Cited Researchers™ from Clarivate™ is an annual list recognizing influential researchers in the sciences and social sciences from around the world.

The 2022 list contains about 4,000 Highly Cited Researchers in 21 fields of the sciences and social sciences and about 3,200 Highly Cited Researchers identified as having exceptional performance across several fields. The list focuses on contemporary research achievement: only highly cited papers in the sciences and social sciences journals indexed in the Web of Science Core Collection™ during the 11-year period 2011 to 2021 were surveyed. Highly cited papers are defined as those that rank in the top 1% by citations for field and publication year.

The data derived from Essential Science Indicators™ (ESI), a component of InCites™. The fields are also those employed in ESI – 21 broad fields defined by sets of journals and exceptionally, in the case of multidisciplinary journals such as Nature and Science, by a paper-by-paper assignment to a field based on an analysis of the cited references in the papers. This percentile-based selection method removes the citation advantage of older papers relative to recently published ones, since papers are weighed against others in the same annual cohort.

Researchers who, within an ESI-defined field, publish highly cited papers are judged to be influential, so the production of multiple top 1% papers is interpreted as a mark of exceptional impact. Relatively younger researchers are more likely to emerge in such an analysis than in one dependent on total citations over many years. To be able to recognize early and mid-career as well as senior researchers is one of the goals in generating the list of Highly Cited Researchers. The determination of how many researchers to include in the list for each field is based on the population of each field, as represented by the number of disambiguated author names on all highly cited papers in that field, 2011 to 2021.

The ESI fields vary greatly in size, with Clinical Medicine being the largest and Economics and Business being the smallest in terms of researchers. The square root of the number of authors in each field indicated how many individuals should be selected.

Authors with the greatest number of highly cited papers in an ESI field at the threshold for inclusion and above are putative selectees. However, another criterion for selection is that the researcher must have enough citations to their highly cited papers to rank among all authors in the top 1% by total citations in the ESI field in which that person is considered. All who published highly cited papers at the threshold level and with citations in the field at the level of top 1% are admitted to the list, even if the final list then exceeds the number given by the square root calculation.

In addition, and as concession to the somewhat arbitrary cut-off, any researcher with one fewer highly cited paper than the threshold number is also admitted to the list if total citations to their highly cited papers ranks that individual in the top 50% by total citations of those at the threshold level or higher. The justification for this adjustment is that it seems to work well in identifying influential researchers, in the judgment of the Clarivate citation analysts.

Of course, there are many highly accomplished and influential researchers who are not recognized by the method described above and whose names do not appear in the 2022 list. This outcome would hold no matter what specific method were chosen for selection. Each measure or set of indicators, whether total citations, h-index, relative citation impact, mean percentile score, etc., accentuates different types of performance and achievement. Here we arrive at what many expect from such lists but what is unobtainable: that there is some optimal or ultimate method of measuring performance. The only reasonable approach to interpreting a list of top researchers such as ours is to fully understand the method behind the data and results and why the method is used. With that knowledge, in the end, the results may be judged by users as relevant or irrelevant to their needs or interests.

The data used in the analysis and selection of Highly Cited Researchers derives from Essential Science Indicators (ESI), 2011 to 2021, which then included approximately 179,000 highly cited papers. Each of these papers ranked in the top 1% by total citations according to its ESI field assignment and year of publication. For more information on the identification of highly cited papers in Essential Science Indicators, see the ESI help file at Essential Science Indicators.

Essential Science Indicators surveys the Science Citation Index Expanded™ and components of the Web of Science™, meaning journal articles in the sciences and social sciences. The analysis is further limited to items indexed as articles or reviews only and does not include letters to the editor, correction notices, and other marginalia.

In Essential Science Indicators, all papers, including highly cited papers, are assigned to one of 22 broad fields (the 22nd is Multidisciplinary, on which see below). Each journal in Essential Science Indicators is assigned to only one field and papers appearing in that title are similarly assigned. In the case of multidisciplinary journals such as Science, Nature, and others, a special analysis is undertaken. Each article in such publications is individually reviewed, including an examination of the journals cited in its references. The paper is then reclassified to the most frequently occurring field represented by the article’s cited references.

For each ESI field, author names are disambiguated through advanced clustering methods and the number of clusters is counted, each cluster representing a unique individual. Based on the number of clusters (individuals) for each field, the square root of that number is calculated. That number is used to decide approximately how many researchers to include in each ESI field. From the list of authors in a field ranked by number of highly cited papers, the number of papers at the rank represented by the square root score determines the threshold number of highly cited papers required for inclusion.

In addition, citations to an individual’s highly cited papers must meet the threshold for total citations used in the 2011 to 2021 (6th bimonthly) version of ESI for including a researcher in the top 1% (highly cited list) for an ESI field.

If an author has one fewer highly cited paper than this threshold, but citations to their highly cited papers ranks them in the top 50% by citations among those with highly cited papers at or above the threshold, these individuals are also selected.


Example

ESI field First name Last name Highly cited papers Citation to highly cited papers Field paper threshold Field citation threshold Field citation threshold if one fewer paper than threshold number Status
Field 9 Mary Pandit 17 2,838 11 1,112 2,920 Selected
Field 9 William Clever 10 3,677 11 1,112 2,920 Selected
Field 9 Judith Sage 10 1,338 11 1,112 2,920 Not selected

A criticism of past Highly Cited Researchers lists was that the methodology systematically neglected to identify researchers with cross-field influence: a researcher might contribute multiple highly cited papers in several different fields but would not register enough highly cited papers in any one field for selection. The criticism was valid and welcome. To find individuals with impact equivalent to those we select in a single field, we normalize the highly cited paper counts across fields, so that a paper in Clinical Medicine has the same ‘weight’ as one in Pharmacology/Toxicology.

To do this we fractionate the count for each highly cited paper according to the threshold number used in each field. The fraction for a paper is larger in Pharmacology/Toxicology than in Medicine. If, after collecting all the highly cited papers of an author in all fields we find that the sum of the fractionated paper counts is 1 or more, this demonstrates that the individual has as much influence as those chosen in a single field. A similar procedure is employed for the citation counts, the second criteria for selection.

Example

ESI field First name Last name Highly cited papers Citation to highly cited papers Field citation threshold Field paper threshold Field paper score Field citation score Cross-field paper score Cross-field citation score
Field 3 Joseph Savant 1 98 1,857 22 0.045 0.053 1.670 5.666
Field 6 Joseph Savant 7 2937 946 8 0.875 3.105 1.670 5.666
Field 14 Joseph Savant 3 663 676 6 0.5 0.981 1.670 5.666
Field 16 Joseph Savant 4 3,397 2,223 16 0.25 1.528 1.670 5.666

The fictional researcher Joseph Savant published 15 highly cited papers in four ESI fields during the period 2011 to 2021. Seven papers in Field 6, with a threshold number of eight for selection, earned Savant a credit of 0.875 (or 7/8ths). Three papers in Field 14, with a threshold number of six for selection, were worth. 5. The sum of the fractional paper counts in each field yielded a total Cross-Field paper score of 1.67. A score of 1 or more indicates that the individual achieved impact equivalent to a researcher chosen in a specific ESI field. The second criterion for selection as a Highly Cited Researcher is enough citations to rank in the top 1% by citations for a field. Again, citations in different fields were fractionated in a similar manner to the treatment of papers. In the example above, Professor Savant earned more than five times the number of citations needed for selection as an influential cross-field researcher.

To award credit to a single author among many tens or hundreds listed on a paper strains reason. Therefore, any highly cited paper with more than 30 authors or explicit group authorship was eliminated from our analysis.

We exclude from our analysis highly cited papers that have been retracted.

This year we extended our analysis of retracted papers because we were concerned about cases in which a putative Highly Cited Researcher’s publications that were not highly cited may have been retracted for reasons of misconduct (such as plagiarism, image manipulation, fake peer review). With the assistance of Retraction Watch and its unparalleled database of retractions, Clarivate analysts searched for evidence of misconduct in all publications of those on the preliminary list of Highly Cited Researchers. This extended analysis proved valuable in identifying researchers to exclude, so this exercise will continue in future years. Beyond this, researchers found to have committed scientific misconduct in formal proceedings conducted by a researcher’s institution, a government agency, a funder or a publisher are excluded from our list of Highly Cited Researchers.

In 2019 we began to exclude authors whose collection of highly cited papers revealed unusually high levels of self-citation. For each ESI field, a distribution of self-citation was obtained, and extreme outliers (a very small fraction) were identified and evaluated. For a description of the methodology used to exclude authors with very high levels of self-citation, please see: Adams, J., Pendlebury, D. and Szomszor, M., “How Much is Too Much? The Difference between Research Influence and Self-Citation Excess,” Scientometrics, 123 (2):1119–1147, May 2020.

Inordinate self-citation and unusual collaborative group citation (citation circles or cabals) can seriously undermine the validity of the data analyzed for Highly Cited Researchers. These activities may represent efforts to game the system and create self-generated status. Unfortunately, it appears to us that such activity is increasing. The incentives to achieve Highly Cited Researcher status are in some nations and research systems quite high. Highly Cited Researcher status often results in rewards for a researcher such as higher renumeration, recruitment to other institutions (which benefit in the Academic Ranking of World Universities, since the number of Highly Cited Researchers represents 20% of an institution’s score for ranking) and sometimes offers to become affiliated researchers at other institutions in exchange for large payments and a researcher’s agreement to preferentially list the contracting institution regularly on publications (this represents a shortcut to higher placement in the Academic Ranking of World Universities). This year we extended the identification of these affiliated or guest researchers, designating these as Research Fellows or Associates. These individuals were not counted in our own ranking of nations or institutions.

More ingenious gaming methods require more scrutiny of the publication and citation records of putative Highly Cited Researchers. For example, outsized output, in which individuals publish two or three papers per week over long periods, by relying on international networks of coauthors, raise the possibility that an individual’s high citation counts may result from coauthors alone when publishing without the individual in question. If more than half of a researcher’s citations derive from coauthors, for example, we consider this narrow rather than community-wide influence, and that is not the type of evidence we look for in naming Highly Cited Researchers. And let it be said that any author publishing two or three papers per week strains our understanding of normative standards of authorship and credit.

Clarivate analysts use other filters to identify and exclude researchers whose publication and citation activity is unusual and suspect. We will not enumerate all the checks and filters being deployed in the interest of staying ahead of those attempting to game our identification of Highly Cited Researchers. We can report, with the implementation of more filters this year, the number of putative Highly Cited Researchers excluded from our final list increased from some 300 in 2021 to about 550 this year.

It is discouraging to anticipate that in a few years perhaps up to 10% of those we are identifying through our algorithms may be engaged in publication and citation gaming or misconduct. This then is an explicit call for the research community to police itself through more thorough peer review and other internationally recognized procedures to ensure integrity in research and its publication.