Introduction
This paper develops a novel typology for comparing national research assessment and funding systems worldwide, then analyses significant shifts over the past fifteen years. The typology reflects patterns observed across national systems from thirteen countries. It forms part of AGORRA: A Global Observatory of Responsible Research Assessment, a project initiated by the Research on Research Institute (RoRI) in 2023[1]. AGORRA’s outputs include the RoRI Atlas of Assessment, an online observatory which monitors national assessment and funding systems (informed by the typology introduced in this paper); records changes over time; and provides a platform for assessment system design, experimentation and mutual learning among researchers, policymakers, research funders and institutional leaders.
Our focus is on national research assessment and funding systems, defined as “organized sets of procedures for assessing the merits of research undertaken in publicly funded organisations that are implemented on a regular basis, usually by state or state-delegated agencies” (Whitley, 2007, 6). Our typology and study does not include organisational procedures for the recruitment or promotion of researchers, or the assessment of research grant proposals by funding agencies. This work continues an important tradition of comparative analysis of national research assessment and funding systems (Geuna and Martin, 2003, Hicks, 2012, Debackere et al., 2018, Zacharewicz et al., 2019, Oschner et al., 2021, Sivertsen, 2023). A recent systematic review of this literature (Thomas et al., 2020), which also included more than 300 opinion pieces, reveals that most contributions (including the review itself) implicitly assume that assessment and funding are always combined in national systems. An often-used term is therefore performance-based research funding systems (PRFS), which can be defined as “national systems of research output evaluation used to distribute research funding to universities” (Hicks, 2012, 260). Whitley’s broader definition, quoted above, is important, given information provided by international partners in AGORRA reinforced that:
-
Funding can be formally detached from research assessment and vice versa, and there appears to be a trend in this direction.
-
Some institutional funding systems are not built on research evaluation but on indicators representing already-performed assessments in other contexts – such as external funding and peer-reviewed publications.
-
Some national assessment systems have purposes other than funding allocation, including accountability, organisational learning and strategic development.
-
There is an increasing interest in assessing organisations and their procedures, not only their outputs.
-
Some systems include the assessment of individual researchers, rather than only their organisation’s research.
-
National assessment systems may operate at multiple levels: institutions, departments, research groups, and individual researchers – with different implications at each of these levels.
In contrast with other comparative accounts of performance-based funding systems, our typology and analysis also capture national ex post assessment systems where, for instance, periodic evaluations of research performance are used to provide strategic advice. Besides these differences, our criteria for inclusion of national assessment systems overlap with those employed in other comparisons of performance-based funding systems:
-
Research must be the object of assessment.
-
Evaluations focusing only on the quality of degree programmes or teaching are excluded.
-
Research assessment must be ex post. Evaluations of research grant proposals for project or programme funding are ex ante evaluations, and are excluded.
-
Research outputs must be evaluated in some way. Systems that allocate funding based only on external research funding or PhD enrolment numbers are excluded.
-
It must be a national system. University evaluations of their own research standing, even if used to inform internal funding distribution, are excluded.
(adapted from Hicks, 2012)
This study also extends the literature comparing national systems, by being the first to incorporate a dynamic, longitudinal perspective.
In developing and applying this new typology, we were motivated by three research questions:
-
What characteristics differentiate national research assessment and funding systems, and how can these be categorized and compared?
-
What patterns of change can be observed across national research assessment and funding systems over the period 2010-2024?
-
How might agendas of assessment reform play out over the next 5-10 years?
As in earlier comparative studies (e.g. Hicks 2012; Kolarz et al. 2019; Zacharewicz et al. 2019), we employ a desk-based, analytical review method, whereby the typology was co-developed with country-specific experts, comprising a mix of academic researchers and senior staff within the funding and evaluation agencies involved in AGORRA. The collaborative approach involved the core AGORRA research team holding multiple exchanges with experts across the life-cycle of the project: from initial study design, to providing provisional insights into their respective national systems, to applying and refining early drafts of the typology, and offering iterative feedback for improvement. Experts also shared up-to-date information on recent changes to their systems, contributed to the paper’s final draft, and to the accompanying Atlas of Assessment observatory. In recognition of these extensive and multi-layered contributions, country-specific experts are listed as co-authors of this paper.
Our study examines thirteen countries’ national research assessment and funding systems: Argentina, Australia, Brazil, Chile, China, Colombia, India, Italy, Mexico, Netherlands, Norway, Poland, and the United Kingdom (UK). This sample includes seven countries from the Global South (often under-represented in other studies). Existing comparative literature on national research assessment systems is focused more heavily on peer-review oriented, organisation-level evaluations, which are more common in OECD and EU countries – somewhat overlooking the individual-level systems that are more prevalent in regions like Latin America (Vasen et al, 2023). By including less- studied systems such as those of India, China, and five Latin American countries, our study captures characteristics of several under-mapped assessment systems. A short summary of the 13 countries’ respective systems can be found in the Appendix.
The next section addresses our first research question on characteristics, categorizations, and comparisons of national systems by introducing our typology as applied to thirteen countries. We then present a theoretical framework for observing system changes over time by distinguishing between three major performance paradigms in the development of research assessment and funding over the last forty years. We draw on insights from the typology and conceptual framework to analyse thirteen national systems and to inform our concluding exploration of future trajectories in research assessment over the next five-to-ten years. Taken together, the typology and conceptual framework offer a shared language for understanding the diverse landscape of research assessment systems and the various transformations they are undertaking, enabling sophisticated cross-country comparisons and mutual learning.
1.1 Typology of national research assessment and funding systems
Our typology is separated into four core aspects to categorize major differences and four additional aspects to support closer examination and assessment of the systems. Examples from thirteen national systems are given to illustrate each aspect of the typology. National systems may be complex by combining different assessment schemes for different purposes at different levels of aggregation. Each scheme may be described differently according to the typology. A country may therefore appear more than once to exemplify one aspect.
The four core aspects are:
1. Assigned Purpose
Systems may differ according to their primary purpose:
a. Funding allocation and reputation
b. Accountability
c. organisational learning and strategic development
d. Statistics and overview of research activity
d. Promotion of individual researchers
e. Accreditation
e. Other
Where these purposes are combined, the relative importance of each can be weighted.
Examples: Systems may differ in terms of their primary assigned purpose. Accountability – the idea of holding research performers accountable for their use of public funds and the results of research – is a frequently recurring purpose across our sample of systems. Many countries’ systems combine accountability with other objectives like funding allocation and reputation building (funding allocation and reputation). Some, like the UK’s Research Excellence Framework (REF) and Italy’s Evaluation of Research Quality (VQR), directly link evaluation results to funding outcomes, while others, such as Chile’s National Accreditation Commission, place more emphasis on the reputational consequences of research evaluation results. This combination of purposes extends to national systems for individual assessment and promotion in Argentina, China, Colombia, and Mexico. In Latin America such systems routinely assess “the individual performance of academics based on their academic activities and outputs and assigns them a ‘researcher category’, which carries prestige and, in many cases, additional monthly rewards” (Vasen et al, 2023, 244). Indicator-based systems, like those of Norway and India, also incorporate funding and reputational elements to varying degrees (with India’s NIRF characterized by low reliance on funding and high emphasis on reputation). By contrast, a smaller number of systems, exemplified by the Dutch Strategy Evaluation Protocol, Argentinian PEI and Norwegian Disciplinary Evaluation, prioritize organisational learning and strategic development (alongside accountability in the Dutch case).
2. Unit of assessment and scope
Systems may collect information about and assess:
a. Disciplines across organisations
b. The organisation as a whole
c. Units within the organisation/Research groups
d. Individual researchers
The scope may include all eligible candidates representing the unit of assessment or a certain selection among them. A unit of assessment is often a university department and may represent a discipline at the same time. The level of assessment may differ from the level of funding. Both need to be categorized. As an example, the United Kingdom assesses the quality of research of units of assessment within the organisations while the outcome influences the funding of the organisation as a whole.
More than one unit may be addressed in the same system. Italy combines a) and b). Norway combines a,) b), and c) in the system for disciplinary evaluation. In China, the word ‘Double’ in the system called ‘Double First-Class Evaluation’ means that it combines university-level and disciplinary- level assessments. A country may have more than one system, each of them addressing different units of assessment and purposes, e.g. Argentina, China, Colombia, and Norway.
3. Focus of the assessment
Systems may focus on and collect evidence on different aspects of research performance:
a. Scholarly outputs
b. Scientific impact (citations)
c. Societal interaction (collaboration and co-creation, public engagement, impact cases, technology transfer)
d. Competitive grants
e. organisational performance
f. Research culture
g. Performance of individuals
h. Other
Examples: The UK initially only focused on scholarly outputs and environment. Societal interaction (impact cases) and organisational performance (more weight on environment) have been included in the two most recent exercises in 2014 and 2021, and research culture (as an expansion of environment assessment) will be included in 2029. Revenues from competitive grants are often an indicator of organisational performance in indicator-based systems. Norway’s and Poland’s funding system are examples. Argentina, China, Colombia, and Mexico all have specific systems for assessing the performance of individuals.
4. Effects on funding and reputation
Systems are often discussed as having intended and unintended effects. Dahler-Larsen (2014) points out that systems always have ‘constitutive effects’ just by being implemented.
By their official aims and practical implications, systems may determine:
a. Funding and reputation
b. Only reputation
c. Other significant effects
The relative importance – and the strength of the influences on funding and reputation – of each can be weighted (Strong-Medium-Weak) by users of the typology.
Examples: Most national systems influence both funding and reputation. Examples where the assessments only influence reputation are Chile, Colombia (2 of 3 systems), India, the Netherlands, and Norway (1 of 2 systems). However, reputation may in turn indirectly influence funding. The ERA system in Australia initially influenced both funding and reputation. After the effect on funding was removed, reputation was still experienced as important for the universities.
Other significant effects than funding and reputation may be the influence on (and purpose of) strategic development and learning at the local level. This is the case in Chile, Colombia (1 of 3 systems), the Netherlands and Norway (1 of 2 systems). Systems with strong influence on funding and reputation will generally also affect strategic development. Individual level assessments systems may influence careers, salaries, and resources for performing research. Argentina, Brazil, China, Colombia, and Mexico have national systems for career assessment. Accreditation can be another effect: the outcome of the assessment may influence the right of an institution to grant certain degrees, provide certain courses, or establish new professorships. Poland is an example.
Four additional dimensions of the typology to support further examination of national systems now follows.
5. Methods
The national systems often rely on expert panels informed by peer review and by statistics and indicators, but they may differ in how peer-review and/or expert advice are organized and to what degree the assessment is informed by data and indicators. The types of indicators and their data sources may also differ. The relative importance of each specific method can be weighted.
Examples: The balance between quantitative performance indicators and qualitative peer review remains a persistent tension in national assessment and funding systems. Many evaluation-based systems aspire to operate according to principles of ‘informed peer review’ whereby quantitative indicators inform – but do not replace – expert assessments (Butler, 2007). Some systems differentiate methods by field: in Australia’s ERA and Italy’s VQR, social sciences and humanities have been assessed using peer review, with STEM and other disciplinary panels more likely to use bibliometrics. While the ERA has ceased, the role of bibliometrics in the current VQR is reduced. The UK’s REF allows expert panels some discretion in deciding to use bibliometric measures, but expert review remains the primary means of assessing outputs.
Indicator-based systems offer standardized approaches, exemplified by the Norwegian Indicator, which serves as a nationwide system for partly allocating resources based on a comprehensive publication database, a publication indicator, and a performance-based funding model (Sivertsen, 2018). Similarly, Poland’s Parametric Exercise (EJDN) aggregates points on a range of quantitative performance criteria, including bibliometrics, PhD graduations, and educational indicators, which inform core funding distribution and may affect accreditations.
The use of algorithms and formulas is not limited to indicator-based systems. In Italy’s VQR, panel scores given to units of assessment feed into competitive performance rankings and formula-based distribution of research funding.
Given their primary assigned purposes of generating strategic advice to research performing units, indicator-informed peer review-based evaluations like the Dutch Strategy Evaluation Protocol and the Norwegian disciplinary evaluations, work differently. There, expert panels use peer review to deliver narrative reports on performance and ongoing organisational improvements.
Pressures to ensure appropriate uses of quantitative indicators have been heightened in recent years by the rise of ‘responsible metrics’ as a trans-national professional reform movement across research policy and universities (discussed below in Section 1.2).
Systems that affect institutional funding directly may appear in three main types: evaluation-based funding (the use of peer review and expert panels), indicator-based funding (direct use of performance indicators), and funding contingent on performance agreements between the funder and the individual research organisations.
Examples: Among the thirteen countries included in this overview only Norway and Poland have indicator-based organisational funding systems, and none have organisational funding based on performance agreements. These types are however more widespread than our sample indicates (Sivertsen, 2023). Indicators may also have an important role in informing evaluation-based funding systems. This is the case in Argentina, Chile, China, Italy, and Mexico. The expert panels in Australia’s ERA used to be strongly informed by supporting contextual indicators in select disciplines (e.g. STEM).
Different purposes may result in different main directions for the assessment. A formative evaluation learns from the past and looks forward, serving organisational learning and strategic development. A summative evaluation looks at past performance, checks whether goals or expectations have been reached, and serves decisions and/or resource allocation. Where these directions are combined, the relative importance of each can be weighted according to expert judgement.
Examples: Systems tend to be summative if they rely mainly on empirical evidence for the assessment of past performance. Also, if they determine funding, the outcome of the assessment has to be translated into a quantitative formula. With the exception of Chile, the Netherlands, one of Australia’s (now ceased) systems, and one of the Norwegian systems, most systems covered by our overview will tend to have a summative direction. However, one should not neglect the possible combination of the two purposes of summative and formative, particularly if the effect on funding and reputation is weak (Sivertsen, 2023). Both purposes may appear in official statements. An example is the three purposes that are stated for the REF in the UK in 2029:
-
Inform the allocation of block-grant research funding to HEIs based on research quality;
-
provide accountability for public investment in research and produce evidence of the benefits of this investment;
-
provide insights into the health of research in HEIs in the UK (REF, 2029)
The first purpose can be characterized as summative and the third as formative while the second may express both purposes. The example shows that this classification can in some cases be more a matter of emphasis than a strict binary distinction.
8. Governance
Governance is about how the systems are designed, implemented, and organized on a continuing basis with distributed responsibilities. Systems may differ according to: the role of the agency responsible for operating the system; the transparency and predictability of the results; the degree to which a system is made mandatory, voluntary, or incentivized.
Governing agencies
Systems develop within specific national and historical contexts, with varying degrees of political and administrative centralisation. They tend to be under formal control of a central government agency or arms-length body like a research council, but the role of these agencies varies across countries, particularly in their relationship to government and the academic community. The involvement of and collaboration with the national academic community in the design, implementation, management and evaluation of systems also varies.
Examples: Some systems are directly controlled by government agencies, such as in Poland, India, and Colombia, where ministries of science, education, or equivalent administer assessments.
However, a growing trend over the past three decades is the growth of intermediary organisations like funding agencies and research councils (Braun, 1998, Braun and Guston, 2003) overseeing ex- post assessment and funding, as seen in Brazil, Norway, the UK, and the Netherlands.
The extent of academic community involvement in system design also varies significantly between nations. Countries like the UK and the Netherlands regularly consult with academics, research managers, and sector groups when making periodic adjustments to their systems, while others, such as Colombia and India, offer limited opportunities for input. Even within countries, the level of academic consultation can differ across systems. The presence or absence of meta-evaluations of assessment and funding systems serves as another indicator of accountability to the research community and other stakeholders. At present few countries routinely commission independent meta-evaluations and make them publicly available. The UK REF (Stern, 2016, Digital Science, 2015) and Italian VQR (Expert Review Panel, n.d., Galderisi et al., 2019) though have had meta-evaluation reviews commissioned periodically.
The transparency and predictability of the methods and results (High-Medium-Low)
The transparency and predictability of methods and results in national assessment systems are variously influenced by the choice of method, purpose of assessment, and the consequences of results for funding and reputation.
Examples: Indicator-based systems, due to their standardized nature, theoretically offer higher reproducibility and transparency, as exemplified by the Norwegian Indicator. However, this principle is not universally upheld. Peer review-based evaluations, being less standardized, generally offer lower reproducibility and transparency. The purpose of the evaluation plays a role, as evaluations linked to funding may intentionally limit transparency due to potential litigation risks, as seen in the UK REF’s closed-door panel discussions where panel scoring of individual items are destroyed and not made public in accordance with confidentiality and data protection principles. Peer review evaluations oriented towards delivering narrative reports and strategic advice, like the Dutch Strategy Evaluation Protocol (SEP), exhibits low transparency in how results are produced and used. A distinction can be made here between the public availability of the overall results, versus how and to what degree feedback is utilized by the evaluated units (Gras, 2022).
Mandatory vs. voluntary vs. incentivized participation
Finally, the extent to which assessment exercises are mandatory, incentivized or voluntary varies across national assessment systems. The Brazilian CAPES, the Dutch SEP and the Italian VQR are examples of systems for which participation has effectively been mandatory through regulation or laws. Systems in which research performing organisations or researchers are incentivized to
participate but not legally mandated (through the ‘carrot’ of financial or reputational rewards or potential negative costs of not participating), include Argentina’s PRINUAR, the UK REF, and China’s national assessment and selection systems for elite individual researchers. Other systems where participation is incentivized, include India’s NIRF, Chile’s CNA, Colombia’s High-Quality Accreditation Model, and Mexico’s SNI. In some nations with multiple ex-post assessment systems, certain systems may be mandatory, while others are incentivized or voluntary. For example, the Norwegian indicator-based funding system for research in the higher education sector is mandatory for universities until 2025 (and still mandatory in hospitals and public research institutes), but subject specific evaluations are only normatively expected (and thus classed as voluntary in our typology); and in China, the National Disciplinary Evaluation is voluntary, but the Double-First Class Evaluation is incentivized.
1.1 A framework for understanding ongoing developments in national research assessment and funding systems
Our multi-case comparison of the evolution of national assessment and funding systems 2010-2024, combined with insights from earlier comparative studies, reveals the following broad pattern: the purposes, main focuses and methods of research assessment, have changed, albeit to varying degrees globally, over the last four decades. Changes can be mapped according to three broad paradigms of research performance popularized within research policy at various points in time:
-
The performance paradigm of professional-disciplinary evaluation.
-
The performance paradigm of excellence and competition.
-
The performance paradigm of responsible research assessment.
Originating from Thomas Kuhn, and developed for policy analysis by Peter Hall, the term paradigm can be defined as “a shared model of reality that guides policy makers’ problem-solving activities” (Carson et al., 2009, 18). Crucially, one performance paradigm does not straightforwardly replace another: in practice, there is ‘layering’ (Aagaard, 2017, Capano, 2023) of different aspects of these paradigmatic ideas, which emerge and evolve in context specific ways and shape assessment systems in different ways over time.
In giving our paradigms labels such as ‘excellence’ and ‘responsible research assessment’, we acknowledge we are giving them normative ‘member categories’ used by actors within the field. We consider such labels however to be useful analytical tools for naming and interpreting broad developments in assessment systems. Such labels typically emerge once developments are already underway, serving to capture a sense of transformation within research assessment. At the same time, these labels are performative: they assist actors in agenda building and help to justify, validate, or strengthen certain positions and decisions (Rip, 2000). We now elaborate on each paradigm.
The first paradigm: professional-disciplinary evaluation
The first paradigm of research performance, particularly dominant before the advent of formalized, external assessment, was largely driven by the professional-disciplinary tradition (Elzinga, 2012).
Here academic departments operated with relatively high levels of autonomy, relying on internal disciplinary standards to guide decision-making around hiring, promotion and internal allocation of resources (Whitley, 2000).
Up until the 2010s, much had been written about changes in the policy rationales, design, and methods of national public science systems. In 1994, John Ziman described the end of an era of ‘big bang’ investment in public sciences that had defined the post-war period for publicly funded research in many Western countries until the 1970s (Ziman, 1994). For Ziman, such states were entering a ‘dynamic steady state’ era of science policymaking, defined by moderate increases, stabilization, or cuts to research funding, with economic crises, ideological policy shifts, and changing cultural conceptions of science and technology making politicians generally less willing to continue meeting demands for levels of funding enjoyed in the rapid expansionist post-war period. In Ziman’s book, the clearest manifestation of this new incrementalist regime was the then-recent emergence of the world’s first periodic ex-post national assessment and funding system, the Research Assessment Exercise (RAE) in the UK. Together with decreases in block grant funding in favour of project funding and increased setting of priorities around public policy goals, the RAE promised to help rationally administer allocation of dwindling funds, provide research quality assurance mechanisms, ensure accountability for how research performers were spending scarce public funds, and promote the UK’s global competitiveness in research. Even before the first version of the UK’s exercise was conducted in 1986, national research policymakers in the UK and elsewhere had become increasingly concerned with the competitiveness and economic performance of their national ‘knowledge economies’, influenced in part by evaluation reports published by international organisations like the OECD, World Bank, and the European Union (Wagner et al., 2001, Henriques and Larédo, 2013, Lepori et al., 2007). The faith that increasing investment would improve the quality of research and the economic and social benefits that flow from it was giving way to a belief that funds should best be invested strategically in promising areas of science and technology.
This period was particularly advantageous and attractive for certain research domains like biomedical and health sciences, with others (e.g. social sciences and humanities) experiencing stressful estrangement and cutbacks. In the UK and many other national systems, a new performance paradigm in research policy was thus beginning. Despite such developments, the 1970s and 1980s were for many academics still largely an era where research assessment was dominated by disciplinary standards within individual fields, rather than explicit consideration of interdisciplinary or societal needs (Elzinga, 2012, Whitley, 2000).
The second paradigm: excellence and competition
From the early 1990s onwards came a growing performance paradigm with a strong focus on ‘excellence’ in research assessment and funding. The UK’s RAE system would even later change its name to the Research Excellence Framework (REF). In other countries, excellence also became a common denominator that could include all areas of research (regardless of contribution to economic growth) and bridge with different societal expectations. However, excellence also turned out to be a strong instrument of prioritization: we will only fund the best, and the funded will get more (Scholten et al., 2021). A competition not only within areas of research, but across them as well, was implemented. The focus on measuring research performance coincided with the increasing availability of bibliometric information and its usage in performance-based funding.
The concept of excellence was initially seen as a unifying standard that could encompass all research fields and align with societal expectations. However, in practice, such approaches were not well- aligned with research and evaluation traditions of certain fields, including much of social sciences and humanities. Within academic literatures, the rise of the excellence regime has been attributed to wider currents of globalization (pressures for competition between national knowledge economies) and New Public Management (less centralized command-and-control government combined with periodic accounting and auditing, performance and outcome measurement, and promises of increased efficiency through competition) (Elzinga 2012, Hicks 2012). Periodic assessments, performance-based funding, quantitative performance indicators, and increasing implementation of public policy goals into evaluation criteria of universities and funding agencies are characteristic of such developments (Rip, 2004, Whitley, 2019). A notable example of the ascendance of the excellence performance paradigm can be found in the official research policy aim at the European level to establish ‘more effective national research systems – including increased competition within national borders and sustained or greater investment in research’, as expressed by the European Commission (2012) in a communication with guidelines for A Reinforced European Research Area Partnership for Excellence and Growth.
As implied in the concept of ‘layering’, the excellence paradigm does not replace the disciplinary paradigm per se, but builds on and reconfigures it. Thus, systems that promote accountability and competition, including those with the word ‘Excellence’ in their title (like Excellence in Research Australia and the Research Excellence Framework), have continued to uphold certain practices synonymous with the disciplinary paradigm (e.g. appointing expert peer review panels to administer the evaluations along disciplinary lines). Likewise, systems that utilize bibliometrics and rankings often transform peer reviewed scholarly outputs into quantifiable data points for aggregation (Osterloh and Frei, 2014).
The third paradigm: responsible research assessment (RRA)
The evolution of the performance paradigm as manifest in national research assessment and funding systems is arguably entering a third stage (in some countries), with the emergence of the ‘responsible research assessment’ (RRA) agenda (Curry et al., 2020, Curry et al., 2022, IAP-GYA-ISC, 2023, Benamara et al., 2024). These calls for policy changes can be read as a response to perceived ‘public value failures’ in existing paradigms of research assessment (Bozeman and Sarewitz, 2011), featuring discontent towards both the traditional professional disciplinary evaluation paradigm (for excessive self-referentiality, specialization and esotericism of the ‘ivory tower’, manifest in perceived overemphasis on scientific publications and citation impact indicators) and with the excellence paradigm (for excessive emphasis on competition and selectivity). Critics have held excellence- oriented national assessment and funding systems responsible for creating unintended consequences in research, such as hyper-competition and poor academic work cultures, task reduction, goal displacement, and deterring inter- and trans-disciplinarity (de Rijcke et al., 2016).
RRA attempts to shift focus towards promoting a range of priorities, including Equity, Diversity, and Inclusion (EDI) and improving research culture, open science, promoting collaboration, and addressing societal challenges through impact-oriented research and alignment with public policy goals. The convergence of various science reform movements from the mid-2010s onwards has generated more visible, organised momentum for research assessment reform. Trans-national and national initiatives have provided visibility and engagement platforms for this broad agenda. Perhaps the best-known reform initiative, Coalition for Advancing Research Assessment (CoARA), advocates for research assessment to begin recognizing more alternative research quality criteria around open science and research integrity, equity and inclusion, and research culture in research assessments of various kinds. CoARA – a coalition of research funding organisations, research performing organisations, and assessment authorities – builds on calls for responsible uses of quantitative indicators, advanced for instance by the San Francisco Declaration on Research Assessment (DORA, 2013), Leiden Manifesto (Hicks et al., 2015), and Metric Tide (Wilsdon, 2016), which challenge the power of journal-based indicators and repositions bibliometric tools as means of supporting rather than replacing expert qualitative peer judgment (Rushforth and Hammarfelt, 2023). Similarly, the Global Research Council working group on research assessment recently set out a framework of eleven dimensions of responsible research assessment (Benamara et al., 2024). CoARA and GRC Dimensions incorporate but go beyond calls for responsible metrics, by outlining a number of hitherto marginalized qualities they would like to see rewarded and recognized in various levels of research assessments (CoARA recommends its commitments and principles be enacted at individual, proposal, and organisational assessment levels). Marginalized quality criteria they promote include professional service work, open science contributions, research integrity, equity and inclusion, and societal engagement. The RRA umbrella thus brings together existing and emerging ideas, practices, and criteria, seen as promising alternatives to shortcomings of the disciplinary and excellence paradigms. This may prompt greater awareness of alternative possibilities – even if not all the practices or ideas are new per se.
Twelve years after the European Commission (2012) had published its guidelines for A Reinforced European Research Area Partnership for Excellence and Growth, arguably a high watermark of the excellence paradigm, the same organisation in 2024 published an Action Plan to implement the Agreement on Reforming Research Assessment, in collaboration with CoARA (European Commission, 2024). The plan emphasizes qualitative assessment with the avoidance of inappropriate metrics and clearly shows that the general trends in research assessment have entered into the early stages of a third paradigm in which broader criteria for research assessment are promoted.
Can the national systems be expected to change at the same time and in the same direction?
CoARA is presently an example of how research performing and research funding organisations may collaborate across nations to change the criteria and procedures for research assessment in the same direction. However, CoARA mainly focuses on individual level assessments for recruitment, promotion and external project funding. Sivertsen and Rushforth have argued CoARA’s Agreement on Reform of Research Assessment does not approach organisational assessment and funding with the same clear understanding and guidelines (Sivertsen and Rushforth, 2024). The reason may be that national systems for assessment and funding of research organisations are more rooted in the ‘larger’ and more different national traditions for policies for spending and organisation in the total public R&D sector. The question is thereby to what degree can we expect countries to change their national assessment and funding systems in the same direction in accommodating the recent influence of the RRA performance paradigm?
Based on experiences with the designs, implementations, developments, and discussions of performance-based organisational funding systems in 26 countries, Sivertsen explains why the systems may differ and change at different rates with different influences:
Although some systems may seem similar across countries, they are never the same and they are modified all the time. PBFS [performance-based funding systems] differ because they are anchored in the local traditions and mechanisms of state budgeting and embedded in the local negotiations about priorities and developments in the higher education sector. They are dynamic because they are continuously contested and thereby often adjusted.
Countries also mutually learn from each other and inspire changes in their PBFS. The systems are conservative as well. Once implemented, they become games with rules and specific terminologies and infrastructures that are difficult to change. Also, they need to be predictable because they influence budgets and the spending of tax revenues on the funding side. There is a need to ensure some stability of budgets at the institutions. (Sivertsen, 2023, 90)
A further complication is that even when countries are influenced by the same performance paradigm, ‘the final result is not convergence but different interpretations of the same general recipe’ (Capano, 2023). As we will now show, layering of the nascent third paradigm – RRA – plays out very differently and to varying degrees in national systems across our sample of countries.
Preliminary observations of an emerging third paradigm with responsible research assessment paradigm
In highlighting junctures and patterns of change across our sample of countries, 2010-2024, we argue: a) different aspects of the three performance paradigms are ‘layered’ onto one another to greater or lesser degrees, meaning wholescale replacement of one by another does not occur (c.f. Aagaard, 2017, Capano, 2023), and b) the extent to which a third RRA paradigm has begun to affect national assessment and funding systems (or is even visible at all) varies considerably across our featured countries.
Early manifestations of an emerging third phase came with the introduction of societal contributions as a new criterion, in Australia (Williams and Grant, 2018) and most famously in the UK’s REF 2014 (Martin, 2011). This showed dissatisfaction with the emphasis on research excellence defined largely in disciplinary terms and can be read as a move to increase the importance of social and economic goals. The REF’s adoption of societal impact also inspired other countries, such as the Netherlands, Brazil, and Poland, to utilize similar criterion, albeit in distinctive ways. The Netherlands and Norway, like the UK, incorporated societal impact or relevance into existing systems, and in Poland, societal impact criterion has been added to the EJDN evaluation system, one of many changes made by the 2018 Law on Higher Education (Wróblewska, 2025). Australia set up the EI as a new, separate system to the already existing ERA (which continued to serve the rationale of promoting disciplinary-based excellence), with the EI dedicated to research engagement and societal impact. The EI ran only once, in 2018. Societal contributions are arguably the most prominently developed and formally integrated dimensions of the RRA umbrella to feature across national assessment and funding systems to date, with a growing (though not universal) number of systems accommodating some variation of this criterion.
Other RRA components, like open science, multilingualism, and responsible metrics, appear to have had varying levels of visible impact on designs and methods of systems within our sample. Though Poland’s EJDN has incorporated societal impact, they have so far not accommodated additional items that have emerged under the RRA umbrella. Ongoing sectoral discussions in Australia suggest widespread awareness of the RRA agenda internationally, although the discontinuation of the ERA in 2023 and uncertainties over what will succeed it, means it is unclear presently how exactly the RRA agenda might play out in future national assessments and funding systems. While quantitative, indicator-based systems and excellence criteria feature prominently in many Latin American research assessment systems, there have also been efforts to diversify assessment approaches.
Efforts to counter English-language bias and support publishing in major regional languages like Spanish and Portuguese have grown, mainly by integrating regional indexes like SciELO, RedALyC, and Latindex into national assessment systems (Beigel, 2025). In 2014, Argentina’s CONICET approved a resolution for the social science and humanities equating journals indexed in regional databases with those in international indexes like Web of Science and Scopus. CONICET’s researcher career system also has qualitative, narrative components integrated into its evaluations, alongside bibliometrics and interviews. As mentioned, Brazil’s 2021–2024 CAPES evaluation cycle also integrated ‘impact of society’ as an explicit criterion. The above changes have been promoted and amplified by regional advocacy networks such as CLACSO, whose 2022 principles sought to raise awareness, adapt, and extend global frameworks like DORA and the Helsinki Initiative on Multilingualism in a Latin American and Caribbean context (CLACSO, 2022).
In 2018, five ministries and institutions in China issued a special action, calling on Chinese science to address concerns over the dominance of ‘the four onlys’ (‘only papers, only titles, only educational background, and only awards’) as the primary criteria for research evaluation and talent recognition across the system (Xiaoxuan and Fang, 2020). This coincided with critiques growing across national research policies towards ‘Science Citation Index worship’, echoing statements from North America and Europe, such as the DORA Declaration and Leiden Manifesto (Zhang and Sivertsen, 2020). Even before the special action by the five ministries and institutions began, some well-known research institutes had already started moving away from largely quantitative, output-based definitions of academic performance, with the Chinese Academy of Sciences (CAS) among those taking a leading role in the reform. From 2011 onward, the Chinese Academy of Science’s (CAS) Research Institute Evaluation was reformed into a major achievement, output-oriented evaluation system primarily using qualitative peer review, with quantitative indicators serving only as supporting information.
This shift was the culmination of longer-term efforts by CAS to increase the role of expert judgment and counter-balance systemic effects of the metrics-dominated approach its Research Institute Evaluation system had initially adopted in the early 1990s (Xiaoxuan and Fang, 2020).
One feature of an emerging RRA paradigm is a shift in assessment methods from an exclusive focus on outputs (heavily emphasized in Hicks’s 2012 definition of PBRF systems) to include more process- oriented indicators and narrative-based methods of research performance. Of the performance based funding systems we studied, the UK’s REF has experimented the most so far with moves away from traditional output and results-based forms of evaluation. This is particularly evident in the introduction of ‘institutional-level environment statements’ in REF 2021, which continued a shift in emphasis towards process-oriented indicators, rather than only research outputs and results, that started in RAE 2008. The environment statement required institutions to submit strategic statements that focused on, for instance, support for inclusive research cultures for staff and research students, and steps to ensure sustainability of the unit of assessment through investing in people and infrastructure (Inglis et al., 2024). The elevation of process and input-oriented factors is a qualitatively new development for performance-based funding systems. Where process indicators were mentioned at all in earlier comparative studies, they were treated as unrelated to research performance (e.g. Hicks’s 2012 account explicitly positioned process indicators as separate from this type of system). The growing importance of more process-oriented indicators over the past decade and a half in the UK REF has gone largely under the radar in existing comparative studies. This trend is being intensified still further via the expanded weighting of ‘People, Culture, and Environment’ in REF 2029 and currently ongoing development of new indicators. While earlier accounts stating that process indicators of research management and infrastructure fall outside the scope of assessing research performance reflected prevailing views at the time they were written, an evolving cultural framing of assessment is seeking to reposition research culture and process indicators as important elements in assessing research performance.
Open science – itself a diverse umbrella term (UNESCO, 2021) – has begun in diverse ways to feature as a requirement in some national assessment systems. In Latin America, SciELO, RedALyC, and Latindex’s infrastructures have long supported integration of Diamond Open Access journals (free to read and publish in), facilitating their inclusion in formal research assessment processes. These indexing systems have also implemented quality control mechanisms to screen so-called predatory or spurious journals (Beigel, 2025). Meanwhile, open access publishing was prioritized in Colombia’s 2024 Research Groups Assessment Model. The UK REF 2021 passed an open access mandate as a condition for eligibility of some output types, as has the most recent version of the Italian VQR 2020- 2024, but not all countries have such requirements yet. Furthermore, within the period covered, none of the studied countries’ committed to adhering to use of open information on which to base evaluations of performance in their national assessment and funding systems, as advocated by the Barcelona Declaration on Open Research Information (Barcelona Declaration 2024). Elsewhere
under the open science umbrella, the Italian VQR 2024’s assessment is based on three criteria: originality, methodology, and the impact of the submitted research products. Unlike in past VQR exercises, the methodological criterion has been re-defined to encourage evaluators to pay attention not only to the rigor of research steps but also to certain aspects such as reproducibility, transparency, accessibility and the reuse of data (when applicable) in the publications. This again is reflective of national assessment systems’ gradual responsiveness towards pressures by trans- national science reform movements on issues of open science, research integrity, and metascience.
Several national assessment system authorities have furthermore signed declarations such as DORA, to commit to responsible uses of (particularly quantitative) research indicators. These include NWO and KNAW (overseeing the Dutch SEP) and UKRI (of which Research England – one of the UK’s four higher education funding bodies who govern the REF – is a council), while ANVUR (overseeing the Italian VQR) is a signatory of CoARA. In 2023, Chile’s National Research and Development Agency (ANID) became the first Latin American institution to sign up to the CoARA agreement. Typically, there is no single agency that oversees all research assessments (including ex-ante and ex-post) in an entire national context and voluntary action initiatives like DORA and CoARA can be signed by different actors in a given national research context. In Argentina, for instance, CONICET signed DORA in 2022, and while it is one of the most important agencies in the country, it is the only national agency responsible for overseeing a national assessment that has done so. Some universities also sign up to these agreements, without the national assessment system agency doing so, as is the case in Poland. It is at this point unclear how signing these voluntary action initiatives translates into concrete reforms of research assessment practices in a given national system, especially as there are not typically strict enforcement or compliance checks by the initiatives themselves. Of course, there are other types of assessment beyond national ex-post assessment systems (e.g. career or project selection assessments) where such initiatives can be taken up, though these are beyond the scope of this study.
Notable across the 2010-2024 period is that no new strategic advice-oriented assessment systems have been established, save for Australia’s EI (which was discontinued after running once) and the Norwegian Disciplinary Evaluation. For the latter, since 2011, one of the main tasks of the evaluation committees for biology and for medicine and health has been to deliver advice to the research council and the government. This ambition was followed up in the 2017 and 2018 evaluations of social science and humanities, and in the third round of STEM-evaluations starting in 2021. Across the thirteen countries, though, the overall picture is one of consolidation of summative-oriented systems that do not explicitly deliver strategic advice.
Combinations of disciplinary and excellence-oriented ideas are visible in all cases, albeit how these are combined and play out in different systems varies considerably. Principles of excellence (such as competition through selective performance-based funding derived from evaluating traditional research output and results evaluations) continue to be valorized. So far, efforts to mitigate or counter perceived dysfunctions in the disciplinary and excellence regimes have led to notable, visible changes in only some countries’ national assessment systems.
Possible developments over the next 5-10 years
How might elements of the Responsible Research Assessment (RRA) agenda translate differently, according to the methods, rationales, and focuses of the respective national assessment and funding systems in our sample? Generally, the principles and criteria of RRA may be more readily absorbed in formative, advisory assessment systems. In summative, competitive systems, by contrast, changes in assessment criteria can directly alter evaluation outcomes, which may affect institutional reputation or access to funding, making such changes more consequential and therefore more likely to be contested. Before recent adaptations of performance-based funding systems like the UK REF to include more process indicators, evaluation of research inputs and processes was typically the preserve of advisory-oriented assessments. As far back as 2009-2015, the Dutch SEP included ‘vitality and feasibility’ (e.g. staffing, prioritization and project management as research performance criteria). By 2021 additional criteria for evaluating research performance in the SEP included: open science, PhD policy and training, academic culture, and human resources policy. Although incorporating RRA criteria into formative frameworks like the Dutch SEP and Norwegian Disciplinary Evaluation may be less contested, they are potentially less likely to serve as strong levers for inducing behaviour and culture change. ‘Strong’ systems, by contrast, seem more likely to shift priorities of the assessed, given the significant resource and reputational importance of scoring well in their outcomes (Whitley, 2007). Arguably the introduction of societal impact in REF 2014 has contributed to shifts (albeit unevenly) in the sector, in ways that more discretionary, advisory
procedures could not. Though not without controversy, the inclusion of ‘People, Culture and Environment’ in REF 2029 will likely provide strong stimulus for organisations in the sector to take more seriously this particular interpretation of the RRA agenda.
One uncertainty with respect to RRA-oriented reforms in the context of national assessment and funding systems is the extent to which indicator-based systems might play a role in furthering this reform agenda. Ostensibly, statements such as CoARA’s Agreement may appear sceptical towards what our typology terms indicator-based funding systems (where standardized indicators are applied directly to measure performance). However, the potential for indicator-based systems to incorporate and promote certain elements of the RRA agenda, such as open science, team science, or peer review data, should not be overlooked (Sivertsen and Rushforth, 2024). Advances in bibliometric meta-data might, for instance, help indicator-based allocation schemes evaluate and monitor certain kinds of open access publishing activity, or track cross-sector or cross-disciplinary collaborations or research impacts. Caution of course needs to be taken in employing any quantitative assessment system, to ensure data quality, reliability, and monitoring for unintended consequences. However, the potential for advanced bibliometric methods and data to support the RRA agenda (rather than being its nemesis) is an important consideration meriting wider discussion and awareness-raising. This is particularly important for countries with established indicator-based national systems and little appetite to establish large-scale national peer-review based evaluations.
From where we sit now, it appears unlikely there will be a single direction of travel for national research assessment systems. Some systems may consolidate around large-scale competitive peer review based systems, while others may explore alternatives (as hinted by Adams et al, 2022). Some policymakers may look to indicator-based systems as an option, while others could look towards advisory-oriented assessments or systems where accreditation is an important purpose (a recent report by Guillet (2025) calls on research culture and environment to become a major focus in the external quality assurance of research in higher education institutions conducted by accreditation bodies). Interest will no doubt grow in the potential for AI and data-driven assessment methods, especially in light of concerns regarding the administrative burden of large assessment exercises, though care should be taken not to succumb to hype or reproduce existing biases (Adams et al., 2022). An alternative to national research assessment and funding systems are performance agreements, a ‘dialogue-based funding’ model whereby governments or intermediaries negotiate targets with higher education and research organisations, with funding tied partially or fully to organisations then meeting the contractual targets (Salmi and Hauptman, 2006, 40, 58). Supporters of performance agreements argue they can support diversification of organisational profiling, enhance sensitivity of performance criteria towards institutional missions and public values, generate more context-specific indicators, and strengthen trust between government and institutions (Jongbloed et al., 2020). This approach is not without potential drawbacks though, with critics pointing to concerns over erosion of organisational autonomy, increased bureaucracy, and financial penalties for unmet targets occurring for reasons beyond an organisation’s control (Sivertsen, 2023).
Whichever models get selected, the pathways towards RRA reforms in national assessment systems are likely to be manifold and complex. Some systems may largely ignore or buffer RRA ideas altogether, some may shift gear and press accelerate more rapidly towards RRA, while others may start to accept and enact these ideas and practices but at a slower pace. Certain countries may experience ‘pressure from below’, for instance, in 2024, 36% (1147 institutions) of the total organisational signatories of DORA’s Declaration were from Latin America; but equally academic elites and policymakers in such a regional context could close ranks and valorise the excellence paradigm. There may also be perceptions that the RRA paradigm is a largely European movement.
This scenario points towards a fundamental fork in the road in the coming 5-10 years: some national systems begin or continue to accommodate an RRA-oriented agenda, while others further consolidate along disciplinary-excellence lines. Such developments in national systems are of course not immune to larger systems shocks: financial crises, austerity, pandemics, backlashes against universities, and geopolitical developments, may yet shift landscapes for publicly-funded research in very unpredictable directions. An emerging performance paradigm like RRA may benefit from such disruptions or be held back by these shifting fault-lines.
Conclusion
This paper has two objectives: first, to develop and apply a comprehensive typology of national research assessment and funding systems across thirteen countries (including some from the Global South); and second, to analyse changes in these systems from 2010 to 2024.
Our typology confirms the considerable diversity in national assessment and funding systems reported in existing comparative literature (Geuna and Martin, 2003, Whitley and Gläser, 2007, Hicks, 2012, Zacharewicz et al., 2019, Oschner et al., 2021), reinforcing that there is no universal recipe. Our typology makes several important new contributions. First, it provides an up-to-date tool that enables the comparison of diverse systems that have evolved very differently along the three performance paradigms forming our conceptual framework. This makes it easier to compare systems that would otherwise be difficult to compare directly. Second, our study included countries with at least one of the following types of ex-post national assessment systems currently or recently in place: indicator-based funding; peer review linked to funding; peer review linked to organisational improvement; and/or individual-level national assessment systems. By expanding our inclusion criteria beyond performance-based funding systems to include, for instance, ex-post evaluation systems with the primary assigned purpose of generating strategic advice, we present a more comprehensive overview of alternative options for administering research funding and quality assurance across national research spaces (c.f. Sivertsen, 2023). Third, our sample allows us to shed further light on ex-post systems where the unit of assessment is individual-level research performance. These predominate in Latin America and have tended to be relatively under-studied in both comparative analyses and research evaluation literature more generally (Vasen et al., 2023).
Fourth, our collaborative model has allowed our typology to draw on and synthesize an extensive range of information that otherwise can only be found across a large number of documents and in different languages.
Our longitudinal perspective has led us to theorize changes in rationales and expectations surrounding national research assessment and funding systems from 2010-2024. This is a departure from the cross-sectional designs typically informing comparisons of national assessment and funding systems. This approach led us to argue that over time three paradigms of performance assessment have shaped these systems, to varying degrees. First, was the disciplinary paradigm (peer autonomy, internal disciplinary standards), then the excellence paradigm (competition, selectivity of funding, performance indicators) and most recently the nascent ‘responsible research assessment’ (RRA) paradigm (broadening assessment beyond traditional outputs, diversifying research quality, widening public values). Importantly, we contend national systems have not moved seamlessly from one phase to another: rather paradigmatic ideas accumulate over time, creating ‘layering’ effects where elements of earlier paradigms persist alongside newer paradigms and combine to generate systems with complex, hybrid characteristics.
Our analysis also shows that the RRA agenda is a paradigm still in its early stages, featuring in some but not all national systems, meaning it is not yet a global trend. Non-linear and geographically diverse responses to RRA so far suggest a potential fork in the evolution of national assessment and funding systems could emerge in the coming five-to-ten years, where only some national systems accommodate and adapt to this paradigm in a pro-active fashion. Furthermore, where the RRA agenda is gathering momentum, it remains heterogeneous, still in its infancy, and context- dependent.
Our analysis leads us to caution against oversimplification when describing trends in national assessment systems. While systems have been discontinued in some countries, others have been introduced or are under consideration for the first time. In our view, the overall picture is one largely of consolidation of such systems under dynamic steady state conditions, and of gradual, uneven changes in rationales, designs, and focuses of assessments, rather than wholescale transformation or withdrawal. Looking to the near future, the key challenge for both reforming and newly established systems alike, will be selecting appropriate criteria, design, and methods for assessing research performance and successfully translating aspirational RRA principles into workable assessment practices. How this emerging paradigm gets translated into large-scale periodic peer review-based assessments, versus advisory assessments, indicator-based systems, or perhaps even national individual assessment systems or performance agreements, and how the emerging paradigm interacts with existing ideas and practices of disciplinarity and excellence, are urgent policy questions demanding additional empirical and theoretical research.
Future research should extend our current analysis to an even broader range of countries not included in this study. Additionally, investigating the practical implementation and outcomes of RRA principles in various national contexts would provide valuable insights for policymakers, institution leaders and researchers worldwide. It is essential that the evolution of these systems reflects each country’s unique strategic needs rather than uncritically pursuing a single ‘best practice’ model. In this respect, our typology and analysis of system changes offers valuable opportunities for cross- national mutual learning and further underlines the importance of a comparative research agenda. Indeed with the exception of the important studies cited throughout this paper, comparative perspectives are rare, and longitudinal comparative studies still rarer.
In conclusion, this study contributes a comparative lens and understanding of the complex landscape of national research assessment and funding systems. By providing a comprehensive typology and conceptual framework to compare systems’ evolutions longitudinally, we hope to inform more effective and context-appropriate policy decisions in the domain of research performance evaluation and funding allocation.
Endnote
[1] https://researchonresearch.org/project/agorra/
Acknowledgements
This paper forms part of AGORRA: A Global Observatory of Responsible Research Assessment, a collaboration led by the Research on Research Institute (RoRI). RoRI’s second phase (2023–2027) is funded by an international consortium of partners, including: Australian Research Council (ARC); Canadian Institutes of Health Research (CIHR); Digital Science; Dutch Research Council (NWO); FWF-Austrian Science Fund; Gordon and Betty Moore Foundation [Grant number GBMF12312; DOI 10.37807/GBMF12312]; King Baudouin Foundation; La Caixa Foundation; Leiden University; Luxembourg National Research Fund (FNR); Michael Smith Health Research BC; National Research Foundation of South Africa; Novo Nordisk Foundation [Grant number NNF23SA0083996]; Research England (part of UK Research and Innovation); Social Sciences and Humanities Research Council of Canada (SSHRC); Swiss National Science Foundation (SNSF); University College London (UCL); Volkswagen Foundation; and Wellcome Trust [Grant number 228086/Z/23/Z]. The Research Council of Norway has funded the research of Gunnar Sivertsen in this project through grant number 256223 R-
Quest. Thank you to Emanuel Kulczycki for helpful feedback on the Polish national research assessment and funding systems. Sincere thanks to all AGORRA partners for their engagement and support:
-
Research England/UKRI
-
Australian Research Council
-
ANVUR (Agenzia Nazionale Di Valutazione Del Sistema Universitario E Della Ricerca)
-
Evaluation Research Centre, Chinese Academy of Sciences (CAS), Beijing
-
Dutch Research Council (NWO)
-
The National Research Foundation of South Africa (NRF-SA)
-
Canadian Institutes of Health Research (CIHR)
-
Social Sciences and Humanities Research Council (SSHRC)
-
Swedish Research Council (SRC)
-
Research Council of Norway (RCN)
-
Volkswagen Foundation
We would also like to record our gratitude to members of the AGORRA working group for advice and guidance at every stage. Responsibility for the content of RoRI outputs lies with the authors and RoRI CIC. Any views expressed do not necessarily reflect those of our partners. RoRI is committed to open research as an important enabler of our mission, as set out in our Open Research Policy. Any errors or omissions remain our own.
CREDIT Authorship Statement
Conceptualization: Alexander Rushforth, Gunnar Sivertsen, James Wilsdon
Writing – Original Draft: Alexander Rushforth, Gunnar Sivertsen, James Wilsdon.
Writing – Review and Editing: Alexander Rushforth, Gunnar Sivertsen, James Wilsdon, Ana Arango, Adriana Bin, Catriona Firth, Claire Fraser, Nino Gogadze, Natalia Gras, Lee Harris, Jon Holm, Peter Kolarz, Moumita Koley, Jorge Maldonado, Marie-Helene Nienaltowski, Laura Rovelli, Sergio Salles-Filho, Scipione Sarlo, Nerina Sarthou, Arne Sjostedt, Federico Vasen, Nicole Ward-Boot, Marta Wroblewska, Fang Xu, and Lin Zhang.
Country-specific desk-work performed by: Alexander Rushforth, Gunnar Sivertsen, James Wilsdon, Ana Arango, Adriana Bin, Catriona Firth, Claire Fraser, Nino Gogadze, Natalia Gras, Lee Harris, Jon Holm, Peter Kolarz, Moumita Koley, Jorge Maldonado, Marie-Helene Nienaltowski, Laura Rovelli, Sergio Salles-Filho, Scipione Sarlo, Nerina Sarthou, Arne Sjostedt, Federico Vasen, Nicole Ward-Boot, Marta Wroblewska, Fang Xu, and Lin Zhang.
Supervision: Alexander Rushforth and James Wilsdon.
Openness statement
This is a desk-based review and does not involve empirical data collection. However, in the resource-sharing spirit of open science, further details provided by co-authors – particularly in classifying national assessment systems with the typology and reporting developments over time – is available as a supplementary file on Figshare.
In addition, the Atlas of Assessment observatory aims to serve as a ‘living document’ of national assessment systems. Curated up-to-date information contained in the observatory is publicly available for download from: https://researchonresearch.org/atlas/about-agorra/
References
Aagaard, Kaare. 2017. ‘The Evolution of a National Research Funding System: Transformative Change through Layering and Displacement’. Minerva 55 (3): 279–97.
Abramo, Giovanni. (2024). The forced battle between peer-review and scientometric research assessment: Why the CoARA initiative is unsound, Research Evaluation, 2024; rvae021, https://doi.org/10.1093/reseval/rvae021
ACOLA & Office of the Chief Scientist. 2023. Research Assessment in Australia: Evidence for Modernisation. https://www.chiefscientist.gov.au/ResearchAssessment
Adams, Jonathan, Ryan Beardsley, Lutz Bornmann, Jonathan Grant, Martin Szomszor, and Kate Williams. 2022. ‘Research Assessment: Origins, Evolution, Outcomes’. Clarivate. https://clarivate.com/academia-government/wp-content/uploads/sites/3/dlm_uploads/XBU968048850-ISI-Research-Assessment-Report-v5b-Spreads.pdf.
Barcelona Declaration. 2024. ‘Barcelona Declaration on Open Research Information’. 2024.
Beigel, Fernanda. 2025. ‘The Transformative Relation between Publishers and Editors: Research Quality and Academic Autonomy at Stake’. Quantitative Science Studies, 1–17.
Benamara, Abdelmajid;, Ahmed; Fahal, Alicia; Kowaltowski, Anh-Khoi; Trinh, Anne; Cody, Catriona; Firth, and et al. 2024. ‘Dimensions of Responsible Research Assessment (Full Report and Summary)’. Figshare: Online Resource. https://doi.org/10.6084/m9.figshare.26064223.v3.
Bozeman, Barry, and Daniel Sarewitz. 2011. ‘Public Value Mapping and Science Policy Evaluation’. Minerva 49:1–23.
Braun, Dietmar. 1998. ‘The Role of Funding Agencies in the Cognitive Development of Science’. Research Policy 27 (8): 807–21.
Braun, Dietmar, and David H. Guston. 2003. ‘Principal-Agent Theory and Research Policy: An Introduction’. Science and Public Policy 30 (5): 302–8
Bush, Vannevar. 1945. Science The Endless Frontier: A Report to the President by Vannevar Bush, Director of the Office of Scientific Research and Development, July 1945. United States Government Printing Office, Washington DC
Butler, Linda. 2007. ‘Assessing University Research: A Plea for a Balanced Approach’. Science and Public Policy 34 (8): 565–74.
Capano, Giliberto. 2023. ‘Ideas and Instruments in Public Research Funding’. In Handbook of Public Funding of Research, 73–89. Cheltenham: Edward Elgar Publishing.
Carson, Marcus, Tom R. Burns, and Dolores Calvo. 2009. ‘Introduction’. In Paradigms in Public Policy: Theory and Practice of Paradigm Shifts in the EU, edited by Marcus Carson, Tom R. Burns, and Dolores Calvo, 11–28. Berlin: Peter Lang.
CLACSO. 2022. ‘Declaration of Principles’. Presented at the CLACSO´s XXVII General Assembly, Mexico. https://biblioteca-repositorio.clacso.edu.ar/bitstream/CLACSO/169747/1/Declaration-of-Principes.pdf.
Curry, S., S. De Rijcke, A. Hatch, Dorsamy (Gansen) Pillay, I. van der Weijden, and James Wilsdon. 2020. ‘The Changing Role of Funders in Responsible Research Assessment’.
Curry, S., E. Gadd, and James Wilsdon. 2022. ‘Harnessing the Metric Tide: Indicators, Infrastructures & Priorities for UK Responsible Research Assessment. Report of The Metric Tide Revisited Panel’.
Dahler-Larsen, P. 2014. Constitutive effects of performance indicators: Getting beyond unintended consequences. Public Management Review, 16(7), 969–986.
Debackere, Koenraad, E. Arnold, G. Sivertsen, J. Spaapen, and D. Sturn. 2018. ‘Mutual Learning Exercise: Performance-Based Funding of University Research’. Directorate-General for Research and Innovation, European Commission, B-1049 Brussels ISBN, 978–92.
Digital Science. n.d. ‘The Nature, Scale and Beneficiaries of Research Impact: An Initial Analysis of Research Excellence Framework (REF) 2014 Impact Case Studies.’ http://dera.ioe.ac.uk/22540/1/Analysis_of_REF_impact.pdf.
DORA. 2013. ‘The Declaration’. 2013.
Elzinga, Aant. 2012. ‘Features of the Current Science Policy Regime: Viewed in Historical Perspective’. Science and Public Policy 39 (4): 416–28.
European Commission. 2012. ‘A Reinforced European Research Area Partnership for Excellence and Growth’.
European Commission. 2024. ‘Action Plan by the Commission to implement the ten commitments of the Agreement on Reforming Research Assessment (ARRA)’.
Expert Review Panel. n.d. ‘VQR 2015-2019’. https://www.anvur.it/sites/default/files/2025-02/Expert-Review-Panel_Report-on-VQR-2015-2019.pdf.
Galderisi, Claudio, Mauro Perretti, Nura Galles, and Thed N van Leeuwen. 2019. ‘Report of the Group of Experts Charged by ANVUR to Advice on the Process “Valutazione Della Qualità Della Ricerca (VQR)”’. https://www.anvur.it/sites/default/files/2025-02/High-Experts-Report-on-VQR.pdf.
Geuna, Aldo, and Ben R. Martin. 2003. ‘University Research Evaluation and Funding: An International Comparison’. Minerva 41 (4): 277–304.
Gras, Natalia. n.d. ‘Forms of Research Assessment Oriented at Development Problems. Practices and Perspectives from National Science and Technology organisations and Higher Education Institutions in Latin America and the Caribbean.’ https://doi.org/10.5281/zenodo.6607850.
Guillet, Sophie. 2025. ‘External Quality Assurance of Research in Higher Education Institutions. Taking Stock of the Practices of European Quality Assurance Agencies’. European Association for Quality Assurance in Higher Education. https://www.enqa.eu/wp-content/uploads/ENQA-QA-of-R-report.pdf.
Hall, Peter. 1993. ‘Policy Paradigms, Social Learning, and the State: The Case of Economic Policymaking in Britain’ Comparative Politics, 25 (3): 275-296. http://www.jstor.org/stable/422246
Henriques, Luisa, and Philippe Larédo. 2013. ‘Policy-Making in Science Policy: The “OECD Model” Unveiled’. Research Policy 42 (3): 801–16.
Hicks, Diana. 2012. ‘Performance-Based University Research Funding Systems’. Research Policy 41 (2): 251–61.
Hicks, Diana, Paul Wouters, Ludo Waltman, Sarah De Rijcke, and Ismael Rafols. 2015. ‘Bibliometrics: The Leiden Manifesto for Research Metrics’. Nature 520 (7548): 429–31.
IAP-GYA-ISC. 2023. ‘The Future of Research Evaluation’. Paris: Center for Science Futures.
Inglis, Matthew, Elizabeth Gadd, and Elizabeth Stokoe. 2024. ‘What Is a High-Quality Research Environment?
Evidence from the UK’s Research Excellence Framework’. Research Evaluation. https://doi.org/10.1093/reseval/rvae010.
Jones, Richard and Wilsdon, James. (2018). The biomedical bubble: Why UK research and innovation needs a greater diversity of priorities, politics, places and people. London, Nesta.
Jong, Lisette; Franssen, Thomas; Pinfield, Stephen (2021). ‘Excellence’ in the Research Ecosystem: A Literature Review. (RoRI Working Paper No. 5). Research on Research Institute. Report. https://doi.org/10.6084/m9.figshare.16669834.v2
Jongbloed, Ben, and Harry de Boer. 2020. ‘Performance Agreements in Denmark, Ontario and the Netherlands’.
Kuhn, Thomas. 1962. The Structure of Scientific Revolutions. pp. 54.University of Chicago Press: Chicago Lepori, Benedetto, Peter van den Besselaar, Michael Dinges, Bianca Potì, Emanuela Reale, Stig Slipersæter, Jean Thèves, and Barend van der Meulen. 2007. ‘Comparing the Evolution of National Research Policies: What Patterns of Change?’ Science and Public Policy 34 (6): 372–88. https://doi.org/10.3152/030234207X234578.
Leslie, Ian (2025). ‘Notes on the Great Vibe Shift’. Substack, 18 January 2025. https://www.ian- leslie.com/p/notes-on-the-great-vibe-shift
Martin, Ben R. 2011. ‘The Research Excellence Framework and the “Impact Agenda”: Are We Creating a Frankenstein Monster?’ Research Evaluation 20 (3): 247–54.
Ochsner, Michael, Emanuel Kulczycki, Aldis Gedutis, and Ginevra Peruginelli. 2020. ‘National Research Evaluation Systems’. In Handbook Bibliometrics, edited by Rafael Ball, 99–106. De Gruyter Saur. https://doi.org/10.1515/9783110646610-011.
REF. (2029). Research Excellence Framework. https://2029.ref.ac.uk/?detectflash=false
Rijcke, Sarah de, Paul Wouters, Alex Rushforth, Thomas Franssen, and Björn Hammarfelt. 2016. ‘Evaluation Practices and Effects of Indicator Use—a Literature Review’. Research Evaluation 25 (2): 161–69.
Rip, Arie. 2000. ‘Fashions, Lock-Ins and the Heterogeneity of Knowledge Production’. In The Future of Knowledge Production in the Academy, 28–39. Society for Research into Higher Education and Open Universi.
———. 2004. ‘Strategic Research, Post-Modern Universities and Research Training’. Higher Education Policy 17:153–66.
Rushforth, Alexander, and Björn Hammarfelt. 2023. ‘The Rise of Responsible Metrics as a Professional Reform Movement: A Collective Action Frames Account’. Quantitative Science Studies 4 (4): 879–97. https://doi.org/10.1162/qss_a_00280.
Salmi, Jamil, and Arthur Hauptman. 2006. ‘Innovations in Tertiary Education Financing: A Comparative Evaluation of Allocation Mechanisms’. World Bank, Education Working Paper Series Number 4. Scholten, Wout, Thomas P. Franssen, Leonie van Drooge, Sarah de Rijcke, and Laurens K. Hessels. 2021. ‘Funding for Few, Anticipation among All: Effects of Excellence Funding on Academic Research Groups’. Science and Public Policy 48 (2): 265–75.
Sivertsen, Gunnar. 2018. ‘The Norwegian Model in Norway’. Journal of Data and Information Science 3 (4): 3– 19.
———. 2023. ‘Performance-Based Research Funding and Its Impacts on Research organisations’. In Handbook of Public Funding of Research, 90–106. Edward Elgar Publishing.
Sivertsen, Gunnar, and Alexander Rushforth. 2024. ‘The Ongoing Reform of Research Assessment’. In Challenges in Research Policy, edited by Gunnar Sivertsen and Liv Langfeldt. Dordrecht: Springer.
Stern, Lord Nicholas. 2016. ‘Building on Success and Learning from Experience. An Independent Review of the Research Excellence Framework’. London: Department for Business, Energy and Industrial Strategy. https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/541338/ind-16-9-ref-stern-review.pdf.
Thomas, Duncan A., Maria Nedeva, Mayra M. Tirado, and Merle Jacob. 2020. ‘Changing Research on Research Evaluation: A Critical Literature Review to Revisit the Agenda’. Research Evaluation 29 (3): 275–88.
UNESCO. 2021. ‘Recommendation on Open Science’.
Vasen, Federico, Nerina F. Sarthou, Silvina A. Romano, Brenda D. Gutiérrez, and Manuel Pintos. 2023. ‘Turning Academics into Researchers: The Development of National Researcher Categorization Systems in Latin America’. Research Evaluation 32 (2): 244–55.
Wagner, Caroline S, Irene Brahmakulam, Brian Jackson, Anny Wong, and Tatsuro Yoda. 2001. ‘Science and Technology Collaboration: Building Capacity in Developing Countries: Prepared for the World Bank’. RAND Corporation.
Whitley, Richard. 2000. The Intellectual and Social organisation of the Sciences. Oxford: OUP.
———. 2007. ‘Changing Governance of the Public Sciences’. In The Changing Governance of the Sciences, 3–27. Dordrecht: Springer.
———. 2019. ‘Changing Science Policies, Authority Relationships and Innovations in Public Science Systems’. In Handbook on Science and Public Policy, 204–26. Edward Elgar Publishing.
Whitley, Richard, and Jochen Gläser. 2007. The Changing Governance of the Sciences. Dordrecht: Springer.
Williams, Kate, and Jonathan Grant. 2018. ‘A Comparative Review of How the Policy and Procedures to Assess Research Impact Evolved in Australia and the UK’. Research Evaluation 27 (2): 93–105. https://doi.org/10.1093/reseval/rvx042.
Wilsdon, James. 2016. The Metric Tide: Independent Review of the Role of Metrics in Research Assessment and Management. HEFCE.
Wróblewska, Marta Natalia. 2025. ‘One size fits all? A comparative review of policy-making in the area of research impact evaluation in the UK, Poland and Norway’. Research Evaluation, Volume 34, rvaf010, https://doi.org/10.1093/reseval/rvaf010.
Xiaoxuan, L., & Fang, X. U. (2020). How to Break the ‘Siwei’?—Practice and Enlightenment Based on Research Institute Evaluation of Chinese Academy of Sciences. Bulletin of Chinese Academy of Sciences (Chinese Version), 35(12), 1431–1438. https://doi.org/10.16418/j.issn.1000-3045.20201116002
Zacharewicz, Thomas, Benedetto Lepori, Emanuela Reale, and Koen Jonkers. 2019. ‘Performance-Based Research Funding in EU Member States—a Comparative Assessment’. Science and Public Policy 46 (1): 105–15.
Zhang, L., & Sivertsen, G. (2023). The New Research Assessment Reform in China and Its Implementation. In Towards a New Research Era (pp. 239–252). Brill. https://doi.org/10.1163/9789004546035_017
Ziman, John M. 1994. Prometheus Bound. Cambridge: Cambridge University Press.
Appendix
National Research Assessment and Funding Systems
|
Country
|
Name of system(s)
|
Year of introduction / major change(s)
|
Census period
|
|
Argentina
|
Program for Researcher Teachers (PRINUAR, previously PROINCE)
|
1994
|
2 yrs (admission and promotion) / 4 years (permanence)
|
|
Institutional Evaluation Program
/Programa de Evaluación Institucional (PEI)
|
2005
|
Ad hoc
|
|
CONICET Career of Scientific and Technological Researcher (CICYT)
|
1961
|
1 or 2 years according to seniority
|
|
Australia
|
Excellence in Research for Australia (ERA)
|
2010, Ceased
since 2023
|
3 years
|
|
Engagement and Impact (EI) Assessment
|
2018
|
Not repeated
|
|
Brazil
|
CAPES Evaluation System for Graduate Programs
|
1976
|
4 years
|
|
Chile
|
National Accreditation Commission / Comisión Nacional de Acreditación (CNA)
|
2006
|
3-7 years
|
|
China
|
National Disciplinary Evaluation
|
2002
|
4 years
|
|
Double First-Class Evaluation
|
2017
|
5 years
|
|
National assessment and selection systems for elite individual researchers
|
1994
|
1 year
|
|
Chinese Academy of Science research institute evaluation
|
1998
|
1 or 5 years
|
|
Colombia
|
Research Groups Assessment Model / Modelo de Medición de Grupos
|
1991
|
~2 years
|
|
Decreto 1279 de
2002
|
2002
|
1 year
|
|
High Quality Accreditation Model
|
1992
|
4-10 years
|
|
India
|
National Institutional Ranking Framework (NIRF)
|
2016
|
1 year
|
|
Italy
|
Evaluation of Research Quality (VQR)
|
2011
|
5 years
|
|
Mexico
|
National System of Researchers (SNI)
|
1984
|
Varies per seniority level
|
|
Netherlands
|
Strategy Evaluation Protocol (SEP), previously Standard Evaluation Protocol
|
1994
|
6 years
|
|
Norway
|
Evaluations of specific subjects and thematic areas
|
1990
|
10 years
|
|
Indicator-Based Funding
|
2004, due to cease for universities in 2025
|
1 year
|
|
Poland
|
Evaluation of Quality of Scientific Activity / Ewaluacja Jakości Działalności Naukowej (EJDN)
|
Early 1990s, current format run in 2020/1
|
4 years
|
|
Algorithm Performance Based Funding System
|
Early 1990s Current format 2020
|
1 year
|
|
Research University Program (IDUB)
|
2019
|
6 years
|
|
UK
|
Research Excellence Framework (REF), previously Research Selectivity Exercise (RSE) and Research Assessment Exercise (RAE).
|
1986 (RSE),
1992 (RAE),
2014 (REF)
|
~7-8 years
|