Published at MetaROR

November 20, 2024

Table of contents

Available versions
Cite this article as:

Kochetkov, D. (2024, March 21). Evolution of Peer Review in Scientific Communication. https://doi.org/10.31235/osf.io/b2ra3

Evolution of peer review in scientific communication

Dmitry Kochetkov1,2,3 Email ORCID

1. Centre for Science and Technology Studies, Leiden University, Leiden (Netherlands)
2. Department of Probability Theory and Cybersecurity, RUDN University, Moscow (Russian Federation)
3. Institute for the Study of Science, Russian Academy of Sciences, Moscow (Russian Federation)

Originally published on March 19, 2024 at: 

Abstract

It is traditionally believed that peer review is the backbone of an academic journal and scientific communication, ensuring high quality and trust in the published materials. However, peer review only became an institutionalized practice in the second half of the 20th century, although the first scientific journals appeared three centuries earlier. By the beginning of the 21st century, there emerged an opinion that the traditional model of peer review is in deep crisis. This study aims to synthesize the key characteristics, practices, and outcomes of traditional and innovative peer review models in scholarly publishing. The article discusses the evolution of the institution of scientific peer review and the formation of the current crisis. We analyze the modern landscape of innovations in peer review and scientific communication. Based on this analysis, three main peer review models in relation to editorial workflow are identified: pre-publication peer review (traditional model), registered reports, and post-publication (peer) review (including preprints (peer) review). We argue that the third model offers the best way to implement the main functions of scientific communication.

Introduction 

Scientific peer review can be treated as feedback from scientists (peers) on a particular research study. It is one of the key distinguishing features of academic journals (compared to popular and professional publications). Traditionally, peer review is meant to ensure high quality of scientific research and publications, increasing the level of trust in them within the academic community and among funders (G. D. Smith & Jackson, 2022). Moreover, peer review helps to identify and correct errors in scientific works that may lead to incorrect conclusions and misguided decisions (Kelly et al., 2014). Authors benefit from feedback that helps them to improve their work, though this process can be challenging (Watling et al., 2023).Reviewers typically evaluate a wide range of criteria associated with the quality of scientific research. E.g., the report commissioned by the House of Commons in the UK (Peer Review in Scientific Publications: Eighth Report of Session 2010-12. Vol. 1, 2011) highlighted the following aspects:

  1. Study design and rigor of methodology

  2. Soundness of results,

  3. Transparency of data used in the study,

  4. Interpretation of results,

  5. Whether study objectives are met,

  6. Completeness of the study (are the results final or preliminary),

  7. Scientific novelty and significance,

  8. Ethical issues related to the study and publication.

International journals often require additional assessment of the level of English. However, this is not a comprehensive list, and the list of criteria to be evaluated may vary from journal to journal.

Nevertheless, as early as the beginning of the 21st century, it was argued that the system of peer review is “broken” (McCook, 2006). The main problem noted by McCook is the increasing number of manuscripts and the burden on reviewers. However, this is just the tip of the iceberg. Allen et al. (2022) highlighted the issue of the “black box”: the anonymity of traditional peer review should maintain honesty and ethical norms, but it also can stifle discussion, generate biases, and reduce the overall effectiveness of peer review. In fact, the function of being the “supreme judge” in deciding what is “good” and “bad” science is taken on by peer review, defending the dominant scientific paradigm and stifling the emergence of new ideas that always arise on the periphery. However, as academician L.I. Abalkin once remarked, “no one has the right to usurp the truth” (Sukharev, 2020, p. 44). If we do not change our approach, science will either stagnate or transition into other forms of communication. Moreover, the current system has become an “exploitation machine”: publishers benefit in most cases, while reviewers work voluntarily. There is a point of view that peer review is included in the implicit contract of the researcher. Nevertheless, given that most of the research and, accordingly, research positions are funded from public funds, we nonetheless observe a tendency to “reap where they did not sow.”

R. Smith (2006) strongly criticized the review while at the same time comparing it to democracy: “a system full of problems but the least worst we have” (p. 178). Is this really the case? And can we talk about peer review as a uniform concept, given the variety of existing models? This study aims to synthesize the key characteristics, practices, and outcomes of traditional and innovative peer review models in scholarly publishing. In the following section, we will attempt to demonstrate how the traditional peer review model has developed and how it has come to the current crisis. Furthermore, we will discuss possible ways to overcome the crisis and how the institution of peer review is evolving in the context of global changes in scientific publishing. A separate section discusses modular publishing, which incorporates various innovations in the publishing process, and in particular, the review process. The scope of this piece is limited to peer review in the context of the publication of scientific articles, but its findings are quite applicable to the publication of books or conference proceedings[1]. At the same time, review for other purposes, e.g., evaluation of grant applications, is a topic for a separate discussion.

Evolution and crisis of peer review

The practice of prepublication peer review as we understand it today emerged much later than the founding of first academic journals. E.g., Journal des Sçavans, which was published from 1665 and is considered the first academic journal, printed a warning on the first page “We aim to report the ideas of others without guaranteeing them” (Rennie, 1999, p. 2). However, Kronick (1990), argued that peer review as feedback from peers, in the broad sense of the word, existed as soon as scientists began to exchange research results. Peer review emerged in the form of letters, reviews, and comments that appeared after publication (usually in the case of books). A narrower understanding of peer review, as an evaluation of scientific work by peers before publication (prereview), first appeared in 1731 in the first issue of Medical Essays and Observations, published by the Royal Society of Edinburgh. In 1752, the Royal Society of London took responsibility for publication of Philosophical Transactions and established the “Committee on Papers.” The review process was conducted by the members of the Royal Society with the highest expertise in the topics under consideration. In the late 1890s, the printed peer review report was introduced as a supplement to invitation letter (Fyfe, 2019).

Despite the emergence of pre-publication peer review in the 18th century, it remained a non-institutionalized practice for a long time. In many journals, peer review was not conducted, and the decision on publication was mainly made by the editor. From this point of view, academic journals of the 17th-19th centuries more closely resembled modern newspapers or popular magazines. Peer review became a standard practice only after World War II (Chapelle, 2014). E.g., the well-known British medical journal The Lancet introduced the practice of obligatory peer review in 1976. The rapidly increasing flow of manuscripts played a key role in the institutionalization of peer review, prompting journals to conduct an “entry filtration” of content. Thus, by the second half of the 20th century, the traditional model of pre-publication peer review had been definitively established (Fig. 1).

 

Figure 1. Model 1 – traditional publishing model.

The institutionalization of peer review is manifested in the development of ethical principles which have been adopted by the majority of the academic community. One of the most well-known documents, Ethical Guidelines for Peer Reviewers (2013), was developed by the Committee on Publication Ethics (COPE). This document contains basic principles for reviewers, which have become common practice in the workflow of academic publishers worldwide. Furthermore, most academic journals have a section on their websites that describes the peer review policy applied by the journal.

Despite the progress in editorial policies, which we observe in the formalization of requirements for the peer review process, opinions have been expressed since the end of the 20th century about a crisis in peer review as an institution. In the Introduction, we have already mentioned some of the existing issues, and now we will consider them in more detail. In particular, the following issues can be highlighted:

  1. The rapidly growing volume of manuscripts, coupled with an increasing workload for researchers and faculty, leads to a shortage of reviewers. The primary reason for declining to review is the simple lack of time (Tite & Schroter, 2007; Willis, 2016). This issue causes extension of review periods and frustration of authors.

  2. The shortage of reviewers forces journals to expand their search. Sometimes, this results in manuscripts being reviewed by researchers who do not possess sufficient expertise in the subject. Several studies have noted a low level of consensus among reviewers (Bornmann, 2011), leading some research to refer to peer review as a “game of chance” (Neff & Olden, 2006). The low level of peer review also contributes to the crisis of reproducibility in scientific research (Stoddart, 2016). Although this crisis is due to a multitude of factors, the peer review system bears a significant responsibility for it.

  3. The current peer review system exacerbates inequality in science. Bias often hides behind anonymity, creating a “black box” problem. Despite constant calls for equality and inclusivity in science (COPE, 2021), a few groups still dominate scientific periodicals, such as male authors from the United States and the United Kingdom. O. M. Smith et al. (2023) analyzed 300,000 manuscripts in biological sciences and concluded that authors from historically excluded communities face worse outcomes in peer review, and journal efforts to eliminate reviewer bias have not yet been successful. Nevertheless, we must recognize the debatable nature of this issue. For instance, Squazzoni et al. (2021) found no systematic bias against manuscripts submitted by women in the peer review process, with some evidence of favorable treatment for women in certain fields. Walker et al. (2015) found that the gender of the author and the characteristics of the author’s institution had a significant impact on the review outcomes. However, it is impossible to determine whether this was due to objective differences in scientific merit or to biases.

  4. Continuing from point 3, peer review is also often seen to protect widely accepted approaches and concepts to the detriment of novelty. Peer review can inadvertently stifle innovation and radical new ideas (Steinhauser et al., 2012). The process tends to favor established concepts and discourage the publication of unusual or disparate discoveries (Hess, 1975). As a result, it may limit opportunities for game-changing scientific discoveries (Braben & Dowler, 2017)[2]. The neoclassical school in economics may be seen as an example of this phenomenon. The crisis of the neoclassical school began in the early 21st century (Williams & McNeill, 2005), partly due to the inability to explain the global financial crisis of 2008 (Keen, 2015). However, a paradigm shift has still not occurred – the neoclassical school still occupies a central position in the economic science (and the policies of many countries).

  5. Finally, the current form of peer review is simply inefficient. On the one hand, long peer review slows down the process of disseminating new knowledge (see point 1), and on the other hand, often a large number of reviews are required for a single article. The reason for this is that when authors receive a rejection from one journal, they often submit the same article to another journal, starting the entire process anew. Aczel et al. (2021) found that in 2021, reviewers worldwide spent over 100 million hours, equivalent to more than 15,000 years. If we evaluate this time in terms of money, the cost for reviewers in the USA amounted to over $1.5 billion, in China over $600 million, and in the UK around $400 million. Therefore, peer review is a quite costly activity, and currently, doubts arise regarding the efficiency of its utilization.

Innovations in peer review

We have demonstrated the current crisis of the traditional peer review model. In this regard, the question arises about the possible ways to overcome the crisis. Recently, a lot of literature has been published on innovations in the field of peer review (see reviews by Kaltenbrunner et al., 2022; Woods et al., 2022). Waltman et al. (2023) classified innovations in peer review into four “schools of thought.” We propose adding a parameter to this typology that will characterize innovations relative to the currently dominant publication workflow (incremental / radical). It enables creation of a two-factor matrix (Table 1).

Table 1. Matrix of innovations in peer review.

Course of changes 

Character of innovations 

Incremental 

Incremental

Radical

Quality and reproducibility 

– Training of reviewers

– Software for “automatic” reviewing

– Reviewing of data/code

– Registered reports

Democracy and transparency 

– Review of methodological quality and rigor only

– Open peer review

– Post-publication (peer) review (including preprints (peer) review)

Equity and inclusion 

– Diversity of editorial boards – Reducing biases – Double blind peer review

Efficiency and incentives 

– Recognition of the reviewer’s work – “Portable” peer review

– Journal-independent peer review

Source: compiled by the author based on Waltman et al. (2023).

We should acknowledge that the above-mentioned innovations can simultaneously be placed in different groups. For example, registered reports not only aim to improve the quality of peer review, but also aim to contribute to its efficiency. Now, let’s consider each of the directions in detail.

 Quality and reproducibility

Training of reviewers through seminars and online courses is part of the strategies of many publishers[3]. At the same time, we have not been able to find statistical data or research to assess the effectiveness of such training. Software for automatic evaluation of scientific papers based on artificial intelligence (AI) has emerged relatively recently (StatReviewer[4], UNSILO[5])[6]. We can also allocate here the package for checking statistical analysis statcheck[7]. Currently, these are just auxiliary tools that cannot replace human labor (Baker, 2015; Heaven, 2018), but considering the pace of development of generative AI technologies, these tools have a great future. The increasing role of data in scientific research has led some publishers to recognize the need for review of datasets (e.g., PLOS (A Reviewer’s Quick Guide to Assessing Open Datasets, n.d.). This also applies to review of code in research papers[8].

We have identified registered reports as a radical innovation because it changes the view of the publication workflow and the object of peer review (Registered Reports: Peer Review before Results Are Known to Align Scientific Values and Practices., n.d.). Registered reports are a special type of empirical publication that reflects a hypothetico-deductive approach in science (Fig. 2). Studies are registered and undergo the first stage of review at the early stages of research process. In this case, the research question and methodological approach are evaluated directly. If the peer review results are positive, the study is provisionally accepted for publication, after which data collection, analysis, and interpretation are carried out. These steps are followed by the second stage of peer review, during which the conducted study is compared to the previously registered methodological approach (study protocol).

Figure 2. Registered reports – publication workflow (Model 2).

It should be noted that most initiatives aimed at improving the quality of peer review simultaneously increase the costs.

Democracy and transparency

The approach to peer review in which only the rigor and soundness of the methodology are reviewed (as applied, for example, in PLOS ONE and Scientific Reports) somewhat resembles registered reports, with the difference that the review is conducted in a single stage. This preserves the traditional publication workflow but changes the object of review. The key motivation in this case is that the broader academic community will be better able to assess the significance and contribution of the study than just editor and peer reviewers (Spezi et al., 2017). The next level of “openness” is open peer review, where the reviews are available to readers along with the published article (biomedical journals such as BMJ and BMC were pioneers in adopting this innovation). Wolfram et al. (2020) identified 617 journals that published at least one article with open identities or open peer review reports as of 2019. Though a steady growth of open peer review adoption has been observed recently, publishers have implemented this practice in different ways, resulting in different levels of transparency. Another issue is that in the case of rejection, only authors see the reviews.

Post-publication review, which is most often implemented in the form of open review of preprints, can be considered the most transparent approach. This approach radically changes the essence of peer review. It is no longer a tool for deciding whether to publish a paper or not, but rather a platform for discussion. Publication is no longer the final stage of work; it becomes its starting point. Platforms such as eLife[9], Peer Community in[10], and F1000Research[11] use a model called “Publish-Review-Curate” (PRC). The MetaROR project using this model of review is expected to launch this year (Kaltenbrunner et al., 2023). The PRC model is shown in Fig. 3. It is important to note that for each specific case it will be slightly different. E.g., in the case of MetaROR, the publication is initially hosted on preprint servers such as arXiv, MetaArXiv, SocArXiv, bioRxiv, or OSF Preprints.

Figure 3. “Publish-Review-Curate” model (Model 3). Model 3a involves uploading the manuscript directly to a platform (e.g., F1000Research). Model 3b, on the other hand, involves initially posting a preprint on a preprint server followed by peer review on a peer review platform (e.g., eLife and Peer Community in).

In addition to the projects mentioned, there are other platforms, for example, PREreview[12], which departs even more radically from the traditional review format due to the decentralized structure of work.

Equity and inclusion

The principles of equity and inclusion, as well as the inappropriateness of biases of different origins (geographic, gender, ethnicity), are reflected in numerous recommendations (e.g., COPE, 2021; Royal Society of Chemistry, 2020) and policies of most major academic publishers. However, as mentioned above, the results of implementing these policies are still far from successful, and perhaps these processes require more time. Double-blind peer review is intended to protect the identity of the author and thereby prevent bias in the review. This practice has been used for quite a time in the social sciences and humanities (Horbach & Halffman, 2020; Karhulahti & Backe, 2021). However, anonymity is very conditional – there are still many “keys” left in the manuscript, by which one can determine, if not the identity of the author, then his country, research group, or affiliated organization. On the other hand, the reviewer’s identity is much more securely protected. This issue is especially evident in localized communities: in Russia we often encounter deliberately positive or deliberately negative reviews (Sukharev, 2020). The same is true in specialized fields where reviewers may have conflicts of interest (Rühli et al., 2009). Thus, “closeness” is not a good way to address biases.

 Efficiency and incentives

Any work requires not only an internal motive, but also external incentives. Peer review, as one of the key activities in science, requires appropriate recognition. This practice is implemented in the form of certificates of recognition from academic publishers, as well as records reflected in the profiles of researchers on various platforms (Web of Science, ORCID). Unfortunately, at the moment, peer review is practically not taken into account in the systems of reward and recognition of researchers and faculty adopted at universities and at the national level. Note that open review increases the visibility of reviewers’ work, which should potentially affect recognition.

As mentioned, traditional peer review faces efficiency issues. This is largely due to the fact that the same article, having been rejected in one journal, is submitted to another, where peer review process begins from scratch. One way to solve this problem would be to transfer reviews between journals, also known as “portable peer review.” At the moment, this model is used by large publishing houses (manuscript transfer to another journal of the same publishing house). There are also consortia of journals, such as the Neuroscience Peer Review Consortium (Saper et al., 2009), as well as Manuscript Exchange Common Approach (MECA), an initiative that supports the exchange of manuscripts and reviews between journals and platforms, including preprint servers (NISO RP-30-2023, Manuscript Exchange Common Approach (MECA) (Version 2.0.1), 2023). Although review exchange reduces peer review costs, it doesn’t significantly change the editorial workflow; thus, it is simply an add-on to Model 1 (the traditional model).

The idea of exchanging reviews has evolved into journal-independent peer review. The Reviewer Commons[13], a consortium of 23 life sciences journals, brought this idea into practice. A manuscript is published on a preprint server and undergoes independent review, following which the author can revise the paper and submit it to one of the consortium members. Improving the quality of peer review is achieved by ensuring that reviewers focus on the manuscript itself, rather than the question whether it fits a particular journal. However, we believe that journal-independent peer review is a special case of Model 3 (“Publish-Review-Curate”).

Modular Publishing

Strictly speaking, modular publishing is primarily an innovative approach for the publishing workflow in general rather than specifically for peer review. This approach allows for a more detailed and in-depth exploration of the research process. Besides, modular publishing, which is a type of deconstructed publication[14], combines different models of peer review. This is why we have placed this innovation in a separate category. Nevertheless, modular publications can potentially have a significant impact on peer review practices. Modular publication platforms are like preprint servers, except that they publish not an entire manuscript, but individual significant fragments of it (hypotheses, methodologies, datasets, program code, etc.). These items essentially represent the different stages of a research process. This approach could potentially allow for feedback on each stage completed.

The most well-known initiatives at the moment are ResearchEquals[15] and Octopus[16]. ResearchEquals allows to upload 37 research modules, one of which is “Other”. There is a separate “Review” item. The research modules can be uploaded in any order. Octopus assumes uploading eight research elements in a certain sequence (one of which is peer review), which is more consistent with empirical research. Thus, both platforms offer open post-publication review. Octopus assumes the possibility of revising previously published modules; in ResearchEquals, there is no possibility of versioning. Based on this, we can conclude that the review in the modular publishing resembles Model 3, while the idea itself may be seen as an extension of Model 2. Currently, some of the features are not being implemented due to the technical limitations of the platforms.

A related initiative is currently being developed by the Centre for Open Science, which plans to launch a new approach to scientific communication called Lifecycle Journals (Lifecycle Journals, n.d.). These journals will combine the idea of modular publishing with features such as post-publication peer review and registered reports. Although the initiative is currently in the planning stage, it seems promising.

Discussion and conclusion

In the previous sections, we briefly examined the evolution of the peer review and its current crisis in relation to scientific communication. Next, we explored the main innovations in peer review, which can be classified according to the course of proposed changes and the degree of influence on the editorial workflow, incremental and radical. As a result, we can conclude that, at present, there are three major models of peer review and related editorial workflow:

  • Model 1: traditional model (pre-publication peer review),

  • Model 2: registered reports,

  • Model 3: “Publish-Review-Curate” (post-publication review).

Table 2 presents comparative characteristics of these models.

Table 2 – Comparative analysis of the three review models in terms of editorial workflow

Comparison options 

Model 1 

Model 2 

Model 3 

Content available to the reader 

– Final version

– Peer reviews (optional)

– Preprint (optional)

– Registration (study design)

– Final version of the manuscript on the platform – Peer reviews (optional)

– Preprint (optional)

– Preprint (multiple versions)

– Reviews

– Editorial assessment (optional)

– Final version in traditional journal (optional)

Editor’s role 

Decision making

Decision making (limited)

Ensuring the quality of scientific communication (does not require making accept/reject decisions)

Aim of review 

Manuscript evaluation in order to identify its strengths and weaknesses, help authors improve their work, and finally make a decision on acceptance for publication (1 stage)

Stage 1: evaluation of the strengths and weaknesses of the study design, its rigor, and making final decision on acceptance

Stage 2: evaluation of the conducted study in accordance with the preregistered design and making final decision on acceptance

(2 stages)

Manuscript evaluation aimed at identifying its strengths and weaknesses

Object of review 

Manuscript as a whole (methodology, relevance, novelty, results, etc.), sometimes only the methodology

Manuscript in terms of study design and execution

Manuscript as a whole (methodology, relevance, novelty, results, etc.)

Types of research 

Any

Empirical

Any

We can also compare the three models in terms of the main functions of science communication (Table 3). The four main functions of scientific communication are registration, dissemination, certification and archiving (Roosendaal & Geurts, 1997; Taubert, 2017).

Table 3 – Comparative analysis of the three models of review in terms of functions of scientific communication

Functions of scientific communication

Model 1

Model 2 

Model 3

Registration 

After final version is published*

After final report is published*

Immediately after publication of the preprint

Dissemination 

With a time lag (editorial processes, review, production) *; in the case of a subscription distribution model, there are financial costs of accessing information

With a time lag (editorial processes, review, production) *; in the case of a subscription distribution model, there are financial costs of accessing information

Immediately after the publication of the preprint; with time lag, readers have access not only to the publication itself, but also to reviews and editor’s assessment (optional)

Certification 

Provided by the opinion of the editor and peer reviewers (most often anonymous)

Provided by the opinion of the editor and peer reviewers (most often anonymous)

The readers form their own assessment based on open reviews and the editor’s assessment (optional)

Archiving 

Journal/publisher server

Journal/publisher server

Preprint servers, publishing platform; further, the article may also be published in a traditional journal (optional)

* Preprint is optional for Models 1 and 2.

Thus, in Model 3, all functions of scientific communication are implemented most quickly and transparently. The additional costs arising from the independent assessment of information based on open reviews are more than compensated by the emerging opportunities for scientific pluralism. Model 3 corresponds to the vision of the International Science Council (ISC) on “more efficient and effective modes of peer review that are inspired by open norms” (International Science Council, 2023, p. 12).

The traditional publication process model with a “black box” peer review inside is increasingly proving its inadequacy. Registered reports are promising but are exclusively focused on empirical research. The “Publish-Review-Curate” model is universal that we expect to be the future of scientific publishing. The transition will not happen today or tomorrow, but in the next 5-10 years, the number of projects such as eLife, F1000Research, Peer Community in, or MetaROR will rapidly increase. We should also note that the constructive elements of Model 3 can be transferred to Model 2 (in terms of openness of the review process, especially on the first stage).

At the same time, we must recognize the complexity of institutional change. The possibilities for normative regulation here are quite limited – much depends on the traditions embedded in the academic community, and it will take a lot of time to change them. Openness is a complex process that requires three conditions:

  1. group of people willing to take responsibility for the quality of scientific communication in this academic community,

  2. authors and reviewers willing to accept this practice,

  3. appropriate infrastructure.

Avissar-Whiting et al. (2024) provided a useful toolbox of recommendations for all parties potentially involved in the preprint review process.

A significant development was the announcement by the Bill & Melinda Gates Foundation of its new Open Access Policy (Brembs & Drury, 2024). The foundation has discontinued Article Processing Charges (APCs) and introduced a mandatory requirement for grantees to publish preprints. In addition, the foundation is supporting the development of an open infrastructure for preprinting. Later, it was announced that the foundation would be launching a new verified preprint platform in collaboration with F1000Research (“Gates Foundation Collaborates with F1000 to Launch Verified Preprint Platform,” 2024). This is another step towards normalizing post-publication reviews, while the publication of preprints is increasingly becoming the norm (Drury, 2022).

Post-publication review is a return to the roots of scientific communication. This model will allow all actors involved to take greater responsibility for their work, authors for their articles, reviewers for their assessments, and editors for supporting the process of scientific communication. This is the atmosphere of scientific discussion that we need very much.

However, it is important to acknowledge that current peer review practices vary significantly across fields. While pre-publication peer review is prevalent in almost all fields, there are numerous variations in terms of its openness or anonymity. Additionally, while some fields have successfully incorporated innovations, others continue to resist. E.g., preprint publishing has been an essential form of publication in physics since 1990s (Ginsparg, 2011), and open peer review was introduced in biomedical journals prior to other domains (e.g., BMJ and BMC). At the same time, the social sciences and humanities (SSH), in terms of their peer review and publication process, remain relatively closed (Ross-Hellauer & Horbach, 2024).

We should also acknowledge that the list of innovations in peer review presented in this paper is not comprehensive. For instance, we could mention ranking papers instead of reviewing them or bidding for papers (Birukou et al., 2011). However, these initiatives are not yet widely adopted, so they do not significantly affect the publishing landscape.

Acknowledgements

The author gratefully acknowledges the peers’ contributions. Ludo Waltman and Wolfgang Kaltenbrunner reviewed draft versions of this paper and provided valuable suggestions for improvement.

Competing interests

The author is affiliated with the Centre for Science and Technology Studies of Leiden University, which is involved in the development of the MetaROR project.

Contributions

Visualization: DK

Writing – original draft: DK

Data availability statement

Data sharing is not applicable to this article as no new data were created in this study.

References

A Reviewer’s Quick Guide to Assessing Open Datasets. (n.d.). PLOS. Retrieved January 23, 2024, from https://plos.org/resource/peer-reviewing-data/

Aczel, B., Szaszi, B. & Holcombe, A. O. (2021). A billion-dollar donation: estimating the cost of researchers’ time spent on peer review. Research Integrity and Peer Review6(1), 14. https://doi.org/10.1186/s41073-021-00118-2

Allen, K., Reardon, J., Lu, Y., Smith, D. V., Rainsford, E. & Walsh, L. (2022). Towards improving peer review: Crowd-sourced insights from Twitter. Journal of University Teaching & Learning Practice19(3). https://ro.uow.edu.au/jutlp/vol19/iss3/02

Avissar-Whiting, M., Belliard, F., Bertozzi, S. M., Brand, A., Brown, K., Clément-Stoneham, G., Dawson, S., Dey, G., Ecer, D., Edmunds, S. C., Farley, A., Fischer, T. D., Franko, M., Fraser, J. S., Funk, K., Ganier, C., Harrison, M., Hatch, A., Hazlett, H., … Williams, M. (2024). Recommendations for accelerating open preprint peer review to improve the culture of science. PLOS Biology22(2), e3002502. https://doi.org/10.1371/journal.pbio.3002502

Baker, M. (2015). Smart software spots statistical errors in psychology papers. Nature. https://doi.org/10.1038/nature.2015.18657

Birukou, A., Wakeling, J. R., Bartolini, C., Casati, F., Marchese, M., Mirylenka, K., Osman, N., Ragone, A., Sierra, C. & Wassef, A. (2011). Alternatives to Peer Review: Novel Approaches for Research Evaluation. Frontiers in Computational Neuroscience5. https://doi.org/10.3389/fncom.2011.00056

Bornmann, L. (2011). Scientific peer review. Annual Review of Information Science and Technology45(1), 197–245. https://doi.org/10.1002/aris.2011.1440450112

Braben, D. & Dowler, R. (2017, 17. September). Peer review processes risk stifling creativity and limiting opportunities for game-changing scientific discoveries. LSE Impact Blog. https://blogs.lse.ac.uk/impactofsocialsciences/2017/09/17/peer-review-processes-risk-stifling-creativity-and-limiting-opportunities-for-scientific-discoveries/

Brembs, B. & Drury, L. (2024, 27. March). The Open Access rising tide: Gates Foundation ends support to Article Processing Charges. International Science Council. https://council.science/current/blog/the-open-access-rising-tide-gates-foundation-ends-support-to-article-processing-charges/

Chapelle, F. H. (2014). The History and Practice of Peer Review. Groundwater52(1), 1–1. https://doi.org/10.1111/gwat.12139

COPE. (2013). Ethical guidelines for peer reviewers (English). https://doi.org/10.24318/cope.2019.1.9

COPE. (2021). Diversity and inclusivity. https://doi.org/10.24318/RLqSoVsZ

Drury, L. (2022). The Normalization of Preprints. SRELS Journal of Information Management, 79–85. https://doi.org/10.17821/srels/2022/v59i2/169462

Fyfe, A. (2019, 15. September). Quality in peer review: a view through the lens of time. The Royal Society. https://royalsociety.org/blog/2019/09/quality-in-peer-review-a-view-through-the-lens-of-time/

Gates Foundation Collaborates with F1000 to Launch Verified Preprint Platform. (2024, 4. April). Taylor & Francis News. https://newsroom.taylorandfrancisgroup.com/gates-foundation-collaborates-with-f1000-to-launch-verified-preprint-platform/

Ginsparg, P. (2011). ArXiv at 20. Nature476(7359), 145–147. https://doi.org/10.1038/476145a

Heaven, D. (2018). AI peer reviewers unleashed to ease publishing grind. Nature563(7733), 609–610. https://doi.org/10.1038/d41586-018-07245-9

Hess, E. L. (1975). Effects of the review process. IEEE Transactions on Professional CommunicationPC18(3), 196–199. https://doi.org/10.1109/TPC.1975.6591188

Horbach, S. P. J. M. & Halffman, W. (2020). Journal Peer Review and Editorial Evaluation: Cautious Innovator or Sleepy Giant? Minerva58, 139–161. https://doi.org/10.1007/s11024-019-09388-z

International Science Council. (2023). The Case for Reform of Scientific Publishing. https://doi.org/10.24948/2023.14

Johnson, R. (2024, 28. June). Beyond the journal: The future of scientific publishing. FEBS. https://network.febs.org/posts/beyond-the-journal-the-future-of-scientific-publishing?channel_id=728-viewpoints

Kaltenbrunner, W., Pinfield, S., Waltman, L., Woods, H. B. & Brumberg, J. (2022). Innovating peer review, reconfiguring scholarly communication: An analytical overview of ongoing peer review innovation activities. SocArXiv. https://doi.org/10.31235/osf.io/8hdxu

Kaltenbrunner, Wolfgang, Waltman, L., Barnett, A., Byrne, J., Chin, J. M., Holcombe, A., Pinfield, S., Vazire, S. & Wilsdon, J. (2023). MetaRoR – a new form of scholarly publishing and peer review for STS. EASST Review421. https://easst.net/easst-review/easst-review-volume-421-july-2023/metaror-a-new-form-of-scholarly-publishing-and-peer-review-for-sts/

Karhulahti, V.-M. & Backe, H.-J. (2021). Transparency of peer review: a semi-structured interview study with chief editors from social sciences and humanities. Research Integrity and Peer Review6(1), 13. https://doi.org/10.1186/s41073-021-00116-4

Keen, S. (2015). Post Keynesian Theories of Crisis. The American Journal of Economics and Sociology74(2), 298–324. https://doi.org/10.1111/ajes.12099

Kelly, J., Sadeghieh, T. & Adeli, K. (2014). Peer review in scientific publications:benefits, critiques, & a survival guide. The Journal of the International Federation for Clinical Chemistry and Laboratory Medicine25(3), 227–243.

Kronick, D. A. (1990). Peer review in 18th-century scientific journalism. JAMA263(10), 1321–1322. http://www.ncbi.nlm.nih.gov/pubmed/2406469

Lifecycle Journals. (n.d.). Center for Open Science. Retrieved April 19, 2024, from https://www.cos.io/lifecyclejournals

McCook, A. (2006. February). Is peer review broken? The Scientist. https://www.the-scientist.com/uncategorized/is-peer-review-broken-47872

Neff, B. D. & Olden, J. D. (2006). Is Peer Review a Game of Chance? BioScience56(4), 333–340. https://doi.org/10.1641/0006-3568(2006)56%5B333:IPRAGO%5D2.0.CO;2

NISO RP-30-2023, Manuscript Exchange Common Approach (MECA) (Version 2.0.1). (2023). https://doi.org/10.3789/niso-rp-30-2023

Nosek, B. A., Spies, J. R. & Motyl, M. (2012). Scientific Utopia. Perspectives on Psychological Science7(6), 615–631. https://doi.org/10.1177/1745691612459058

Open Science Collaboration. (2015). Estimating the reproducibility of psychological science. Science349(6251). https://doi.org/10.1126/science.aac4716

Registered Reports: Peer review before results are known to align scientific values and practices. (n.d.). Center for Open Science. Retrieved January 22, 2024, from https://www.cos.io/initiatives/registered-reports

Rennie, D. (1999). Editorial peer review: Its development and rationale. In F. Godleeand & T. Jefferson (Eds.), Peer Review in Health Sciences (pp. 1–13). BMJ Books.

Roosendaal, H. E. & Geurts, P. A. T. M. (1997). Forces and functions in scientific communication: an analysis of their interplay. Co-Operative Research in Information Systems in Physics, 1–32.

Ross-Hellauer, T. & Horbach, S. P. J. M. (2024). Additional experiments required: A scoping review of recent evidence on key aspects of Open Peer Review. Research Evaluation. https://doi.org/10.1093/reseval/rvae004

Royal Society of Chemistry. (2020). Joint commitment for action on inclusion and diversity in publishing. https://www.rsc.org/policy-evidence-campaigns/inclusion-diversity/joint-commitment-for-action-inclusion-and-diversity-in-publishing/#:~:text=The Joint commitment for action,a workshop with other publishers

Rühli, F. J., Finnegan, M., Hershkovitz, I. & Henneberg, M. (2009). Peer‐review for the peer‐review system. Human_ontogenetics3(1), 3–6. https://doi.org/10.1002/huon.200900004

Saper, C. B., Maunsell, J. H. R. & Sagvolden, T. (2009). The Neuroscience Peer Review Consortium. Behavioral and Brain Functions5, 4. https://doi.org/10.1186/1744-9081-5-4

Smith, G. D. & Jackson, D. (2022). Integrity and trust in research and publication: The crucial role of peer review. Journal of Advanced Nursing78(11). https://doi.org/10.1111/jan.15438

Smith, O. M., Davis, K. L., Pizza, R. B., Waterman, R., Dobson, K. C., Foster, B., Jarvey, J. C., Jones, L. N., Leuenberger, W., Nourn, N., Conway, E. E., Fiser, C. M., Hansen, Z. A., Hristova, A., Mack, C., Saunders, A. N., Utley, O. J., Young, M. L. & Davis, C. L. (2023). Peer review perpetuates barriers for historically excluded groups. Nature Ecology & Evolution7(4), 512–523. https://doi.org/10.1038/s41559-023-01999-w

Smith, R. (2006). Peer Review: A Flawed Process at the Heart of Science and Journals. Journal of the Royal Society of Medicine99(4), 178–182. https://doi.org/10.1177/014107680609900414

Spezi, V., Wakeling, S., Pinfield, S., Creaser, C., Fry, J. & Willett, P. (2017). Open-access mega-journals. Journal of Documentation73(2), 263–283. https://doi.org/10.1108/JD-06-2016-0082

Squazzoni, F., Bravo, G., Farjam, M., Marusic, A., Mehmani, B., Willis, M., Birukou, A., Dondio, P. & Grimaldo, F. (2021). Peer review and gender bias: A study on 145 scholarly journals. Science Advances7(2). https://doi.org/10.1126/sciadv.abd0299

Steinhauser, G., Adlassnig, W., Risch, J. A., Anderlini, S., Arguriou, P., Armendariz, A. Z., Bains, W., Baker, C., Barnes, M., Barnett, J., Baumgartner, M., Baumgartner, T., Bendall, C. A., Bender, Y. S., Bichler, M., Biermann, T., Bini, R., Blanco, E., Bleau, J., … Zwiren, N. (2012). Peer review versus editorial review and their role in innovative science. Theoretical Medicine and Bioethics33(5), 359–376. https://doi.org/10.1007/s11017-012-9233-1

Stoddart, C. (2016). Is there a reproducibility crisis in science? Nature. https://doi.org/10.1038/d41586-019-00067-3

Sukharev, O. S. (2020). Topos of Russian peer review (on peer review as creativity, subject to amateurism). Investments in Russia10(309), 43–48.

Taubert, N. (2017). Formale wissenschaftliche Kommunikation. In Forschungsfeld Wissenschaftskommunikation (pp. 125–139). Springer Fachmedien Wiesbaden. https://doi.org/10.1007/978-3-658-12898-2_7

Tite, L. & Schroter, S. (2007). Why do peer reviewers decline to review? A survey. Journal of Epidemiology & Community Health61, 9–12.

Walker, R., Barros, B., Conejo, R., Neumann, K. & Telefont, M. (2015). Bias in peer review: a case study. F1000Research4, 21. https://doi.org/10.12688/f1000research.6012.1

Waltman, L., Kaltenbrunner, W., Pinfield, S. & Woods, H. B. (2023). How to improve scientific peer review: Four schools of thought. Learned Publishing36(3), 334–347. https://doi.org/10.1002/leap.1544

Waltman, L., Mulati, B., Ni, R., Wang, J., Lai, K. H. (Adrian), Luwel, M., Noyons, E., van Leeuwen, T. & Weimer, V. (2023). Preprinting and open peer review at the STI 2023 conference: Evaluation of an open science experiment. Leiden Madtrics. https://www.leidenmadtrics.nl/articles/preprinting-and-open-peer-review-at-the-sti-2023-conference-evaluation-of-an-open-science-experiment

Watling, C., Shaw, J., Field, E. & Ginsburg, S. (2023). ‘For the most part it works’: Exploring how authors navigate peer review feedback. Medical Education57(2), 151–160. https://doi.org/10.1111/medu.14932

Williams, J. B. & McNeill, J. M. (2005). The Current Crisis In Neoclassical Economics and the Case for an Economic Analysis Based on Sustainable Development. SSRN Electronic Journal. https://doi.org/10.2139/ssrn.1606342

Willis, M. (2016). Why do peer reviewers decline to review manuscripts? A study of reviewer invitation responses. Learned Publishing29(1), 5–7. https://doi.org/10.1002/leap.1006

Wolfram, D., Wang, P., Hembree, A. & Park, H. (2020). Open peer review: promoting transparency in open science. Scientometrics125(2), 1033–1051. https://doi.org/10.1007/s11192-020-03488-4

Woods, H. B., Brumberg, J., Kaltenbrunner, W., Pinfield, S. & Waltman, L. (2022). Innovations in peer review in scholarly publishing: A meta-summary. SocArXiv. https://doi.org/10.31235/osf.io/qaksd

Notes

[1] E.g., an open science experiment during the recent Science, Technology and Innovation Indicators (STI2023) conference (Waltman, Mulati, et al., 2023).

[2] Here, we must acknowledge that this effect varies significantly across fields. In some disciplines, critical observers worry that journals on the contrary tend to give incentives for inflated claims prioritizing the publication of novel positive results (Nosek et al., 2012). This often forces authors to place scientific novelty over the reproducibility of the results. For instance, such debates have been ongoing in psychology for a long time (Open Science Collaboration, 2015).

[3] E.g., Certified Peer Reviewer Course by Elsevier. URL: https://researcheracademy.elsevier.com/navigating-peer-review/certified-peer-reviewer-course (date of access: 22.01.2024).

[4] StatReviewer. URL: http://statreviewer.com/ (date of access: 22.01.2024).

[5] UNSILO. URL: https://site.unsilo.com/site/ (date of access: 22.01.2024).

[6] At the same time, plagiarism detection systems have existed much longer, for example, “Antiplagiat,” a well-known system in Russia, originated in 2005.

[7] statcheck. URL: https://michelenuijten.shinyapps.io/statcheck-web/ (date of access: 22.01.2024), also R package.

[8] Among recent initiatives, we can mention CODECHECK. URL: https://codecheck.org.uk/process/ (date of access: 22.01.2024).

[9] eLife. URL: https://elifesciences.org/ (date of access: 22.01.2024).

[10] Peer Community in. URL: https://peercommunityin.org/ (date of access: 22.01.2024).

[11] F1000Research. URL: https://f1000research.com/ (date of access: 22.01.2024).

[12] PREreview. https://prereview.org/ (date of access: 22.01.2024).

[13] Reviewer Commons. URL: https://www.reviewcommons.org/ (date of access: 22.01.2024).

[14] Deconstructed publication is a type of scholarly output, where the research is communicated in separate stages, rather than as a single, traditional journal article. This term was coined by Johnson (2024), and he also includes registered reports in this category.

[15] ResearchEquals. URL: https://www.researchequals.com/ (date of access: 28.02.2024).

[16] Octopus. URL: https://www.octopus.ac/ (date of access: 28.02.2024).

Editors

Kathryn Zeiler
Editor-in-Chief

Jason Chin
Handling Editor

Editorial assessment

by Jason Chin

DOI: 10.70744/MetaROR.9.1.ea

This article provides a brief history and review of peer review. It evaluates peer review models against the goals of scientific communication, expressing a preference for publish, review, curate (PRC) models. The review and history are useful. However, the article’s progression and arguments, along with what it seeks to contribute to the literature need refinement and clarification. The argument for PRC is under-developed due to a lack of clarity about what the article means by scientific communication. Clarity here might make the endorsement of PRC seem like less of a foregone conclusion.

As an important corollary, and in the interest of transparency, I declare that I am a founding managing editor of MetaROR, which is a PRC platform. It may be advisable for the author to make a similar declaration because I understand that they are affiliated with one of the universities involved in the founding of MetaROR.

Recommendations from the editor

I strongly endorse the main theme of most of the reviews, which is that the progression and underlying justifications for this article’s arguments needs a great deal of work. In my view, this article’s main contribution seems to be the evaluation of the three peer review models against the functions of scientific communication. I say ‘seems to be’ because the article is not very clear on that and I hope you will consider clarifying what your manuscript seeks to add to the existing work in this field.

In any case, if that assessment of the three models is your main contribution, that part is somewhat underdeveloped. Moreover, I never got the sense that there is clear agreement in the literature about what the tenets of scientific communication are. Note that scientific communication is a field in its own right.

I also agree that paper is too strongly worded at times, with limitations and assumptions in the analysis minimised or not stated. For example, all of the typologies and categories drawn could easily be reorganised and there is a high degree of subjectivity in this entire exercise. Subjective choices should be highlighted and made salient for the reader.

Note that greater clarity, rigour, and humility may also help with any alleged or actual bias.

Some more minor points are:

  1. I agree with Reviewer 3 that the ‘we’ perspective is distracting.

  2. The paragraph starting with ‘Nevertheless’ on page 2 is very long.

  3. There are many points where language could be shortened for readability, for example:

    • Page 3: ‘decision on publication’ could be ‘publication decision’.

    • Page 5: ‘efficiency of its utilization’ could be ‘its efficiency’.

    • Page 7: ‘It should be noted…’ could be ‘Note that…’.

  4. Page 7: ‘It should be noted that..’ – this needs a reference.

  5. I’m not sure that registered reports reflect a hypothetico-deductive approach (page 6). For instance, systematic reviews (even non-quantitative ones) are often published as registered reports and Cochrane has required this even before the move towards registered reports in quantitative psychology.

  6. I agree that modular publishing sits uneasily as its own chapter.

  7. Page 14: ‘The “Publish-Review-Curate” model is universal that we expect to be the future of scientific publishing. The transition will not happen today or tomorrow, but in the next 5-10 years, the number of projects such as eLife, F1000Research, Peer Community in, or MetaROR will rapidly increase’. This seems overly strong (an example of my larger critique and that of the reviewers).

Competing interests: I declare that I am a founding editor of MetaROR, which is a PRC platform.

Peer review 1

Balazs Aczel

DOI: 10.70744/MetaROR.9.1.rv1

The work ‘Evolution of Peer Review in Scientific Communication’ provides a concise and readable summary of the historical role of peer review in modern science. The paper categorises the peer review practices into three models: (1) traditional pre-publication peer review; (2) registered reports; (3) post-publication peer review. The author compares the three models and draws the conclusion that the “third model offers the best way to implement the main function of scientific communication”.

I would contest this conclusion. In my eyes the three models serve different aims – with more or less drawbacks. For example, although Model 3 is less chance to insert bias to the readers, it also weakens the filtering function of the review system. Let’s just think about the dangers of machine-generated articles, paper-mills, p-hacked research reports and so on. Although the editors do some pre-screening for the submissions, in a world with only Model 3 peer review the literature could easily get loaded with even more ‘garbage’ than in a model where additional peers help the screening.

Compared to registered reports other aspects can come to focus that Model 3 cannot cover. It’s the efficiency of researchers’ work. In the care of registered reports, Stage 1 review can still help researchers to modify or improve their research design or data collection method. Empirical work can be costly and time-consuming and post-publication review can only say that “you should have done it differently then it would make sense”.

Finally, the author puts openness as a strength of Model 3. In my eyes, openness is a separate question. All models can work very openly and transparently in the right circumstances. This dimension is not an inherent part of the models.

In conclusion, I would not make verdict over the models, instead emphasise the different functions they can play in scientific communication.

A minor comment: I found that a number of statements lack references in the Introduction. I would have found them useful for statements such as “There is a point of view that peer review is included in the implicit contract of the researcher.”

Competing interests: None.

Peer review 2

Martin Bush

DOI: 10.70744/MetaROR.9.1.rv2

In “Evolution of Peer Review in Scientific Communication”, Kochetkov provides a point-of-view discussion of the current state of play of peer review for scientific literature, focussing on the major models in contemporary use and recent innovations in reform. In particular, they present a typology of three main forms of peer review: traditional pre-publication review; registered reports; and post-publication review, their preferred model. The main contribution it could make would be to help consolidate typologies and terminologies, to consolidate major lines of argument and to present some useful visualisations of these. On the other hand, the overall discussion is not strongly original in character.

The major strength of this article is that the discussion is well-informed by contemporary developments in peer-review reform. The typology presented is modest and, for that, readily comprehensible and intuitive. This is to some extent a weakness as well as a strength; a typology that is too straightforward may not be useful enough. As suggested at the end it might be worth considering how to complexify the typology at least at subordinate levels without sacrificing this strength. The diagrams of workflows are particularly clear.

The primary weakness of this article is that it presents itself as an ‘analysis’ from which they ‘conclude’ certain results such as their typology, when this appears clearly to be an opinion piece. In my view, this results in a false claim of objectivity which detracts from what would otherwise be an interesting and informative, albeit subjective, discussion, and thus fails to discuss the limitations of this approach. A secondary weakness is that the discussion is not well structured and there are some imprecisions of expression that have the potential to confuse, at least at first.

This primary weakness is manifested in several ways. The evidence and reasoning for claims made is patchy or absent. One instance of the former is the discussion of bias in peer review. There are a multitude of studies of such bias and indeed quite a few meta-analyses of these studies. A systematic search could have been done here but there is no attempt to discuss the totality of this literature. Instead, only a few specific studies are cited. Why are these ones chosen? We have no idea. To this extent I am not convinced that the references used here are the most appropriate. Instances of the latter are the claim that “The most well-known initiatives at the moment are ResearchEquals and Octopus” for which no evidence is provided, the claim that “we believe that journal-independent peer review is a special case of Model 3” for which no further argument is provided, and the claim that “the function of being the “supreme judge” in deciding what is “good” and “bad” science is taken on by peer review” for which neither is provided.

A particular example of this weakness, which is perhaps of marginal importance to the overall paper but of strong interest to this reviewer is the rather odd engagement with history within the paper. It is titled “Evolution of Peer Review” but is really focussed on the contemporary state-of-play. Section 2 starts with a short history of peer review in scientific publishing, but that seems intended only to establish what is described as the ‘traditional’ model of peer review. Given that that short history had just shown how peer review had been continually changing in character over centuries – and indeed Kochetkov goes on to describe further changes – it is a little difficult to work out what ‘traditional’ might mean here; what was ‘traditional’ in 2010 was not the same as what was ‘traditional’ in 1970. It is not clear how seriously this history is being taken. Kochetkov has earlier written that “as early as the beginning of the 21st century, it was argued that the system of peer review is ‘broken'” but of course criticisms – including fundamental criticisms – of peer review are much older than this. Overall, this use of history seems designed to privilege the experience of a particular moment in time, that coincides with the start of the metascience reform movement.

Section 2 also demonstrates some of the second weakness described, a rather loose structure. Having moved from a discussion of the history of peer review to detail the first model, ‘traditional’ peer review, it then also goes on to describe the problems of this model. This part of the paper is one of the best – and best -evidenced. Given the importance of it to the main thrust of the discussion it should probably have been given more space as a Section all on its own.

Another example is Section 4 on Modular Publishing, in which Kochetkov notes “Strictly speaking, modular publishing is primarily an innovative approach for the publishing workflow in general rather than specifically for peer review.” Kochetkov says “This is why we have placed this innovation in a separate category” but if it is not an innovation in peer review, the bigger question is ‘Why was it included in this article at all?’.

One example of the imprecisions of language is as follows. The author also shifts between the terms ‘scientific communication’ and ‘science communication’ but, at least in many contexts familiar to this reviewer, these are not the same things, the former denoting science-internal dissemination of results through publication (which the author considers), conferences and the like (which the author specifically excludes) while the latter denotes the science-external public dissemination of scientific findings to non-technical audiences, which is entirely out of scope for this article.

A final note is that Section 3, while an interesting discussion, seems largely derivative from a typology of Waltman, with the addition of a consideration of whether a reform is ‘radical’ or ‘incremental’, based on how ‘disruptive’ the reform is. Given that this is inherently a subjective decision, I wonder if it might not have been more informative to consider ‘disruptiveness’ on a scale and plot it accordingly. This would allow for some range to be imagined for each reform as well; surely reforms might be more or less disruptive depending on how they are implemented. Given that each reform is considered against each model, it is somewhat surprising that this is not presented in a tabular or graphical form.

Beyond the specific suggestions in the preceding paragraphs, my suggestions to improve this article are as follows:

  1. Reconceptualize this as an opinion piece. Where systematic evidence can be drawn upon to make points, use that, but don’t be afraid to just present a discussion from what is clearly a well-informed author.

  2. Reconsider the focus on history and ‘evolution’ if the point is about the current state of play and evaluation of reforms (much as I would always want to see more studies on the history and evolution of peer review).

  3. Consider ways in which the typology might be expanded, even if at subordinate level.

Competing interests: I have no competing interests in the compilation of this review, although I do have specific interests as noted above.

Peer review 3

Olmo R. van den Akker

DOI: 10.70744/MetaROR.9.1.rv3

In this manuscript, the author provides a historical review of the place of peer review in the scientific ecosystem, including a discussion of the so-called current crisis and a presentation of three important peer review models. I believe this is a non-comprehensive yet useful overview. My main contention is that the structure of the paper could be improved. More specifically, the author could expand on the different goals of peer review and discuss these goals earlier in the paper. This would allow readers to better interpret the different issues plaguing peer review and helps put the costs and benefits of the three models into context. Other than that, I found some claims made in the paper a little too strong. Presenting some empirical evidence or downplaying these claims would improve the manuscript in my opinion. Below, you can find my comments:

  1. In my view, the biggest issue with the current peer review system is the low quality of reviews, but the manuscript only mentions this fleetingly. The current system facilitates publication bias, confirmation bias, and is generally very inconsistent. I think this is partly due to reviewers’ lack of accountability in such a closed peer review system, but I would be curious to hear the author’s ideas about this, more elaborately than they provide them as part of issue 2.

  2. I’m missing a section in the introduction on what the goals of peer review are or should be. You mention issues with peer review, and these are mostly fair, but their importance is only made salient if you link them to the goals of peer review. The author does mention some functions of peer review later in the paper, but I think it would be good to expand that discussion and move it to a place earlier in the manuscript.

  3. Table 1 is intuitive but some background on how the author arrived at these categorizations would be welcome. When is something incremental and when is something radical? Why are some innovations included but not others (e.g., collaborative peer review, see https://content.prereview.org/how-collaborative-peer-review-can-transform-scientific-research/)?

  4. “Training of reviewers through seminars and online courses is part of the strategies of many publishers. At the same time, we have not been able to find statistical data or research to assess the effectiveness of such training.” (p. 5)  There is some literature on this, although not recent. See work by Sara Schroter for example, Schroter et al., 2004; Schroter et al., 2008)

  5. “It should be noted that most initiatives aimed at improving the quality of peer review simultaneously increase the costs.” (p. 7)  This claim needs some support. Please explicate why this typically is the case and how it should impact our evaluations of these initiatives.

  6. I would rephrase “Idea of the study” in Figure 2 since the other models start with a tangible output (the manuscript). This is the same for registered reports where they submit a tangible report including hypotheses, study design, and analysis plan. In the same vein, I think study design in the rest of the figure might also not be the best phrasing.  Maybe the author could use the terminology used by COS (Stage 1 manuscript, and Stage 2 manuscript, see Details & Workflow tab of https://www.cos.io/initiatives/registered-reports). Relatedly, “Author submits the first version of the manuscript” in the first box after the ‘Manuscript (report)’ node maybe a confusing phrase because I think many researchers see the first version of the manuscript as the stage 1 report sent out for stage 1 review.

  7. One pathway that is not included in Figure 2 is that authors can decide to not conduct the study when improvements are required. Relatedly, in the publish-review-curate model, is revising the manuscripts based on the reviews not optional as well? Especially in the case of 3a, authors can hardly be forced to make changes even though the reviews are posted on the platform.

  8. I think the author should discuss the importance of ‘open identities’ more. This factor is now not explicitly included in any of the models, while it has been found to be one of the main characteristics of peer review systems (Ross-Hellauer, 2017). More generally, I was wondering why the author chose these three models and not others. What were the inclusion criteria for inclusion in the manuscript? Some information on the underlying process would be welcome, especially when claims like “However, we believe that journal-independent peer review is a special case of Model 3 (“Publish-Review-Curate”).” are made without substantiation.

  9. Maybe it helps to outline the goals of the paper a bit more clearly in the introduction. This helps the reader to know what to expect.

  10. The Modular Publishing section is not inherently related to peer review models, as you mention in the first sentence of that paragraph. As such, I think it would be best to omit this section entirely to maintain the flow of the paper. Alternatively, you could shortly discuss it in the discussion section but a separate paragraph seems too much from my point of view.

  11. Labeling model 3 as post-publication review might be confusing to some readers. I believe many researchers see post-publication review as researchers making comments on preprints, or submitting commentaries to journals. Those activities are substantially different from the publish-review-curate model so I think it is important to distinguish between these types.

  12. I do not think the conclusions drawn below Table 3 logically follow from the earlier text. For example, why are “all functions of scientific communication implemented most quickly and transparently in Model 3”? It could be that the entire process takes longer in Model 3 (e.g. because reviewers need more time), so that Model 1 and Model 2 lead to outputs quicker. The same holds for the following claim: “The additional costs arising from the independent assessment of information based on open reviews are more than compensated by the emerging opportunities for scientific pluralism.” What is the empirical evidence for this? While I personally do think that Model 3 improves on Model 1, emphatic statements like this require empirical evidence. Maybe the author could provide some suggestions on how we can attain this evidence. Model 2 does have some empirical evidence underpinning its validity (see Scheel, Schijen, Lakens, 2021; Soderberg et al., 2021; Sarafoglou et al. 2022) but more meta-research inquiries into the effectiveness and cost-benefits ratio of registered reports would still be welcome in general.

  13. What is the underlaying source for the claim that openness requires three conditions?

  14. “If we do not change our approach, science will either stagnate or transition into other forms of communication.” (p. 2)  I don’t think this claim is supported sufficiently strongly. While I agree there are important problems in peer review, I think would need to be a more in-depth and evidence-based analysis before claims like this can be made.

  15. On some occasions, the author uses “we” while the study is single authored.

  16. Figure 1: The top-left arrow from revision to (re-)submission is hidden

  17. “The low level of peer review also contributes to the crisis of reproducibility in scientific research (Stoddart, 2016).” (p. 4)  I assume the author means the low quality of peer review.

  18. “Although this crisis is due to a multitude of factors, the peer review system bears a significant responsibility for it.” (p. 4)  This is also a big claim that is not substantiated

  19. “Software for automatic evaluation of scientific papers based on artificial intelligence (AI) has emerged relatively recently” (p. 5)  The author could add RegCheck (https://regcheck.app/) here, even though it is still in development. This tool is especially salient in light of the finding that preregistration-paper checks are rarely done as part of reviews (see Syed, 2023)

  20. There is a typo in last box of Figure 1 (“decicion” instead of “decision”). I also found typos in the second box of Figure 2, where “screns” should be “screens”, and the author decision box where “desicion” should be “decision”

  21. Maybe it would be good to mention results blinded review in the first paragraph of 3.2. This is a form of peer review where the study is already carried out but reviewers are blinded to the results. See work by Locascio (2017), Grand et al. (2018), and Woznyj et al. (2018).

  22. Is “Not considered for peer review” in figure 3b not the same as rejected? I feel that it is rejected in the sense that neither the manuscript not the reviews will be posted on the platform.

  23. “In addition to the projects mentioned, there are other platforms, for example, PREreview12, which departs even more radically from the traditional review format due to the decentralized structure of work.” (p. 11)  For completeness, I think it would be helpful to add some more information here, for example why exactly decentralization is a radical departure from the traditional model.

  24. “However, anonymity is very conditional – there are still many “keys” left in the manuscript, by which one can determine, if not the identity of the author, then his country, research group, or affiliated organization.” (p.11)  I would opt for the neutral “their” here instead of “his”, especially given that this is a paragraph about equity and inclusion.

  25. “Thus, “closeness” is not a good way to address biases.” (p. 11)  This might be a straw man argument because I don’t believe researchers have argued that it is a good method to combat biases. If they did, it would be good to cite them here. Alternatively, the sentence could be omitted entirely.

  26. I would start the Modular Publishing section with the definition as that allows readers to interpret the other statements better.

  27. It would be helpful if the Models were labeled (instead of using Model 1, Model 2, and Model 3) so that readers don’t have to think back what each model involved.

  28. Table 2: “Decision making” for the editor’s role is quite broad, I recommend to specify and include what kind of decisions need to be made.

  29. Table 2: “Aim of review” – I believe the aim of peer review differs also within these models (see the “schools of thought” the author mentions earlier), so maybe a statement on what the review entails would be a better way to phrase this.

  30. Table 2: One could argue that the object of the review’ in Registered Reports is also the manuscript as a whole, just in different stages. As such, I would phrase this differently.

Good luck with any revision!

Olmo van den Akker (ovdakker@gmail.com)

References

Grand, J. A., Rogelberg, S. G., Banks, G. C., Landis, R. S., & Tonidandel, S. (2018). From outcome to process focus: Fostering a more robust psychological science through registered reports and results-blind reviewing. Perspectives on Psychological Science, 13(4), 448-456.

Ross-Hellauer, T. (2017). What is open peer review? A systematic review. F1000Research, 6.

Sarafoglou, A., Kovacs, M., Bakos, B., Wagenmakers, E. J., & Aczel, B. (2022). A survey on how preregistration affects the research workflow: Better science but more work. Royal Society Open Science, 9(7), 211997.

Scheel, A. M., Schijen, M. R., & Lakens, D. (2021). An excess of positive results: Comparing the standard psychology literature with registered reports. Advances in Methods and Practices in Psychological Science, 4(2), 25152459211007467.

Schroter, S., Black, N., Evans, S., Carpenter, J., Godlee, F., & Smith, R. (2004). Effects of training on quality of peer review: randomised controlled trial. Bmj, 328(7441), 673.

Schroter, S., Black, N., Evans, S., Godlee, F., Osorio, L., & Smith, R. (2008). What errors do peer reviewers detect, and does training improve their ability to detect them?. Journal of the Royal Society of Medicine, 101(10), 507-514.

Soderberg, C. K., Errington, T. M., Schiavone, S. R., Bottesini, J., Thorn, F. S., Vazire, S., … & Nosek, B. A. (2021). Initial evidence of research quality of registered reports compared with the standard publishing model. Nature Human Behaviour, 5(8), 990-997.

Syed, M. (2023). Some data indicating that editors and reviewers do not check preregistrations during the review process. PsyArXiv Preprints.

Locascio, J. J. (2017). Results blind science publishing. Basic and applied social psychology, 39(5), 239-246.

Woznyj, H. M., Grenier, K., Ross, R., Banks, G. C., & Rogelberg, S. G. (2018). Results-blind review: A masked crusader for science. European Journal of Work and Organizational Psychology, 27(5), 561-576.

Competing interests: None.

Peer review 4

Anonymous User

DOI: 10.70744/MetaROR.9.1.rv4

Overall thoughts: This is an interesting history piece regarding peer review and the development of review over time. Given the author’s conflict of interest and association with the Centre developing MetaROR, I think that this paper might be a better fit for an information page or introduction to the journal and rationale for the creation of MetaROR, rather than being billed as an independent article. Alternatively, more thorough information about advantages to pre-publication review or more downsides/challenges to post-publication review might make the article seem less affiliated. I appreciate seeing the history and current efforts to change peer review, though I am not comfortable broadly encouraging use of these new approaches based on this article alone.

Page 3: It’s hard to get a feel for the timeline given the dates that are described. We have peer review becoming standard after WWII (after 1945), definitively established by the second half of the century, an example of obligatory peer review starting in 1976, and in crisis by the end of the 20th century. I would consider adding examples that better support this timeline – did it become more common in specific journals before 1976? Was the crisis by the end of the 20th century something that happened over time or something that was already intrinsic to the institution? It doesn’t seem like enough time to get established and then enter crisis, but more details/examples could help make the timeline clear.

Consider discussing the benefits of the traditional model of peer review.

Table 1 – Most of these are self-explanatory to me as a reader, but not all. I don’t know what a registered report refers to, and it stands to reason that not all of these innovations are familiar to all readers. You do go through each of these sections, but that’s not clear when I initially look at the table. Consider having a more informative caption. Additionally, the left column is “Course of changes” here but “Directions” in text. I’d pick one and go with it for consistency.

3.2: Considering mentioning your conflict of interest here where MetaROR is mentioned.

With some of these methods, there’s the ability to also submit to a regular journal. Going to a regular journal presumably would instigate a whole new round of review, which may or may not contradict the previous round of post-publication review and would increase the length of time to publication by going through both types. If someone has a goal to publish in a journal, what benefit would they get by going through the post-publication review first, given this extra time?

There’s a section talking about institutional change (page 14). It mentions that openness requires three conditions – people taking responsibility for scientific communication, authors and reviewers, and infrastructure. I would consider adding some discussion of readers and evaluators. Readers have to be willing to accept these papers as reliable, trustworthy, and respectable to read and use the information in them. Evaluators such as tenure committees and potential employers would need to consider papers submitted through these approaches as evidence of scientific scholarship for the effort to be worthwhile for scientists.

Based on this overview, which seems somewhat skewed towards the merits of these methods (conflict of interest, limited perspective on downsides to new methods/upsides to old methods), I am not quite ready to accept this effort as equivalent of a regular journal and pre-publication peer review process. I look forward to learning more about the approach and seeing this review method in action and as it develops.

Competing interests: None.

Author response

DOI: 10.70744/MetaROR.9.1.ar

Response to the Editors and the Reviewers

I am sincerely grateful to the editors and peer reviewers at MetaROR for their detailed feedback and valuable comments and suggestions. I have addressed each point below.

Handling editor

1. “However, the article’s progression and arguments, along with what it seeks to contribute to the literature need refinement and clarification. The argument for PRC is under-developed due to a lack of clarity about what the article means by scientific communication. Clarity here might make the endorsement of PRC seem like less of a foregone conclusion.”

The structure of the paper (and discussion) has changed significantly to address the feedback.

2. “I strongly endorse the main theme of most of the reviews, which is that the progression and underlying justifications for this article’s arguments needs a great deal of work. In my view, this article’s main contribution seems to be the evaluation of the three peer review models against the functions of scientific communication. I say ‘seems to be’ because the article is not very clear on that and I hope you will consider clarifying what your manuscript seeks to add to the existing work in this field. In any case, if that assessment of the three models is your main contribution, that part is somewhat underdeveloped. Moreover, I never got the sense that there is clear agreement in the literature about what the tenets of scientific communication are. Note that scientific communication is a field in its own right.”

I have implemented a more rigorous approach to argumentation in response. “Scientific communication” was replaced by “scholarly communication.”

3. “I also agree that paper is too strongly worded at times, with limitations and assumptions in the analysis minimised or not stated. For example, all of the typologies and categories drawn could easily be reorganised and there is a high degree of subjectivity in this entire exercise. Subjective choices should be highlighted and made salient for the reader. Note that greater clarity, rigour, and humility may also help with any alleged or actual bias.”

I have incorporated the conceptual framework and description of the research methodology. However, the Discussion section reflects my personal perspective in some points, which I have explicitly highlighted to ensure clarity.

4. “I agree with Reviewer 3 that the ‘we’ perspective is distracting.”

This has been fixed.

5. “The paragraph starting with ‘Nevertheless’ on page 2 is very long.”

The text was restructured.

6. “There are many points where language could be shortened for readability, for example:

Page 3: ‘decision on publication’ could be ‘publication decision’.

Page 5: ‘efficiency of its utilization’ could be ‘its efficiency’.

Page 7: ‘It should be noted…’ could be ‘Note that…’.”

I have proofread the text.

7. “Page 7: ‘It should be noted that..’ – this needs a reference.”

This statement has been moved to the Discussion section, paraphrased, and reference added

“It should be also noted that peer review innovations pull in opposing directions, with some aiming to increase efficiency and reduce costs, while others aim to promote rigor and increase costs (Kaltenbrunner et al., 2022).”

8. “I’m not sure that registered reports reflect a hypothetico-deductive approach (page 6). For instance, systematic reviews (even non-quantitative ones) are often published as registered reports and Cochrane has required this even before the move towards registered reports in quantitative psychology.”

I have added this clarification.

9. “I agree that modular publishing sits uneasily as its own chapter.”

Modular publishing has been combined with registered reports into the deconstructed publication group of models, now Section 5.1.

10. “Page 14: ‘The “Publish-Review-Curate” model is universal that we expect to be the future of scientific publishing. The transition will not happen today or tomorrow, but in the next 5-10 years, the number of projects such as eLife, F1000Research, Peer Community in, or MetaROR will rapidly increase’. This seems overly strong (an example of my larger critique and that of the reviewers).”

This part of the text has been rewritten.

Reviewer 1

11. “For example, although Model 3 is less chance to insert bias to the readers, it also weakens the filtering function of the review system. Let’s just think about the dangers of machine-generated articles, paper-mills, p-hacked research reports and so on. Although the editors do some pre-screening for the submissions, in a world with only Model 3 peer review the literature could easily get loaded with even more ‘garbage’ than in a model where additional peers help the screening.”

I think that generated text is better detected by software tools. At the same time, I tried and described the pros and cons of different models in a more balanced way in the concluding section.

12. “Compared to registered reports other aspects can come to focus that Model 3 cannot cover. It’s the efficiency of researchers’ work. In the care of registered reports, Stage 1 review can still help researchers to modify or improve their research design or data collection method. Empirical work can be costly and time-consuming and post-publication review can only say that ‘you should have done it differently then it would make sense’.”

Thank you very much for this valuable contribution, I have added this statement at P. 11.

13. “Finally, the author puts openness as a strength of Model 3. In my eyes, openness is a separate question. All models can work very openly and transparently in the right circumstances. This dimension is not an inherent part of the models.”

I think that the model, providing peer reviews to all the submissions, ensures maximum transparency. However, I have made effort to make the wording more balanced and distinguish my personal perspective from the literature.

14. “In conclusion, I would not make verdict over the models, instead emphasize the different functions they can play in scientific communication.”

This idea has been reflected now in the concluding section.

15. “A minor comment: I found that a number of statements lack references in the Introduction. I would have found them useful for statements such as ‘There is a point of view that peer review is included in the implicit contract of the researcher.’”

Thank you for your feedback. I have implemented a more rigorous approach to argumentation in response.

Reviewer 2

16. “The primary weakness of this article is that it presents itself as an ‘analysis’ from which they ‘conclude’ certain results such as their typology, when this appears clearly to be an opinion piece. In my view, this results in a false claim of objectivity which detracts from what would

otherwise be an interesting and informative, albeit subjective, discussion, and thus fails to discuss the limitations of this approach.”

I have incorporated the conceptual framework and description of the research methodology. However, the Discussion section reflects my personal perspective in some points, which I have explicitly highlighted to ensure clarity.

17. “A secondary weakness is that the discussion is not well structured and there are some imprecisions of expression that have the potential to confuse, at least at first.”

The structure of the paper (and discussion) has changed significantly.

18. “The evidence and reasoning for claims made is patchy or absent. One instance of the former is the discussion of bias in peer review. There are a multitude of studies of such bias and indeed quite a few meta-analyses of these studies. A systematic search could have been done here but there is no attempt to discuss the totality of this literature. Instead, only a few specific studies are cited. Why are these ones chosen? We have no idea. To this extent I am not convinced that the references used here are the most appropriate.”

I have reviewed the existing references and incorporated additional sources. However, the study does not claim to conduct a systematic literature review; rather, it adopts an interpretative approach to literature analysis.

19. “Instances of the latter are the claim that ‘The most well-known initiatives at the moment are ResearchEquals and Octopus’ for which no evidence is provided, the claim that ‘we believe that journal-independent peer review is a special case of Model 3’ for which no further argument is provided, and the claim that ‘the function of being the “supreme judge” in deciding what is “good” and “bad” science is taken on by peer review’ for which neither is provided.

Thank you for your feedback. I have implemented a more rigorous approach to argumentation in response.

20. “A particular example of this weakness, which is perhaps of marginal importance to the overall paper but of strong interest to this reviewer is the rather odd engagement with history within the paper. It is titled “Evolution of Peer Review” but is really focussed on the contemporary state-of-play. Section 2 starts with a short history of peer review in scientific publishing, but that seems intended only to establish what is described as the ‘traditional’ model of peer review. Given that that short history had just shown how peer review had been continually changing in character over centuries – and indeed Kochetkov goes on to describe further changes – it is a little difficult to work out what ‘traditional’ might mean here; what was ‘traditional’ in 2010 was not the same as what was ‘traditional’ in 1970. It is not clear how seriously this history is being taken. Kochetkov has earlier written that “as early as the beginning of the 21st century, it was argued that the system of peer review is ‘broken'” but of course criticisms – including fundamental criticisms – of peer review are much older than this. Overall, this use of history seems designed to privilege the experience of a particular moment in time, that coincides with the start of the metascience reform movement.”

While the paper addresses some aspects of peer review history, it does not provide a comprehensive examination of this topic. A clarifying statement to this effect has been included in the methodology section.

“… this section incorporates elements of historical analysis, it does not fully qualify as such because primary sources were not directly utilized. Instead, it functions as an interpretative literature review, and one that is intentionally concise, as a comprehensive history of peer review falls outside the scope of this research”.

21. “Section 2 also demonstrates some of the second weakness described, a rather loose structure. Having moved from a discussion of the history of peer review to detail the first model, ‘traditional’ peer review, it then also goes on to describe the problems of this model. This part of the paper is one of the best – and best – evidenced. Given the importance of it to the main thrust of the discussion it should probably have been given more space as a Section all on its own.”

This section (now Section 4) has been extended, see also previous comment.

22. “Another example is Section 4 on Modular Publishing, in which Kochetkov notes “Strictly speaking, modular publishing is primarily an innovative approach for the publishing workflow in general rather than specifically for peer review.” Kochetkov says “This is why we have placed this innovation in a separate category” but if it is not an innovation in peer review, the bigger question is ‘Why was it included in this article at all?’.”

Modular publishing has been combined with registered reports into the deconstructed publication group of models, now Section 5.1.

23. “One example of the imprecisions of language is as follows. The author also shifts between the terms ‘scientific communication’ and ‘science communication’ but, at least in many contexts familiar to this reviewer, these are not the same things, the former denoting science-internal dissemination of results through publication (which the author considers), conferences and the like (which the author specifically excludes) while the latter denotes the science-external public dissemination of scientific findings to non-technical audiences, which is entirely out of scope for this article.”

Thank you for your remark. As a non- native speaker, I initially did not grasp the distinction between the terms. However, I believe the phrase ‘scholarly communication’ is the most universally applicable term. This adjustment has now been incorporated into the text.

24. “A final note is that Section 3, while an interesting discussion, seems largely derivative from a typology of Waltman, with the addition of a consideration of whether a reform is ‘radical’ or ‘incremental’, based on how ‘disruptive’ the reform is. Given that this is inherently a subjective decision, I wonder if it might not have been more informative to consider ‘disruptiveness’ on a scale and plot it accordingly. This would allow for some range to be imagined for each reform as well; surely reforms might be more or less disruptive depending on how they are implemented. Given that each reform is considered against each model, it is somewhat surprising that this is not presented in a tabular or graphical form.”

Ultimately, I excluded this metric due to its current reliance on purely subjective judgment. Measuring ‘disruptiveness’, e.g., through surveys or interviews remains a task for future research.

25. “Reconceptualize this as an opinion piece. Where systematic evidence can be drawn upon to make points, use that, but don’t be afraid to just present a discussion from what is clearly a well-informed author.”

I cannot definitively classify this work as an opinion piece. In fact, this manuscript synthesizes elements of a literature review, research article, and opinion essay. My idea was to integrate the strengths of all three genres.

26. “Reconsider the focus on history and ‘evolution’ if the point is about the current state of play and evaluation of reforms (much as I would always want to see more studies on the history and evolution of peer review).”

I have revised the title to better reflect the study’s scope and explicitly emphasize its focus on contemporary developments in the field.

“Peer Review at the Crossroads”

27. “Consider ways in which the typology might be expanded, even if at subordinate level.”

I have updated the typology and introduced the third tier, where it is applicable (see Fig.2).

Reviewer 3

28. “In my view, the biggest issue with the current peer review system is the low quality of reviews, but the manuscript only mentions this fleetingly. The current system facilitates publication bias, confirmation bias, and is generally very inconsistent. I think this is partly due to reviewers’ lack of accountability in such a closed peer review system, but I would be curious to hear the author’s ideas about this, more elaborately than they provide them as part of issue 2.

I have elaborated on this issue in the footnote.

29. “I’m missing a section in the introduction on what the goals of peer review are or should be. You mention issues with peer review, and these are mostly fair, but their importance is only made salient if you link them to the goals of peer review. The author does mention some functions of peer review later in the paper, but I think it would be good to expand that discussion and move it to a place earlier in the manuscript.”

The functions of peer review are summarized in the first paragraph of Introduction.

30. “Table 1 is intuitive but some background on how the author arrived at these categorizations would be welcome. When is something incremental and when is something radical? Why are some innovations included but not others (e.g., collaborative peer review, see https://content.prereview.org/how-collaborative-peer-review-can-transform-scientific-research/)?”

Collaborative peer review, namely, Prereview was mentioned in the context of Model 3 (Publish-Review-Curate). However, I have extended this part of the paper.

31“‘Training of reviewers through seminars and online courses is part of the strategies of many publishers. At the same time, we have not been able to find statistical data or research to assess the effectiveness of such training.’ (p. 5)  There is some literature on this, although not recent. See work by Sara Schroter for example, Schroter et al., 2004; Schroter et al., 2008)”

Thank you very much, I have added these studies and a few more recent ones.

32. “‘It should be noted that most initiatives aimed at improving the quality of peer review simultaneously increase the costs.’ (p. 7) This claim needs some support. Please explicate why this typically is the case and how it should impact our evaluations of these initiatives.”

I have moved this part to the Discussion section.

33. “I would rephrase “Idea of the study” in Figure 2 since the other models start with a tangible output (the manuscript). This is the same for registered reports where they submit a tangible report including hypotheses, study design, and analysis plan. In the same vein, I think study design in the rest of the figure might also not be the best phrasing. Maybe the author could use the terminology used by COS (Stage 1 manuscript, and Stage 2 manuscript, see Details & Workflow tab of https://www.cos.io/initiatives/registered-reports). Relatedly, “Author submits the first version of the manuscript” in the first box after the ‘Manuscript (report)’ node maybe a confusing phrase because I think many researchers see the first version of the manuscript as the stage 1 report sent out for stage 1 review.”

Thank you very much. Stage 1 and Stage 2 manuscripts look like suitable labelling solution.

34. “One pathway that is not included in Figure 2 is that authors can decide to not conduct the study when improvements are required. Relatedly, in the publish-review-curate model, is revising the manuscripts based on the reviews not optional as well? Especially in the case of

3a, authors can hardly be forced to make changes even though the reviews are posted on the platform.”

All the four models imply a certain level of generalization; thus, I tried to avoid redundant details. However, I have added this choice to the PRC model (now, Model 4).

35. “I think the author should discuss the importance of ‘open identities’ more. This factor is now not explicitly included in any of the models, while it has been found to be one of the main characteristics of peer review systems (Ross-Hellauer, 2017).”

This part has been extended.

36. “More generally, I was wondering why the author chose these three models and not others. What were the inclusion criteria for inclusion in the manuscript? Some information on the underlying process would be welcome, especially when claims like ‘However, we believe that journal-independent peer review is a special case of Model 3 (‘Publish-Review-Curate’).’ are made without substantiation.”

The study included four generalized models of peer review that involved some level of abstraction.

37. “Maybe it helps to outline the goals of the paper a bit more clearly in the introduction. This helps the reader to know what to expect.”

The Introduction has been revised including the goal and objectives.

38. “The Modular Publishing section is not inherently related to peer review models, as you mention in the first sentence of that paragraph. As such, I think it would be best to omit this section entirely to maintain the flow of the paper. Alternatively, you could shortly discuss it in the discussion section but a separate paragraph seems too much from my point of view.”

Modular publishing has been combined with registered reports into the fragmented publishing group of models, now in Section 5.

39. “Labeling model 3 as post-publication review might be confusing to some readers. I believe many researchers see post-publication review as researchers making comments on preprints, or submitting commentaries to journals. Those activities are substantially different from the publish-review-curate model so I think it is important to distinguish between these types.”

The label was changed into Publish- Review-Curate model.

40. “I do not think the conclusions drawn below Table 3 logically follow from the earlier text. For example, why are “all functions of scientific communication implemented most quickly and transparently in Model 3”? It could be that the entire process takes longer in Model 3 (e.g. because reviewers need more time), so that Model 1 and Model 2 lead to outputs quicker. The same holds for the following claim: ‘The additional costs arising from the independent assessment of information based on open reviews are more than compensated by the emerging opportunities for scientific pluralism.’ What is the empirical evidence for this? While I personally do think that Model 3 improves on Model 1, emphatic statements like this require empirical evidence. Maybe the author could provide some suggestions on how we can attain this evidence. Model 2 does have some empirical evidence underpinning its validity (see Scheel, Schijen, Lakens, 2021; Soderberg et al., 2021; Sarafoglou et al. 2022) but more meta-research inquiries into the effectiveness and cost-benefits ratio of registered reports would still be welcome in general.”

The Discussion section has been substantially revised to address this point. While I acknowledge the current scarcity of empirical studies on innovative peer review models, I have incorporated a critical discussion of this methodological gap. I am grateful for the suggested literature on RRs, which I have now integrated into the relevant subsection.

41. “What is the underlaying source for the claim that openness requires three conditions?”

I have made effort to clarify within the text that this reflects my personal stance.

42. “‘If we do not change our approach, science will either stagnate or transition into other forms of communication.’ (p. 2) I don’t think this claim is supported sufficiently strongly. While I agree there are important problems in peer review, I think would need to be a more in-depth and evidence-based analysis before claims like this can be made.”

The sentence has been rephrased.

43. “On some occasions, the author uses ‘we’ while the study is single authored.”

This has been fixed.

44. “Figure 1: The top-left arrow from revision to (re-)submission is hidden”

I have updated Figure 1.

45. “‘The low level of peer review also contributes to the crisis of reproducibility in scientific research (Stoddart, 2016).’ (p. 4) I assume the author means the low quality of peer review.”

This has been fixed.

46. “‘Although this crisis is due to a multitude of factors, the peer review system bears a significant responsibility for it.’ (p. 4) This is also a big claim that is not substantiated”

I have paraphrased this sentence as “While multiple factors drive this crisis, deficiencies in the peer review process remain a significant contributor.” and added a footnote.

47. “‘Software for automatic evaluation of scientific papers based on artificial intelligence (AI) has emerged relatively recently” (p. 5) The author could add RegCheck (https://regcheck.app/) here, even though it is still in development. This tool is especially salient in light of the finding that preregistration-paper checks are rarely done as part of reviews (see Syed, 2023)”

Thank you very much, I have added this information.

48. “There is a typo in last box of Figure 1 (‘decicion’ instead of ‘decision’). I also found typos in the second box of Figure 2, where ‘screns’ should be ‘screens’, and the author decision box where ‘desicion’ should be ‘decision’”

This has been fixed.

49. “Maybe it would be good to mention results blinded review in the first paragraph of 3.2. This is a form of peer review where the study is already carried out but reviewers are blinded to the results. See work by Locascio (2017), Grand et al. (2018), and Woznyj et al. (2018).”

Thanks, I have added this (now section 5.2)

50. “Is ‘Not considered for peer review’ in figure 3b not the same as rejected? I feel that it is rejected in the sense that neither the manuscript not the reviews will be posted on the platform.”

Changed into “Rejected”

51. “‘In addition to the projects mentioned, there are other platforms, for example, PREreview12, which departs even more radically from the traditional review format due to the decentralized structure of work.’ (p. 11) For completeness, I think it would be helpful to add some more information here, for example why exactly decentralization is a radical departure from the traditional model.”

I have extended this passage.

52. “‘However, anonymity is very conditional – there are still many “keys” left in the manuscript, by which one can determine, if not the identity of the author, then his country, research group, or affiliated organization.’ (p.11) I would opt for the neutral ‘their’ here instead of ‘his’, especially given that this is a paragraph about equity and inclusion.”

This has been fixed.

53. “‘Thus, “closeness” is not a good way to address biases.’ (p. 11) This might be a straw man argument because I don’t believe researchers have argued that it is a good method to combat biases. If they did, it would be good to cite them here. Alternatively, the sentence could be

omitted entirely.

I have omitted the sentence.

54. “I would start the Modular Publishing section with the definition as that allows readers to interpret the other statements better.”

Modular publishing has been combined with registered reports into the deconstructed publication group of models, now in Section 5, general definition added.

55. “It would be helpful if the Models were labeled (instead of using Model 1, Model 2, and Model 3) so that readers don’t have to think back what each model involved.”

All the models represent a kind of generalization, which is why non-detailed labels are used. The text labels may vary depending on the context.

56. “Table 2: ‘Decision making’ for the editor’s role is quite broad, I recommend to specify and include what kind of decisions need to be made.”

Changed into “Making accept/reject decisions”

57. “Table 2: ‘Aim of review’ – I believe the aim of peer review differs also within these models (see the ‘schools of thought’ the author mentions earlier), so maybe a statement on what the review entails would be a better way to phrase this.”

Changed into “What does peer review entail?”

58. “Table 2: One could argue that the object of the review’ in Registered Reports is also the manuscript as a whole, just in different stages. As such, I would phrase this differently.

Current wording fits your remark: “Manuscript in terms of study design and execution”

Reviewer 4

59. “Page 3: It’s hard to get a feel for the timeline given the dates that are described. We have peer review becoming standard after WWII (after 1945), definitively established by the second half of the century, an example of obligatory peer review starting in 1976, and in crisis by the end of the 20th century. I would consider adding examples that better support this timeline – did it become more common in specific journals before 1976? Was the crisis by the end of the 20th century something that happened over time or something that was already intrinsic to the institution? It doesn’t seem like enough time to get established and then enter crisis, but more details/examples could help make the timeline clear. Consider discussing the benefits of the traditional model of peer review.”

This section has been extended.

60. “Table 1 – Most of these are self-explanatory to me as a reader, but not all. I don’t know what a registered report refers to, and it stands to reason that not all of these innovations are familiar to all readers. You do go through each of these sections, but that’s not clear when I initially look at the table. Consider having a more informative caption. Additionally, the left column is “Course of changes” here but “Directions” in text. I’d pick one and go with it for consistency.”

Table 1 has been replaced by Figure 2. I have also extended text descriptions, added definitions.

61. “With some of these methods, there’s the ability to also submit to a regular journal. Going to a regular journal presumably would instigate a whole new round of review, which may or may not contradict the previous round of post-publication review and would increase the length of time to publication by going through both types. If someone has a goal to publish in a journal, what benefit would they get by going through the post-publication review first, given this extra time?”

Some of these platforms, e.g., F1000, Lifecycle Journal, replace conventional journal publishing. Modular publishing allows for step-by-step feedback from peers. An important advantage of RRs over other peer review models lies in their capacity to enhance research efficiency. By conducting peer review at Stage 1, researchers gain the opportunity to refine their study design or data collection protocols before empirical work begins. Other models of review can offer critiques such as “the study should have been conducted differently” without actionable opportunity for improvement. The key motivation for having my paper reviewed in MetaROR is the quality of peer review – I have never received so many comments, frankly! Moreover, platforms such as MetaROR usually have partnering journals.

62. “There’s a section talking about institutional change (page 14). It mentions that openness requires three conditions – people taking responsibility for scientific communication, authors and reviewers, and infrastructure. I would consider adding some discussion of readers and evaluators. Readers have to be willing to accept these papers as reliable, trustworthy, and respectable to read and use the information in them. Evaluators such as tenure committees and potential employers would need to consider papers submitted through these approaches as evidence of scientific scholarship for the effort to be worthwhile for scientists.”

I have omitted these conditions and employed the Moore’s Technology Adoption Life Cycle. Thank you very much for your comment!

63. Based on this overview, which seems somewhat skewed towards the merits of these methods (conflict of interest, limited perspective on downsides to new methods/upsides to old methods), I am not quite ready to accept this effort as equivalent of a regular journal and pre-publication peer review process. I look forward to learning more about the approach and seeing this review method in action and as it develops.

The Discussion section has been substantially revised to address this point. While I acknowledge the current scarcity of empirical studies on innovative peer review models, I have incorporated a critical discussion of this methodological gap.

Leave a comment