Skip to main content

A new approach to grant review assessments: score, then rank

Abstract

Background

In many grant review settings, proposals are selected for funding on the basis of summary statistics of review ratings. Challenges of this approach (including the presence of ties and unclear ordering of funding preference for proposals) could be mitigated if rankings such as top-k preferences or paired comparisons, which are local evaluations that enforce ordering across proposals, were also collected and incorporated in the analysis of review ratings. However, analyzing ratings and rankings simultaneously has not been done until recently. This paper describes a practical method for integrating rankings and scores and demonstrates its usefulness for making funding decisions in real-world applications.

Methods

We first present the application of our existing joint model for rankings and ratings, the Mallows-Binomial, in obtaining an integrated score for each proposal and generating the induced preference ordering. We then apply this methodology to several theoretical “toy” examples of rating and ranking data, designed to demonstrate specific properties of the model. We then describe an innovative protocol for collecting rankings of the top-six proposals as an add-on to the typical peer review scoring procedures and provide a case study using actual peer review data to exemplify the output and how the model can appropriately resolve judges’ evaluations.

Results

For the theoretical examples, we show how the model can provide a preference order to equally rated proposals by incorporating rankings, to proposals using ratings and only partial rankings (and how they differ from a ratings-only approach) and to proposals where judges provide internally inconsistent ratings/rankings and outlier scoring. Finally, we discuss how, using real world panel data, this method can provide information about funding priority with a level of accuracy in a well-suited format for research funding decisions.

Conclusions

A methodology is provided to collect and employ both rating and ranking data in peer review assessments of proposal submission quality, highlighting several advantages over methods relying on ratings alone. This method leverages information to most accurately distill reviewer opinion into a useful output to make an informed funding decision and is general enough to be applied to settings such as in the NIH panel review process.

Peer Review reports

Background

Current practice and challenges in peer review

Many advances in biomedical knowledge and development of therapeutic treatments in the last 75 years rest on the bedrock of public research funding [1]. Funding is often distributed in the form of short-term grants for which scientists compete by submitting proposals, which are in turn often reviewed by their peers and evaluated for the most meritorious and scientifically fruitful ideas [2, 3]. To do this, research funders have a variety of evaluation mechanisms. A 2013 study compiled a list of the 55 world’s largest funders [4]. Among the largest, ordered by the amount allocated, are the US National Institutes of Health (NIH) [5], the European Commission [6], the United Kingdom Medical Research Council [7], the French National Institute of Health and Medical Research (INSERM) [8], the US Department of Defense [9], the UK-based Wellcome Trust [10], the Canadian Institutes of Health Research [11], the Australian National Health and Medical Research Council (NHMRC) [12], the US-based Howard Hughes Medical Institute (HHMI) [13], the German Research Foundation (DFG) [14] and the National Natural Science Foundation of China (NSFC) [15]. Of these, 64 percent clearly used a review process in which experts discussed and scored all of the qualifying proposals at a meeting (online or face-to-face). Other mechanisms included reviews of individual proposals (where independent evaluations are generated for a specific project) as well as the utilization of research boards, where sometimes a sub-set of qualifying applicants present their ideas directly to the board.

For funders that use panel review, like the NIH and the American Heart Association (AHA), peer reviewers are often asked to rate each proposal on a numeric scale, independently scoring proposals based on specific review criteria [16, 17]. However, the resulting scores from peer review ratings can depend on epistemological differences [18], personal preference [19] and varying levels of reviewer leniency or harshness [20]. Low inter-rater reliability is a long-standing concern for interpreting peer review results ([21,22,23,24]) although a recent study suggests that this measure is “not straightforward to use in practice in a typical peer review setting, and can be misleading” in assessing the quality of peer review ([25]). Moreover, some research suggests reviewer experiences can reduce disagreement in some instances ([26, 27]). However, even if one ignores reviewer characteristics that may influence their rating, the guidance for such rating systems often includes a great deal of ambiguity, allowing participants to produce a result that may have multiple interpretations. For instance, the NIH scoring guidance for research applications indicates that for an overall impact rating of 5, this could mean either that the application, if funded, “may make a contribution of high importance to the field, but weaknesses bring down the overall impact” OR that it “may make a contribution of moderate importance to the field, with some or no weaknesses [28].” Thus, the meaning of an individual rating of a proposal relative to it’s estimated worthiness of funding is not always completely clear, which is further muddied when incorporated with other reviewers’ ratings as a summary statistic and used as a basis for funding priority.

When rating the quality of a proposal, reviewers are also generally instructed to assess the likelihood a research project will advance a scientific field in an impactful way, and to consider weaknesses that may affect this likelihood [28]. This process often forces reviewers to penalize a project based on a series of identified but minor weaknesses, despite the fact that they may find the application overall to be potentially significant. Indeed, some reports suggest listed weaknesses in the text of reviewer’s critiques are better correlated with overall scores than that of listed strengths [21]. Thus, it may be that the rating process is not capturing a complete picture of a reviewer’s evaluation of a given proposal. The difficulty in interpreting these individual ratings as well as the summary statistics of the panel ratings (often used as the priority order) and translating them into a funding priority is problematic for research funders, especially when distinguishing between similarly rated proposals, especially for projects of high quality, where only a few minor weaknesses can make the difference between a funded project and a rejection. In fact, given current budgetary limitations, agencies find themselves pressed more and more to do just this [29]. Moreover, the lack of estimates of uncertainty in the scientific priority order means funders may be tempted to re-review similarly rated proposals, introducing potentially additional sources of variability to the funding decision. An exploration of alternate approaches to this decision-making process is needed not only to provide more accuracy and consistency but also to improve the credibility of an important gatekeeping function for science, grant review, which has been under fire [30, 31].

As mentioned above, while there are a variety of peer review styles and procedures used by different funders ([32]), many peer review processes use ratings of ‘proposal quality’ by all review panelists not in conflict (NIH estimates 20-40 reviewers; [33]) to inform the final funding decisions of the next level of committee, which are based on direct comparisons of top-tier proposals (e.g. Congressionally Directed Medical Research Programs [9]). NIH’s “Advice for Reviewers” specifies that when at review panel meetings, reviewers “[d]on’t compare one application to another – they should each be evaluated independently based on the review criteria” [16]. NSF instructs reviewers “to evaluate all proposals against two criteria,” intellectual merit and broader impacts, where excellence along these criteria are not characterized as being relative to other proposals [34]. However, psychological studies across a number of areas – job performance measurement, person perception, and attitude measurement – suggest that rankings are less noisy than ratings in the sense that people are less inconsistent when they are tasked with comparing two specific objects than when they need to separately assign each object a score on a scale [35]. Some have also suggested that the use of a comparative process may improve the resolution of the peer review system [25]. Along these lines, some funders include a ranking process in assessment. For instance the Netherlands Organisation of Health Research and Development (ZonMw) instructs their commission to rate proposals based on both quality and relevance, and then rank them based on a matrix of these two scores [36]. The Canadian Institutes of Health Research (CIHR) has proposals rated by one set of reviewers, then “binned” (yes/no categories) for consideration for funding by another set of reviewers who discuss the final funding priorities at a face-to-face meeting [36]. While some limited data suggests that ranking proposals against one another in review panels may help to reduce the impact of extreme reviewers as compared to rating systems [37], none have examined combining both processes of rating and ranking proposals by the same set of reviewers to augment funding decision making.

Ranking and Rating in Assessment

From a psychological perspective, the assessment tasks of ranking versus scoring are quite different. Research on performance and trait evaluation demonstrates that when assessing two objects along some dimension, evaluators can rank more highly the object that receives the lower score [38,39,40,41]. This happens in cases where evaluators score the objects relative to reference classes that do not explicitly include each other, and hence do not involve the direct comparisons required in ranking tasks. The same individual can arrive at what appear to be discrepant judgments about the same object: for example, a proposal may receive a good but not exceptional score and also receive a relatively high rank (top of the list), where this outcome is not psychologically inconsistent because the two evaluative tasks are different in kind.

Given the contrasting types of decision making in evaluative rating and comparative ranking, and given the discriminatory constraints of current review panels to assess proposals, it seems it would be important to capture both rating and ranking information from the same reviewer about a given proposal. Moreover, while collecting ratings provides information about the level of scientific excellence of an individual proposal relative to the goals of the funding program, and allows global comparisons of the magnitude of difference between proposals (e.g. between panels), collecting rankings allows for direct comparative statements about quality in a local sense (e.g. within a panel). Having both types of data could provide both a clearer separation between proposals in funding priority as well as an assessment on the global excellence of the science submitted for review and strives to glean as much information from a reviewer as possible to make the most informed funding decision for a given panel.

However, while the collection of rating data is already formalized at many funding agencies, the collection of ranking data is less common and thought must be given to the feasibility of the implementation strategy; for instance, some research suggests that ranking too many choices becomes difficult for assessors [42], which suggests the need for collecting partial rankings. Once collected, the data must be then aggregated and presented in an interpretable way that is useful for those making funding decisions. The statistical aggregation of ranking data is much more nuanced and involved than typical summary statistics associated with ratings, in part because there is provably no method which satisfies a collection of desirable criteria [43]. Furthermore, fitting statistical ranking models can be laborious and computationally expensive [44, 45]. Rating and ranking data must be combined to create a uniform, equitable approach across all proposals and to ensure the output is comprehensible, especially in situations where reviewers are internally inconsistent in their ratings and rankings. But this requires a sophisticated statistical method, especially in the case of incorporating partial rankings. Here we present a previously described methodology [46] for collecting both rating and ranking data from panel reviewers and for statistically modeling this information to create an ordered, funding priority list of proposals informed by both data sources. In this paper, we demonstrate how to apply this model to real data with procedures in data gathering that are relatively seamless with current review processes, and presentations of results and their interpretation that would be useful to aid funding decision makers.

Methods

Mallows-Binomial Model

The Mallows-Binomial model is a statistical model for identifying preferences and the level of consensus given both rankings and ratings and is described in detail in a previous publication [46]. The model was the first to jointly combine preferences from ordinal rankings and cardinal ratings into a single statistical analysis without performing data conversion. As such, the Mallows-Binomial provides a principled method for learning preferences with uncertainty from both rankings and ratings when available.

As suggested by its name, the Mallows-Binomial combines a Mallows ranking distribution with independent Binomial rating distributions. Information is shared between the two model components via shared parameters. Specifically, the model contains two parameters. The first is the vector-valued parameter \(p\in [0,1]^J\), where J is the number of proposals. We call p the integrated scores, which we seek to estimate. Each component \(p_j\), \(j=1,\dots ,J\) corresponds to the perceived quality of proposal j on the unit interval, where values closer to 0 indicate better quality and values closer to 1 indicate lower quality. As a result, a simple ordering of the proposals via their integrated scores allows one to identify a preference ordering (sometimes called the consensus ranking). We can interpret the integrated scores as representing global comparisons among the proposals and the induced preference ordering as representing local comparisons among the proposals. As such, the model parameter p summarizes both global and local preferences, just as do ratings and rankings, respectively. The second parameter of interest is \(\theta >0\), which we call the consensus scale parameter. Higher values indicate greater consensus among the rankings, while lower values indicate less consensus. The scale of this parameter may be hard to interpret and depends on both the number of proposals to be ranked and the size of the observed rankings [46]. Therefore, \(\theta\) will not be directly interpreted in this work. However, consensus is still reflected in the amount of uncertainty surrounding our estimates of the integrated scores, p. Among panels of similar sizes, the estimated parameters may still be compared to understand relative levels of consensus. Uncertainty associated with the estimated model parameters may be estimated via the nonparametric bootstrap [47]. The assumed data-generating process and model likelihood is provided in the Appendix.

A convenient aspect of the Mallows-Binomial is its ability to handle missing data. First, missing ratings and rankings for reasons unrelated to the perception of quality of those proposals (often referred to as “missing completely at random”) may be simply ignored without biasing parameter estimation. Furthermore, the model is able to handle both complete and top-k rankings, where \(1\le k\le J\). Furthermore, given the presence of conflicts of interest in the peer review process, the model is able to still jointly learn preferences given that different judges may be able to rank and rate slightly different sets of proposals. In such cases, the parameter \(\theta\) should not be directly interpreted but the key parameter of interest, p, may still be used for understanding preferences.

Additionally, the model assumes that ratings and rankings are conditionally independent given the integrated scores parameter p and consensus scale parameter \(\theta\). That is, the rankings and ratings of each judge need not be in alignment. Although this may initially seem to be an unnecessary feature of the model, internally inconsistent rankings and ratings arise frequently in practice (as seen in the case study presented in the following section). When such inconsistencies are present, other methods for jointly learning from rankings and ratings (such as the non-statistical approach presented in [48]) are not applicable. Nonetheless, this conditional independence assumption does not mean that ratings and ranking are completely independent as both of these preference measures are fully informed by the integrated scores p and the variability as ascribed by the Mallows-Binomial model.

The Mallows-Binomial model may be efficiently estimated using the publicly available R package rankrate [49]. For example, point estimation of model parameters took approximately 20 seconds on a standard laptop computer in the real data example presented herein, with additional uncertainty estimation requiring less than 10 minutes to complete with some parallelization. Notably, this method does not require MCMC as in some related work [50], and thus reduces the computational burden in many realistic settings. Additional technical information on the Mallows-Binomial distribution, such as the probability density function, model assumptions, goodness-of-fit tests, and estimation procedures, can be found in [46, 47]. Furthermore, we will soon implement a user-friendly interface based on the rankrate package to assist practitioners in using our model.

Developing theoretical examples and modeling review data

Using the Mallows-Binomial model, we constructed several theoretical review scenarios of potential reviewer voting behavior where the approaches of (i) ratings only, (ii) rankings only and (iii) the combined model could be compared directly. In the following three toy examples, we explored the comparative usefulness of approaches (i)-(iii) in tie-breaking similarly rated proposals (example 1), dealing with partial rankings of all proposals (example 2), and robustness against reviewer inconsistencies between ratings and rankings (example 3). Finally, we carried out a case study of the panel of grant review data from the AIBS review described above, applying the Mallows-Binomial model and producing integrated scores and priority lists for the proposals to exemplify the type of output this modeling approach produces.

Results

Toy Examples

We provide 3 toy examples below which demonstrate the concept of the integrated score, as well as the key advantages of the Mallows-Binomial model in relation to score-only or ranking-only models.

Toy Example 1: Tie-Breaking Equally Rated Proposals using Rankings

The first toy example demonstrates how adding rankings may help break ties between equally or similarly-ranked proposals in a principled manner. Suppose there are 3 proposals and 16 judges, who rate each proposal using a 5-point scale (the integers between 0 and 4) and subsequently rank all proposals. Their ratings and rankings can be found in Table 1.

Table 1 Ratings (left) and rankings (right) from toy example 1

We see that proposals 1 and 2 have the mean rating of 0.5, yet all judges prefer proposal 1 to proposal 2. Next, we display what a ratings-only model, rankings-only model, and the Mallows-Binomial model would output:

  1. 1

    Ratings-Only Model: \(\{1 = 2\}\prec 3\) on the basis of the mean ratings. There is no way of distinguishing proposals 1 and 2.

  2. 2

    Rankings-Only Model: \(1\prec 2\prec 3\) since all judges provided this same ranking. There is no method of discerning that proposals 1 and 2 are essentially tied.

  3. 3

    Mallows-Binomial Model: Integrated scores \(p=[0.125, 0.125 + 10^{-8}, 0.750]\) and induced preference ordering \(1\prec 2\prec 3\).Footnote 1 This result allows us to see both a reasonable preference order and that proposals 1 and 2 are essentially tied.

Key Takeaway: Integrated scores estimated by the Mallows-Binomial model break a tie between proposals 1 and 2 by incorporating rankings. Although the preference order provides a local comparison between objects to demarcate their quality (\(1\prec 2\prec 3\)), the integrated scores simultaneously suggest the global comparison that proposals 1 and 2 are essentially tied.

Toy Example 2: Improved Decision-Making Even with Partial Rankings

The second toy example demonstrates the practicality of the proposed method in that even partial rankings may help discern proposals accurately and reliably while minimally increasing the difficulty of assessing proposals: Given many research proposals, it can be cognitively challenging to provide a complete ranking. Furthermore, it is usually more important to make accurate distinctions between the best proposals as opposed to the worst proposals. Suppose there are 8 proposals and 16 judges, who rate each proposal using a 5-point scale (the integers between 0 and 4) and subsequently rank their top-3 proposals. Their ratings and rankings can be found in Table 2.

Table 2 Ratings (left) and rankings (right) from toy example 2

We see that all judges are internally consistent and exhibit a variety of preferences. For many judges, rankings help to break ties between equally-rated proposals. On the basis of all available data, it is clear that proposal 1 is the most-preferred but the preference order of proposals 2, 3, and 4 is unclear. The remaining proposals are clearly in the bottom half and are unlikely to be funded. We now consider what a ratings-only model, rankings-only model, and Mallows-Binomial model would output:

  1. 1

    Ratings-Only Model: \(1\prec \{2=3=4\}\prec 5\prec 6\prec 7\prec 8\) on the basis of the mean ratings. There is no way of distinguishing proposals 2, 3, and 4.

  2. 2

    Rankings-Only Model: \(1\prec 2\prec 3\prec 4\prec \{5,6,7,8\}\) on the basis of the available rankings. There is no way of distinguishing proposals 5, 6, 7, and 8.

  3. 3

    Mallows-Binomial Model: Integrated scores \(p=[0.000 , 0.125, 0.125 + 10^{-8} , 0.125 + 2\times 10^{-8}, 0.438, 0.750, 0.875, 0.937]\) and induced preference ordering \(1\prec 2\prec 3\prec 4\prec 5\prec 6\prec 7\prec 8\).Footnote 2 This result allows us to distinguish proposals 2, 3, and 4 while noting that they are essentially tied.

Additionally, we display confidence-based ranking summaries for the Mallows-Binomial model and the Ratings-Only Binomial model. In the table, entries correspond to the estimated probability that each proposal is truly ranked in a given rank place. Results are calculated via the bootstrap and are limited to the first four places and first four proposals (Table 3).

Table 3 Probabilities of proposals in the top four rank places based on the Mallows-Binomial (left) and Ratings-Only Binomial (right) models in toy example 2

We draw attention to the bootstrap ranking summary for proposal 2, which seems appropriate in the joint model (approximate tie for 2nd or 3rd place) but odd in the ratings-only model (approximate tie between 2nd and 4th place, but little weight for 3rd place). This strange behavior likely stems from the ratings of judges 9-12.

Key Takeaway: Integrated scores estimated by the Mallows-Binomial and their induced preference ordering draw nuanced distinctions among proposals using both ratings and partial rankings. Specifically, the integrated scores exhibit global comparisons, such as the approximate equivalence in quality between proposals 2, 3, and 4, while the induced preference ordering clarifies the local comparison that \(2\prec 3\prec 4\). Using partial rankings makes the additional ranking task cognitively easier and still allows for separation of the top proposals, which is normally the most important task for the reviewers. Furthermore, the bootstrap ranking summaries for the joint model are much more sensible since they are “anchored” by the rankings, which distinguish similarly-rated proposals.

Toy Example 3: Analyzing Data with Conflicting Ratings and Rankings

The third toy example demonstrates the ability of the model to appropriately capture ratings and rankings even when reviewers provide conflicting information. That is, situations in which the ranking induced by the ordering of ratings is different from the observed ranking. In real data collected by the AIBS, we frequently observe such patterns. At the same time, this example includes a small minority of judges who provide “outlier” ratings and rankings, which differ from the group and heavily influence the mean ratings.

Suppose we have 3 proposals and 16 judges, who rank all proposals and rate each using a 5-point scale (the integers between 0 and 4). Their ratings and rankings can be found in Table 4.

Table 4 Ratings (left) and rankings (right) from toy example 3

We see that judges 1-14 (the vast majority) give essentially equal ratings to proposals 1 and 2 and rate proposal 3 far below them. However, judges 8-14 are inconsistent in that they each give proposal 1 a rating of 1 and proposal 2 a rating of 0, yet rank \(1\prec 2\). Judges 15-16 think very poorly of proposal 1, however, and increase its mean rating significantly. Next, we display what a ratings-only model, rankings-only model, and Mallows-Binomial model would output:

  1. 1

    Ratings-Only Model: \(2\prec 1\prec 3\) on the basis of the mean ratings. The small minority of judges who give proposal 1 a rating of 3 heavily skew the mean ratings and thus affect the outcome.

  2. 2

    Ranking-Only Model: \(1\prec 2\prec 3\) since 14 of the 16 judges provided this same ranking.

  3. 3

    Mallows-Binomial Model: Integrated scores \(p=[0.156 , 0.156 +10^{-8} , 0.750]\) and induced preference ordering \(1\prec 2\prec 3\).Footnote 3 The integrated scores suggest that proposals 1 and 2 are essentially tied in the global sense, yet through the induced ordering appropriately suggest locally that \(1\prec 2\). The outlier judges do not alter the preference ordering.

Additionally, we display confidence-based ranking summaries for the Mallows-Binomial model and the Ratings-Only Binomial model. In the table, entries correspond to the estimated probability that each proposal is truly ranked in a given rank place. Results are calculated via the bootstrap and are limited to the first three places and first three proposals (Table 5).

Table 5 Probabilities of proposals in the top three rank places based on the Mallows-Binomial (left) and Ratings-Only Binomial (right) models in toy example 3

Key Takeaway: Integrated scores estimated by the Mallows-Binomial model and induced preference ordering are able to appropriately resolve judges who provide internally inconsistent ratings/rankings by recognizing that ratings of 0 and 1 for proposals 1 and 2 are essentially equal, given that 14 of the 16 judges ranked proposal 1 above proposal 2. This holds true even in the presence of two “outlier” judges who distort the mean ratings by rating proposal 1 very poorly. Additionally the joint model is more confident that \(1\prec 2\prec 3\), where the ratings-only model is less confident and gives much more probability to \(2\prec 1\prec 3\).

Case Study: A panel grant review data analysis

AIBS Ranking Procedure

In an effort to explore the usefulness of both rating and ranking in real-world funding decisions, the American Institute of Biological Sciences (AIBS) implemented a new procedure in the review of proposals submitted to a biomedical research funding agency. In this annual competition, AIBS reviewed proposals submitted to a 2020 funding announcement describing 2 year awards that are 100-150K in budget. The historic success rates for funding hover around 10 percent. As in previous years, reviewers were recruited based on expertise levels relative to the proposals, as well as on previous review experience and diversity balance. Reviewers were given access to proposal files and evaluation forms via an online system several weeks before the panel meeting and were required to enter preliminary comments and scoring into the system in advance of a teleconference review meeting. Each application was evaluated by two reviewers in advance of the meeting, who were asked to provide a score for the overall scientific merit based on the following application criteria: Impact/Significance, Innovation, Approach, Feasibility, and Investigators/Facilities. The overall scientific merit was scored on a scale from 1 (best) to 5 (worst); one decimal place is allowed in the scores (Table 6).

Table 6 Scoring Definitions

At the meeting, assigned reviewers presented their initial critiques to the panel, then the panel discussed (discussion is inclusive of all panelists who don’t have a conflict), and then all panelists made their final scores in the system after discussion was ended. These procedures have been the standard for the history of the program while AIBS was reviewing these proposals.

In 2020, AIBS added an additional ranking procedure to the assessment process. To collect ranking data, at the end of all proposal discussion, reviewers were provided with a link in the scoring system with a list of all the final average panel scores associated with each proposal (reviewers were blinded to any proposals where they had a conflict of interest). Thus, the list of proposals was different for each reviewer, depending on their conflicts in the review system. Reviewers were then given a link to a GoogleTM form, allowing them to look at all the proposals on the panel and select their “top six” that they would like to see funded. The question was constrained in that only one proposal can be chosen for each ranking position (e.g. first place) and only 6 choices were allowed. It should be noted that the scoring process was not altered in any way; the ranking process occurred after all proposals were scored and access to online scoresheets were locked. Only six rankings per reviewer were collected, as the focus was on ranking projects that each reviewer deemed worthy of funding if they were allowed to choose; and it was deemed impractical to rank all of the proposals. The number of ranked proposals was determined by looking at the historical success rate for this program (3 proposals for a panel of this size) and doubling it so we could examine rankings of both proposals likely to be funded as well as those slightly farther from the funding threshold.

To create the final proposal priority list, both scores and rankings needed to be considered. As mentioned previously, while scores are important indicators of the global scientific quality relative to the goals of the funding program, the rankings are more valid for indicating local proposal quality relative to the other proposals [35]. While rankings alone can be used to determine funding priorities, as they are zero sum and allow for clear discrimination between proposals, without ratings it is not known whether any of the proposals approach the standard of excellence. If only ratings are used (as is often the case), some scores can be close or identical, making it difficult to determine priority order. In order to combine these two information sources to create a funding priority list, a statistical model was chosen to apply to the data to facilitate interpretation.

Data Analysis

Panel 1 from the 2020 AIBS program has 12 reviewers and 28 proposals. Of the 12 reviewers, 11 were “full” reviewers and 1 was a “telecon” (TCON) reviewer, meaning he/she/they was asked to rate only 1 proposal and not rank. Ratings were provided on a 1 to 5 scale in single decimal point increments, which we have converted to the integers between 0 and 40 (a 41-point scale). Subsequently, reviewers were asked to provide a top-6 ranking.

The data have a few intricacies. First, some reviewers had conflicts of interest (COI) with one or more proposals. Specifically, one reviewer had a single COI while two reviewers each had two COI; 23 proposals had no COI while 5 each had one COI. Reviewers were not allowed to rate or rank proposals with which they had a COI. Beyond COI, some ratings and rankings were missing. There were 25 instances of missing ratings and one missing ranking among the “full” reviewers; the “TCON” reviewer provided only one rating and no ranking. In this analysis, we ensure COI missingness does not influence the likelihood of a proposal ranking and treat other missingness as missing completely at random.

Fig. 1
figure 1

Exploratory plots for AIBS Panel 1 (2020). Top: Boxplots of scores by proposal. Bottom: Stacked barcharts of ranking places assigned across judges by proposal

Figure 1 displays exploratory plots of the ratings and rankings from this panel. We notice a variety of rating patterns among the proposals. Some proposals have consistent ratings, while others exhibit wide variance. There are a few proposals which clearly have the best ratings, while others can be immediately seen as being unlikely to receive funding. Overall, the reviewers did not use the full rating scale, instead limiting themselves to the range [3, 30], which corresponds to the range [1.3, 4] on the original scale. For rankings, we notice that only 11 of the 28 total proposals made any of the 10 provided top-6 rankings. There is no clear consensus by looking at the top few rank places. However, we see that proposal 17, 19, and 25 frequently appear in first, second, and third places; proposal 4 appears in 5th place for over half the reviewers who provided rankings.

We now display integrated scores estimated by applying a Mallows-Binomial (MB) model to the AIBS data. In order to draw attention to the utility of the model, we additionally provide results from a traditional method, which we call the “Mean Ratings” (MR) model. In this model, we simply take the mean ratings from each proposal and standardize them to the unit interval. The order of proposals based on their mean ratings is thus the estimated ranking of the proposals. We display results in Table 7 and Fig. 2.

Table 7 Estimates of integrated scores from the Mallows-Binomial (MB) and Mean Ratings (MR) models. A small positive value is added to the integrated scores of proposals 22 and 6 in order to signify that \(p_{20}\prec p_{22}\) and \(p_3\prec p_6\) in the MB model. The order of proposals is based on the MB model
Fig. 2
figure 2

Proposals by estimated rank level in each model

We see in Table 7 that the estimates of integrated scores between the MB and MR models are similar. However, the preference ordering deviates in the MB model in a few cases. This distinction is made clear in Fig. 2, in which we can see directly that the MB model breaks ratings ties in 7th/8th place and 11th/12th places. Additionally, we see a reordering among proposals 3, 6, 13, and 16 between the Mean Ratings and Mallows-Binomial models: Although proposal 6 receives a slightly better mean rating than proposals 3, 13, and 16, its rankings are comparatively worse enough to make it receive a worse integrated score in the joint Mallows-Binomial model. We note that in the case of proposals 27 and 28, each received the same mean rating and neither was ranked by any judge. As such, neither model is able to break their tie. The model is also unable to break a tie in mean ratings between proposals 19 and 25. These proposals received unique rankings among the judges, yet precisely half of the reviewers preferred 19 to 25 while the other half preferred 25 to 19. As a result, the data do not allow for demarcation between these proposals on the basis of ratings or rankings. We turn to uncertainty estimation in order to make funding decisions between these two proposals.

Next, we estimate uncertainty in rank place among the proposals in Tables 8 and 9. Table 8 displays the probabilities of proposals entering first, second, third, or fourth place in the Mallows-Binomial and Mean Ratings models. In the Mallows-Binomial model, there is more certainty in proposal 19 being ranked above proposal 25, which may help us break the tie in integrated scores between the proposals. In comparison to the Mean Ratings model, the Mallows-Binomial provides more evidence that proposals 17, 19, and 25 have similarly high quality. Based on the original data, these results seem probable. For the purpose of making decisions, Table 9 displays the probabilities that proposals should receive funding conditional on the number of proposals the funding agency can support.

Table 8 Probabilities of proposals in each of the top-4 ranking places based on the Mallows-Binomial (left) and Ratings-Only Binomial (right) models in AIBS grant panel review data
Table 9 Probability of proposals being funded conditional on varying numbers of total proposals funded for the Mallows-Binomial model

Discussion

The evaluative method described here (collecting both ratings and partial rankings) and the combination of these streams of information into one integrated score provides a coherent way to assess a group of research proposals. While funding schemas like partial lotteries embrace the seemingly stochastic nature of grant review scores and outcomes ([51]), our approach conversely works to decrease the noise in review outcomes by extracting more meaningful information from reviewers ([52]). Moreover, the output of this process is a priority list for project funding that is structured optimally for the decision type of many secondary, programmatic funding panels and offers a confidence interval or a probability surrounding a given project’s priority. This output can be catered to the expected funding levels, creating probabilities of funding preference conditional on the number of total proposals that are planned to be funded (Table 9). These features are not typically incorporated in grant review output but are crucial for making appropriate funding judgements with confidence. In addition, we show that this method allows for the usage of partial rankings of proposals by reviewers, limiting the burden on panelists, allowing for conflict of interest scenarios, and providing a practical solution for data collection by review organizers. Hesitancy at funding agencies to adopt new decision making tools may be potentially overcome by the convenience of this format, the appeal of estimates of uncertainty in the priority order as well as the availability of software to ease any burden from these computations [49], which should be further alleviated by the development of a user-friendly interface based on this software. Future work should survey funding agencies to determine their perceptions of its usefulness.

This method is likely to be especially impactful for the tie breaking of similarly rated proposals, as scores are often compressed to a limited range in peer review [53]. As in the NIH example [28], this approach may help to discriminate between projects with similar ratings but different potential importance, more so than using criteria scores as tie-breakers, as they are often correlated with the overall score [33]. As reviewer-identified weaknesses have been observed to be more closely associated with ratings of proposal quality than strengths [21], it may be that including zero sum ranking could force a more balanced decision making process by valuing these strengths; this should be a focus of future research. However, it should be mentioned that our model assumes that proposals have a true underlying level of quality, an assumption implied in peer review studies that use different proxies for proposal quality (e.g., citations, patents; [54]). Despite this common assumption, true measures of quality remain elusive.

Collecting the ranking data at the end of all proposal discussion has an added benefit, as it may help to mitigate any temporal effects, as sometimes the decision-making norms of a panel can evolve over the length of the review meeting [55, 56]. For example, the panel may reach agreement on how to interpret the review criteria only after discussion at the panel meeting [57]. The model described here could incorporate such consistent interpretation across the panel while still remaining robust to conflicts in individual reviewer ratings and rankings, as well as to outlier reviewers. Thus, this method provides for a more appropriately weighted priority list that, while often somewhat similar to lists derived from the ratings alone, is resistant to internal inconsistencies and outliers. However, given the variety of epistemological views and expertise present on most review panels ([19, 58, 59]), even when reviewers are given identical scoring instructions and allowed an opportunity to discuss these criteria during the panel meeting, there may still be subgroups of reviewers with divergent views on proposal rating and ranking. In particular, little work has explored whether reviewer preferences for sets of proposals (e.g. high risk proposals) would affect reviewer rankings; this is an area in need of further examination. While our model assumes that individual raters are using the same criteria in similar ways, this is a commonly made assumption in the inter-rater reliability literature [59]; future research should do more to explore this assumption.

Limitations

As mentioned, research funders utilize a wide variety of peer review processes and procedures, and so the usefulness of this method will depend on this context. We have summarized the limitations and strengths of this method against common review processes in Table 10. For example, the presence of bias or strategic voting (e.g. based on connections to the applicant) impact all of these methods at some level, either directly or indirectly during the process. The addition and integration of rankings in grant review will not improve or intervene on problematic social dynamics to which scoring frameworks are already vulnerable. Like ratings, rankings are not immune from implicit bias [60, 61] or the influence of other panelists [62]. Like ratings, rankings permit reviewers to game the system by either engaging in strategically motivated voting [63] or by horse-trading favorite proposals with other panelists [19]. And, like ratings, rankings may reflect diverging interpretations of review criteria [19, 58, 59]. Rankings and our model inherit these forms of social influence, which may be inflected in the model’s results. Thus, while our integrated score provides more capability to discriminate proposals from one another, the thought process behind individual decisions is important, and still requires the use of orientations and training to counteract factors like implicit bias. It should also be noted that scores are often accompanied by written critiques, providing insight into reviewers’ motivations for their scoring decisions. More research is needed to examine whether additional guidance is needed at the ranking stage, to ensure reviewers are using similar criteria to rank proposals, or whether the allowance of different interpretations by reviewers is a feature and not a bug.

The usage of our model does have clear advantages over other methods in that it provides a tie-breaking mechanism for similarly rated proposals, of which none of the other methods are capable (although the random selection in partial lotteries avoids the issue of ties). Our model is also not easily influenced by outliers (e.g. Toy Example 3) compared to the other listed methods, especially those where only a few ratings are used to create the summary statistic that is used to determine funding. As we mention, our method is unique in utilizing more than one source of information to determine funding order and is the only method listed that directly provides a recommended funding priority list other than partial lotteries, which produce random priority lists after the initial rating phase is complete. Finally, our method provides estimates of the uncertainty in the recommended funding priority list – a feature that other review mechanisms currently lack.

It should be noted that while we list in Table 10 an example of the model being applied to a typical unblinded panel review, this type of output would even prove useful for agencies that adopt partial-lottery funding schemas, which include a proposal-rating phase ([64]). If the process contains a rating phase, the amount of extra preparation to implement our methodology would be small yet would likely provide better discrimination between proposals. Again, this method simply extracts more information from reviewers at the time of assessment and integrates it into a meaningful output, hopefully improving the interpretability of the review results, and does not abrogate the need for reviewer training, appropriate recruitment of expertise, vetting for conflicts of interest and monitoring of the process for consistency and equity. Studies have clearly shown significant strides can be made in the consistency of reviewers with even brief training programs that explain review criteria and highlight expectations from participants [65].Moreover, the generation of the proposal preference order through the use of this model does not serve as a replacement for secondary funding committees, lotteries or alternate grant allocation models that may have their own limitations ([66]). Finally, the model was applied here to review data from one panel of 28 proposals; some approximate estimation algorithms that have high accuracy can be used when there are more than 30 proposals [46].

Table 10 Limitations of Current Grant Review Models

Conclusions

To reserach funders, the gathering of both rating and ranking data and the application of the Mallows-Binomial model to this data to generate an integrated score can have many advantages over methods relying on ratings alone. These include providing a ranked priority list with confidence metrics, a higher degree of discrimination between similarly rated proposals, and a robustness to outliers and reviewer inconsistencies. While this method may not provide a panacea for problematic forms of social influence and may need to be adapted to work well with different funding mechanisms, peer review schemas, and success rates to achieve the goals of a research funding program, it is likely that incorporating information from both local and global evaluative tasks leverages information from reviewers to most accurately distill their opinion into a useful output to make the best, most informed decision.

Availability of data and materials

Anonymized data sets of ranking and rating data are available at 10.6084/m9.figshare.20505747.

Notes

  1. We add a small positive constant to the second value only to symbolize that \(p_1<p_2\), which yields the displayed preference ordering.

  2. We add a small positive constant to the third and fourth values only to symbolize that \(p_2<p_3<p_4\), which yields the displayed preference ordering.

  3. We add a small positive constant to the second value only to symbolize that \(p_1<p_2\), which yields the displayed preference ordering.

Abbreviations

AHA:

American Heart Association

AIBS:

American Institute of Biological Sciences

CIHR:

Canadian Institutes of Health Research

CDMRP:

Congressionally Directed Medical Research Programs

NIH:

National Institutes of Health

NSF:

National Science Foundation

ZonMw:

Netherlands Organisation of Health Research and Development

References

  1. NIH Staff. NIH Impact on Health. 2022. NIH. https://www.nih.gov/about-nih/what-we-do/impact-nih-research/our-health. Accessed 5 Apr 2023.

  2. Randolph MA, Bloom FE. Funding health sciences research: a strategy to restore balance. Washington DC: National Academies Press; 1990.

    Google Scholar 

  3. Heyard R, Hottenrott H. The value of research funding for knowledge creation and dissemination: A study of SNSF Research Grants. Humanit Soc Sci Commun. 2021;8(1):1–16.

    Article  Google Scholar 

  4. Viergever RF, Hendriks TC. The 10 largest public and philanthropic funders of health research in the world: what they fund and how they distribute their funds. Health Res Policy Syst. 2016;14(1):1–15.

    Article  Google Scholar 

  5. NIH. NIH Peer Review: Grants and Cooperative Agreements. https://grants.nih.gov. https://grants.nih.gov/grants/peerreview22713webv2.pdf. Accessed 2023.

  6. Veugelers R. Supporting the next generation of biomedical researchers in Europe: The impact from ERC funding on career development for young biomedical researchers. National Academies. https://sites.nationalacademies.org/cs/groups/pgasite/documents/webpage/pga_184824.pdf. Accessed 2023.

  7. UKRI. MRC two stage assessment process. UKRI.org. https://www.ukri.org/wp-content/uploads/2022/02/MRC-080722-PeerReviewProcessFlowDiagram.pdf. Accessed 2023.

  8. INSERM. Research Programs. www.inserm.fr. https://www.inserm.fr/en/about-us/research-programs/. Accessed 2023.

  9. Congressionally Directed Medical Research Programs C. CDMRP’s Two-Tiered Review Process. 2022. CDMRP. https://cdmrp.army.mil/about/2tierrevprocess. Accessed 5 Apr 2023.

  10. Trust W. Grant Funding. wellcome.org. https://wellcome.org/grant-funding. Accessed 2023.

  11. CIHR. Review guidelines - Priority-driven initiatives. cihr-irsc.gc.ca. https://cihr-irsc.gc.ca/e/4656.html#2.2.2. Accessed 2023.

  12. NHMRC. Investigator Grants 2023 Peer Review Guidelines. www.nhmrc.gov.au. https://www.nhmrc.gov.au/funding/find-funding/investigator-grants#download. Accessed 2023.

  13. HHMI. HHMI Investigator Program 2024 Investigator Competition Frequently Asked Questions (FAQ). www.hhmi.org. https://www.hhmi.org/sites/default/files/programs/investigator/investigator2024-faq.pdf. Accessed 2023.

  14. DFG. The decision-making process for individual grants. www.dfg.de. https://www.dfg.de/en/research_funding/decision_making_process/individual_grants/index.html. Accessed 2023.

  15. NSFC. Application Preparation and Submission. www.nsfc.gov.cn. https://www.nsfc.gov.cn/english/site_1/funding/E1/2022/01-12/25901.html. Accessed 2023.

  16. NIH. Meeting Do’s and Don’ts: Advice for Reviewers. 2022. https://grants.nih.gov/grants/policy/review/meeting.htm. Accessed 5 Apr 2023.

  17. AHA. AHA Peer Reviewer Score Guidance. 2022. https://professional.heart.org/-/media/PHD-Images/Research/Scoring-and-Triage-Guidance-2020_21.pdf. Accessed 5 Apr 2023.

  18. Boudreau KJ, Guinan EC, Lakhani KR, Riedl C. Looking across and looking beyond the knowledge frontier: Intellectual distance, novelty, and resource allocation in science. Manag Sci. 2016;62(10):2765–83.

    Article  Google Scholar 

  19. Lamont M. How Professors Think: Inside the Curious World of Academic Judgment. Cambridge: Harvard University Press; 2009.

    Book  Google Scholar 

  20. Johnson VE. Statistical analysis of the National Institutes of Health peer review system. Proc Natl Acad Sci. 2008;105(32):11076–80.

    Article  Google Scholar 

  21. Pier EL, Brauer M, Filut A, Kaatz A, Raclaw J, Nathan MJ, et al. Low agreement among reviewers evaluating the same NIH grant applications. Proc Natl Acad Sci. 2018;115(12):2952–7.

    Article  Google Scholar 

  22. Marsh HW, Jayasinghe UW, Bond NW. Improving the peer-review process for grant applications: Reliability, validity, bias, and generalizability. Am Psychol. 2008;63(3):160.

    Article  Google Scholar 

  23. Cicchetti DV. The reliability of peer review for manuscript and grant submissions: A cross-disciplinary investigation. Behav Brain Sci. 1991;14(1):119–35.

    Article  Google Scholar 

  24. Cole S, Simon GA, et al. Chance and consensus in peer review. Science. 1981;214(4523):881–6.

    Article  Google Scholar 

  25. Erosheva EA, Martinková P, Lee CJ. When zero may not be zero: A cautionary note on the use of inter-rater reliability in evaluating grant peer review. J R Stat Soc Ser A Stat Soc. 2021;184(3):904–19.

    Article  Google Scholar 

  26. Pina DG, Buljan I, Hren D, Marušić A. Meta-Research: A retrospective analysis of the peer review of more than 75,000 Marie Curie proposals between 2007 and 2018. Elife. 2021;10:e59338.

    Article  Google Scholar 

  27. Seeber M, Vlegels J, Reimink E, Marušić A, Pina DG. Does reviewing experience reduce disagreement in proposals evaluation? Insights from Marie Skłodowska-Curie and COST Actions. Res Eval. 2021;30(3):349–60.

    Article  Google Scholar 

  28. NIH-OER. Overall Impact. 2022. https://grants.nih.gov/grants/peer/guidelines_general/scoring_guidance_research.pdf. Accessed 5 Apr 2023.

  29. Gropp RE, Glisson S, Gallo S, Thompson L. Peer review: A system under stress. BioScience. 2017;67(5):407–10.

    Article  Google Scholar 

  30. Taffe MA, Gilpin NW. Racial inequity in grant funding from the US National Institutes of Health. Elife. 2021;10:e65697.

    Article  Google Scholar 

  31. Fang FC, Casadevall A. Research funding: the case for a modified lottery. MBio. 2016;7(2):e00422–16. https://doi.org/10.1128/mBio.00694-16.

  32. Liaw L, Freedman JE, Becker LB, Mehta NN, Liscum L. Peer review practices for evaluating biomedical research grants: A scientific statement from the American Heart Association. Circ Res. 2017;121(4):e9–19.

    Article  Google Scholar 

  33. Lindner MD, Vancea A, Chen MC, Chacko G. NIH peer review: scored review criteria and overall impact. Am J Eval. 2016;37(2):238–49.

    Article  Google Scholar 

  34. NSF. Chapter III - NSF Proposal Processing and Review. 2022. https://www.nsf.gov/pubs/policydocs/pappg22_1/pappg_3.jsp#IIIA. Accessed 5 Apr 2023.

  35. Goffin RD, Olson JM. Is it all relative? Comparative judgments and the possible improvement of self-ratings and ratings of others. Perspect Psychol Sci. 2011;6(1):48–60.

    Article  Google Scholar 

  36. Guthrie S, Ghiga I, Wooding S. What do we know about grant peer review in the health sciences? An updated review of the literature and six case studies. Santa Monica: RAND Corporation; 2018. https://doi.org/10.7249/RR1822.

  37. Mayo NE, Brophy J, Goldberg MS, Klein MB, Miller S, Platt RW, et al. Peering at peer review revealed high degree of chance associated with funding of grant applications. J Clin Epidemiol. 2006;59(8):842–8.

    Article  Google Scholar 

  38. Biernat M, Vescio TK. Manis M. Judging and behaving toward members of stereotyped groups: A shifting standards perspective. Intergroup cognition and intergroup behavior; 1998. p. 151–75.

    Google Scholar 

  39. Biernat M, Vescio TK. She swings, she hits, she’s great, she’s benched: Implications of gender-based shifting standards for judgment and behavior. Personal Soc Psychol Bull. 2002;28(1):66–77.

    Article  Google Scholar 

  40. Biernat M. Toward a broader view of social stereotyping. Am Psychol. 2003;58(12):1019.

    Article  Google Scholar 

  41. Biernat M, Collins EC, Katzarska-Miller I, Thompson ER. Race-based shifting standards and racial discrimination. Personal Soc Psychol Bull. 2009;35(1):16–28.

    Article  Google Scholar 

  42. Tarlow KR, Brossart DF, McCammon AM, Giovanetti AJ, Belle MC, Philip J. Reliable visual analysis of single-case data: A comparison of rating, ranking, and pairwise methods. Cogent Psychol. 2021;8(1):1911076.

    Article  Google Scholar 

  43. Arrow KJ. A difficulty in the concept of social welfare. J Polit Econ. 1950;58(4):328–46.

    Article  Google Scholar 

  44. Mandhani B, Meila M. Tractable search for learning exponential models of rankings. In: Artificial Intelligence and Statistics. PMLR; 2009. p. 392–399.

  45. Hunter DR, et al. MM algorithms for generalized Bradley-Terry models. Ann Stat. 2004;32(1):384–406.

    Article  Google Scholar 

  46. Pearce M, Erosheva EA. A Unified Statistical Learning Model for Rankings and Scores with Application to Grant Panel Review. J Mach Learn Res. 2022;23(210).

  47. Pearce M, Erosheva EA. On the validity of bootstrap uncertainty estimates in the Mallows-Binomial model. 2022. arXiv preprint arXiv:2206.12365.

  48. Liu Y, Xu Y, Shah NB, Singh A. Integrating Rankings into Quantized Scores in Peer Review. 2022. arXiv preprint arXiv:2204.03505.

  49. Pearce M. rankrate: Statistical Tools for Preference Learning with Rankings and Ratings. R CRAN. https://cran.r-project.org/package=rankrate. Accessed 2022.

  50. Heyard R, Ott M, Salanti G, Egger M. Rethinking the Funding Line at the Swiss National Science Foundation: Bayesian Ranking and Lottery. Stat Public Policy. 2022;(just-accepted):1–27.

  51. Greenberg DS. Peer review: And the winner is.... Lancet. 1999;354(9195):2092.

  52. Grant S, Meilă M, Erosheva E, Lee C. Refinement: Measuring informativeness of ratings in the absence of a gold standard. Br J Math Stat Psychol. 2022;75(3):593–615.

    Article  Google Scholar 

  53. Center for Scientific Review N. A Pilot Half Study of Half-Point Increments in Scoring. 2022. NIH. https://public.csr.nih.gov/AboutCSR/HalfPointPilotStudy.

  54. Li D, Agha L. Big names or big ideas: Do peer-review panels select the best science proposals? Science. 2015;348(6233):434–8.

    Article  Google Scholar 

  55. Langfeldt L. The decision-making constraints and processes of grant peer review, and their effects on the review outcome. Soc Stud Sci. 2001;31(6):820–41.

    Article  Google Scholar 

  56. Pier EL, Raclaw J, Carnes M, Ford CE, Kaatz A. Laughter and the chair: Social pressures influencing scoring during grant peer review meetings. J Gen Intern Med. 2019;34(4):513–4.

    Article  Google Scholar 

  57. Abdoul H, Perrey C, Amiel P, Tubach F, Gottot S, Durand-Zaleski I, et al. Peer review of grant applications: Criteria used and qualitative study of reviewer practices. PLoS ONE. 2012;7(9):e46054.

    Article  Google Scholar 

  58. Luukkonen T. Conservatism and risk-taking in peer review: Emerging ERC practices. Res Eval. 2012;21(1):48–60.

    Article  Google Scholar 

  59. Lee CJ. A Kuhnian critique of psychometric research on peer review. Philos Sci. 2012;79(5):859–70.

    Article  Google Scholar 

  60. Lee CJ, Sugimoto CR, Zhang G, Cronin B. Bias in peer review. J Am Soc Inf Sci Technol. 2013;64(1):2–17.

    Article  Google Scholar 

  61. Lee CJ. Revisiting current causes of women’s underrepresentation in science. In: Saul J, Brownstein M, editors. Implicit bias and philosophy volume 1: metaphysics and epistemology. Oxford: Oxford University Press; 2016. p. 265–81.

    Chapter  Google Scholar 

  62. Pier EL, Raclaw J, Kaatz A, Brauer M, Carnes M, Nathan MJ, et al. ‘Your comments are meaner than your score’: Score calibration talk influences intra-and inter-panel variability during scientific grant peer review. Res Eval. 2017;26(1):1–14.

    Article  Google Scholar 

  63. Jecmen S, Yoon M, Conitzer V, Shah NB, Fang F. A Dataset on Malicious Paper Bidding in Peer Review. 2022. arXiv preprint arXiv:2207.02303.

  64. Liu M, Choy V, Clarke P, Barnett A, Blakely T, Pomeroy L. The acceptability of using a lottery to allocate research funding: a survey of applicants. Res Integr Peer Rev. 2020;5(1):1–7.

    Article  Google Scholar 

  65. Sattler DN, McKnight PE, Naney L, Mathis R. Grant peer review: Improving inter-rater reliability with training. PLoS One. 2015;10(6):e0130450. https://doi.org/10.1371/journal.pone.0130450.

  66. Lee CJ, Grant S, Erosheva EA. Alternative grant models might perpetuate Black-White funding gaps. Lancet. 2020;396(10256):955–6.

    Article  Google Scholar 

Download references

Acknowledgements

The authors would like to thank Richard Nakamura for his comments on the manuscript.

Funding

This work was funded by NSF grant 2019901.

Author information

Authors and Affiliations

Authors

Contributions

EE and CL contributed to the funding acquisition for this work, and EE and MP contributed to the conceptualization and statistical analysis. SG, EE, and MP contributed to the methodology and SG contributed to the data curation. SG, EE, CL and MP contributed to the investigation and writing of the original draft, and review and editing of the final draft.

Corresponding author

Correspondence to Stephen A. Gallo.

Ethics declarations

Ethics approval and consent to participate

Data included in the preparation of this manuscript was generated as a result of a contractual agreement between AIBS and individual scientific reviewers to conduct independent, objective scientific peer review services. AIBS is the sole owner of this original data and reviewers consent to this arrangement by signing the contract. For the analysis in this manuscript, only de-identified data were included and shared among authors.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Additional details of the Mallows-Binomial distribution

We first describe the assumed data-generating process: The Mallows-Binomial distribution assumes that each proposal, j has a latent quality, \(p_j\in [0,1]\). By convention, values of \(p_j\) close to 0 indicate better quality and values close to 1 indicate lower quality. Given \(p_j\), the ratings for proposal j are drawn independently and identically distributed from a Binomial\((M,p_j)\) distribution, where M is the known maximum integer score. At the same time, a simple ordering of proposals based on their latent qualities, denoted by \(\pi _0\), indicates the true order of proposals from best worst. Furthermore, the distribution assumes that the population of reviewers exhibits some level of consensus, \(\theta >0\), in which low values indicate weak consensus and high values indicate strong consensus. Given \(\pi _0\) and \(\theta\), rankings are drawn independently and identically distribution from a Mallows\((\pi _0,\theta )\) distribution. In a Mallows distribution, \(\pi _0\) represents the modal probability ranking and \(\theta\) is a scale parameter that controls how likely rankings that are far away from \(\pi _0\) are to be drawn. Since \(p=[p_1 \ p_2 \ \dots , p_{J-1} \ p_J]\) influences both ratings and rankings, we call p the integrated scores.

We now state the model likelihood. Let \(X = [X_1 \ X_2 \ \dots \ X_J]\) represent the collection of ratings assigned to proposals \(j=1,\dots ,J\) by a reviewer, and \(\Pi\) his/her ranking. Under a Mallows-Binomial\((p,\theta )\) distribution, their joint probability can be written:

$$\begin{aligned} \text {Pr}[X=x,\Pi =\pi | p,\theta ] = \prod _{j=1}^J \left( {\begin{array}{c}M\\ x_j\end{array}}\right) p_j^{x_j}(1-p_j)^{M-p_j} \times \frac{\exp ^{-\theta d_K(\pi ,\pi _0)}}{\psi (\theta )} \end{aligned}$$
(1)

In Eq. 1, \(d_K(\pi ,\pi _0)\) represents the Kendall \(\tau\) distance between the reviewer’s observed ranking \(\pi\) and the true consensus ranking \(\pi _0\). This distance may be defined as the total number of adjacent object pairs which must be swapped in the consensus ranking in order to obtain the observed ranking. \(\psi (\theta )\) is a normalization constant that depends on \(\theta\) and has a closed-form expression. Further details of the distribution can be found in [46].

Additional results from case study

 

Table 11 Results from the Mallows-Binomial and Mean Ratings Models
Fig. 3
figure 3

Tree diagram of proposal selection probabilities. Bolded numbers represent proposals; numbers above lines represent the conditional probability of selecting the next proposal given a current stage in the tree; numbers in circles represent joint probabilities up to that point in the tree

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Gallo, S.A., Pearce, M., Lee, C.J. et al. A new approach to grant review assessments: score, then rank. Res Integr Peer Rev 8, 10 (2023). https://doi.org/10.1186/s41073-023-00131-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s41073-023-00131-7

Keywords