| Literature DB >> 36200060 |
Jacqueline N Lanei1,2, Misha Teplitskiy2,3, Gary Gray4, Hardeep Ranu4, Michael Menietti1,2, Eva Guinan2,4,5, Karim R Lakhani1,2.
Abstract
The evaluation and selection of novel projects lies at the heart of scientific and technological innovation, and yet there are persistent concerns about bias, such as conservatism. This paper investigates the role that the format of evaluation, specifically information sharing among expert evaluators, plays in generating conservative decisions. We executed two field experiments in two separate grant-funding opportunities at a leading research university, mobilizing 369 evaluators from seven universities to evaluate 97 projects, resulting in 761 proposal-evaluation pairs and more than $250,000 in awards. We exogenously varied the relative valence (positive and negative) of others' scores and measured how exposures to higher and lower scores affect the focal evaluator's propensity to change their initial score. We found causal evidence of a negativity bias, where evaluators lower their scores by more points after seeing scores more critical than their own rather than raise them after seeing more favorable scores. Qualitative coding of the evaluators' justifications for score changes reveals that exposures to lower scores were associated with greater attention to uncovering weaknesses, whereas exposures to neutral or higher scores were associated with increased emphasis on nonevaluation criteria, such as confidence in one's judgment. The greater power of negative information suggests that information sharing among expert evaluators can lead to more conservative allocation decisions that favor protecting against failure rather than maximizing success.Entities:
Keywords: information sharing; innovation; knowledge frontier; negativity bias; project evaluation
Year: 2021 PMID: 36200060 PMCID: PMC9531843 DOI: 10.1287/mnsc.2021.4107
Source DB: PubMed Journal: Manage Sci ISSN: 0025-1909 Impact factor: 6.172
Figure 1.Overview of Research Setting, Evaluator Recruitment/Selection, and Treatment Conditions
Note. Inst., institutions.
Summary Statistics of Count of Proposals Reviewed by Evaluator
| Treatment | Control | |||||
|---|---|---|---|---|---|---|
| Statistic | Study 1 | Study 2 | Pooled | Study 1 | Study 2 | Pooled |
| No. of evaluators | 244 | 89 | 333 | 34 | 3 | 37 |
| Mean (s.d.) | 1.59 (1.05) | 3.75 (2.44) | 2.17 (1.82) | 1.00 (0.00) | 1.33 (0.58) | 1.03 (0.16) |
| Min, max | 1, 6 | 1, 8 | 1, 8 | 1, 1 | 1, 2 | 1, 2 |
| No. of pairs | 389 | 334 | 723 | 34 | 4 | 38 |
Distribution of Treatment Scores Valence by Original Score
| Original score | Study 1 | Study 2 | |||
|---|---|---|---|---|---|
| Lower | Higher | Lower | Neutral | Higher | |
| 1 | 0 | 28 | 0 | 1 | 0 |
| 2 | 0 | 38 | 0 | 6 | 3 |
| 3 | 30 | 34 | 0 | 9 | 23 |
| 4 | 29 | 34 | 12 | 15 | 10 |
| 5 | 32 | 38 | 21 | 15 | 12 |
| 6 | 43 | 34 | 24 | 27 | 20 |
| 7 | 44 | 0 | 61 | 24 | 0 |
| 8 | 5 | 0 | 15 | 23 | 0 |
| 9 | — | — | 8 | 5 | 0 |
Covariate Balance Check by Treatment-Score Valence (N = 723)
| Variable | Study 1 ( | Study 2 ( |
|---|---|---|
|
| ||
|
| ||
|
| ||
|
| ||
|
|
Correlation Table of Main Variables (N = 723)
| Variable | Mean | s.d. | Min | Max | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|---|---|---|---|
| 1. | −0.145 | 0.880 | −4 | 4 | 1.000 | |||||
| 2. | 1.068 | 0.909 | 0 | 2 | −0.626 | 1.000 | ||||
| 3. | 0.530 | 0.499 | 0 | 1 | 0.005 | −0.043 | 1.000 | |||
| 4. | 3.374 | 0.943 | 1 | 5 | −0.025 | −0.014 | 0.020 | 1.000 | ||
| 5. | 0.350 | 0.477 | 0 | 1 | −0.001 | 0.015 | 0.016 | −0.033 | 1.000 | |
| 6. | 0.536 | 0.499 | 0 | 1 | 0.051 | −0.056 | −0.013 | 0.036 | −0.207 | 1.000 |
| 7. | 5.057 | 1.906 | 1 | 9 | −0.340 | 0.482 | 0.019 | −0.105 | 0.065 | −0.060 |
Estimated Relationships Between Change in Evaluation Score and Treatment Scores Valence
| Variable | Full sample (all scores) | Restricted sample (middle scores) | ||||||
|---|---|---|---|---|---|---|---|---|
| Model 1: | Model 2: | Model 3: | Model 4: | Model 5: | Model 6: | Model 7: | Model 8: | |
| Randomized: Baseline = neutral treatment scores | ||||||||
| Lower scores | −0.759 | −0.756 | −0.753 | −0.867 | −0.625 | −0.618 | −0.663 | −0.911 |
| Higher scores | 0.449 | 0.452 | 0.434 | 0.527 | 0.561 | 0.567 | 0.503 | 0.512 |
| Intellectual distance | −0.0380 | −0.0297 | 0.00304 | −0.0293 | −0.0180 | −0.00965 | ||
| Covariates | ||||||||
| Expertise | −0.0237 | −0.0360 | −0.00824 | −0.0216 | −0.0464 | 0.00878 | ||
| Female | 0.0246 | 0.0371 | −0.0126 | 0.00868 | ||||
| Tenured | 0.0411 | 0.0540 | 0.0498 | 0.0526 | ||||
| Constant | 0.0242 | 0.0909 | −0.180 | −0.407 | −0.0466 | 0.0138 | 0.413 | 0.150 |
| Original Score FE | N | N | Y | Y | N | N | Y | Y |
| Evaluator FE | N | N | N | Y | N | N | N | Y |
| Observations | 723 | 722 | 722 | 544 | 430 | 430 | 430 | 305 |
|
| 0.433 | 0.433 | 0.452 | 0.461 | 0.425 | 0.426 | 0.466 | 0.254 |
| No. of proposals | 97 | 97 | 97 | 94 | 95 | 95 | 95 | 82 |
| No. of evaluators | 333 | 333 | 333 | 155 | 266 | 266 | 266 | 141 |
Notes. Sample size drops from 723 to 722 in model 2 because of missing expertise in one evaluator-proposal pair from study 2. Robust standard errors are in parentheses.
p < 0.01.
Figure 2.Margins Plot of Change in Evaluation Score and Treatment Scores Valence by Original Score with 95% CIs
Overview of Qualitative Data Taxonomy and Coding for Study 2
| Primary topic | Axial code | Open code examples |
|---|---|---|
| Criteria-specific | Impact | “Proposal has minimal impact if any.” |
| Design and methods | “Lacks description of study participants, data analyses section and etc.” | |
| Feasibility | “Limited information about the feasibility of such a study.” | |
| Novelty | “Not so original.” | |
| Reevaluate proposal | “I was between a 3 and a 4. In reviewing the grant again, a 3 would be appropriate.” “Reconsidered.” | |
| Overall assessment | “Good bioinformatics application.” | |
| Nonspecific | Consistent with others | “My score is within the range.” |
| Lack of expertise | “I attribute my original score to lack of expertise. Changed to reflect enthusiasm of other reviewers.” | |
| Review process | “I realize I was using a higher bar than is optimal for a pilot grant.” | |
| Confident in judgment | “I am confident that my judgment is fair.” |
Figure 3.Distribution of Axial Codes by Valenced Treatment Scores (Study 2)
Figure 4.Distribution of Primary Topics by Valenced Treatment Scores (Study 2)
Figure 5.Scatter Plot of Average Updated (Postupdate) vs. Original (Preupdate) Scores
Figure 6.Comparison of Updated vs. Original Proposal Ranks for Study 1 (Left) and Study 2 (Right)
Figure 7.Percentage Turnover in Winning Proposals as a Function of the Payline (Success Rate)