Daniele Fanelli1, Matteo Schleicher1, Ferric C Fang2, Arturo Casadevall3, Elisabeth M Bik4. 1. Department of Methodology, London School of Economics and Political Science, London, United Kingdom. 2. Department of Laboratory Medicine and Pathology, University of Washington School of Medicine, Seattle, Washington, United States of America. 3. Department of Molecular Microbiology and Immunology, Johns Hopkins Bloomberg School of Public Health, Baltimore, MD, United States of America. 4. Harbers Bik LLC, Sunnyvale, CA, United States of America.
Abstract
Pressures to publish, perverse incentives, financial interest and gender are amongst the most commonly discussed risk factors for scientific misconduct. However, evidence of their association with actual data fabrication and falsification is inconclusive. A recent case-controlled analysis of articles containing problematic image duplications suggested that country of affiliation of first and last authors is a significant predictor of scientific misconduct. The same analysis found null or negative associations with individual proxies of publication rate, impact and gender. The latter findings, in line with previous evidence, failed to support common hypotheses about the prevalence and causes of misconduct, but country-level effects may have confounded these results. Here we extend and complete previous results by comparing, via matched-controls analysis, articles from authors in the same country. We found that evidence for individual-level risk factors may be significant in some countries, and null or opposite in others. In particular, in countries where publications are rewarded with cash incentives, and especially China, the risk of problematic image duplication was higher for more productive, more frequently cited, earlier-career researchers working in lower-ranking institutions, in accordance with a "misaligned incentives" explanation for scientific misconduct. However, a null or opposite pattern was observed in all other countries, and especially the USA, UK and Canada, countries where concerns for misaligned incentives are commonly expressed. In line with previous results, we failed to observe a statistically significant association with industry funding and with gender. This is the first direct evidence of a link between publication performance and risk of misconduct and between university ranking and risk of misconduct. Commonly hypothesised individual risk factors for scientific misconduct, including career status and productivity, might be relevant in countries where cash-reward policies generate perverse incentives. In most scientifically active countries, however, where other incentives systems are in place, these patterns are not observed, and other risk factors might be more relevant. Policies to prevent and correct scientific misconduct may need to be tailored to a countries' or institutions' specific context.
Pressures to publish, perverse incentives, financial interest and gender are amongst the most commonly discussed risk factors for scientific misconduct. However, evidence of their association with actual data fabrication and falsification is inconclusive. A recent case-controlled analysis of articles containing problematic image duplications suggested that country of affiliation of first and last authors is a significant predictor of scientific misconduct. The same analysis found null or negative associations with individual proxies of publication rate, impact and gender. The latter findings, in line with previous evidence, failed to support common hypotheses about the prevalence and causes of misconduct, but country-level effects may have confounded these results. Here we extend and complete previous results by comparing, via matched-controls analysis, articles from authors in the same country. We found that evidence for individual-level risk factors may be significant in some countries, and null or opposite in others. In particular, in countries where publications are rewarded with cash incentives, and especially China, the risk of problematic image duplication was higher for more productive, more frequently cited, earlier-career researchers working in lower-ranking institutions, in accordance with a "misaligned incentives" explanation for scientific misconduct. However, a null or opposite pattern was observed in all other countries, and especially the USA, UK and Canada, countries where concerns for misaligned incentives are commonly expressed. In line with previous results, we failed to observe a statistically significant association with industry funding and with gender. This is the first direct evidence of a link between publication performance and risk of misconduct and between university ranking and risk of misconduct. Commonly hypothesised individual risk factors for scientific misconduct, including career status and productivity, might be relevant in countries where cash-reward policies generate perverse incentives. In most scientifically active countries, however, where other incentives systems are in place, these patterns are not observed, and other risk factors might be more relevant. Policies to prevent and correct scientific misconduct may need to be tailored to a countries' or institutions' specific context.
Understanding the psychological, sociological and structural factors that increase the risk of scientific misconduct is a core objective of research on research integrity and meta-science more generally. Identifying such factors is of theoretical interest, as it contributes to our understanding of causal patterns in human behaviour, as well as pragmatic interest, because it is the basis for designing policies and interventions to prevent research misconduct.One of the most commonly discussed structural risk factors for scientific misconduct are “pressures to publish”–that is, unrealistic productivity standards imposed on academics, which may compel them to “cut corners” in order to sustain high levels of productivity and impact [1]. The “pressures to publish” hypothesis may be considered a special case of a more general “misaligned incentives” hypothesis, according to which scientists deviate from norms of good research practice to the extent that they have more to gain and less to lose from doing so [2]. Whilst the notion of pressures to publish and misaligned incentives is intuitive and plausible, evidence of its role as a causal factor for scientific misconduct is ambiguous. On the one hand, surveys support the widespread perception that academics are under pressure to publish, and that this may increase misconduct [3-5]. On the other hand, however, analyses of the literature have repeatedly failed to observe positive correlations between individual productivity or citation impact and poor research quality, at least when the latter was measured as the likelihood to correct or retract articles [6] or to publish over-estimated effect sizes [7,8].The “funding effect” is another form of “misaligned incentive” very widely discussed and studied. A broad literature in clinical medicine suggests that industry sponsorship is more likely to report outcomes that are favourable to the sponsor, possibly due to a combination of methodological and reporting biases [9]. By extension, it seems plausible to hypothesize that privately funded research is at higher risk of scientific misconduct [10]. However, the opposite argument seems equally plausible: commercial interests linked to a research are a disincentive to research misconduct, because any short-term benefit accrued from misreporting research results is likely to be paid in the long term by financial losses caused by commercializing poorly performing products. Indeed, the nature and exact causes of the “funding effect” itself are still subject to some controversy [11].A third factor that has attracted some research and controversy is gender. Statistics of the US Office of Research Integrity show that males are disproportionally represented amongst findings of scientific misconduct [12]. This could be explained by the fact that, on average, males are more aggressive, competitive and prone to taking risks (including with research misconduct) than females, or it could result from other structural factors and confounding variables, including the fact that men are more likely to hold multiple grants and may be less effective at negotiating their defence before and during an investigation [13]. Indirect support for the latter interpretation was offered by evidence that gender is not a significant predictor of retractions and corrections in the literature, or bias in meta-analysis [6,7].A direct test of these and other causal hypotheses for scientific misconduct was conducted by some of us in [14], where we compared characteristics of publications that had been identified as containing or not containing problematic image duplications [15]. Using a matched-control analysis on a sample of such articles published in PLoS ONE, in [14] we observed strong between-country differences, particularly for articles containing image manipulations that are more likely to be intentional (i.e. duplication categories 2 and 3, see below). Articles of which the first and/or last authors were based in China and India were at significantly higher risk of problematic duplications, but the variance was generally high amongst other countries, too.Further, in [14] we found that the probability of image duplication was significantly higher for countries where researchers receive cash incentives for publishing in high-impact journals (in particular, China, Turkey, South Korea), and in countries that combine a developing economy with a hierarchical academic environment (as hypothesised by [16]). Conversely, the risk was significantly lower amongst countries in which publication incentives were in the form of career advancement (for eample, a tenure-track system), countries that combine developed (highly regulated) economies with a non-hierarchical academic environment, and countries that have national and legally binding policies against research misconduct. These patterns were interpreted as partially supporting a misaligned incentives explanation, in which cash rewards and low regulations and social control are significant risk factors for scientific misconduct [14].Conversely, individual-level factors in the same analysis yielded a null or a negative support for the misaligned incentives and other common hypotheses, because the probability of image duplication was null or negatively associated with characteristics of first and last authors including gender, number of years of publication activity, average number of papers published per year, and average citation rates and journal impact factor. However, results of analyses at the individual-level were not deemed conclusive, because the strong between-country differences constituted a significant confounding factor [14]. Bibliometric data for authors in China and India, for example, were likely to be biased since these authors may publish part of their work in local journals not indexed by the ISI Web of Science database used in these analyses.The present study aimed to resolve the apparent contradictions between country-level and individual-level patterns observed in [14], by completing and improving the analysis in three important ways. First, instead of matching articles by time of publication, we matched articles by country in which the first author operates, which allows to fully control any country-level confounding effect. Second, we focus our analysis on multivariable tests and include in all models an interaction term between career length of authors and their average publication rate per year. This interaction term tests the hypothesis that the career length of an individual modulates the relation between misconduct and publication rates, a hypothesis that [14] had not tested.A third improvement from [14] consists in testing two commonly invoked risk factors:Funding source and employment status: we tested the funding effect on misconduct, by assessing the association between the probability of image duplication and the nature of funding acknowledged in each article or, as an alternative proxy, the kind of institution indicated as main affiliation for the first author.Institution ranking: to further probe the effects of cash incentives policies and pressures to publish, we assessed how the probability of image duplication varied with the institution’s position in international university rankings. We hypothesised that the probability of image duplication might be higher for authors in lower ranking institutions. This hypothesis was particularly relevant to China, where high-ranking universities are known to reward authors with relatively modest cash amounts that are proportional to the actual citation impact of articles, whereas low-ranking universities give higher rewards based on the impact factor of the journal in which the article appears, which arguably encourages misconduct [17]. In particular, for each article published in the journal PLoS ONE, which is the source of our data, “tier 3” institutions (which are the lowest ranking in the Chinese 3-tier academic system) offer on average 1,661 USD, which is over four times the average reward paid by “tier 1” institutions and over half the average salary of a newly-hired professor [18]. A publication in the top journals Nature and Science may yield, in some Chinese institutions, up to 165,000 USD dollars. First authors are typically the main or sole beneficiary of these cash rewards (70% of policies examined in [18] only award cash to the first author) and are therefore most likely to be influenced by them.
Methods
Sample of articles
We re-analysed a subset of the data used in [14], which in turn was a re-analysis of part of the data set produced by [15]. The latter consists in a large data set (N = 20,621) of articles whose images had been individually inspected for the presence of problematic image duplications. This sample was obtained by searching for articles with the keyword “Western blot” in 40 different journals and 14 publishers, and the images consist of molecular biology visual data, primarily photo compositions of Western Blots and other cell and tissue images. Image duplications were coded twice independently and classified in a scale of three categories, where, at one end, Category 1 are simple duplications that are most likely to result from simple error, and, at the other, Category 3 include misleading transformations that are most likely to be intentional (for further details see [15]). The sample used by [14] was a subset of this data, which was restricted to articles published in the journal PLoS ONE between 2013 and 2014, and it included N = 346 articles that were deemed to contain problematic image duplications and N = 868 controls that were deemed to contain none. All but one of these articles had two or more authors.
Case-control matching
The matching was based on the country of the first institution appearing in the affiliation list in the article’s WOS record, which is most likely to correspond to the first author’s main or sole country of affiliation. We then matched each “case” (article that contained an image duplication) with two “controls”, which were selected at random from the pool of articles that had the same first author’s country of affiliation but had not been found to contain image duplications. Articles that lacked two exact country matches (N = 46) were excluded from the analysis.Articles were not matched for any other characteristic. In particular, we did not match based on time of publication because the articles are all published in a relatively narrow time window of 2 years, which makes time an unlikely confounding factor.This case-control matching was done blindly to any article characteristic. To avoid spurious inclusions, articles with the same first author (N = 2), and articles marked in the WOS as ‘retracted publications’ (N = 3) or ‘corrections’ (N = 2) were removed from the pool of potential controls. The final sample of papers therefore consisted of 293 papers with problematic image duplications and 2*293 = 586 controls (total N = 879).
Data collection
Our analyses focused on characteristics of first-authors because first authors are, in biomedical research, typically the individuals most directly involved in all aspects of a research project. Further, previous analyses suggested that first authors exhibited patterns of particular interest [14]. Furthermore, first authors are the main or sole beneficiary of cash publication rewards, at least in China [18]. For each treatment and control article, and for each first-author of these articles, we collected the following variables, either directly from records in the Web of Science (henceforth WOS) or from a bibliometric database that compiles data on individual authors using a name disambiguation algorithm [19]:Team size: number of co-authors listed in the article.10*countries/author: number of distinct countries listed as affiliations of the co-authors of the article, divided by the total number of authors. In the previous analysis, this variable was tested as the total number of countries [14], but we use here the ratio of countries to authors as it reduces the dependence of the two variables, allowing for a more robust test of the hypothesis that large international teams have a differential risk. This variable was then multiplied by ten, in order to improve the readability of the plots that show regression results. The regression estimates of main effects and interactions reported in all results and figures, therefore, represent rates of change associated with an increase in the ratio of countries to ten authors in an article. This transformation does not affect the statistical significance or any other aspect of results.Years active: number of years occurring between an author’s first and last publication as recorded in the Web of Science (WOS), calculated as year(lastpublication)−year(firstpublication)+1.Papers/year: for each first author in our sample, the total number of articles they co-authored, as recorded in the WOS, divided by the number of years they have been active (calculated as above).Citations/paper: total number of citations accrued by each first author, divided by the total number of articles they co-authored.Female author: gender of the first author, classified as female or male or unknown based on first name and country. Most names were classified by a commercial service (genderapi.com), and the records that remained unclassified were subsequently completed by hand, via direct inspection of information available online. In particular, Curricula Vtae and professional webpages were inspected for explicit gendering information and, if this was not available, based on photos, see [14] for details). This hand-coding was blind to the outcome (presence/absence of duplication in the paper). Names that remained unclassified (N = 51/879, 5.8%) were re-coded as “male”. We used male as the default gender, based on the fact that most (i.e. 27) of these 51 unclassified names were Chinese or Taiwanese, and that about 75% of scientists in China are male [20]. The effect of this coding choice on results was checked with robustness tests reported in the Results section.Cash incentives: a binary category that identifies countries in which national or institutional policies offer cash rewards to authors who perform well in terms of publication number and/or impact. This is a recoding of the variable used in [14], which was based on a classification by [21]. In practice, in the present sample, this variable separates China and South Korea from all other countries.German-developmental: a binary category that identifies countries that, according to an independent sociological and historical analysis, exert the least effective social control due to a combination of hierarchical academic culture and low regulatory standards [16]. This variable is a recoding of the analogous variable used in [14], which had distinguished three levels of social control and had excluded all non-classifiable countries. Here the variable is recoded as binary, to separate China, Japan and Korea from all other countries.In addition to the variables above, all of which had been tested previously, the following new variables were recorded.University rank and non-ranked institutions: rank of each first author’s affiliated university in the 2017 Shanghai Academic Ranking of World Universities. The Shanghai ranking was chosen because it includes a greater number of Asian institutions, and the 2017 version because it includes a larger list of institutions compared to previous versions (N = 800 vs. N = 500). Each institution has a numerical score between 0 and 800, but only the top 100 universities have a unique numerical score, whereas lower ranks are classified in percentile tiers: 101–150, 151–200 etc. We recoded the Shanghai score in eight tiers, from tier 0 (1–100) to tier 7 (701–800). Affiliations with universities that were not included in the Shanghai ranking were given a score of 8. Affiliations with non-university institutions were given a score of 0, and marked out with a separate binary variable (“non-ranked inst.”). Therefore, regression estimates on institution rank reflect the increase in the odds of image duplication as the position in the Shanghai ranking decreases (its rank score increases by one step). Regression estimates for the non-ranked institution variable reflect the difference between the odds of image duplication for these non-ranked institutions and those for the top-ranking university institutions. When an author indicated multiple affiliations, we recorded the first (and therefore presumably main) listed affiliation. In addition to the Shanghai ranking, which has global application, we classified Chinese universities according to the national 3-tier system [18]. We used publically available lists of “tier 1” [22] and “tier 2” [23] universities, and classified the remaining as “tier 3”.Industry/mixed funding: the funding sources listed in each articles’ WOS record were retrieved and categorized as either “industry”, “public”, “non-profit research organisation”, “university” and “mixed” by MJS. Since most records fell either in the “public” category or in the “mixed” category, the variable was recoded as a binary variable, separating an “industry/mixed fund” category from purely publically funded (i.e. university, non-profit etc.) research. As an alternative proxy for sponsorship influence, we recorded the nature of the affiliation listed in the affiliation section. Since in WOS records affiliations are listed in the same order as the authors, and since presumably authors list their multiple affiliations in decreasing order of importance, the first affiliation listed is likely to be the most representative for the first author.Type of affiliation: this was classified by MJS as “university”, “industry”, and “hospital” (total sample size, respectively, N = 742, N = 98, N = 39). Since the latter two categories were sparsely represented, they were merged together. Note that this variable overlaps with the previous two (hospitals and industries are non-ranked institutions, and authors affiliated with them may also be more likely to receive industry funding).
Analyses
All main results were derived by running a multivariable conditional logistic regression model of the form:In which Xs are the strata (triplets of case and controls), and Xi are the explanatory variables and interactions included in the model. In practice, the strata component of the regression equation is conditioned out of the maximum likelihood estimation, and the resulting analysis can be succinctly described by the following model:
where Pdup is the probability of observing an image duplication in one of the three articles within each stratum, therefore in relative terms. The coefficients for the non-interacted variables (citations/paper, female, institution rank, non-ranked and industry/mixed funding) express the change of the log odds (or the odds ratio, if exponentiated) corresponding to a one-unit increase in the value of the variable whilst holding all other variables constant (note that some variables were re-scaled to make the plots clearer, so their coefficients reflect the new scale, as indicated above). The coefficients of the interaction terms (i.e. team size*countries/author and years active*papers/year) measure the rate at which the coefficient of either one of the variables changes as the value of the other variable changes. Therefore, the coefficients for the corresponding non-interacted variables (i.e. team size, countries/author, years active, papers/year) correspond to the change in log odds associated with a one-unit increase for that variable when the other variable in the interaction has value zero and all other variables are held constant.This project was reviewed and approved by the Research Ethics Committee of the London School of Economics and Political Science.
Results
Descriptive statistics
Of the 293 papers with problematic images that could be matched with two controls and therefore included in the analysis, N = 72 were of duplication Category 1 (most likely to reflect unintentional error), N = 155 of Category 2, and N = 66 of Category 3 (most likely to represent intentional misrepresentation). Most articles had a first author based in China (N = 136) or USA (N = 59), followed at much distance by Italy (N = 11), South Korea (N = 10), Canada (N = 9), France (N = 7), India (N = 7), the UK (N = 7), and Germany (N = 5), whereas the remaining 19 countries had N≤4.
Risk of category 2 and 3 duplications by country groups
If, following [14], we limited the sample to duplication categories 2 and 3 (most likely to reflect intentional manipulation) and we tested the explanatory factors on a sample with all countries combined, we observed no significant relation with the odds of image duplication, except for a weak positive association for institution rank (Fig 1A). If we compared, again following [14], Western/industrialised countries to the others, the positive association with institution rank was only observed in the latter group, whereas the former showed a significant interaction between years of activity and productivity (Fig 1B and 1C).
Fig 1
Regression estimates of the association between characteristics of articles (first authors) and the odds of containing a problematic (category 2 or 3) image duplication, for all articles (a) or country subsets (b, c). Each panel shows results of a separate conditional multiple regression analysis, run on a matched-control sample in which each paper with image duplications (N = 293) was matched with two controls with the same country of affiliation of first author. Error bars represent 95% confidence intervals, and the dots and stars above error bars indicate rejection of the null hypothesis of no association (i.e. OR = 1, dotted line) at conventional levels of statistical significance (°P<0.1, *P<0.05, **P<0.01, ***P<0.001).
Regression estimates of the association between characteristics of articles (first authors) and the odds of containing a problematic (category 2 or 3) image duplication, for all articles (a) or country subsets (b, c). Each panel shows results of a separate conditional multiple regression analysis, run on a matched-control sample in which each paper with image duplications (N = 293) was matched with two controls with the same country of affiliation of first author. Error bars represent 95% confidence intervals, and the dots and stars above error bars indicate rejection of the null hypothesis of no association (i.e. OR = 1, dotted line) at conventional levels of statistical significance (°P<0.1, *P<0.05, **P<0.01, ***P<0.001).Marked differences were observed if the same regression model was run on subsets of countries that had theoretically relevant differences, which previous analyses showed to be significantly associated with problematic image duplications [14]. In particular, papers with first authors from the two countries within our sample that have a cash-based incentives policy (that is, CN and KR) exhibit positive associations with institution rank score, career length, papers/year and with their interaction, which are not observed at all in the other countries (Fig 2A and 2B).
Fig 2
Regression estimates of the association between characteristics of articles and of first authors and the odds of containing a problematic (category 2 or 3) image duplication, for all articles grouped according to theoretically relevant characteristics of the country of first authors.
Each panel shows results of a separate conditional multiple regression analysis, run on a matched-control sample in which each paper with image duplications (N = 293) was matched with two controls with the same country of affiliation of first author. Error bars represent 95% confidence intervals, and the dots and stars above error bars indicate rejection of the null hypothesis of no association (i.e. OR = 1, dotted line) at conventional levels of statistical significance (°P<0.1, *P<0.05, **P<0.01, ***P<0.001).
Regression estimates of the association between characteristics of articles and of first authors and the odds of containing a problematic (category 2 or 3) image duplication, for all articles grouped according to theoretically relevant characteristics of the country of first authors.
Each panel shows results of a separate conditional multiple regression analysis, run on a matched-control sample in which each paper with image duplications (N = 293) was matched with two controls with the same country of affiliation of first author. Error bars represent 95% confidence intervals, and the dots and stars above error bars indicate rejection of the null hypothesis of no association (i.e. OR = 1, dotted line) at conventional levels of statistical significance (°P<0.1, *P<0.05, **P<0.01, ***P<0.001).If countries were divided by another variable that [14] had identified as significant, i.e. the socio-culturally relevant distinction of German-developmental countries vs. other, the effects of career length and productivity were no longer significant or significantly different between the two groups (Fig 2C and 2D).If the analysis was limited to China on the one hand and USA, UK and Canada on the other (to increase sample size, we tested these three countries together under the assumption that they are relatively homogenous socio-culturally), the profile difference appeared even more extreme. USA, CA and UK only exhibited a weak (not formally significant at the 0.05 level) negative association with citation rates, and null associations with the other factors, most of which were in the opposite direction relative to the China subset (Fig 2E and 2F).The main effects and interactions of team size and number of countries are weak and non-statistically significant in all analyses, but they exhibit effects that are in the same direction observed in [14] (where the total number of countries was used instead of the per-capita number) and show differences between all pairs of subgroups (Fig 2A–2F).
Interpretation of findings
To help interpret the findings in Figs 2 and 3 reports probabilities calculated from the models reported in Fig 2A and 2B. We fixed the values of gender and funding status, and assessed how the estimated relative probability to have a category 2–3 duplication (note: the probability for treatment vs. controls in our sample, not in the general population) changed as we varied the average productivity of the author, their career length, the ranking of their institution, their average citation rate, the number of co-authors in the article and the number of countries. The resulting fitted probabilities may not represent any real combination of characteristics and therefore their value should not be taken literally, but rather in relative terms, i.e. as an illustration of how the risk of image duplication varies with varying conditions, according to our model’s estimates.
Fig 3
Fitted probabilities calculated from the regression results shown in Fig 2A and 2B.
Each panel shows the rate of change of the relative probability of being a “treatment” (article with problematic category 2 or 3 image duplication) rather than a “control”, as the values of some author and team characteristics are held constant (those indicated in the subtitle) whilst other characteristics vary (indicated in the figure or in the panel titles).
Fitted probabilities calculated from the regression results shown in Fig 2A and 2B.
Each panel shows the rate of change of the relative probability of being a “treatment” (article with problematic category 2 or 3 image duplication) rather than a “control”, as the values of some author and team characteristics are held constant (those indicated in the subtitle) whilst other characteristics vary (indicated in the figure or in the panel titles).Amongst cash-incentives countries, the relative probability of reporting a problematic duplication increases as productivity increases for individuals with shorter careers, and decreases for those with longer careers, regardless of other conditions (Fig 3, left panels). The overall probability is higher if the author works in a low-ranking institution and if the author is highly cited, whereas it decreases when the paper was co-authored with a larger, international team (Fig 3B and 3C, left panels).The opposite is observed in the other countries. There is no relevant association between the risk of problematic duplication and any combination of productivity, career length and university ranking (Fig 3, right panels). However, the risk is substantially lower for authors who are highly cited (Fig 3B), and it is higher for large international teams (Fig 3C). Note how these effects are the exact opposite of what is observed in the two cash-incentives countries.
Robustness of findings and secondary analyses
The main variables and analyses presented above had been pre-registered (osf.io/uzykd) as part of a plan to repeat all of the analyses in [14], testing each variable in univariate tests and then running multiple regressions as a secondary analysis. In this article we focused on the multivariable analysis, mainly for brevity and because multivariable analyses constitute more accurate tests of statistical associations. Data and code included with this article can be used to run all univariate analyses.The multivariable model used in this study is analogous to that presented in [14], except for the recoding of number of countries into countries per author and for the inclusion of an interaction term between career length and average publication rate—variables that in [14] were treated as independent effects only. In retrospect, an interaction term should have been included in the previous analysis. If the recoded country/authors variable and interaction are included in a re-analysis of data in [14], similar results to this analysis are obtained: both the interaction terms of team size*countries/author and career length*papers/year yield significant signals, but only in cash-incentives countries (respectively, OR = 5.146, z = 2.158, P = 0.031 and OR = 0.122, z = 1.954, P = 0.051. Full output of these analyses are included in the R code, see SI).We explored the possible presence of other interaction effects between career length and ranking or citations, but we observed no significant interaction effect and/or notable difference between cash-incentives and non-cash incentives, suggesting that the patterns reported in Figs 2 and 3 are specific to this categorization.If a binary variable representing type of affiliation (i.e. university or hospital affiliation vs. other kind of affiliation) was added to the model in place of the “non-ranked institution” and the “industry/mixed funding” variables, results were similar, showing a lower probability of duplication for non-university affiliation in the non-cash incentives countries (b = -0.76±0.45, z = -1.78, P = 0.0759.).Adding a categorical variable that reflects the three “tiers” recognized amongst Chinese universities (total sample size for Tier 1, 2 and 3, respectively N = 175, N = 62 and N = 170) did not change the results nor did it significantly increase the fit of the model (χ2 = 1.1092, df = 2, P = 0.574). Relative to tier 1, tier 2 and tier 3 university affiliation showed a positive but weak (not statistically significant) association with problematic image duplications (b = 0.36±1.43, z = 1.054, P = 0.292 and b = 0.21±1.23, z = 0.784, P = 0.433, respectively), suggesting that the misaligned incentives issue is less directly connected to the Chinese national tier system, and more directly linked to international rankings.If gender was recoded attributing all unclassified individuals as “female” instead of “male”, results were identical, showing very small and non-statistically significant effects of gender in both cash- and non-cash incentives countries (respectively, b = -0.06±0.94, z = -0.328, P = 0.7430, and b = 0.16±1.17, z = 0.785, P = 0.432).In a previous encoding of the “industry and mixed” dummy variable, which incorrectly included all types of mixed combinations of funding, we observed a significantly lower probability of duplication for this category amongst non-cash incentives countries. We explored this possibility further, by testing, in the same multiple regression model, all dichotomizations of funding sources, i.e. dummy variables for “no-profit/mixed fund”, “university/mixed fund”, “government/mixed fund” (see S1 Code). All of these dummy variables tended to be negatively associated with the probability of category 2–3 duplications in non-cash-incentives countries, but the null hypothesis was far from being rejected in all cases. Therefore, we found no evidence for systematic effects of funding of any types, once controlling for the other factors discussed in this analysis.
Risk of category 1 vs. category 2,3 duplications
All results presented above are based on duplications of category 2 and 3, because these are more likely to represent intentional misrepresentations. In agreement with this interpretation, [14] noted that most univariable statistically significant associations with theoretically relevant variables occurred amongst category 2 and 3 duplications. However, the frequency of duplication types is unequally distributed amongst countries. In particular, the probability of producing duplications of more problematic kinds is lower in the USA and higher in China, with other countries having intermediate frequencies (Fig 4). Indeed, a logistic regression model suggests that the risk of producing category 2–3 rather than category 1 duplications is, holding all other variables constant, significantly lower for the USA relative to other countries, whereas that of China is as high or higher (Fig 5). Further, holding country and all other variables constant, the risk is significantly lower for highly-cited authors (Fig 5).
Fig 4
Number and percentage of problematic image duplications in our sample, partitioned by country of first author and by severity of duplication (where category 3 is the most problematic, because it is most likely to reflect intentional misrepresentation of data).
Fig 5
Binary logistic regression estimates of the association between characteristics of “treatment” articles in our sample (N = 293) and the odds of reporting an image duplication most likely to reflect intentional misrepresentation of results (i.e. category 2 or 3).
Error bars represent 95% confidence intervals, and the dots and stars above error bars indicate rejection of the null hypothesis of no association (i.e. OR = 1, dotted line) at conventional levels of statistical significance (°P<0.1, *P<0.05, **P<0.01, ***P<0.001).
Binary logistic regression estimates of the association between characteristics of “treatment” articles in our sample (N = 293) and the odds of reporting an image duplication most likely to reflect intentional misrepresentation of results (i.e. category 2 or 3).
Error bars represent 95% confidence intervals, and the dots and stars above error bars indicate rejection of the null hypothesis of no association (i.e. OR = 1, dotted line) at conventional levels of statistical significance (°P<0.1, *P<0.05, **P<0.01, ***P<0.001).If the same set of variables is fitted with an ordinal model, to assess the risk of category 1 vs. 2 vs. 3, similar results are obtained overall, but the threshold separating categories 2 and 3 is non-significant, which corroborates previous suggestions that category 2 and 3 do indeed represent a distinctive and relatively homogeneous category of problematic image duplications compared to category 1 [14].
Discussion
A previous analysis of the same data set had identified country of author affiliation as a highly significant modulating factor for the risk of scientific misconduct [14]. This study completed and extended that analysis by assessing how the probability of duplicating images varies with commonly hypothesised risk factors for scientific misconduct, once country-level effects are controlled for. Results of this and the previous analysis suggest that some common hypotheses about the social and psychological determinants of scientific misconduct are not or negatively supported, whereas other hypotheses may or may not be supported, depending on the country of affiliation of the author, possibly due to the structure of incentives policies that operates in the country.The hypothesis that gender is a risk factor for misconduct is not supported. In none of our analyses was there any significantly higher risk for men or women. In most analyses the estimated risk was very close to equal (Figs 1 and 2). Precision in gender classification tends to be lower for East Asian countries, and a small percentage of names was not classifiable and coded as male (5.8%). However, classifying these names as female did not alter the results. More importantly, the lack of difference between genders was observed across Western European and North-American countries, where names are typically gendered and therefore errors in gender attribution are extremely low. Along with similar evidence obtained when comparing retractions and corrections [6], and bias in meta-analyses [7], these results do not provide any support for the hypothesis that psychological differences between men and women modulate the risk of misconduct. The over-representation of males in US ORI findings of scientific misconduct remains to be explained, but it may be due to confounding factors such as gender differences in the number of grants held, the number of publications produced, or other traits that affect the likelihood that an ORI investigation is initiated and concludes with a finding of misconduct, as suggested in [13].The hypothesis that private sponsorship may increase the risk of misconduct was inconclusively tested, because our test had very low statistical power. In all analyses we failed to observe the predicted association between risk of duplication and private funding. However, the number of studies in our sample that were fully or partially funded by industry was very small (N = 22 in total, and N = 14 for category 2–3), which severely limited our ability to detect any pattern. Exploratory analyses of other funding categories, in which the numbers were more balanced, also failed to suggest any systematic pattern. Therefore, the possibility of a link between funding source and risk of misconduct, whilst not especially suggested by our findings, remains to be assessed, and our findings do not challenge the established association between funding sources and bias [9]. These observations are consistent with earlier studies that operated a distinction between research bias and fraud [24].Evidence for the other hypotheses is more complex than commonly presumed. In China and Korea, countries in which high-impact publications are rewarded financially, the highest probabilities of manipulating images are observed in articles with authors who are highly productive, early-career researchers, have high citation rates and work in small, local collaborations and in low-ranking institutions (Fig 3). Further, authors from these countries appear generally to be at a higher risk of misconduct (Fig 5), as already suggested in previous analyses [6,14]. However, radically different patterns were observed in the other countries. Here the estimated risk is lower in small local collaborations, regardless of an author’s productivity, career length or institution’s ranking. And being a highly cited author is associated, in these countries, with a significantly lower risk of problematic duplications (Fig 3).Most of the differential effects observed in China and Korea relative to other countries may be plausibly explained as resulting directly or indirectly from the incentives created by a cash-reward policy for high-impact publications. This hypothesis is statistically supported by our results, in which the “cash incentives” country category was associated with significantly more effects than alternative country categorizations (Fig 2). The hypothesis also offers the most coherent explanation of our multiple findings. Cash rewards may encourage authors to “cut corners” in their work, particularly if such authors work in low-ranking institutions, where salaries are low and cash rewards are largest [18], and if they are early-career researchers, and therefore have less to lose and more to gain from taking risks [25]. Working in a small and local team (a team from a low-ranking institution) may also reduce the level of mutual control and criticism, increasing the risk. Further, since cash incentives are typically only given for high-impact publications in international journals, these problematic authors would be expected to publish more articles in English-speaking journals and seek high levels of citation, which would result in the observed positive correlation between productivity, citation rates and risk of image duplication. Conversely, authors in countries with different types of policies have none of these incentives, and instead are likely to see their work scrutinised in proportion to its impact, which would plausibly result in highly cited authors being less likely to produce papers with manipulated results. The risk of image duplications might increase in larger international collaborations if these collaborations are more likely to include co-authors from high-risk countries and conditions.Fully or partially alternative explanations, however, are also possible. For example, we cannot exclude that researchers in USA, UK, Canada and other countries have greater awareness of policies and principles of research integrity and have also access to better resources, which might enable them to either avoid forms of misconduct or, more cynically, to misrepresent data in subtler forms that are harder to discover (for example, they could be manipulating images in subtler and more elusive ways). Other entirely alternative explanations might invoke, for example, general socio-cultural differences between the countries involved. One of the referees on this manuscript, in particular, suggested that countries like China tend to be guided morally by a “culture of shame”, in which individuals strive to meet social expectations by any means necessary, and shame arises primarily from being caught. This is in contrast to other countries, including Western, Judeo-Christian countries, which may be dominated by a “culture of guilt”, in which the moral unit is the individual, not the collective. A “shame culture” vs. “guilt culture” variable could represent an alternative explanation of at least some of the differential patterns of misconduct observed across countries. Whilst plausible, all of these explanations are merely speculative, and may be tested in future studies.What can these results, combined with previous analyses, tell us about the controversial pressures to publish hypothesis? In our interpretation, they do not support it, but they do offer partial support of the hypothesis that misaligned incentives may increase the risk of misconduct. The pressure to publish hypothesis is not supported because the only significant patterns we detected in these and previous analyses support a role for cash incentive policies, which do not constitute real “pressures” on researchers, but rather an incentive that is misaligned with the scientific norms of disinterestedness [26]. Researchers in cash-incentives countries may of course feel additional kinds of institutional pressures, analogous to those reported by scientists in other countries [4,27]. However, most scientifically active countries do not exhibit the patterns predicted by the pressures to publish hypothesis, including countries like the UK or the USA, where genuine “pressures to publish” may be imposed on researchers in the form of bibliometric performance evaluation standards. To any extent that this and previous analyses reflect general patterns of research misconduct, current evidence suggests that in the latter countries, incentives are operating as they should: the higher the impact an author accrues, the less likely they are to be authors of flawed or falsified findings, regardless of their productivity, career level, or institution’s reputation. Further, in these countries, privately sponsored research appears to be at a similar or lower risk of producing fraudulent results.Our results should not be interpreted to mean that the perception of pressures to publish is not relevant in individual cases of scientific misconduct. To the contrary, the history of individual cases of misconduct often suggests that stress and performance expectations are contributing factors [28,29]. However, findings in this and multiple previous analyses fail to support the claim that pressures to publish are a systemic and generalized risk factor for misconduct, and they suggest that the relation between institutional incentives and research quality is complex and requires nuanced and contextualised analyses.Before drawing conclusions, however, a few important limitations of our analyses must be emphasised. First, our data set is not representative of all forms of data manipulations and all research areas. We analysed a single form of data manipulation (and/or problematic error), measured in articles that were all published in a single journal. Further, the studies were selected for screening based on their use of relatively similar types of methodologies (i.e. results in the form of cell or gel images), and thus represent a rather narrow range of research types. Therefore, incentives and risk factors that we found to be significant in this study may not apply to other forms of misconduct and other journals and research areas. Furthermore, the study was not designed to sample different countries and, because of the matching procedure, our data were dominated by authors from China and USA. Therefore, our results may not be generalizable to fields and countries not represented in the sample. The type of data manipulation examined in this and previous related analyses is relatively unsophisticated. Future studies might need to assess whether other (and possibly subtler) forms of data misrepresentation are more or less common than this form, and how they might relate to various risk factors.Second, our regression models, whilst including factors that are hypothesised to be relevant to scientific misconduct, may have excluded other relevant but unknown factors. Therefore, our regression estimates, like all regression estimates, are imperfect approximations and might still be biased due to omitted variables. These limitations entail that the specific numbers (odds ratios and probabilities) that we calculate should not be considered in absolute terms, but rather in relative terms, as theoretically meaningful comparisons between the relative importance of common risk factors. In other words, this study mainly assesses how factors make a greater or lesser difference relative to each other.Third, our analysis focuses on data for first authors. The two main justifications for this choice are the fact that first authors are the main or sole beneficiaries of cash reward incentives [18], and the plausible assumption that first authors, who are typically the individuals who most contribute to the article, are most likely to be responsible for irregularities in the data. Our previous studies support the validity of this choice, by showing that predictors of misconduct tend to be more significantly associated with characteristics of first author relative to last author or an average of the team [6,14]. To further assess the validity of this assumption, we randomly sampled 20 articles and checked the contribution statements. In all 20 cases, the first author was indicated as having “performed the experiments”, in 16 cases as also having analysed the data, and in most cases s/he was indicated as having contributed to all or most of the other roles. This strongly supports our assumption that the first author is most likely to be responsible for any manipulations in the data. However, absent a formal investigation, it is not possible to ascribe with certainty a suspected misconduct to a particular individual, and indeed in some cases more than one individual could have been involved. Further, it is conceivable that incentives for misconduct may differ depending on whether one is a trainee, a research scientist, a junior faculty member or a senior faculty member. Therefore, much of the complexity and diversity of phenomena underlying research misconduct may be obscured by imperfect measurement and heterogeneity in this analysis.It should also be emphasized that, although our main results focused on forms of image duplication that are most likely to result from intentional manipulation, the specific nature and causes of these duplications was not ascertained, and a sizable proportion of them may still not constitute scientific misconduct defined as the intention to misrepresent research results. One cannot judge intent in figure construction from visual inspection alone, and categories 1–3 should be considered provisional binning for the severity of the types of problems found in figures rather than categories where the causes of the figure problems are known and understood. The exact proportion of data irregularities that are due to intentional manipulation is unknown and it is generally likely to vary by research area and journal. An in-depth investigation conducted at the journal Molecular and Cell Biology found that the vast majority of duplicated images were the result of honest error, and only 2/59 articles with image duplications (3.3%, 0.2% of all articles examined) were deemed to reflect an intention to deceive [30]. An analogous screening program at the Journal of Clinical Investigation found evidence of probable scientific misconduct in only 2/55 of images (3.6%, 1% of all articles examined) [31].A few key strengths of this study should also be emphasised. First, the matched-control approach we adopted, combined with a conditional logistic regression, is an very powerful method, which allowed to control for crucial confounding factors that are rarely accounted for in studies on misconduct. Despite the non-random sampling, therefore, our results are quasi-experimental in nature. Second, this analysis employed a relatively large sample size. This gave it considerable statistical power: a post hoc analysis based on unconditional logistic regression suggested that the study had a statistical power of 99.9% to detect a small-size effect (i.e. OR = 1.5) when taking into account papers from all duplication categories (N = 879), and 99.7% when limited to categories 2 and 3 papers (N = 663). Therefore, whereas our results must be interpreted with caution with regards to their generalizability, they may be considered internally valid and highly precise comparisons between the relative importance of risk factors for scientific misconduct.In light of their strengths and limitations, and combined with results of a previous analysis, these results suggest that scientific misconduct is a highly complex phenomenon, the nature and causes of which depend on interactions between multiple risk factors, operating at different levels. The probability to engage in misconduct may be highest amongst authors working in countries whose contribution to the international literature is rapidly expanding, such as China and India, and particularly if they are rewarded with cash when they publish in international and high-impact journals. Within these countries, commonly hypothesised predictors of scientific misconduct, including being at an early-career stage and being incentivized by metrics of productivity and impact, may indeed be significant. However, in most other countries these factors may be irrelevant or have the opposite association with misconduct. Our results, like all observational studies, cannot conclusively prove causation, but they strongly suggest that the question of whether many common hypotheses and speculations about the “causes” of scientific misconduct are correct or not may depend on the socio-cultural context.It is encouraging to observe that, whilst complex, these patterns follow a logic that can be partially understood, suggesting that the risk of misconduct and QRP may be reduced with appropriate preventive policies and initiatives. However, such policies and initiatives may need to be designed to address the specific needs and problems faced by a particular country or community.
csv file with all raw data, with UT codes anonymized.
(CSV)Click here for additional data file.
R code with all analyses and secondary tests.
(R)Click here for additional data file.27 Apr 2021PONE-D-20-40614Individual and institutional predictors of misconduct are modulated by national publication incentives policyPLOS ONEDear Dr. Fanelli,Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.Your manuscript was carefully reviewed by experts in the field that made important comments and suggestions that if addressed rigorously will greatly improve with interesting work. Please ready carefully and address every comment and suggestion as best as you can. Make changes, in the manuscript, if you find appropriate. I am looking forward to receive the revised version at your earliest convenience. Thank you. Please submit your revised manuscript by Jun 05 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.Please include the following items when submitting your revised manuscript:A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript.Kind regards,Cesario BianchiAcademic EditorPLOS ONEAdditional Editor Comments:Dear Dr. Fanelli,Your manuscript was reviewed by 3 experts that found your manuscript of interest. Scientific misconduct is a very important topic that has brought more attention with the high number of papers about covid-19 been retracted recently.Please, ready carefully all comments and address thoroughly every single one. They are all very meaningful and answering them will greatly improve your interesting work. I personally would like to see some examples of image misconduct e correctness. I am also I little concerned about addressing misconduct by countries. I agree with one of the reviewers that would like to see some data from other countries ( is the reviewer example us turkey) and Brazil, for example.More than once the reviewers would like more methodological details.Cash incentives may come from direct payment per article but may be an important factor for fellowships and/or grant renewals.Journal Requirements:When submitting your revision, we need you to address these additional requirements.1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found atandhttps://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf2. Please note that according to our submission guidelines (http://journals.plos.org/plosone/s/submission-guidelines), outmoded terms and potentially stigmatizing labels should be changed to more current, acceptable terminology. In order to avoid conflation between gender and sex, "female” or "male" should be changed to "woman” or "man" as appropriate.3. Thank you for stating the following in the Competing Interests section:"The authors have declared that no competing interests exist."We note that one or more of the authors are employed by a commercial company: Harbers Bik LLC.3.1. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.Please also include the following statement within your amended Funding Statement.“The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement.3.2. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc.Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests4. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.[Note: HTML markup is below. Please do not edit.]Reviewers' comments:Reviewer's Responses to Questions
Comments to the Author1. Is the manuscript technically sound, and do the data support the conclusions?The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: YesReviewer #2: YesReviewer #3: Partly********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: YesReviewer #2: I Don't KnowReviewer #3: Yes********** 3. Have the authors made all data underlying the findings in their manuscript fully available?The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: YesReviewer #2: YesReviewer #3: Yes********** 4. Is the manuscript presented in an intelligible fashion and written in standard English?PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: YesReviewer #2: YesReviewer #3: Yes********** 5. Review Comments to the AuthorPlease use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This is an interesting manuscript that re-uses an illustrative dataset in order to disentangle individual and country level predictors for image duplication in research articles. They an show that once country effects are controlled for some factors for which prior evidence is inconclusive do not appear to have a significant impact.The analysis is based on a dataset that was used before. The description of the dataset used for the study at hand on p. 4 is somewhat confusing. Is the same sample as in ref [14] used here or a larger sample? Please also provide some basic information about the data and do not expect the reader to get into ref [14] and [15] to search for this information.The analysis is based on a case-control matching. Unfortunately, the matching approach is described very briefly on p.6. It suggests that the country information is the only information used for the matching. But this seems unlikely. Hence, I would like to see some more information:- Are the matched articles from the same year- Same journal or at least subfieldThe authors only match on the first affiliation mentioned on the article assuming that the lead author is the responsible author. Can the authors provide a robustness check for single-authored studies? If the sample becomes to small then I wonder whether the results hold if the authors match on the country combination of the author team.Institution rankings vary over time. Would it be possible to use the time-varying ranking for your analysis?Reviewer #2: A few general comments the authors may want to consider:1. Although I understand the general principles behind the methods of analyses used, I don't have sufficient expertise to judge whether they have been performed appropriately and rigorously.2. Isn’t it possible that differences in findings of research misconduct are based on factors other than whether the research misconduct occurred? For example, what if researchers in the US, Canada, and the UK have the resources, awareness, etc. to do a better job of cheating (e.g., using images from some other project so there are no duplicates in the same paper)?3. The manuscript recognizes that duplication of images is only one form of misconduct. In theory, if the factors highlighted are correct (i.e., the importance of financial incentives), then this would generalize to many other forms of misrepresentation of research findings as well as simply sloppy research practices. That said, it seems duplicate images in a given paper seems like a very specific kind of misconduct, and one which is pretty unsophisticated (e.g., a moment’s thought would recognize that people often get caught for this misconduct, and being caught is easily avoided by just using other images that aren’t in the same paper). It may warrant some discussion that if inappropriate incentives are driving this kind of misconduct, then it is likely only the tip of the iceberg and/or something very different is going on here then with other kinds of misconduct.4. If a paper has duplicate images, that could be something all authors were complicit in, but it’s at least as plausible that the duplication was committed by one of the authors and the other authors were insufficiently knowledgeable and/or attentive to have noticed the duplication. In that case, if a connection is being made to financial incentives for publication, it would be important to be clear (a) who gets the incentive (all authors, first author, corresponding author) and (b) who did what (e.g., if the paper has an author contributions section it might be possible to determine which author(s) was/were responsible for the images in question).5. Finally, while I don’t think it’s essential for publication of this work, shouldn’t it be possible to run a fairly simple experiment of using the factors identified as predictors of image duplication by selecting 10 examples of duplication of images and 20 controls from each of 2 or 3 other journals?Reviewer #3: Peer ReviewHugh DesmondIHPST (CNRS/Paris I)University of AntwerpIn this article, the authors re-analyze previous data on publications involving image manipulation, in order to gain a more precise understanding of the potential causes of patterns of misconduct. The wider significance of this research is to investigate the “misaligned incentive” view of scientific misconduct. According to this view, the prevalence of scientific misconduct is not simply due to isolated “bad apples” – ill-intentioned and/or anti-social individuals – but also due to the broader incentive system in science. The authors find that this view is only partially supported: only in some countries (i.e., China and South Korea, which are characterized by cash incentives) do (career) incentives lead to increased misconduct. This adds to a previous study by Fanelli et al. 2019.This is a well-constructed study and a valuable contribution to the literature. The misaligned incentive view is indeed still quite prevalent – often it is even the default view. It’s not clear it deserves that dominance. So this study is a welcome contribution to the debate about the true causes of patterns of research misconduct.My comments concern how the study is advertised. A safer but less interesting title would be: “Productive early career academics in low ranking institutions in South Korea and China more likely to duplicate images.” Justification is lacking why the results support conclusions about misconduct in general or about cash incentives in general (and not just cash incentives *in China*). Moreover, there are disclaimers cautioning the reader against reading too much into the study – which is of course fine, but puzzling nonetheless, since the authors’ disclaimers seem to to directly undercut the way the study is advertised.Comment 1 -- Scope of conclusion: misconduct or image duplication?This study is advertised as concerning “misconduct”, even though it only tests for factors that predict image duplication. However, do patterns of behavior regarding image duplication necessarily hold for patterns of misconduct more generally?The disclaimer on lines 515ff (see comment 3) is especially puzzling, since the authors caution against drawing any conclusions about misconduct in general.Moreover, there is a concrete reason why the null result found for Western researchers may not hold for misconduct generally. Image duplication is an egregious form of misconduct, and by 2013/2014 (the years of the analyzed studies) there would have been strong awareness of the norm of FFP for researchers in Western institutions. Moreover, image duplication is relatively easy to detect. So by this reasoning, the incentives to engage in egregious misconduct would not be strong enough: probability of being caught is too high, and researchers are aware of this probability. However, career incentives might still promote more subtle forms of misconduct. For instance: salami slicing, guest authorship, or bullying are all patterns of misconduct/QRP that are difficult to charge scientists with. In fact, in some contexts such as the USA, only FFP is categorized as misconduct. So, it’s not that misaligned incentives do not show up in misconduct, they just show up in different, less detectable forms of misconduct. (Compare the situation to financial incentives and sports doping.)So, there should be additional argumentation why the study holds for misconduct in general, and more emphasis on just how the results are to be interpreted. Or else the scope of the title & abstract should be narrowed.Comment 2 -- Role of cash incentivesThe authors conclude that “Most of the differential effects observed in China and Korea relative to other countries may be plausibly explained as resulting directly or indirectly from the incentives created by a cash” (lines 468-470).While I agree that it is “plausible”, given that this is the headline result of the article, the reader would expect additional justification. There are many other factors that could explain the differential effects observed in China and Korea.a. Cash incentives vs. norms of academic hierarchyThe authors introduce the category of “german-developmental”, but don’t use it in the analysis and interpretation of the findings. Nonetheless, Figure 2c documents how in “german-developmental” countries, low institutional ranking of the first author predicts image duplication with statistical significance. So why could one not conclude that the increased image duplication in China/S-Korea is being caused by strict academic hierarchies – or not *also* being caused by this?Obviously the authors are limited here by the available data. It would be nice to have data from, for instance, Turkey, to test whether we see the same incidence of data manipulation. In absence of such data, what can be inferred about the role of cash incentives *in particular*?b. Cash incentives vs. RI cultureThe role of RI culture is acknowledged in lines 102-109, but it does not shape the discussion of the findings. Nonetheless, by many accounts, China has only in the past few years become more serious about educating researchers on RI norms. So, given that the data is from 2013/14, the observed patterns might reflect the regulatory situation rather than cash incentives in particular. This would also imply that the pattern of image duplication might not hold as strongly – or not at all – in 2021.To test this, one would need data from a scientifically active country with weak RI culture and with no cash incentives.c. Cash incentives vs. non-WEIRD psychologyIf you would introduce cash incentives in Germany, would image duplication shoot up dramatically? This is what the takeaway from this study implies. Yet somehow this scenario strikes seems implausible. There is much more at stake.For instance, one could point to research on how moral psychology in e.g. China seems to work quite differently, with a much stronger emphasis on “shame” rather than “guilt” (cf. Henrich 2020). In a shame culture, there is a very strong incentive to meet social expectations -- whatever those might be, and regardless of the *means* by which one reaches these expectations. The moral unit is the collective, not the individual, and shame only arises from being caught (not from doing the deed). In this view, cash incentives only matter insofar they express social expectations and the value of the collective – and hence only secondarily are about cash. By this line of reasoning, the difference-maker is not cash-incentives policy, but rather shame vs. guilt cultures.Comment 3 -- Conflicting signalsAt the very end, there are some sentences that read like a disclaimer (lines 515-519). The authors caution the reader that the results may not apply to “other forms of misconduct” or to “countries not represented in the sample”. If this disclaimer is true, then shouldn’t the takeaway from this study be revised?Minor CommentsComment 4Line 12 “A third improvement from [14] consists in testing three new variables that measure two commonly invoked risk factors.”This wording might be a bit confusing for the reader – I found myself looking for the “three new variables”. Perhaps clarify that what follows are the two risk factors.Comment 5In the beginning the main result is described as: “the risk of problematic image duplication was higher for highly productive, ***highly cited***, early career researchers working in low-ranking institutions”However, in Figure 2, citation rate is not seen to have any statistically significant relation to image duplication. Therefore, why is it mentioned as a result? Moreover, if it is not a statistically significant predictive factor, then why do a simulation of it at all (Figure 3)?Comment 6line 574-575 : “Therefore, we can conclude that many common hypotheses and speculations about the “causes” of scientific misconduct are likely to be correct or not, depending on the socio-cultural context.”Sentence doesn’t seem quite right. I presume the authors mean this: “Therefore, we can conclude that THE QUESTION OF WHETHER many common hypotheses and speculations about the “causes” of scientific misconduct are correct or not, DEPENDS on the socio-cultural context.”(Also, if this is the actual conclusion, why not have it reflected in title, abstract, introduction? Referring to the “socio-cultural context” is a more safer conclusion that referring to “cash incentive policies” in particular, and covers the roles academic hierarchy norms, RI cultures, or moral psychology might play.)Comment 7Line 576: “It is encouraging to observe that, whilst complex, these patterns follow a logic that can be understood, suggesting that the risk of misconduct and QRP may be reduced with appropriate preventive policies and initiatives.”But what logic though? If upheld, the paper seems to contribute to undermining the dominant logic, namely the “misaligned incentive” view. If this view is false, then research misconduct may prove to be much more intractable (in Western countries) than previously hoped?In fact, it would have been interesting to have heard slightly more about what the potential policy implications of this study would be. Should we give up on trying to nudge researcher behaviour through incentive engineering? (For a logic why incentive engineering cannot safeguard trust, see Desmond forthcoming in Philosophy of Science, http://philsci-archive.pitt.edu/18661/)********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: NoReviewer #2: Yes: Michael KalichmanReviewer #3: Yes: Hugh Desmond[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.28 May 2021A rebuttal to each reviewer comment is included in the PdfSubmitted filename: Rebuttal letter.docxClick here for additional data file.22 Oct 2021
PONE-D-20-40614R1
Are individual and institutional predictors of misconduct modulated by national publication incentives policy? Results of a matched-control analysis of problematic image duplications.
PLOS ONE
Dear Dr. Fanelli,Thank you for patience whilst we completed further evaluation of your manuscript. Please accept our apologies once again for the inconvenience caused by the rescinding of your previous Accept decision.After careful consideration, we feel that the manuscript has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Your manuscript has been assessed by a member of PLOS ONE’s statistical advisory board (Reviewer 3) and I am pleased to let you know that no concerns have been raised as part of this review. However, following internal editorial review of the manuscript, there are a number of revisions to the text that you should perform before the manuscript can be accepted for publication. These are noted as follows:Several concerns have been noted over the methodology for assigning gender:We note that you report a high percentage of unclassified authors from Chinese language names. Has the genderapi service been validated for names from all countries assessed in the manuscript? If not, this should be discussed as a potential limitation in the Discussion section.We understand that a gender was assigned to all unclassified names. Please provide additional details concerning how this manual gender assignment was performed, including specific protocols and a validation of this approach. Whilst we note that you have referred readers to Reference 14 for further information on this approach, this reference does not provide sufficient details for how this was done. This information should be included in the Methods section.It is indicated that most unclassified names were Chinese. Please update your Methods section to provide the exact figure.3.The article title implies conclusions/investigation of a causal relationship and frames the research around misconduct. We request a revision to the title for the following reasons:a)The title suggests that predictors of misconduct are modulated by national publication incentives policies. However, as noted in your Discussion, your manuscript cannot prove causation. The title of the manuscript should be updated to clearly indicate this.b)The use of “misconduct” in the title also requires additional nuance, since the manuscript does not directly measure misconduct (it only measures problematic image duplications)c)We suggest the following revision:A matched-control analysis examining what individual and institutional characteristics are associated with increased rates of problematic image duplications in scientific publications.4. Please update Reference 14 to include the issue, page numbers and correct year, as it is currently marked as ‘in press’.Figure 4 requires clarification: It is missing x-axis labels, and either (a) n values for each x-axis category should be specified in the legend, or (b) if “frequency” actually means “# of articles” then n values aren’t needed but the y-axis should be labelled more clearly.6. We understand that you obtained ethical approval to conduct this study. Please update the Methods section of your manuscript indicate this, along with the name of the body that provided ethical approval for the study to take place.7. We note that you have indicated that ‘All relevant data are within the manuscript and its Supporting Information files’. However, we note that not all of the analysis described in the manuscript can be reproduced from the data file provided. The datafile does not include the specific articles included in your analysis. We understand that you may not be able to make these data available with your manuscript. However, are you able to provide these data on request? If so, please include details of how such requests should be made in your Data Availability statement. Please note that it is not acceptable for authors to be the sole named individual responsible for ensuring data access. Your institution’s ethics or data protection committee regulate data access and act as a contact for requests of this data.8. In addition, we would be grateful if you could provide an additional tab in your ‘S1_dataset.csv’ with a legend providing a definition of each of the columns.Please submit your revised manuscript by Dec 06 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.Please include the following items when submitting your revised manuscript:
A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.We look forward to receiving your revised manuscript.Kind regards,Bradford Dubik, Associate EditorOn behalf of:Cesario BianchiAcademic EditorPLOS ONEJournal Requirements:Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.[Note: HTML markup is below. Please do not edit.]Reviewers' comments:Reviewer's Responses to Questions
Comments to the Author1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressedReviewer #3: All comments have been addressedReviewer #4: All comments have been addressed********** 2. Is the manuscript technically sound, and do the data support the conclusions?The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: YesReviewer #3: YesReviewer #4: Yes********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: YesReviewer #3: YesReviewer #4: Yes********** 4. Have the authors made all data underlying the findings in their manuscript fully available?The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: YesReviewer #3: YesReviewer #4: Yes********** 5. Is the manuscript presented in an intelligible fashion and written in standard English?PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: YesReviewer #3: YesReviewer #4: Yes********** 6. Review Comments to the AuthorPlease use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I am satisfied with the changes that have been made during the revision process.To reach the character limit I add: congratulations!Reviewer #3: (No Response)Reviewer #4: (No Response)********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: NoReviewer #3: Yes: Hugh DesmondReviewer #4: No[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.
16 Nov 2021Thank you for your feedback, here below are responses to each query.1. Several concerns have been noted over the methodology for assigning gender:a. We note that you report a high percentage of unclassified authors from Chinese language names. Has the genderapi service been validated for names from all countries assessed in the manuscript? If not, this should be discussed as a potential limitation in the Discussion section.b. We understand that a gender was assigned to all unclassified names. Please provide additional details concerning how this manual gender assignment was performed, including specific protocols and a validation of this approach. Whilst we note that you have referred readers to Reference 14 for further information on this approach, this reference does not provide sufficient details for how this was done. This information should be included in the Methods section.RESPONSE: We agree that errors in gender classifications could affect the results, but we note a number of robustness checks that makes this unlikely. The percentage of unclassified author names was 5.8%, which is not really a "high" percentage. Of these, little over 50% were Chinese/Taiwanese, which is a relative majority but, again, not suggestive of a special imbalance.As originally reported, the results were tested by re-assigning the unclassified names to either gender, and doing so makes not difference in the results. This would obviously be the case, since the percentage is small.More importantly, our conclusions about gender cut across countries, and are observed in US and Eu countries, where disambiguation errors are much rarer because names are typically gendered.The Methods section was altered as follows:"Most names were classified by a commercial service (genderapi.com), and the records that remained unclassified were subsequently completed by hand, via direct inspection of information available online. In particular, CVs and professional webpages were inspected for explicitly gendering information and, if this was not available, based on and photos, see [14] for details). This hand-coding was blind to the outcome (presence/absence of duplication in the paper). Names that remained unclassified (N=51/879, 5.8%) were re-coded as “male”. We used male as the default gender, based on the fact that most (i.e. 27) of these 51 unclassified names were Chinese or Taiwanese, and that about 75% of scientists in China are male [20]. The effect of this coding choice on results was checked with robustness tests reported in the Results section."The Discussion was extended with:"Precision in gender classification tends to be lower for East Asian countries, and a small percentage of names was not classifiable and coded as male (5.8%). However, classifying these names as female did not alter the results. More importantly, the lack of difference between genders was observed in Western-European and North-American countries, where names are typically gendered and therefore errors in gender attribution are extremely low. "2. It is indicated that most unclassified names were Chinese. Please update your Methods section to provide the exact figure.RESPONSE: The number is 27, which is a little above 50% the total. this detail was added in the methods section as shown above.3.The article title implies conclusions/investigation of a causal relationship and frames the research around misconduct. We request a revision to the title for the following reasons:a)The title suggests that predictors of misconduct are modulated by national publication incentives policies. However, as noted in your Discussion, your manuscript cannot prove causation. The title of the manuscript should be updated to clearly indicate this.b)The use of “misconduct” in the title also requires additional nuance, since the manuscript does not directly measure misconduct (it only measures problematic image duplications)c)We suggest the following revision:1. A matched-control analysis examining what individual and institutional characteristics are associated with increased rates of problematic image duplications in scientific publications.RESPONSE: We appreciate the concerns about causal implications, and note that the choice of words like "predictor" in the title was indeed intended to avoid the concept of causation (predictor is a statistical term). It should also be noticed that the first part of the title is intended to describe the substantive hypotheses being tested, which are indeed causal hypotheses and are not our own hypotheses but hypotheses proposed in the literature. These hypotheses are about scientific misconduct in general.Our study is a test of such hypotheses done on image duplications and, whilst observational in nature, it does come close to a causal analysis by using a matched control, and it is intended to produce evidence about such theoretical hypotheses.All that said, we are happy to try to remove any ambiguity or strong causal implications in the title. We appreciate the suggestion made by the editors, but we feel that the title suggested doesn't quite capture what this study was about - rather, it describes the previous analysis of which this study was a follow-up.We changed the title to: "Do individual and institutional predictors of misconduct vary by country? Results of a matched-control analysis of problematic image duplications."4. Please update Reference 14 to include the issue, page numbers and correct year, as it is currently marked as ‘in press’.RESPONSE: amended5. Figure 4 requires clarification: It is missing x-axis labels, and either (a) n values for each x-axis category should be specified in the legend, or (b) if “frequency” actually means “# of articles” then n values aren’t needed but the y-axis should be labelled more clearly.RESPONSE: We appreciate the suggestion to improve clarity. The figure was corrected.Title now reads:"severity of image duplication, by country", y-axis is titled "n of articles"And the legend explains:"Number and percentage of problematic image duplications in our sample, partitioned by country of first author and by severity of duplication (where category 3 is the most problematic, because it is most likely to reflect intentional misrepresentation of data)."6. We understand that you obtained ethical approval to conduct this study. Please update the Methods section of your manuscript indicate this, along with the name of the body that provided ethical approval for the study to take place.RESPONSE: The methods section now includes the sentence:"This project was reviewed and approved by the Research Ethics Committee of the London School of Economics and Political Science."7. We note that you have indicated that ‘All relevant data are within the manuscript and its Supporting Information files’. However, we note that not all of the analysis described in the manuscript can be reproduced from the data file provided. The datafile does not include the specific articles included in your analysis. We understand that you may not be able to make these data available with your manuscript. However, are you able to provide these data on request? If so, please include details of how such requests should be made in your Data Availability statement. Please note that it is not acceptable for authors to be the sole named individual responsible for ensuring data access. Your institution’s ethics or data protection committee regulate data access and act as a contact for requests of this data.RESPONSE: All identifying information from the original manuscripts is removed for ethical reasons, and therefore cannot be shared. However, we confirm that all the data required to reproduce all analyses in the study is available, albeit in an anonymised form.8. In addition, we would be grateful if you could provide an additional tab in your ‘S1_dataset.csv’ with a legend providing a definition of each of the columns.RESPONSE: A "Column legend" tab has been added to the file.Submitted filename: rebuttal.odtClick here for additional data file.2 Feb 2022Do individual and institutional predictors of misconduct vary by country? Results of a matched-control analysis o f problematic image duplications.PONE-D-20-40614R2Dear Dr. Fanelli,We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.Kind regards,George VousdenDeputy Editor-in-ChiefPLOS ONEAdditional Editor Comments (optional):Please note that the original Academic Editor handling your manuscript became unavailable and I have taken over its handling.Reviewers' comments:28 Jul 2021PONE-D-20-40614R1Are individual and institutional predictors of misconduct modulated by national publication incentives policy? Results of a matched-control analysis of problematic image duplications.Dear Dr. Fanelli:I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.If we can help with anything else, please email us at plosone@plos.org.Thank you for submitting your work to PLOS ONE and supporting open access.Kind regards,PLOS ONE Editorial Office Staffon behalf ofDr. Cesario BianchiAcademic EditorPLOS ONE