Literature DB >> 24028480

Understanding research misconduct: a comparative analysis of 120 cases of professional wrongdoing.

James M DuBois1, Emily E Anderson, John Chibnall, Kelly Carroll, Tyler Gibb, Chiji Ogbuka, Timothy Rubbelke.   

Abstract

We analyzed 40 cases of falsification, fabrication, or plagiarism (FFP), comparing them to other types of wrongdoing in research (n=40) and medicine (n=40). Fifty-one variables were coded from an average of 29 news or investigative reports per case. Financial incentives, oversight failures, and seniority correlate significantly with more serious instances of FFP. However, most environmental variables were nearly absent from cases of FFP and none were more strongly present in cases of FFP than in other types of wrongdoing. Qualitative data suggest FFP involves thinking errors, poor coping with research pressures, and inadequate oversight. We offer recommendations for education, institutional investigations, policy, and further research.

Entities:  

Mesh:

Year:  2013        PMID: 24028480      PMCID: PMC3805450          DOI: 10.1080/08989621.2013.822248

Source DB:  PubMed          Journal:  Account Res        ISSN: 0898-9621            Impact factor:   2.622


INTRODUCTION

Research and experience tell us that people find things interesting when they are new, unexpected, and complex (Silvia, 2005). Discussing why research misconduct is wrong meets none of these criteria. Falsification and fabrication of data constitute a form of lying and plagiarism a kind of stealing. As children most of us learned that lying and stealing are wrong. It is also relatively uninteresting to discuss the negative consequences of research misconduct because their harmful effects are so immediately evident: Fabrication and falsification contaminate scientific literature with incorrect information, which wastes funds, hampers medical and technological development, and poses risks to patients and consumers, while plagiarism deprives authors of credit for their work. However, we propose that the question “why do investigators engage in research misconduct?” is interesting because misconduct cases are typically shrouded in secrecy, and studies on professional wrongdoing suggest that the causes of misconduct are complex. These characteristics of research misconduct—secrecy and complexity—also make it very difficult to study the phenomenon.

THE PROBLEM OF SECRECY

A meta-analysis of surveys on research misconduct found that almost 2% of investigators surveyed admitted to fabricating or falsifying data at least once; when asked about the behavior of their colleagues, approximately 14% reported observing a case (Fanelli, 2009). Our team's recent survey of the 194 comprehensive doctoral or medical schools in the United States found that 88% of research integrity officers had investigated a credible case of wrongdoing over the past 2 years (range 1–16 or more cases); 54% of these cases involved research misconduct (DuBois et al., 2013a). With such prevalence rates, one would expect that research misconduct would be well studied and well understood. However, most cases are not reported to federal authorities—even cases involving federal funding (Titus et al., 2008). Cases that are reported are investigated confidentially. Although findings are published, federal investigators currently retain little detailed information on wrongdoers’ reported motives or their research environments. When investigation reports are provided in response to a Freedom of Information Act request, the quality and quantity of information is highly varied depending on the quality of the institutional investigation and degree of redacting deemed appropriate (personal communication with John Dahlberg, Office of Research Integrity, January 29, 2013). Even when an investigation finds the accused to be guilty of misconduct, our extensive literature reviews over the past four years have found that most such cases receive scant attention from investigative reporters.

THE PROBLEM OF COMPLEXITY

Wrongdoing in research is complex in at least two important ways. First, wrongdoing in research takes many different forms (DuBois et al., 2012; Martinson et al., 2005). In “Phase I” of our project, described in detail elsewhere (DuBois et al., In press), we researched and wrote synopses of 100 highly heterogeneous cases of wrongdoing in healthcare delivery and research. Five experts rated the severity or seriousness of wrongdoing found in each case, and we identified variables that correlated with more severe forms of wrongdoing. One lesson learned was that wrongdoing is not a homogeneous phenomenon; different kinds of wrongdoing are associated with different variables. Thus, it is misleading to speak of “organizational deviance” or “professional misbehavior” as though it is a unified phenomenon. Accordingly, we need to inquire separately into different forms of wrongdoing: Why would someone fabricate research data? Why would a clinical investigator enroll patients who do not meet the inclusion criteria approved by an institutional review board (IRB)? Why would an animal researcher not follow a protocol that requires administration of anesthesia before performing painful procedures? Second, there are many reasons why a researcher might behave unprofessionally. Some reasons concern personality traits. For example, an overly robust sense of self-entitlement and cynicism have been associated with poor ethical decision making (Mumford et al. 2006). Other explanations concern individual psychological responses to environmental factors. For example, several studies have found that stress—whether cognitive overload or exposure to stressful life events—is associated with diminished ethical decision-making (Mumford et al., 2001) and research misconduct (Davis et al., 2007). Other researchers have found that feeling unfairly treated by a system may contribute to skirting rules within that system (Martinson et al., 2006; Martinson et al., 2010; Keith-Spiegel and Koocher, 2005). Davis (2003) hypothesizes that culture may play a role in research misconduct: cultural norms inculcated in one's nation of birth or training may contribute to behavior that is deviant within U.S. culture. Finally, a review of recent criminological studies suggests that environmental factors that create opportunity for research misconduct deserve closer attention (Adams and Pimple, 2005). Consistent with this view, our team conducted a systematic literature review and identified ten environmental factors that may predict wrongdoing in research: Playing conflicting professional roles (Grover, 1993; Levine, 1992); Having financial rewards for wrongdoing (Jennings, 2006; Rodwin, 1993); Others benefitting from wrongdoing (Victor et al., 1993); Being penalized for following the rules (Hegarty and Sims, 1978; Jansen and Von Glinow, 1985); Others being penalized for doing what is right (e.g., reporting wrongdoing) (Schuchman, 2008); Unjust treatment of employees or unfair processes (Greenberg, 1993; Keith-Spiegel and Koocher, 2005; Martinson et al., 2006); Ambiguous legal or ethical norms (Davis, 2003; Meyer and Bernier, 2002; Shah et al., 2004); Vulnerable victims (Bandura, 1999; Zimbardo, 2007); Oversight failures; Occupying a position of authority over peers (Bramstedt and Kassimatis, 2004; Marshall, 1999). As we show in this previously published review, each of these ten factors can be understood as contributing motive, means, or opportunity (or a combination thereof) for wrongdoing in research, each is supported by some empirical studies, and each is evidenced in one or more cases our team has studied (DuBois et al., 2012).

RATIONALE FOR A HISTORIOMETRIC APPROACH

Historiometry can be defined as the statistical analysis of coded data from historical narratives. Our use of the method involves identifying appropriate cases, reading all the published literature we can find on a case, coding the presence of more than 50 variables, producing descriptive statistics, and analyzing data to identify which variables are significantly associated with a specific kind of wrongdoing. The limitations of such an approach are fairly obvious. For example, it is difficult to control for confounding variables when examining non-randomly assigned variables, and published reports may omit important information. However, historiometry offers several major advantages over more common methodologies: It can examine the relationship of a large number of variables within one study; its data sources foster ecological validity; and it is feasible.

Multivariable Friendly

Given the rich array of theories and data on the causes of research misconduct, it is important to identify a method that can track the influence of more than one variable at a time. Yet most experimental designs test the influence of a very small set of independent variables. In contrast, historiometry allows us to extract numerous variables from published reports and build complex statistical models (limited primarily by sample sizes rather than participant time).

Ecological Validity

Social psychology abounds with rigorous, prospective, experimental data on ethical decision-making obtained from university students with no history of wrongdoing. It is unclear whether such data—despite its methodological elegance—generalizes well to real world cases of wrongdoing among established researchers. Apart from differences in the relevant populations, the recent movement toward effectiveness research rests on the observation that randomized controlled trials often leave much to be desired when it comes to knowledge of real world applications (Depp and Lebowitz, 2007). We have very little data obtained directly from those who have engaged in wrongdoing, and much of what we do have derives from self-reports from individuals who may have very little insight into social factors that contributed to their behavior: Data suggest that self-serving biases operate below the level of awareness, and individuals may easily ignore the ways that system failures contributed to their behavior (Dana and Loewenstein, 2003; Katz et al., 2003). Historiometric data—being drawn from real world cases—has a much greater chance of being ecologically valid than alternative designs (Brewer, 2000).

Feasibility

It is not likely that anyone could conduct a study of research misconduct that involves a large random sample. Falsification, fabrication and plagiarism are relatively rare events that occur in the privacy of offices and laboratories. They are also events that we do not want to induce in a prospective manner, for example, through random manipulation of variables that may predict misconduct because the consequences are so serious. In contrast, historiometric studies are feasible.

Rationale for a Comparative Approach

As noted above, we cannot treat wrongdoing in research as though it is a unified phenomenon such as “organizational deviance.” Accordingly, in this article we examine three different sets of relatively homogeneous cases: research misconduct, other wrongdoing in research, and wrongdoing in medical practice. Although many cases of professional misbehavior involve multiple forms of wrongdoing, cluster analysis of our first 100 cases indicated that research misconduct does not overlap with human subjects or animal care violations or with violations in the domain of medical practice. This solves a major problem for the historiometric study of research misconduct, namely, that no one publishes detailed accounts of “good research,” which could serve as comparison or baseline cases.

METHODS

Our method involved identifying appropriate cases; conducting literature reviews on cases; reading all relevant literature; extracting and rating the presence of key variables; and statistically analyzing data to produce descriptive statistics and to examine the relationship between research misconduct and independent variables. Each element of our method is described below.

Case Sampling

As is common in historiometric research (Simonton, 2003; Mumford, 2006), we used criterion-based sampling. Table 1 presents our sampling criteria with a rationale for each. To identify which variables characterize research misconduct in a statistically significant manner, we sampled three distinct kinds of wrongdoing: research misconduct (FFP), other research wrongdoing (such as violations of animal care or of human subjects protections), and wrongdoing in medical practice (such as boundary violations, abuse of prescribing privileges, or fraudulent unnecessary procedures). All kinds of wrongdoing represented in our sample appear in a taxonomy we developed to guide sampling; the taxonomy has been used by our team with a free-marginal multirater Kappa coefficient of .85 (DuBois et al., 2012).
Table 1:

Case Inclusion Criteria

CriterionRationale
Location: USA– Provides a relatively clear and stable set of laws and cultural norms surrounding the 3 sampling areas
Date of occurrence: Jan. 1980–Dec. 2011– Case conclusion must be at least 1 year old to allow adequate reporting of information on the case
– Because widely published cases of misconduct are rare, it is important to work with a broad range of dates
– By 1980 norms prohibited all of the behaviors under investigation
Kind of wrongdoing: (1) misconduct, (2) other research violations, or (3) wrongdoing in medical practice– All involve serious professional violations
– Cases are distinct and may serve as comparison cases
Case coverage: Case is described in at least 5 published sources, including articles and reports of agencies and boards– Indicates social significance of case
– Ensures adequate availability of information on the case
Case involves a clear individual wrongdoer upon whom we can focus– We aim to understand and predict individual behavior, rather than corporate behavior, which may have different dynamics
Case Inclusion Criteria We decided to sample 40 cases of each kind of wrongdoing to create samples of equal size that, based on effect sizes observed in our past studies (DuBois et al., 2013b), would be large enough to detect significant differences if such differences existed. Once we established our sampling criteria, we conducted extensive reviews of the literature. Many cases were excluded because they were not described in a minimum of 5 publications. On the one hand, this means that our cases may not be representative of all such cases occurring in the United States during our sampling timeframe; on the other hand, Simonton argues that historiometric methodology does well to focus on the most well-known, or in our study, notorious cases: “The most eminent individuals in a domain are not only the most representative of the phenomenon of interest, but information about such subjects is likely to be more extensive and reliable” (Simonton, 2003, p. 625). Our literature reviews were conducted by developing search criteria in consultation with a research librarian and using search terms in LexisNexis Law (searching federal and state cases, newspapers, and magazines), Medline, Google, and Google Scholar. Additionally, for misconduct cases we examined all findings published by the Office of Research Integrity (ORI); for cases of other research violations we examined all findings published by the Office of Human Research Protections; and for cases of wrongdoing in medical practice we examined reports from the HHS Office of Inspector General and state medical boards.

Data Extraction

A research assistant (RA) read all publications relevant to a case (an average of 29 articles per case). Using a case datasheet and scoring guide, RAs extracted data on 51 variables pertaining to the setting (n = 5), the wrongdoer (n = 8), the wrongdoer's history of misbehavior and illegal activity (n = 4), how the case was reported and investigated (n = 7), consequences to the wrongdoer (n = 6) and wrongdoer's institution (n = 3), environmental variables (n = 12, scale items), case duration, complexity, and media coverage (n = 3), and the RA's theory of the case, including notable social dynamics and wrongdoer motives (n = 3, open-ended). While our primary independent variables of interest are the 10 environmental factors enumerated in the introduction, many of these variables attempt to address at least indirectly some of the individual variables mentioned in the introduction. For example, we tracked the individual's nation of origin and training to take into account the possible role of culture, and we tracked personal problems such as divorce or declarations of bankruptcy as markers of stressors. Continuous variables (such age, date, number of sources consulted) were entered as such; bivariate variables (such as foreign-trained or evidence of insanity plea) were entered as yes/no; and scaled items were scored on a scale from 1 to 3 using a benchmark scoring guide described in DuBois et al. (2013b). The meaning of the scale was defined for each variable, e.g., “1 = no evidence of conflicting roles; 2 = conflicting roles were played but managed; and 3 = conflicting roles were played and not managed.” The guide provides examples of each score for each variable. The team ensured the quality of case research through several means: RAs were all Ph.D. students in health care ethics who read a series of 21 articles explaining the relevance of each primary variable we tracked and then received 8 hours of formal training prior to researching their first case; individual RA's researched anywhere from 7 to 24 cases across the life of the project. All cases were reviewed by a Ph.D. co-investigator who read at least three articles on each case (Anderson); and the team met weekly with the principal investigator (DuBois) to discuss case research, questions on the scoring guide, and procedures. For this dataset only one RA per case produced all ratings because we previously established very high inter-rater reliabilities (ICC = .84–1.0) for all scaled values using a set of 100 cases (DuBois et al., 2013b).

Data Analysis

Data analysis involved generating descriptive statistics (frequencies and means) and analyzing the relationship of the kind of wrongdoing to all independent variables using Chi square to test for differences in frequency and ANOVA to test for differences in mean values on scaled items.

RESULTS

In what follows, we present descriptive statistics for the three forms of wrongdoing with tests of significance aimed at identifying when a given variable is more strongly associated with a specific kind of wrongdoing. We use tables to present statistical details; our text summarizes results. We conclude with a presentation of qualitative data on wrongdoer motives and social dynamics.

Frequency of Wrongdoer and Setting Variables

Table 2 presents frequencies for wrongdoer and setting variables. Chi square (X2) values are presented to indicate when the frequency of a variable (such as being male or being trained abroad) differs significantly across kinds of wrongdoing.
Table 2:

Wrongdoer and Setting Variables: Comparison of Frequencies

Independent VariableResearch MisconductOther ResearchMedical PracticeX2
Setting variables
 Academic medical setting90%a75%a3%b71.03∗∗∗
 Government funding70%a58%a0%b45.63∗∗∗
 Private funding33%a, b63%a0%b36.12∗∗∗
 Had an accomplice10%a40%b35%b10.18∗∗
 Others found guilty20%33%35%2.50
Wrongdoer Variables
 Male75%a90%b93%b5.89
 Born outside United States10%20%20%1.92
 Trained outside United States15%18%25%1.40
 Plea of insanity0%0%5%4.07
 Found unfit to stand trial0%0%0%
 Evidence of addiction0%a0%a10%b8.28
 Significant personal problems5%a0%a23%b13.41∗∗∗
 Claimed following orders/policy3%10%0%5.43
History of Wrongdoing
 Wrongdoing was repeated68%a80%a, b95%b9.79∗∗
 Different kinds of wrongdoing43%a65%a, b75%b9.30∗∗
 Wrongdoing in multiple institutions18%20%28%1.28
 Felony arrests in personal life3%a0%a13%b7.37

∗ = p < .05, ∗∗ = p < .01, ∗∗∗ = p < .001. N = 120 cases: 40 Research Misconduct, 40 Other Research, 40 Medical Practice. When chi-squared is significant, percentages that do not share subscripts are significantly different by standardized residual.

Highly publicized cases of research misconduct are characterized by very few setting traits: They take place in academic medical centers (90%) with government funding (70%). However, this is likely because non-government funded cases need not be reported to oversight agencies and can be handled relatively quietly. Whereas ORI and NSF publish outcomes of their investigations, private industry and universities do not. Most individuals who engage in research misconduct are male (75%), but the rate is not higher than the prevalence of males receiving NIH research grants (82%) at the median time of the cases (Office of Statistical Analysis and Reporting, 2012). The individuals’ history of wrongdoing is not as pronounced in research misconduct cases as in other kinds of cases: Most individuals repeated their misconduct prior to being investigated formally (68%); but fewer than half engaged in different kinds of wrongdoing (43%) or committed violations at multiple institutions (18%). Wrongdoer and Setting Variables: Comparison of Frequencies ∗ = p < .05, ∗∗ = p < .01, ∗∗∗ = p < .001. N = 120 cases: 40 Research Misconduct, 40 Other Research, 40 Medical Practice. When chi-squared is significant, percentages that do not share subscripts are significantly different by standardized residual. Five percent or fewer attributed their research misconduct to significant personal problems, addictions, or mental disorders, and none had been convicted of felonies in their personal lives; most of these factors are significantly more likely to be found in cases of medical practice violations. Individuals who are born or trained outside of the United States are not represented in our sample of highly publicized cases of research misconduct with any greater frequency than in other kinds of wrongdoing, and they are represented in proportion to the percentage working in the field around the median date in our study (Davis, 2003). Research misconduct cases involve highly focused kinds of wrongdoing; to the extent they involve another wrongdoing, it is usually an extension of the primary wrongdoing—e.g., a violation of publication ethics.

Frequency of Reporting, Investigation, and Consequences Variables

Table 3 presents frequencies and Chi square results on the reporting (whistleblowing), investigation, and consequences of cases. In 28% of cases there was a failed attempt at reporting research misconduct (that is, the wrongdoing continued for some time following an initial report); however, this did not occur at a greater rate than in other kinds of wrongdoing.
Table 3:

Reporting, Investigation, and Consequences Variables: Comparison of Frequencies

Independent VariablesResearch MisconductOther ResearchMedical PracticeX2
 Failed reporting attempts28%35%43%1.98
Whistleblower Description+
 Patient/participant/family0%a18%a, b45%b38.36∗∗∗
 Subordinate23%a10%a, b5%b
 Peer: institutional15%a3%b3%b
 Peer: external5%8%8%
 Oversight personnel: institutional15%a8%a, b3%b
 Oversight personnel: external5%13%10%
 Others (e.g., reporter)18%25%15%
 Unknown20%18%13%
Investigation
 Institutional investigation88%a53%a, b25%b31.72∗∗∗
 Board or professional body investigation20%a23%a85%b44.40∗∗∗
 Regulatory oversight body investigation85%a75%a53%b10.73∗∗
 Criminal investigation33%a20%a73%b24.75∗∗∗
 Civil investigation10%a43%a, b60%b21.97∗∗∗
Consequences to Wrongdoer
 Loss of job, opportunities, funding …90%a50%b85%a20.27∗∗∗
 Loss of licensure/credentialing20%a15%a85%b50.83∗∗∗
 Financial penalties33%a30%a75%b20.61∗∗∗
 Prison, probation, house arrest25%a, b8%a65%b31.68∗∗∗
 Treatment, rehabilitation, education18%13%30%4.06
 Wrongdoer still working in the field48%a, b25%a78%b22.20∗∗∗
Consequences to Institution
 Publication of oversight failure33%a70%b28%a17.58∗∗∗
 Financial penalties, fines, settlements18%30%18%2.45
 Impact on mission/increased audits …13%a48%b13%a17.82∗∗∗

= p < .05, ∗∗ = p < .01, ∗∗∗ = p < .001. N = 120 cases: 40 Research Misconduct, 40 Other Research, 40 Medical Practice. When chi-squared is significant, percentages that do not share subscripts are significantly different by standardized residual.

Whistleblower role was run as 1 variable with 8 roles, thus one X2 value is reported for all roles.

Cases of research misconduct are far more likely to have an institutional whistleblower than other research wrongdoing or wrongdoing in medical practice, with subordinates comprising the largest group of whistleblowers (23%). Not surprisingly, most cases involved an institutional (88%) and federal investigation (85%)—rates significantly higher than found in cases involving other kinds of wrongdoing. Consequences to the wrongdoer were also more severe, typically including loss of job or funding opportunities (90%).

Environmental Variables

Very few of the variables that were hypothesized to predict professional misbehavior were present in cases of research misconduct. Only two variables were rated higher than 1.5 on a scale from 1 to 3: Oversight failures, which in the context of research misconduct can include failures of research integrity officers to investigate a case in a timely manner or of collaborators to investigate suspicious data (m = 1.65) and position of authority (m = 1.90). However, oversight failures play less of a role in cases of research misconduct than other kinds of wrongdoing and the prevalence of individuals in positions of authority was likely due to the fact that principal investigators are ordinarily held accountable for the integrity of data published in papers funded by their grants. Reporting, Investigation, and Consequences Variables: Comparison of Frequencies = p < .05, ∗∗ = p < .01, ∗∗∗ = p < .001. N = 120 cases: 40 Research Misconduct, 40 Other Research, 40 Medical Practice. When chi-squared is significant, percentages that do not share subscripts are significantly different by standardized residual. Whistleblower role was run as 1 variable with 8 roles, thus one X2 value is reported for all roles. Two variables were nearly absent across all kinds of misbehavior: wrongdoers who were reacting to organizational injustice, and wrongdoers who were penalized for doing what is right. All other variables were nearly absent in cases of research misconduct, but more strongly present in other areas of wrongdoing, including especially ambiguous norms, vulnerable victims, conflicting roles, and financial rewards. We may speak of cases that are more “serious” insofar as they last longer and involve multiple kinds of wrongdoing. Several variables correlated with more severe cases of misconduct. Financial rewards to the wrongdoer (rho = .55, p < .01) and to others involved in the case (rho = .57, p < .01), as well as oversight failures (rho = .32, p < .05), were correlated with an increased number of violations. The duration of a case was correlated with the seniority (authority level) of the investigator (rho = .43, p < .01) with cases involving more senior investigators lasting longer. These are all moderately strong correlations. Environmental Variables: Comparison of Mean Scores (ANOVA) = p < .05, ∗∗ = p < .01, ∗∗∗ = p < .001. N = 120 cases: 40 Research Misconduct, 40 Other Research, 40 Medical Practice. When ANOVA is significant, means that do not share subscripts are significantly different by Tukey post hoc pairwise comparison. Scores are means with standard deviations in parentheses. MANOVA including these 10 variables was statistically significant, Wilks’ Lambda = .309, p < .001. Across all categories of wrongdoing, cases last a surprisingly long time—between 3.8 and 4.2 years.

Qualitative Theory

In contrast to cases of other kinds of wrongdoing, very little seems distinctive about the research misconduct cases. A wide variety of environmental and individual factors that are frequently hypothesized to correlate with professional misbehavior were nearly absent from our dataset of highly publicized cases; the remaining variables were only moderately present. So how is misconduct to be explained? Drawing upon their extensive reading on the cases, RAs were asked to respond to three qualitative questions that would allow us to build a theory of each case. They typically wrote 3–5 sentences in response to each question. The following are the questions with the instructions provided to RAs: Describe Notable Social Dynamics. Insofar as they are observed, note at least the following things: highly competitive environment; high profile work; high pressure, high quotas; prominent non-financial rewards; modeling of bad behavior by peers or supervisors; generally poor moral climate. Describe the Wrongdoer's Intentions and Character. Describe the wrongdoer's intentions as you understand them, and his/her response to the wrongdoing. Address whether she/he expressed remorse or apologized. Did he/she lie or try to cover up wrongdoing? Blame others? You may also address personality factors mentioned by in the literature. What is Your Theory of the Case? Provide your common sense theory of how the case occurred, using the theoretical framework embraced by this project (the interaction of individual and environmental factors in providing motive, means or opportunity). You may want to note when desirable information is missing, e.g., “no motive is apparent.” We used an open-ended format for these items in part because we believed that it would be difficult to quantify accurately variables such as the degree of competitiveness in an environment or the investigator's intentions given our data sources, which do not consistently address such matters. Accordingly, the following should be read as a tentative, preliminary exploration of the data that were available to us through our historiometric study of cases. Table 5 presents the principal investigator's (DuBois) coding of these responses—treating responses to all three questions as generating one overarching “theory of the case.”
Table 5:

Qualitative Themes from the Research Misconduct Cases (n = 40)∗

1. Self-centered personality or thinking  a. Bright, arrogant, narcissistic  b. Thought they would get away with it—it seemed easy  c. Lust for scientific success—highly ambitious  d. Greed—increase wealth (rather than just keep job with next grant)48%
2. Unwarranted certainty—making data fit hypotheses that they thought were correct, making images more clearly represent ‘the truth’13%
3. Pressure—intense desire for a quick publication or next grant33%
4. Careless—either with data, images, or supervision25%
5. Unclear—uncertain about what they were thinking or what contributed to behavior20%

Percentages add up to more than 100% because a second code was used in some cases

Qualitative Themes from the Research Misconduct Cases (n = 40)∗ Percentages add up to more than 100% because a second code was used in some cases

DISCUSSION

We examined 40 high profile cases of research misconduct and compared them to 40 cases of other violations of research ethics (heavily representing human subjects violations) and 40 cases of wrongdoing in medical practice. To summarize our findings: Cases of research misconduct generally involved repeat offenses (68%) by an individual who was acting alone (90%) across an average of 3.8 years. In no cases did the individual plead insanity or blame the behavior on an addiction; only one case involved the claim to be following orders. Twenty-eight percent of cases involved failed initial attempts to report misconduct—that is, suspicions were reported, but either not investigated or no finding was made initially and the behavior continued. In contrast to other kinds of wrongdoing, institutional colleagues comprise the largest group of whistleblowers in misconduct cases. Only two environmental variables scored higher than 1.5 on a scale from 1 to 3: Oversight failures and being in a position of authority. Both of these factors—as well as financial rewards—were correlated with more serious cases of misconduct—but none of these factors was more strongly present in misconduct cases than other research wrongdoing cases. When compared to other kinds of wrongdoing in research or to violations of medical ethics, research misconduct is associated with very few variables. This is somewhat surprising given the number of variables that this project tracked, and the fact that we explored some novel environmental factors—including “opportunity” factors—that might explain how research misconduct arises. In what follows we discuss the implications of our findings for responsible conduct of research (RCR) education, for research integrity officer (RIO) behavior, and for further research.

RCR Education

By studying cases of research misconduct, we have learned important lessons for RCR education, lessons that would justify recent shifts from philosophically to psychologically driven programs. The norms involved in misconduct cases were never ambiguous; this contrasted significantly with other kinds of wrongdoing in research. Accordingly, it is unlikely that traditional ethics education—with its focus on promulgating specific rules as well as identifying general principles and strategies for deducing specific rules—will help prevent or remediate misconduct. Individual factors appear to play a significant role in research misconduct. The “history of wrongdoing” results did not suggest that misconduct is the product of antisocial personalities (e.g., only 3% had felony arrests). However, qualitative data suggest that narcissistic thinking plays a role: We repeatedly observed senior investigators who thought selfishly, thought they could get away with something, or thought they were justified in making data fit their hypotheses. We also found wrongdoers reporting pressure to publish and obtain grant funding; in some cases, however, this also tied into narcissistic thinking insofar as they did not aim merely to keep their jobs with publications and grants, but to become superstars in their fields. Finally, we found some cases that involved carelessness—either with data or oversight of personnel. Given these factors, we recommend that education focus more on sense-making strategies and mental models or bias reduction. Sense-making strategies teach people to question their own judgments, examine their motives, think of others, forecast consequences, and seek help (Mumford et al., 2008). Bias reduction education often requires group work aimed at identifying the distortions in selfish thinking (Gibbs, 2009). While we offer these recommendations for education addressing research misconduct, they seem sensible regarding all areas of the responsible conduct of research. However, more traditional ethics education may have a greater role to play in areas with ambiguous norms, for example, the negotiation of order of authorship or best practices for enrolling participants who lack decisional capacity.

RIO Behavior

Given that most whistleblowers are institutional colleagues or subordinates, RIOs should make themselves known and accessible to investigators. The fact that 28% of cases involved failed attempts at reporting suggests that individuals should be encouraged to report to those best qualified and most interested in following up on reports of suspicious data (that is, RIOs), rather than those most convenient to them within a research division. Additionally, our data suggest that cases of research misconduct are fairly heterogeneous. Given that most RIO's will investigate only a few cases of misconduct across a 5-year period, we recommend caution in generalizing about the causes and correlates of research misconduct.

Further Research

In some domains of wrongdoing, for example, wrongdoing in medical practice, we have identified strong environmental correlates to wrongdoing. Under such circumstances, it makes sense to complete a larger set of cases (e.g., 100 or more) to enable regression analyses and causal modeling. However, the failure to identify multiple moderate correlates to research misconduct suggests that further research with larger samples and regression modeling may yield little new information. We believe larger sample historiometric research is indicated to understand cases of human subjects protection failures and other kinds of wrongdoing in research. However, research on research misconduct might more fruitfully focus on individual variables such as cognitive biases, forecasting skills, and personality traits.

Policy and Prevention

We supported Adams and Pimple's (2006) suggestion to focus on environmental factors that provide opportunity for crime. However, our efforts to identify such correlates have proved far more fruitful in domains other than research misconduct. In part, this may be due to the fact that most investigators perceive themselves to have the opportunity to engage in research misconduct: They often enjoy high levels of autonomy, have access to databases, and decide on the final dataset that biostaticians and others see. But when such factors are ubiquitous they are also unlikely to emerge as statistical predictors. In this sense, some of the most prominent environmental factors that provide opportunity may be analogous to oxygen: You cannot have a fire without oxygen, but the presence of oxygen alone does not predict fire. From a policy and prevention perspective, it may be that the key to prevention has more to do with increasing the likelihood of detection (that is, reducing the ubiquity of perceived opportunity) than with adjusting other features of the environment (such as financial rewards, social norms, or conflicting roles).
Table 4:

Environmental Variables: Comparison of Mean Scores (ANOVA)

Environmental VariablesResearch Misconduct +Other ResearchMedical PracticeF
Conflicting roles++1.15a (.53)1.75b (.95)1.05a (.32)13.28∗∗∗
Financial reward++1.33a (.69)2.00b (.93)2.23b (.95)11.70∗∗∗
Others benefit++1.30a (.61)1.98b (.80)1.83b (.90)8.26∗∗∗
Penalized for doing right1.00 (.00)1.00 (.00)1.02 (.16)n/a
Others penalized++1.30 (.65)1.18 (.55)1.20 (.46).56
Mistreatment of wrongdoer1.08 (.27)1.02 (.16)1.00 (.00)n/a
Ambiguous norms++1.02a (.15)1.43b (.55)1.05a (.22)16.04∗∗∗
Vulnerable victims++1.07a (.35)1.70b (.79)1.68b (.69)12.22∗∗∗
Oversight failures++1.65a (.77)2.13b (.79)1.78ab (.77)4.03
Position of authority++1.90a (.74)2.45b (.68)1.55a (.81)14.73∗∗∗
Collaboration1.20 (.46)1.33 (.47)1.00 (.00)n/a
Other Variables
History of wrongdoing++1.30a (.88)1.65ab (.86)2.10b (1.03)7.45∗∗∗
Duration3.83 (1.28)4.30 (1.47)4.22 (1.48)1.31
Number of violations++2.03a (1.48)3.63b (1.64)3.13b(1.49)11.33∗∗∗
Number of Sources19.30a (17.52)32.23b (23.98)34.38b (31.19)4.30

= p < .05, ∗∗ = p < .01, ∗∗∗ = p < .001. N = 120 cases: 40 Research Misconduct, 40 Other Research, 40 Medical Practice. When ANOVA is significant, means that do not share subscripts are significantly different by Tukey post hoc pairwise comparison.

Scores are means with standard deviations in parentheses.

MANOVA including these 10 variables was statistically significant, Wilks’ Lambda = .309, p < .001.

  25 in total

1.  Shutdown of research at Duke sends a message.

Authors:  E Marshall
Journal:  Science       Date:  1999-05-21       Impact factor: 47.728

2.  A social science perspective on gifts to physicians from industry.

Authors:  Jason Dana; George Loewenstein
Journal:  JAMA       Date:  2003-07-09       Impact factor: 56.272

3.  The role of culture in research misconduct.

Authors:  Mark S Davis
Journal:  Account Res       Date:  2003 Jul-Sep       Impact factor: 2.622

4.  Moral disengagement in the perpetration of inhumanities.

Authors:  A Bandura
Journal:  Pers Soc Psychol Rev       Date:  1999

5.  Scientists behaving badly.

Authors:  Brian C Martinson; Melissa S Anderson; Raymond de Vries
Journal:  Nature       Date:  2005-06-09       Impact factor: 49.962

6.  Scientists' perceptions of organizational justice and self-reported misbehaviors.

Authors:  Brian C Martinson; Melissa S Anderson; A Lauren Crain; Raymond de Vries
Journal:  J Empir Res Hum Res Ethics       Date:  2006-03       Impact factor: 1.742

Review 7.  Clinical trials: bridging the gap between efficacy and effectiveness.

Authors:  Colin Depp; Barry D Lebowitz
Journal:  Int Rev Psychiatry       Date:  2007-10

Review 8.  Qualitative and quantitative analyses of historical data.

Authors:  Dean Keith Simonton
Journal:  Annu Rev Psychol       Date:  2002-06-10       Impact factor: 24.137

9.  How do institutional review boards apply the federal risk and benefit standards for pediatric research?

Authors:  Seema Shah; Amy Whittle; Benjamin Wilfond; Gary Gensler; David Wendler
Journal:  JAMA       Date:  2004-01-28       Impact factor: 56.272

10.  Understanding the Severity of Wrongdoing in Health Care Delivery and Research: Lessons Learned From a Historiometric Study of 100 Cases.

Authors:  James M DuBois; Emily E Anderson; John T Chibnall
Journal:  AJOB Prim Res       Date:  2013-07-22
View more
  15 in total

1.  Perceptions of Work-Related Stress and Ethical Misconduct Amongst Non-tenured Researchers in Italy.

Authors:  Oronzo Parlangeli; Stefano Guidi; Enrica Marchigiani; Margherita Bracci; Paul M Liston
Journal:  Sci Eng Ethics       Date:  2019-02-04       Impact factor: 3.525

2.  In Their Own Words: Research Misconduct from the Perspective of Researchers in Malaysian Universities.

Authors:  Angelina P Olesen; Latifah Amin; Zurina Mahadi
Journal:  Sci Eng Ethics       Date:  2017-12-16       Impact factor: 3.525

3.  Five Dimensions of Research Ethics: A Stakeholder Framework for Creating a Climate of Research Integrity.

Authors:  James M DuBois; Alison L Antes
Journal:  Acad Med       Date:  2018-04       Impact factor: 6.893

4.  How to Conduct Responsible Research: A Guide for Graduate Students.

Authors:  Alison L Antes; Leonard B Maggi
Journal:  Curr Protoc       Date:  2021-03

5.  Compliance Disengagement in Research: Development and Validation of a New Measure.

Authors:  James M DuBois; John T Chibnall; John Gibbs
Journal:  Sci Eng Ethics       Date:  2015-07-15       Impact factor: 3.525

Review 6.  The visibility of scientific misconduct: A review of the literature on retracted journal articles.

Authors:  Felicitas Hesselmann; Verena Graf; Marion Schmidt; Martin Reinhart
Journal:  Curr Sociol       Date:  2016-10-13

7.  The Professionalism and Integrity in Research Program: Description and Preliminary Outcomes.

Authors:  James M DuBois; John T Chibnall; Raymond Tait; Jillon S Vander Wal
Journal:  Acad Med       Date:  2018-04       Impact factor: 6.893

8.  Sexual Violation of Patients by Physicians: A Mixed-Methods, Exploratory Analysis of 101 Cases.

Authors:  James M DuBois; Heidi A Walsh; John T Chibnall; Emily E Anderson; Michelle R Eggers; Mobolaji Fowose; Hannah Ziobrowski
Journal:  Sex Abuse       Date:  2017-06-19

9.  Exploring unnecessary invasive procedures in the United States: a retrospective mixed-methods analysis of cases from 2008-2016.

Authors:  James M DuBois; John T Chibnall; Emily E Anderson; Heidi A Walsh; Michelle Eggers; Kari Baldwin; Kelly K Dineen
Journal:  Patient Saf Surg       Date:  2017-12-18

10.  Quality of randomized controlled trials of new generation antidepressants and antipsychotics identified in the China National Knowledge Infrastructure (CNKI): a literature and telephone interview study.

Authors:  Zheng Tong; Fangzhou Li; Yusuke Ogawa; Norio Watanabe; Toshi A Furukawa
Journal:  BMC Med Res Methodol       Date:  2018-09-24       Impact factor: 4.615

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.