The pool of talented, well-trained scientists is growing and increasingly sophisticated technology is becoming available, thus creating the potential to achieve truly significant advances in knowledge. Yet, paradoxically, the funds available to support such efforts are inadequate. The resulting heightened competition for these limited dollars has created an environment that is highly conducive to research misconduct. Therefore, there could be no better time for the academic research community to address this issue.
A Persistent Pernicious Problem
A 2010 New York Times headline read, “Expert on morality is on leave after research inquiry.”1 The story described alleged research misconduct by a distinguished investigator at an elite university. Ironically, his area of study? The origins of morality.
Research misconduct is indeed very much with us but is not just a contemporary concern. In 1830, the mathematician Charles Babbage wrote Reflections on the Decline of Science in England, and on Some of Its Causes, in which he noted the problems created by the “frauds of observers.”2 More recent surveys of scientists and trainees in the United States and Europe have found significant numbers of individuals who acknowledged being personally guilty of misconduct or who were aware of its being committed by others.3–11 In 2008, Titus and colleagues12 estimated that as many as 1,000 instances of research misconduct go unreported annually in the United States.
Nonetheless, science is said to be self-correcting, which in principle is true. However, fraudulent results can persist uncorrected in the literature for many years. For example, it took more than 100 years to establish that Haeckel’s illustrations of embryonic development, the basis of his thesis that ontogeny recapitulates phylogeny, had been altered to fit his hypothesis.2
Publications based on fraudulent data may ultimately be detected and retractions published in the journals in which they had appeared. However, unfortunately, flawed research continues to be cited long after those retractions appear.13–18 Therefore, the damage that is inflicted on others in wasted time, effort, funds, and ineffective or dangerous clinical care is incalculable.
In 1991, the federal government established the Office of Research Integrity (ORI), with the responsibility to investigate charges of research misconduct, defined as fabrication and/or falsification of data and plagiarism.19 (In the 19th century, research misconduct was more colorfully defined as “hoaxing, forging, trimming, and cooking.”)2
Under ORI regulations, federally funded research training programs are required to provide instruction in the responsible conduct of research (RCR), with a component devoted specifically to research misconduct.20 However, surveys of trainees and early- and midcourse scientists who have taken such courses have found that although these programs increased the participants’ knowledge of ethical issues and their ability to identify them, they were unlikely to affect their future behavior.3,21–24 As simply stated by a participant in one of these studies: “In most cases, I believe a person is ethical or not. I feel that if a person chooses to do something unethical it’s not because they never took an ethics class.”21
An Institute of Medicine report similarly acknowledged the limitations of such training, stating, “The scientific community must appeal to the consciences of individual scientists and the scientific community as a whole to invoke the highest possible standards of research behavior.”25
Is that not obvious? Fabrication, falsification, and plagiarism are the academic equivalents of lying, cheating, and stealing. The ethical standards proscribing such behavior are established long before one enters graduate training in science. Therefore, we should not be surprised to find that RCR courses do not influence the behavior of these young adult trainees.
What, then, can be done? I reviewed the 146 individual narratives contained in ORI reports of those found guilty of misconduct (1992–2003). The number of findings of research misconduct reported in 2009 was no different from the average from 1992–2009.26 It was my hope that, as a psychiatrist, I could identify patterns of behavior that could serve as the basis for recommendations to mitigate the problem.
Approximately one-third of the respondents (the accused) were support staff, one-third were postdoctoral fellows and graduate students, and one-third were faculty. Accusations of fabrication represented 45% of the offenses, falsification 66%, and plagiarism 12%. The first two offenses frequently occurred together. Approximately three-quarters of the respondents admitted their guilt or did not provide a defense. None claimed that the offense of which they were accused should not be considered research misconduct. They frequently attributed their behavior to extenuating circumstances.
One hundred twenty-four articles based on faulty data were published in 72 journals. These included Cell (5 articles), Diabetes (5), Hypertension (5), Proceedings of the National Academy of Sciences (5), The Journal of Biological Chemistry (4), Nature (4), Science (4), and the New England Journal of Medicine (3). Consider the effect that any one of these articles, published in these high-impact journals, might have had on its respective field.
Beyond these statistics, though, are 146 individuals, each with his or her own story to tell. These vignettes provided the data for this study. The patterns of personal characteristics and circumstances, which I identified and the subsequent recommendations are based on 50 years of clinical experience in psychiatry and 19 years as the chairman of two institutional review boards.
Who Were the Respondents?
These individuals were usually not members of the scientific community. They were not bound by a standard of professional ethics and usually did not have any personal investment in the validity of the study’s outcome. They were also less likely to understand the significance of their misconduct on the research objectives and the subsequent impact of their actions. They were frequently pressured to increase the intake of new subjects or to generate more data. In some cases, their income was directly based on their productivity. Consider the following examples:
Three individuals were hired to conduct neighborhood interviews. The subjects were to receive cash payments for participating in the follow-up interviews. The signatures on the receipts for the cash payments differed from the consent form signatures obtained at the time of enrollment. The three individuals acknowledged that they had forged the signatures, taken the payments, and fabricated the interview data.
A technician admitted that the times of day he recorded for blood samples were not the actual times that the samples were collected. He said that he could not follow protocol schedules and also provide as many samples as were required. The ORI investigating committee concluded that he had been assigned responsibility for more protocols than he could reasonably have been expected to perform. The technician also stated that he was not made aware of the significance of the timing of the blood sampling to the research objectives.
Postdoctoral fellows and graduate students
These individuals were under great pressure to produce publishable results. The postdoctoral fellows were competing for faculty positions, and the graduate students for postdoctoral fellowships. They also may have been working under limited supervision. Consider the following examples:
One respondent acknowledged that he had falsified data “to make it fit the hypothesis.” He had recently been notified that he was to be terminated and believed that he needed additional publishable research to get another appointment.
Another respondent acknowledged that she had fabricated data in an article which had been accepted for publication. She stated that she had been under pressure from a superior to generate data and felt that her action was justifiable because she had observed a senior scientist in her laboratory “clean up” data to make them more acceptable for publication.
A third respondent acknowledged that he had manipulated data to yield a desired result. He was described by one superior as “one of the brightest and hardworking students anyone has ever seen.” The student had previously published articles in Cell and Nature based on legitimate findings. He reported “intense self-imposed pressure” to replicate his early successes and stated: “My behavior has been a heavy burden on me. I can offer no excuse. I cannot begin to express my regrets for my actions.”
The need to publish exists at all stages of an academic career. To qualify for tenure, assistant professors must publish and demonstrate that they have the potential to obtain research funding. Associate professors must publish to support their requests for research funding and to justify promotion to professor. Full professors must publish to garner support for their laboratories and to demonstrate that they remain leaders in their fields. Consider the following examples:
Assistant professors. One respondent was the senior author of an article published in Science. A senior colleague questioned the validity of his findings. The respondent could not produce the data on which the article was based. He initially claimed that the data had been lost when they were entered into the computer and that bound data books had not been used. He subsequently acknowledged that the data were fabricated.
Another respondent was asked to provide the data to support the figures that were included in her publications. She stated that she could not provide the data because a mouse had destroyed the tapes.
Associate professors. One respondent was accused of fabricating data in a grant application. He defended his action as “data projection” and not “data falsification.” He acknowledged that he included the “projected data” to increase his chances of securing grant funding.
Another respondent was accused of failing to follow the protocol as stated in the method section of a published article. He acknowledged that certain measurements were based on his personal observations and not on measurements by caliper or ruler as required in the protocol. The review committee did not believe that the respondent was motivated by a desire to enhance results but, rather, by the conviction that his own subjective judgments were as valid as the specific requirements of the protocol’s method.
Full professors. One respondent, the chairman of a medical school basic science department, was found guilty of submitting a grant application that contained the plagiarized text of 11 paragraphs and one-third of the references from a National Institutes of Health (NIH) grant application that he had reviewed while serving as a member of an NIH study section. In his own defense, the respondent submitted a handwritten grant application that he alleged antedated the grant he was accused of plagiarizing. That document, however, was found to have been fabricated.
I also found three similar instances of senior faculty who plagiarized material from grants that they had reviewed as consultants or as members of NIH study sections.
Another respondent was found guilty of encouraging a subordinate who had completed one experiment to inflate the size of the sample by three to provide the number of tests required by the protocol. He explained, “If we had done it three times, we would have gotten the same result anyway.” He referred to the criticism of his artificially inflating the sample size as “nitpicking.”
Clinical investigators. These individuals were primarily involved in the enrollment of subjects according to a study’s guidelines. They may have held clinical faculty appointments at university-affiliated hospitals but usually had not participated in the design of the protocol.
One respondent was found guilty of falsifying and/or fabricating data in the histories and screening exams of patients recruited for a cancer treatment study. He acknowledged having done so. In his own defense, he stated that he believed that the patients would benefit from participating in the protocol and should not be denied treatment on the basis of “trivial details.”
Why Did the Respondents Violate the Rules?
These acts of research misconduct seemed to be the result of the interaction of psychological traits and/or states and the circumstances in which these individuals found themselves. The respondents could be categorized as follows:
* The desperate, whose fear of failure overcame a personal code of conduct,
* The perfectionist, for whom any failure was a catastrophe,
* The ethically challenged, who succumbed to temptation,
* The grandiose, who believed that his or her superior judgment did not require verification,
* The sociopath, who was totally absent a conscience (and, fortunately, was rare),
* The nonprofessional support staff, who were unconstrained by the ethics of science, unaware of the scientific consequences of their actions, and/or tempted by financial rewards.
In a similar study of comparable ORI reports, researchers scanned the texts and performed multidimensional scaling and cluster analysis. They also found that personal characteristics and the circumstances in which the respondents found themselves were major contributing factors to research misconduct.27
We should also consider the role of our current academic culture. In his book, The Cheating Culture: Why More Americans Are Doing Wrong to Get Ahead, Callahan28 noted the high prevalence of cheating found among U.S. high school and college students. Perhaps that change in societal norms has made “cheating” later in a scientific career more acceptable.
What Remedies Are Available?
My observations are not surprising. However, the scientific community has not addressed them realistically. RCR instruction cannot be expected to establish basic ethical standards in a classroom of young adult graduate students. However, variations on such a course might be effective for the nonprofessional staff, for whom such training is not now required. Members of this group might be less likely to fabricate or falsify data if they have a better understanding of the goals of the research in which they are involved. They should know how their findings could contribute to advances in science and/or improved medical care and the serious consequences of publishing fraudulent data.
All potential respondents might be deterred from misconduct by the knowledge that the submission of false information in a government agency grant application is a violation of federal law, punishable by imprisonment.29
However, establishing remedies for the psychological characteristics and the life circumstances of potential respondents poses a much more difficult problem. Grandiosity, perfectionism, and sociopathy cannot be eradicated from the scientific community, or any other, and little can be done to reduce the reality of the need to publish or perish.
Nonetheless, I believe that there are institutional policy initiatives that can have an impact on the prevalence of research misconduct:
1. Improvement in the quality of mentoring in training programs, and
2. A policy that acknowledges the important contributions of whistleblowers and establishes truly effective means of protecting them from retaliation.
Improving the quality of mentoring
Another 2010 New York Times headline read, “Three Harvard researchers retract a claim on the aging of stem cells.”30 The article contained the following comment: “It is not clear whether the competitive nature of science puts pressure on [postdoctoral] students to cut corners, or whether the laboratory chief creates an atmosphere that induces cheating.” Unfortunately, either, or both, could be true.
Twenty-five years ago, in an article entitled “The Pathogenesis of Fraud in Medical Science” Petersdorf31 spoke of the impact of an era of “big science,” in which trainees were inadequately supervised. Unfortunately, “big science” has become even bigger, yet our mentoring system is essentially unchanged.
A study of mentoring in a series of closed ORI misconduct cases found that 62% of mentors had not established procedural standards (e.g., the need to maintain a laboratory notebook), and 73% had not reviewed the raw data generated by their trainees.32 Another survey of 98 medical schools found that only 23.5% had published guidelines for mentorship.33 Appropriate mentoring, which establishes high ethical standards for trainees and provides meaningful personal support and supervision, can decrease the likelihood of research trainees engaging in what one study referred to as problematic behavior.34
Effective mentors must be prepared to play many roles. First, they must address their trainees’ intense fear of failure. A study of ORI reports found that over 50% of graduate students and postdoctoral fellows reported stress as a contributing factor to their acts of research misconduct and 62% to a self-imposed need to perform well.33 Therefore, mentors must redefine the concept of failure for this group of highly competitive individuals, for many of whom anything less than perfection is a catastrophe. For example, mentors could share their own experiences with failed experiments that obviously did not destroy their academic careers. Trainees can benefit for a lifetime if they can learn to put such “failures” in a realistic perspective. Institutions should also make professional counseling services available because supportive mentoring alone may be inadequate to achieve this goal.27
A close mentoring relationship also would provide a model of scientific integrity for trainees to follow. One senior scientist put it well: “Although my graduate students have taken university training re: scientific integrity, I think they learn more from my personal insistence that they do proper controls and ‘tell it like it is.’”35
A close mentoring relationship also would provide an opportunity for mentors to maintain a watchful eye on trainees’ conduct of their research. This would require that a realistic ratio of trainees per mentor be established. Mentors would also pay closer attention to the work of their trainees if they were made to share responsibility for any publications based on a trainee’s research.
Finally, universities would pay greater attention to the quality of mentorship if that were made a significant factor in evaluating training grant applications. For example, the ratio of trainees to mentor should be made a measure of the quality of a program.
I recommend two publications for mentors and mentors-to-be: Advisor, Teacher, Role Model, Friend: On Being a Mentor to Students in Science and Engineering, a monograph published by the National Academy of Sciences36; and a report from the University of California, San Francisco, on their comprehensive mentor development program.37
Acknowledging and protecting whistleblowers
Because the total prevention of research misconduct is impossible, the scientific community must depend on whistleblowers to minimize the presence and/or persistence of flawed data in the scientific literature.
Research studies have found that the fear of retaliation causes many students and faculty in the United States and abroad to refrain from reporting cases of suspected research misconduct.2–7,38,39 Understandably so, because an ORI study found that more than two-thirds of whistleblowers did indeed suffer adverse consequences as a result of their actions,40 despite current Public Health Service regulations prohibiting retaliation.41
Therefore, institutional leaders must acknowledge the whistleblowers’ courage and their gratitude for the significant contributions that whistleblowers have made to the scientific community and to their institutions’ integrity. It is also essential that institutional leaders deal with the whistleblowers’ realistic fears of retaliation. Whistleblowers should be reassured that their institutions are prepared to take swift and effective action to protect them. Increased reporting by potential whistleblowers will not occur until they are acknowledged for their contributions and convinced that they will receive truly adequate protection from retaliation.
It would be helpful if institutions also provided an experienced administrator to initially receive a whistleblower’s allegations. This individual can serve as an ombudsman with the ability and authority to resolve some problems before a formal inquiry is made.42,43 The ability to resolve disputes informally is essential because individuals who are ultimately exonerated of a research misconduct charge, nonetheless can suffer detrimental consequences.44 For example, two months after the 2010 New York Times front page report cited earlier1 named a scientist alleged to have committed research misconduct, the newspaper published a follow-up report on page D3 with the following headline: “Difficulties in defining errors in case against Harvard researcher.”45 Should this scientist ultimately be exonerated, it would be impossible to fully repair the damage done to his reputation. Therefore, research institutions bear an obligation to establish a mechanism that protects the interests of all parties involved in accusations of research misconduct.
It might appear that the most effective protection for whistleblowers would be for them to remain anonymous.46 ORI does accept anonymous allegations that are then referred for investigation to the institutions responsible for the research. However, such investigations are often hampered by the unavailability of the anonymous whistleblower to provide additional information. Some institutions have instituted compliance hotlines to receive anonymous complaints, but there are no data available on their effectiveness.
Research misconduct is the product of a combination of individual character traits, an intense fear of failure, or the lure of academic and/or financial rewards. Character traits do not lend themselves readily to remediation, and the anxiety induced by the reality of publish or perish cannot be abolished. Therefore, to continue to expect any significant impact from RCR courses for trainees is wishful thinking. However, such courses developed specifically for support staff might prove to be effective.
I believe that there are other means to more effectively address the problem. I recommend that institutional policies be instituted to improve the quality of mentorship—e.g., establishing realistic ratios of trainees to mentors. The quality of mentoring should be made a factor in the evaluation of applications for the funding of training grants and mentors made to share responsibility for the published work of their trainees.
The contributions of whistleblowers should be appropriately acknowledged by their institutions, and meaningful protection from retaliation should be provided. Such policies would encourage more individuals to expose fraudulent research and would serve as a deterrent to others. If adopted, these policy changes will not eradicate research misconduct, which is an impossible task. However, I do believe that they can have a significant impact on its prevalence.
Acknowledgments: The author wishes to thank his colleagues Professors Richard Kessin, Richard Mayeux, David Schachter, Andrew Wit, and Samuel Silverstein for their helpful constructive criticism of earlier drafts of this article.
Other disclosures: None.
Ethical approval: This study was approved by Columbia University Department of Psychiatry institutional review board.
1. Wade N. Expert on morality is on leave after research inquiry. New York Times. 2010:A17
2. Judson HF. The Great Betrayal: Fraud in Science. 2004 New York, NY Harcourt
3. Kalichman MW, Friedman PJ. A pilot study of biomedical trainees’ perceptions concerning research ethics. Acad Med. 1992;67:769–775
4. Steneck NH. Confronting misconduct in science in the 1980s and 1990s: What has and has not been accomplished? Sci Eng Ethics. 1999;5:161–176
5. Swazey J, Anderson M, Lewis K. Ethical problems in academic research. Acad Sci. 1993;81:542–543
6. Eastwood S, Derish P, Leash E, Ordway S. Ethical issues in biomedical research: Perceptions and practices of postdoctoral research fellows responding to a survey. Sci Eng Ethics. 1996;2:89–114
7. Geggie D. A survey of newly appointed consultants’ attitudes toward research fraud. J Med Ethics. 2001;27:344–346
8. Ranstam J, Buyse M, George SL, et al. Fraud in medical research: An international survey of biostatisticians. Control Clin Trials. 2000;21:415–427
9. Martinson BC, Anderson MS, de Vries R. Scientists behaving badly. Nature. 2005;435:737–738
10. Anderson MS, Ronning EA, de Vries R, Martinson BC. The perverse effects of competition on scientists’ work and relationships. Sci Eng Ethics. 2007;13:437–461
11. Gallup Organization. . Observing and Reporting Suspected Misconduct in Biomedical Research. 2008 Washington, DC Office of Research Integrity
12. Titus SL, Wells JA, Rhoades LJ. Repairing research integrity. Nature. 2008;453:980–982
13. Friedman PJ. Correcting the literature following fraudulent publication. JAMA. 1990;263:1416–1419
14. Pfeifer MP, Snodgrass GL. The continued use of retracted, invalid scientific literature. JAMA. 1990;263:1420–1423
15. Budd JM, Sievert M, Schultz TR. Phenomena of retraction: Reasons for retraction and citations to the publications. JAMA. 1998;280:296–297
16. Sox HC, Rennie D. Research misconduct, retraction, and cleansing the medical literature: Lessons from the Poehlman case. Ann Intern Med. 2006;144:609–613
17. Foo JY. A retrospective analysis of the trend of retracted publications in the field of biomedical and life sciences. Sci Eng Ethics. 2011;17:459–468
18. Zimmer C. It’s science, but not necessarily right. New York Times. June 26, 2011:SR12
19. Department of Health and Human Services, Public Health Service. . Policies and procedures for dealing with possible scientific misconduct in extramural research procedures. Fed Regist. July 13, 1991;56:27383–27394
20. National Institutes of Health. . Reminder and update: Requirement for instruction in the responsible conduct of research in the National Research Service Award Institutional Training Grants. NIH Guide. 1992:212–3
21. Plemmons DK, Brody SA, Kalichman MW. Student perceptions of the effectiveness of education in the responsible conduct of research. Sci Eng Ethics. 2006;12:571–582
22. DuBois JM, Dueker JM, Andersen EE, Campbell J. The development and assessment of an NIH-funded research ethics training program. Acad Med. 2008;83:596–603
23. Powell ST, Allison MA, Kalichman MW. Effectiveness of a responsible conduct of research course: A preliminary study. Sci Eng Ethics. 2007;13:249–264
24. Antes AL, Wang X, Mumford MD, Brown RP, Connelly S, Devenport LD. Evaluating the effects that existing instruction on responsible conduct of research has on ethical decision making. Acad Med. 2010;85:519–526
25. Board on Health Sciences Policy. . Institute of Medicine. Integrity in Scientific Research: Creating an Environment That Promotes Responsible Conduct. Executive Summary. 2002 Washington, DC National Academies Press
27. Davis MS, Riske-Morris M, Diaz SR. Causal factors implicated in research misconduct: Evidence from ORI case files. Sci Eng Ethics. 2007;13:395–414
28. Callahan D. The Cheating Culture: Why More Americans Are Doing Wrong to Get Ahead. 2004 New York, NY Harcourt
30. Wade N. 3 Harvard researchers retract a claim on the aging of stem cells. New York Times. October 15, 2010:A22
31. Petersdorf RG. The pathogenesis of fraud in medical science. Ann Intern Med. 1986;104:252–254
32. Wright DE, Titus SL, Cornelison JB. Mentoring and research misconduct: An analysis of research mentoring in closed ORI cases. Sci Eng Ethics. 2008;14:323–336
33. Rhoades LJ. Beyond conflict of interest: The responsible conduct of research. Sci Eng Ethics. 2002;8:459–468
34. Anderson MS, Horn AS, Risbey KR, Ronning EA, De Vries R, Martinson BC. What do mentoring and training in the responsible conduct of research have to do with scientists’ misbehavior? Findings from a national survey of NIH-funded scientists. Acad Med. 2007;82:853–860
36. Committee on Science, Engineering, and Public Policy. . National Academy of Sciences. Advisor, Teacher, Role Model, Friend: On Being a Mentor to Students in Science and Engineering. 1997 Washington, DC National Academies Press
37. Feldman MD, Huang L, Guglielmo BJ, et al. Training the next generation of research mentors: The University of California, San Francisco, clinical and translational science institute mentor development program. Clin Transl Sci. 2009;2:216–221
38. Miceli MP, Near JP. The incidence of wrongdoing, whistle-blowing, and retaliation: Results of a naturally occurring field experiment. Employee Responsibilities Rights J. 1989;2:91–108
39. Nylenna M, Andersen D, Dahlquist G, Sarvas M, Aakvaag A. Handling of scientific dishonesty in the Nordic countries. National Committees on Scientific Dishonesty in the Nordic Countries. Lancet. 1999;354:57–61
40. Research Triangle Institute. . Consequences of whistleblowing for the whistleblower in misconduct in science cases. 1995 Washington, DC Office of Research Integrity
41. Department of Health and Human Services. . Public health service standards for the protection of research misconduct whistleblowers. 42 CFR Part 94. 65 Fed Reg 70830 and 65 Fed Reg 82972
42. Fischbach RL, Gilbert DC. The ombudsman for research practice: A proposal for a new position and an invitation to comment. Sci Eng Ethics. 1995;1:389–402
43. Gunsalus CK. Preventing the need for whistleblowing: Practical advice for university administrators. Sci Eng Ethics. 1998;4:75–94
44. Lubalin JS, Matheson JL. The fallout: What happens to whistleblowers and those accused but exonerated of scientific misconduct? Sci Eng Ethics. 1999;5:229–250
45. Wade N. Difficulties in defining errors in case against Harvard researcher. New York Times. October 26, 2010:D3
46. Price AR. Anonymity and pseudonymity in whistleblowing to the US Office of Research Integrity. Acad Med. 1998;73:467–472