- Feb 2021
Joe. (2021, January 26). Crystal Prison Zone: I tried to report scientific misconduct. How did it go? Crystal Prison Zone. http://crystalprisonzone.blogspot.com/2021/01/i-tried-to-report-scientific-misconduct.html
- Mar 2016
The results presented here suggest that competition among researchers haspronounced effects on the way science is done. It affects the progress of sciencethrough secrecy and sabotage and interferes with peer review and other universal-istic merit-review systems. It twists relationships within a field and can increase thelikelihood of a scientist engaging in misconduct. None of the focus-groupparticipants made reference to positive effects of competition on their work,despite the fact that the focus-group questions dealt in a general way with scientists’work and the norms of conduct that govern that work. If the protocol questions hadasked explicitly about competition, doubtless there would have been somediscussion about the positive aspects of science. In the context of the generalquestions, though, the scientists referred to competition as a constant and negativeforce that interferes with the way science is done. It is disconcerting to ponder theconsequences of competition, such as mistrust and defensive posturing, for acommunity that has long been committed—in principle—to shared ideas andcollegiality.
Conclusions on the negative results of competition
Competition is also manifested in scientists’ pressured haste, leading tocarelessness, which can verge on questionable behavior. One discussant talkedabout scientists ‘‘cutting a little corner’’ in order to get a paper out before others orto get a larger grant, and another said that she once published a result that she gotthree times in one week but could not replicate the following week
How competitiveness also results in error.
But there’s, I think there is a question of how you interpret the data, even ... ifthe experiments are very well designed. And, in terms of advice—not that I’mgoing to say that it’s shocking—but one of my mentors, whom I very muchrespect as a scientist—I think he’s extraordinarily good—advised me to alwaysput the most positive spin you can on your data. And if you try to present, like,present your data objectively, like in a job seminar, you’re guaranteed tonotgetthe job
Importance of "spinning" data
You are. And you know what the problems are in doing the experiments. And ifyou, in your mind, think that there should be one more control—because youknow this stuff better than anybody else because you’re doing it, you know—you decided not to do that, not to bring up what the potential difficulties are, youhave a better chance of getting that paper published. But it’s—I don’t think it’sthe right thing to do.
deliberate positive bias
dishonesty occursmore with the postdoc. Because they want to get the data—whereas if they don’tpublish, they don’t move on. And they, I think, are more likely to sort of fudge alittle bit here and there if they need to get the data done. Unless, like you say, youwatch them.
senior faculty over-estimate the likelihood of juniors committing misconduct.
One mid-career scientist told a story of how he and others in his lab counteractedan abuse of power by his mentor, a senior scientist, while he was in training. Hismentor received a manuscript to review that was authored by a ‘‘quasi-competitor.’’It presented results of experiments similar to those that were going on in thementor’s lab. The scientist continued, ‘‘That paper ... basically would have beat usto the punch. They would have published these results before us, and they wouldhave gotten credit, and not us. And my mentor, God bless him, sat on the paper.’’The mentor not only delayed writing the review but asked someone working in thelab to write it (a move of questionable ethicality in itself). That lab person and ourrespondent decided, in response, to stall their own work, so that their lab would nothave an unfair advantage over the group who submitted the paper for review. In theend, the original group got credit for the findings, while the respondent’s lab wasalso able to publish their slightly different findings. He ended his story with,‘‘Sometimes you’re in an awkward position, and you try to do the best thing you canunder the circumstances, within your own internal ethical clock or whatever. Andsometimes it’s ugly and it’s imperfect, but it’s the only thing you can do. If we hadgone to the mentor and voiced this objection, our careers would have been over. Ifwe had approached the journal—God forbid, forget it.’’ The speaker qualified thisstory by saying that it made him sound much more ethical than he actually is.
peer review deliberately delayed in order to slow competitor
The focus-group discussions showed, however, that scientists see peer review asaffording a unique, even protected opportunity for competitors to take advantage ofthem. In this sense, competition infects the peer review process, not only throughscientists’ competition with other applicants, but also through scientists’ distrust ofthe reviewers themselves, as competitors. The following exchange among mid-career discussants shows their sense of vulnerability
Evidence that peer-reviewers are competitors.
By contrast, others use it in thepublication process solely to maintain their competitive bid for priority in a line ofresearch inquiry, as a way to sabotage others’ progress. A scientist in an early-career group acknowledged the need to make results reproducible by telling people‘‘the whole recipe of the whole method’’ if asked directly, but then she talkedabout the ‘‘little trick’’ of not including all the details in a publication orpresentation. Like the scientists in a different group quoted above, she mentionedothers’ practice of taking photographs of poster presentations in order then topublish the results first. She said that people, in defense, ‘‘omit tiny little details’’:‘‘But sometimes in the publication, people, just to protect themselves, will not giveall the details. It’s always right, but maybe it’s not totally complete—to protectthemselves. Because your ideas get stolen constantly, and it’s so competitive ifyou’re a small lab.
A more deliberate form of not sharing is the omission of critical details inpresentations, papers and grant proposals so that others will have difficultyreplicating and extending one’s own research.
You submit the first grant, youpropose the novel thing. You know damn well any study section that’s evenmildly conservative is going give you, ‘‘Well, it sounds promising.’’ Theymight give you a good score, you hope for a good score, but it’s not going toget funded, because it’s too novel, it’s too risky, it’s too blah blah. But youalready have the damn data. You know on the second resubmit, you’re goingto say, ‘‘Good point! We took that to heart. Oh, what a wonderful suggestion!We will worry about this too. Guess what? Here’s the data!’’ Shove it downtheir throat. And then it’s funded. Because, wow, you flagged them, yousucker-punched them. They said, ‘‘This is really novel, blah, blah. Boy if youcould only do that, that would be a great grant.’’ Well, you alreadydiddo it,and that’s the point. And you basically sucker-punch the study section intogiving you the money by default. They have to at that point. They don’t havea choice.
On the need to have results before funding is given.
Competition for funding, publications, scientific priority and overall career successleads scientists to describe their work in terms of strategies one might use in a game.Focus-group participants revealed that, like it or not, working within the scientificcommunity requires artful maneuvering and strategizing.
relationship of competition to explicit game-playing.
In short, there are many people (the oversupply factor) competing for prestigious,desirable and scarce rewards and resources (the funding factor), in a struggle thatbestows those rewards disproportionately on those of marginally greater achieve-ment (the tournament factor). This situation is supported to the detriment of that‘‘legion of the discontented’’ and to the benefit of senior investigators, because it‘‘generates good research by employing idealistic young graduate students andpostdoctoral fellows at low cost’’ . In other words, the benefits accrue to fundingand employing institutions. This paper explores some of the costs that accompanythese benefits
Economic structure of competition at universities.
Richard B. Freeman and colleagues  havecharacterized the problem as follows: ‘‘Research in the biosciences fits a tournamenteconomic structure. A tournament offers participants the chance of winning a bigprize—an independent research career, tenure, a named chair, scientific renown,awards—through competition.... It fosters intense competition by amplifying smalldifferences in productivity into large differences in recognition and reward. Well-structured tournaments stimulate competition. Because the differences in rewardsexceed the differences in output, there is a disproportionate incentive to ‘win’’’(p.2293). Research environments in which only small numbers of scientists have theopportunity to gain significant attention increase the competitive stakes: playing thegame may be a gamble, but the payoff for winning is significant [28,36]
The tournament structure of biosciences.
Another perspective sees competition as a function not just of funding, but of thebalance between supply and demand of resources, particularly human resources. Inthe current competitive system, young scientists are pitted against one another forattractive career opportunities that are becoming increasingly scarce .Researchers, feeling the pressure to be first to present findings in their fields,employ armies of graduate students and postdoctoral fellows and strive to maketheir laboratory groups the smartest and the fastest. The result is a ‘‘postdocbottleneck’’  where the supply for highly educated and trained researchers farexceeds the demand [30–33]. In concrete terms, Donald Kennedy and colleagues have described the structural problem as a source of excess supply of humancapital: ‘‘We’ve arranged to produce more knowledge workers than we can employ,creating a labor-excess economy that keeps labor costs down and productivity high’’(p. 1105). The system produces, they claim, a ‘‘legion of the discontented’’ .They argue that institutional and policy decisions about training scientists should becoupled to placement histories of recent graduates, numbers of intellectual offspringof faculty, and job markets for scientists. Roger L. Geiger  has suggested thatthe imbalance between supply and demand is due in part to deficiencies in graduate
Role of lack of positions. Interestingly, this has been shown by Fang et al to be not reflected in misconduct stats: i.e. the vast majority of scientific fraud is conducted by senior (male) scientists, not job hungry post-docs or grad students.
Analysts differ as to the reasons why competition has intensified. Some see thesituation in terms of money. Tempering the effects of competition is not a primeimpetus behind calls by the National Science Board  and by a recent coalition of140 college presidents and other leaders  for more federal funding for scientificresearch; however, some scientists see such advocacy movements in terms of easingcertain aspects of competition that are worsened by tight dollars. More money, morepositions, and overall expansion of the research enterprise would improve thesituation
role of funding
here are indications, however, that the natureof competition has changed in recent years. Goodstein  argues that this shift islinked to negative outcomes:Throughout most of its history, science was constrained only by the limits ofits participants’ imagination and creativity. In the past few decades, however,that state of affairs has changed dramatically. Science is now held back mainlyby the number of research posts and the amount of research funds available.What had been a purely intellectual competition has become an intensestruggle for scarce resources. In the long run, this change, which is permanentand irreversible, will probably have an undesirable effect on ethical behavioramong scientists. Instances of scientific fraud will almost surely become morecommon, as will other forms of scientific misconduct (p. 31)
relationship of negative aspects of competition to change in funding model that promotes scarcity. See Goodstein, D. (2002). Scientific misconduct.Academe, 88, 28–31
It is negatively correlated with subscription tonormative systems (either traditional or alternative) and sense of community
Scientific competition is is negatively correlated to eithical systems and sense of community.
Melissa S.Anderson  furthermore found that a competitive departmental environment inscience is positively correlated with exposure to misconduct, fears of retaliation forwhistle-blowing, and conflic
More evidence of correlation between competition and "exposure to misconduct".
Empirical findings show a strong, positive relationship between the level ofperceived competition in an academic department and the likelihood that depart-mental members will observe misconduct among their colleagues 
Higher the level of perceived competition in academic departments, the greater the liklihood that people will see misconduct among peers.
David Blumenthal and colleagues  found that university geneticists and otherlife scientists who perceive higher levels of competition in their fields are morelikely to withhold data or results. Such withholding took the form of omittinginformation from a manuscript or delaying publication to protect one’s scientificlead, maintaining trade secrets, or delaying publication to protect commercial valueor meet a sponsor’s requirements. John P. Walsh and Wei Hong  have reportedsimilar finding
Evidence that competition causes scientists to withhold results and/or data
Theirdiscussions suggest clearly that the downside of competition has been underesti-mated and that it may have more prominent effects now than in past years. Asreputation, respect and prestige are increasingly connected to resources and tosuccess in the competitions that distribute those resources, scientists find more oftheir work and careers caught up in competitive arenas. The six categories ofcompetition’s effects that emerged in our analyses suggest reason for concern aboutthe systemic incentives of the U.S. scientific enterprise and their implications forscientific integrity.
Implications of competition for scientific integrity.
- positive bias
- results decay
- scientific competition
- tournament economic structure
- scientific funding
- scientific integrity
- peer review
- scientific misconduct