A year ago, I received a message from Anna Powell-Smith about a research paper written by two doctors from Cambridge University that was a mirror image of a post I wrote on my personal blog1 roughly two years prior. The structure of the document was the same, as was the rationale, the methods, and the conclusions drawn. There were entire sentences that were identical to my post. Some wording changes were introduced, but the words were unmistakably mine. The authors had also changed some of the details of the methods, and in doing so introduced technical errors, which confounded proper replication. The paper had been press-released by the journal,2 and even noted by Retraction Watch.3
I checked my site’s analytics and found a record of a user from the University of Cambridge computer network accessing the blog post in question three times on 2015 December 7 and again on 2016 February 16, ten days prior to the original publication of the paper in question on 2016 February 26.4
At first, I was amused by the absurdity of the situation. The blog post was, ironically, a method for preventing certain kinds of scientific fraud. I was flattered that anyone noticed my blog at all, and I believed that academic publishing would have a means for correcting itself when the wrong people are credited with an idea. But as time went on, I became more and more frustrated by the fact that none of the institutions that were meant to prevent this sort of thing were working.
The journal did not catch the similarities between this paper and my blog in the first place, and the peer review of the paper was flawed as well. The journal employs an open peer review process in which the reviewers’ identities are published. The reviewers must all make a statement saying, “I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.” Despite this process, none of the reviewers made an attempt to analyse the validity of the methods used.
After the journal’s examination of the case, they informed us that updating the paper to cite me after the fact would undo any harm done by failing to credit the source of the paper’s idea. A new version was hastily published that cited me, using a non-standard citation format that omitted the name of my blog, the title of my post, and the date of original publication. The authors did note that the idea had been proposed in “the grey literature,” so I re-named my blog to “The Grey Literature” to match.
I was shocked by the journal’s response. Authorship of a paper confers authority in a subject matter, and their cavalier attitude toward this, especially given the validity issues I had raised with them, seemed irresponsible to me. In the meantime, the paper was cited favourably by the Economist5 and in the BMJ6, crediting Iriving and Holden.
I went to Retraction Watch with this story,7 which brought to light even more problems with this example of open peer review. The peer reviewers were interviewed, and rather than re-evaluating their support for the paper, they doubled down, choosing instead to disparage my professional work and call me a liar. One reviewer wrote, “It is concerning that this blogger would be attempting a doctorate and comfortably ascribe to a colleague such falsehoods.”
The journal refused to retract the paper. It was excellent press for the journal and for the paper’s putative authors, and it would have been embarrassing for them to retract it. The journal had rolled out the red carpet for this paper after all,2 and it was quickly accruing citations.
The case was forwarded to the next meeting of the Committee on Publication Ethics (COPE) for their advice. Three months later, at the August 2016 COPE meeting, the case was presented and voted on.8 It was surreal for me to be forced to wait for a seemingly unaccountable panel of journal editors to sit as a de facto court, deciding whether or not someone else would be credited with my words, all behind locked doors, with only one side of the case—the journal editors’—represented. In the end, they all but characterised my complaints as “punitive,” and dismissed them as if my only reason for desiring a retraction was that I was hurt and wanted revenge. The validity issues that I raised were acknowledged but no action was recommended. Their advice was to send the case to the authors’ institution, Cambridge University, for investigation. I do not know if Cambridge did conduct an investigation, and there has been no contact with me.
There is, to my knowledge, no way to appeal a decision from COPE, and I know of no mechanism of accountability for its members in the case they advise a journal with the wrong answer. As of January 2017, the journal officially considered the case closed.
It is very easy to become disheartened and jaded when things like this happen—as the Economist article citing Irving and Holden says, “Clinical trials are a murky old world.”5 The institutions that are supposed to protect the integrity of the academic literature sometimes act in ways that miss the lofty standards we expect from modern science.
Fortunately, the scientific community turned out to be a bigger place than I had given it credit for. There are people like Anna, who let me know that this was happening in the first place and Ben Goldacre, who provided insight and support. My supervisor and my colleagues in the STREAM research group were incredibly supportive and invested in the outcome of this case. A number of bloggers (Retraction Watch,7,9 Neuroskeptic,10 Jordan Anaya11—if I missed one, let me know!) picked up this story and drew attention to it, and in the end, the paper was reviewed by Daniel Himmelstein,12 whose persistence and thoroughness convinced the journal to re-open the case and invite Dr Knottenbelt’s decisive review.
While it is true that the mistakes introduced into the methods are what finally brought about its retraction, those mistakes happened in the first place because the authors did not come up with the idea themselves. It is a fallacy to think that issues of scientific integrity can be considered in isolation from issues of scientific validity, and this case very clearly shows how that sort of thinking could lead to a wrong decision.
Of course, there are still major problems with academic publishing. But there are also intelligent and conscientious people who haven’t given up yet. And that is an encouraging thought.
References
1. Carlisle, B. G. Proof of prespecified endpoints in medical research with the bitcoin blockchain. The Grey Literature (2014).
2. F1000 Press release: Doctors use Bitcoin tech to improve transparency in clinical trial research. (2016). Available at: http://f1000.com/resources/160511_Blockchain_FINAL.pdf. (Accessed: 23rd June 2016)
3. In major shift, medical journal to publish protocols along with clinical trials. Retraction Watch (2016).
4. Irving, G. & Holden, J. How blockchain-timestamped protocols could improve the trustworthiness of medical science. F1000Research 5, 222 (2017).
5. Better with bitcoin | The Economist. Available at: http://www.economist.com/news/science-and-technology/21699099-blockchain-technology-could-improve-reliability-medical-trials-better. (Accessed: 23rd June 2016)
6. Topol, E. J. Money back guarantees for non-reproducible results? BMJ 353, i2770 (2016).
7. Plagiarism concerns raised over popular blockchain paper on catching misconduct. Retraction Watch (2016).
8. What extent of plagiarism demands a retraction vs correction? | Committee on Publication Ethics: COPE. Available at: http://publicationethics.org/case/what-extent-plagiarism-demands-retraction-vs-correction. (Accessed: 16th August 2016)
9. Authors retract much-debated blockchain paper from F1000. Retraction Watch (2017).
10. Neuroskeptic. Blogs, Papers, Plagiarism and Bitcoin – Neuroskeptic. (2016).
11. Anaya, J. Medical students can’t help but plagiarize, apparently. Medium (2016). Available at: https://medium.com/@OmnesRes/medical-students-cant-help-but-plagiarize-apparently-f81074824c17. (Accessed: 21st July 2016)
12. Himmelstein, Daniel. Satoshi Village. The most interesting case of scientific irreproducibility? Available at: http://blog.dhimmel.com/irreproducible-timestamps/. (Accessed: 8th March 2017)
BibTeX
@Manual{stream2017-1280, title = {Recapping the recent plagiarism scandal}, journal = {STREAM research}, author = {Benjamin Gregory Carlisle}, address = {Montreal, Canada}, date = 2017, month = jun, day = 2, url = {http://www.translationalethics.com/2017/06/02/recapping-the-recent-plagiarism-scandal/} }
MLA
Benjamin Gregory Carlisle. "Recapping the recent plagiarism scandal" Web blog post. STREAM research. 02 Jun 2017. Web. 11 Dec 2024. <http://www.translationalethics.com/2017/06/02/recapping-the-recent-plagiarism-scandal/>
APA
Benjamin Gregory Carlisle. (2017, Jun 02). Recapping the recent plagiarism scandal [Web log post]. Retrieved from http://www.translationalethics.com/2017/06/02/recapping-the-recent-plagiarism-scandal/