November 2, 2012

Scientific fraud is rife: it's time to stand up for good science - The Guardian

The way we fund and publish science encourages fraud. A forum about academic misconduct aims to find practical solutions
 A meeting room 
Peer review happens behind closed doors, with anonymous reviews only seen by editors and authors. This means we have no idea how effective it is. Photo: Alamy
Science is broken. Psychology was rocked recently by stories of academics making up data, sometimes overshadowing whole careers. And it isn't the only discipline with problems - the current record for fraudulent papers is held by anaesthesiologist Yoshitaka Fujii, with 172 faked articles.
These scandals highlight deeper cultural problems in academia. Pressure to turn out lots of high-quality publications not only promotes extreme behaviours, it normalises the little things, like the selective publication of positive novel findings – which leads to "non-significant" but possibly true findings sitting unpublished on shelves, and a lack of much needed replication studies.
Why does this matter? Science is about furthering our collective knowledge, and it happens in increments. Successive generations of scientists build upon theoretical foundations set by their predecessors. If those foundations are made of sand, though, then time and money will be wasted in the pursuit of ideas that simply aren't right.
A recent paper in the journal Proceedings of the National Academy of Sciences shows that since 1973, nearly a thousand biomedical papers have been retracted because someone cheated the system. That's a massive 67% of all biomedical retractions. And the situation is getting worse - last year, Nature reported that the rise in retraction rates has overtaken the rise in the number of papers being published.
This is happening because the entire way that we go about funding, researching and publishing science is flawed. As Chris Chambers and Petroc Sumner point out, the reasons are numerous and interconnecting:
• Pressure to publish in "high impact" journals, at all research career levels;
• Universities treat successful grant applications as outputs, upon which continued careers depend;
• Statistical analyses are hard, and sometimes researchers get it wrong;
• Journals favour positive results over null findings, even though null findings from a well conducted study are just as informative;
• The way journal articles are assessed is inconsistent and secretive, and allows statistical errors to creep through.
Problems occur at all levels in the system, and we need to stop stubbornly arguing that "it's not that bad" or that talking about it somehow damages science. The damage has already been done – now we need to start fixing it.
Chambers and Sumner argue that replication is critical to keeping science honest, and they are right. Replication is a great way to verify the results of a given study, and its widespread adoption would, in time, act as a deterrent for dodgy practices. The nature of statistics means that sometimes positive findings arise by chance, and if replications aren't published, we can't be sure that a finding wasn't simply a statistical anomaly.
But replication isn't enough: we need to enact practical changes at all levels in the system. The scientific process must be as open to scrutiny as possible – that means enforcing study pre-registration to deter inappropriate post-hoc statistical testing, archiving and sharing data online for others to scrutinise, and incentivising these practices (such as guaranteeing publications, regardless of findings).
The peer-review process needs to be overhauled. Currently, it happens behind closed doors, with anonymous reviews only seen by journal editors and manuscript authors. This means we have no real idea how effective peer review is – though we know it can easily be gamed. Extreme examples of fake reviewers, fake journal articles, and even fake journals have been uncovered.
More often, shoddy science and dodgy statistics are accepted for publication by reviewers with inadequate levels of expertise. Peer review must become more transparent. Journals like Frontiers already use an interactive reviewing format, with reviewers and authors discussing a paper in a real-time, forum-like setting.
A simple next step would be to make this system open and viewable by everyone, while maintaining the anonymity of the reviewers themselves. This would allow young researchers to be critical of a senior academic's paper without fear of career suicide.
On 12 November, we are hosting a session on academic misconduct at SpotOn London, Nature's conference about all things science online.
The aim of the session is to find practical solutions to these problems that science faces. It will involve scientific researchers, journalists and journal editors. We've made some suggestions here, but we want more from you. What would you like to see discussed? Do you have any ideas, opinions or solutions?
We'll take the best points and air them at the session, so speak up now! Let's stop burying our heads in the sand and stand up for good science.
Pete Etchells is a biological psychologist and Suzi Gage is a translational epidemiology PhD student. Both are at the University of Bristol

Random Posts


  • Responding to Possible Plagiarism

    SCIENCE, 6 March 2009: Vol. 323. no. 5919, pp. 1293 - 1294 DOI: 10.1126/science.1167408Tara C. Long,1 Mounir Errami,2 Angela C. George,1 Zhaohui Sun,2 Harold R. Garner1,2*The peer-review process is the best mechanism to ensure the high quality of scientific publications. However, recent studies have... READ MORE>>

  • Study finds plenty of apparent plagiarism (Science News)

    Data mining reveals too many similarities between papers By Janet Raloff Web edition : Thursday, March 5th, 2009 Enlarge IS THIS PLAGIARISM? Yellow highlights aspects of this paper that copy material published in a previous paper — by other authors. UT Southwestern Medical CenterIf copyi... READ MORE>>

  • Borrowing words, or claiming them?

    EditorialNature Immunology 10, 225 (2009)doi:10.1038/ni0309-225Journals are taking steps to stem of the practice of plagiarism.Have you ever experienced a sense of déjà vu after reading a colleague's manuscript or researching a topic of interest? A paragraph or entire section sounds eerily familiar—... READ MORE>>

  • It's Culture, Not Morality - INSIDE HIGHER ED

    Scott JaschikWhat if everything you learned about fighting plagiarism was doomed to failure? Computer software, threats on the syllabus, pledges of zero tolerance, honor codes -- what if all the popular strategies don't much matter? And what if all of that anger you feel -- as you catch students cle... READ MORE>>

  • Problems with anti-plagiarism database

    Mauno VihinenNATURE|Vol 457|1 January 2009 SIR — Sophisticated tools have been developed to detect duplicate publication and plagiarism, as noted in M. Errani and H. Garner’s Commentary ‘A tale of two citations’ (Nature 451, 397–399; 2008) and in your News story ‘Entire-paper plagiarism caught by so... READ MORE>>

  • Entire-paper plagiarism caught by software - NATURE

    Thousands of 'similarities' found between papers. Declan Butler Nature 455, 715 (2008) >>>> Many of the duplicates in Deja Vu come from non-English-speaking countries, and some scientists have asserted that a degree of plagiarism is justified as a way of improving the English of their te... READ MORE>>

  • Ethics in science: Are we losing the moral high ground?

    Faisal M Sanai Associate Editor, Saudi J Gastroenterol 2008;14:107-8In the competitive world of academia, a person's worth is often ostensibly gauged by one's scientific contribution, wherein the 'article count' has become the simplistic measure of this contribution. The number and frequency of... READ MORE>>

.

.
.

Popular Posts