Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
Science

Fake Scientific Papers Are Alarmingly Common 64

From a Science magazine report, shared by schwit1: When neuropsychologist Bernhard Sabel put his new fake-paper detector to work, he was "shocked" by what it found. After screening some 5000 papers, he estimates up to 34% of neuroscience papers published in 2020 were likely made up or plagiarized; in medicine, the figure was 24%. Both numbers, which he and colleagues report in a medRxiv preprint posted on 8 May, are well above levels they calculated for 2010 -- and far larger than the 2% baseline estimated in a 2022 publishers' group report. "It is just too hard to believe" at first, says Sabel of Otto von Guericke University Magdeburg and editor-in-chief of Restorative Neurology and Neuroscience. It's as if "somebody tells you 30% of what you eat is toxic." His findings underscore what was widely suspected: Journals are awash in a rising tide of scientific manuscripts from paper mills -- secretive businesses that allow researchers to pad their publication records by paying for fake papers or undeserved authorship.

"Paper mills have made a fortune by basically attacking a system that has had no idea how to cope with this stuff," says Dorothy Bishop, a University of Oxford psychologist who studies fraudulent publishing practices. A 2 May announcement from the publisher Hindawi underlined the threat: It shut down four of its journals it found were "heavily compromised" by articles from paper mills. Sabel's tool relies on just two indicators -- authors who use private, noninstitutional email addresses, and those who list an affiliation with a hospital. It isn't a perfect solution, because of a high false-positive rate. Other developers of fake-paper detectors, who often reveal little about how their tools work, contend with similar issues. Still, the detectors raise hopes for gaining the advantage over paper mills, which churn out bogus manuscripts containing text, data, and images partly or wholly plagiarized or fabricated, often massaged by ghost writers.

Some papers are endorsed by unrigorous reviewers solicited by the authors. Such manuscripts threaten to corrupt the scientific literature, misleading readers and potentially distorting systematic reviews. The recent advent of artificial intelligence tools such as ChatGPT has amplified the concern. To fight back, the International Association of Scientific, Technical, and Medical Publishers (STM), representing 120 publishers, is leading an effort called the Integrity Hub to develop new tools. STM is not revealing much about the detection methods, to avoid tipping off paper mills. "There is a bit of an arms race," says Joris van Rossum, the Integrity Hub's product director. He did say one reliable sign of a fake is referencing many retracted papers; another involves manuscripts and reviews emailed from internet addresses crafted to look like those of legitimate institutions. Twenty publishers -- including the largest, such as Elsevier, Springer Nature, and Wiley -- are helping develop the Integrity Hub tools, and 10 of the publishers are expected to use a paper mill detector the group unveiled in April.
This discussion has been archived. No new comments can be posted.

Fake Scientific Papers Are Alarmingly Common

Comments Filter:
  • The the release of ChatGPT is only going to accelerate the problem.
    • Re: (Score:1, Insightful)

      by cats-paw ( 34890 )

      yep. and probably make them more difficult to catch.

      Also, incentive problem.

      Your entire academic career depends on publishing, so you have an incentive to cheat.

      • by Chas ( 5144 ) on Friday May 12, 2023 @01:00PM (#63517067) Homepage Journal

        The real bitch in ferreting it out is also the circular reference problem.
        You get a bogus paper referring to another bogus paper and another and another. And after enough layers, it becomes EXCEEDINGLY difficult to filter BS as it gets entrenched in the paper-sphere.

        • The real bitch in ferreting it out is also the circular reference problem.
          You get a bogus paper referring to another bogus paper and another and another. And after enough layers, it becomes EXCEEDINGLY difficult to filter BS as it gets entrenched in the paper-sphere.

          Until you try to implement something based on the papers' findings.

        • This is a good thing. The more the bullshit is exposed the sooner the problem gets fixed.

          • by vivian ( 156520 )

            Only if all the bullshitters involved face consequences like getting kicked out of their field.

      • I've found that this all varies depending upon the field. Ie, Chemistry has lots of names on each paper (or did) with the senior member always listed first. For CS, the number of names were smaller, 2 to 4, with the student's name often going first. Though it probably depends upon the prof too. Junior Professors are indeed under a lot of pressure.

        I also think it's not just now that times have changed, but there's been change for the last century possibly. I know when I was in grad school in early 90s t

        • I'm not sure of how this will fall out here, but long term research a few years ago found that half of research results are wrong. If some are faked, then that might explain why that number was so high.

    • ChatGPT can't write a coherent paper. A future version might be able to do so, but even then, it's going to have trouble achieving something novel.
  • Fraud (Score:4, Interesting)

    by The Evil Atheist ( 2484676 ) on Friday May 12, 2023 @12:39PM (#63517011)
    Time to treat scientific fraud like real fraud. But not only should the authors involved get punished, but their institution. That would teach them for continuing to use the failed publish-or-die criteria.
    • Comment removed based on user account deletion
      • Comment removed based on user account deletion
        • Re:Fraud (Score:5, Interesting)

          by HiThere ( 15173 ) <charleshixsn.earthlink@net> on Friday May 12, 2023 @03:25PM (#63517535)

          While true, that's a horribly oversimplified model. So oversimplified that acting on it is incorrect.

          I think that the BASIC problem may be the way the patent system is broken. If two groups independently invent something, only one can get the patent. Independent replication should be what is *demanded*. And both (all?) should get equal compensation.

          There's a problem here, of course, when the cost of the experiment is so large that it can't be duplicated, but such results should not be considered resolved. (What's the mass of a W-boson? ... We've to two independent answers, and they don't quite agree. So if we only had one, should be believe it?)

  • Paper Mills (Score:5, Informative)

    by Sperbels ( 1008585 ) on Friday May 12, 2023 @12:55PM (#63517047)
    I'd just like to point out that a 'paper mill' is actually a factory where they manufacture paper. This summary was quite confusing to me a moment, as if the paper industry was somehow conspiring to pump out as many fake scientific papers as possible....which I guess makes sense if these scientific papers were printed and had mass distribution.
    • When academics have to go by "publish or perish" then publish they will. Hence a whole lot of bullshit gets published.

    • ... a paper-manufacturing plant run by an agricultural university [wikipedia.org]??? Who knew?

    • I'd just like to point out that a 'paper mill' is actually a factory where they manufacture paper.

      Yeah, American idioms are one of the hardest things for non-native speakers to understand. Recently I used the phrase "kick the can down the road" and was met with a blank look until I explained the meaning. Then I had to explain that "kick the can down the road", "kick the can", and "kick the bucket" were all totally different concepts.

      There are idioms in other languages that don't translate into English very well either. I was recently told to "shake the fleas" in French and I had that same blank look

  • It is a for profit ende or meant only to maximize short term profit, not enhance human knowledge

    Legitimate science, though, has no defense against fake papers. And it should not. Finding have to taken at face value, with a grain of salt. Ongoing research will determine if the findings have validity or use.

    It is the mainstream belief that science and research discover facts, rather than being a process to determine what might be real, that causes all these problems. We see this in education where the maj

    • by phantomfive ( 622387 ) on Friday May 12, 2023 @01:07PM (#63517095) Journal
      Science is a tool.

      Legitimate science, though, has no defense against fake papers.

      The scientific defense against fake papers is experiment and reproducibility. Science is built on the idea that you can't trust people.

      • MAKE REPRODUCTION SEXY AGAIN

        er wait

        Seriously though, how do we get more attempts to reproduce results?

        • From a societal perspective, the answer is to fund it.

          From a minimalist perspective, it will happen automatically (slowly and haphazardly) as people try to use those results.
      • by fermion ( 181285 )
        A few years ago I was working at a very famous private research university. Over the course of a month we tried to fabricate a novel material from published research. We even called the authors. Couldnâ(TM)t do it

        Does this mean the authors were lying or we just couldnâ(TM)t do it. We had talent and money. But maybe not skill.

        • Does this mean the authors were lying or we just couldnâ(TM)t do it

          Could be either one. That sucks though.

        • by vlad30 ( 44644 )

          We had talent and money. But maybe not skill.

          I'll give a woodworking example/Analogy For several years a very skilled young woodworker worked at a company and became known for his excellent work, so well that the customers started saying he should start out on his own so he decided he would.

          He had talent and money so he bought all the best tools, mitre saws, and table saws, and set up shop and the customers came to him with the orders yet not a single piece he made was perfect as it was in the last company. After a few months he had to close his b

      • by vlad30 ( 44644 )

        Science is built on the idea that you can't trust people.

        Especially don't trust scientists until the results/theorems/models are independently verified

    • "Tell me you know nothing about publishing scientific research without telling me you know nothing about publishing scientific research."
      • by fermion ( 181285 )
        An interesting story. 30 years ago or so I was involved in very expensive cutting edge research. This means that not many were doing it, and we were building circuit boards and writing code from scratch. One result was counter intuitive, but has analogies in nature, think an ant hill, so was repeatedly published as we did experiments. Further on as more people got involved it was proven to be a software error. But that is the process of science.

        The funny thing was all our professors were saying donâ(

    • The authors of the paper are in the neuroscience field, so they had an interesting in splitting their own field out and lumping all others together. In the paper, "medicine" and "general medicine" is used to refer to all medical sciences with the exception of neuroscience.

      Being pedantic does your argument no favors.

      • by fermion ( 181285 )
        Not my experience. One friend who works in a major big city hospital that everyone would recognize and has patients flown in from all over the state, country, and world, is a neuroscientists with more letters after their name than anyone I I know.

        He was telling me of one truly fascinating case. He also told me that he was behind on his publications, so one one trip he wrote 5 papers on the case. Three were accepted. One data point, three papers. Not rigorous science.

  • It should be illegal for any psychologists advise any persons unless that advice is based on psychological research that has been replicated at least twice.

    Those charlatans spew way too much dangerous shit.
    • Psychology's really hard research. The problem is that people want acclaim and easy answers, but it's one of the most imprecise fields because people keep moving around and don't behave as you'd expect. Never trust a psych researcher whose priors always conveniently match their outputs.
      • Obligatory addition: we'd get more replication studies if 1.) replications were funded and 2.) replications had any affect on the market. I know people who've clearly, roundly failed to replicate studies, but get less support and attention than the study they're critiquing, which continues to be cited.
        • Note that people have been medicated (and literally lobotomized) based on that kind of stuff.
          • Not to mention people murdered by people let out of jail by psychiatrists who do not have science to tell them who to release or not.
  • Of course one quarter of it is fraud, but hey those are still good odds, unless this study is fraudulent then....

    • by XXongo ( 3986865 )

      Of course one quarter of it is fraud, but hey those are still good odds, unless this study is fraudulent then....

      Actually, they said fraud or plagiarized.

      Unless the paper plagiarized was fraud, the plagiarized papers may still be good results.

    • by HiThere ( 15173 )

      Did you notice that they mentioned a high false positive finding? And they listed their criteria for deciding that a paper is fraudulent in some way ("Sabel's tool relies on just two indicators -- authors who use private, noninstitutional email addresses, and those who list an affiliation with a hospital."), so the high false positive rate isn't surprising.

      This may be a decent tool for gathering statistics (with large error bars), but it would be reckless to use it to judge an individual paper.

  • Publish or perish (Score:5, Insightful)

    by bradley13 ( 1118935 ) on Friday May 12, 2023 @01:34PM (#63517181) Homepage

    Two problems.

    First, thete are too many marginal people in research. Schools want grad students, and let people in who aren't really capable of contributing to their field. But they make for cheap labor to grade undergrad homework and to produce reams if crappy papers.

    Which brings us to the second problem: "publish or perish". Universities can't evaluate research on its merits, so they reach for a simple yardstick: how many papers have you published this year?

    Combine the two, and most published research is crap. Even good research is broken up into "MPIs" - minimum publishable increments - so that you get as many papers as possible out of each project.

    • by Anonymous Coward

      >reams [of] crappy papers
      I wonder if Einstein's paper theorizing mass-energy equivalence could even get published today. Its text can fit on a single page, and he didn't provide experimental data.

    • Re:Publish or perish (Score:4, Informative)

      by Areyoukiddingme ( 1289470 ) on Friday May 12, 2023 @04:46PM (#63517709)

      Schools want grad students, and let people in who aren't really capable of contributing to their field. But they make for cheap labor to grade undergrad homework and to produce reams if crappy papers.

      Which brings us to the second problem: "publish or perish". Universities can't evaluate research on its merits, so they reach for a simple yardstick: how many papers have you published this year?

      Here's the problem. Neither of those problems apply to this result. From The Fine Summary:

      Sabel's tool relies on just two indicators -- authors who use private, noninstitutional email addresses, and those who list an affiliation with a hospital.

      If this flood of trash papers was being done to counter the "publish or perish" mandate from universities, they would use their academic email addresses in their by-lines to make damn sure they got credit. That's not happening. They're using Yahoo addresses and such-like. Personally I think scientists should have switched to using addresses they personally control for their publications a long time ago because no academic affiliation is forever, even before tenure was gutted, and being able to have a constant email address for all of your publications would help establish your personal brand, but to make that stick you have to do really good work. Instead academic scientists have been leaning heavily on the cachet of their specific academic affiliations for many many years, and now the well has been poisoned.

      In any case, because the habit of clinging to their affiliations is so strong, papers published without them are legitimately suspect. No one playing the academic game would hide their affiliation. That's quite the opposite of their need.

      • by Uecker ( 1842596 )

        Not sure what was done here, but email address and affiliation do not need to correspond. I know many people who use their private email on papers if they expect to change institution in the future.

  • So a neuropsychologist publishes a paper saying that up to 34% of neuroscience papers were likely made up or plagiarized. So, logically, this means that there's about a 1 in 3 chance that his own paper was made up or plagiarized. (Also, logically, if Bernhard Sabel weighs the same as a duck, he's made of wood, and therefore ... A WITCH!)
  • by MobyDisk ( 75490 ) on Friday May 12, 2023 @02:03PM (#63517257) Homepage

    We need to start putting significant money and resources toward replicating important published science. And they can pay for it off the torts earned when they find fake research.

    We also need to end the "publish or perish" mindset [slashdot.org] in academia, and to acknowledge the value of neutral or negative results [slashdot.org] instead of expecting every paper to be an enormous breakthrough. Science needs to proceed like a tortoise, slow and steady. No fast but with decades-long setbacks due to falsified information.

  • It seems like this would be a good place to deploy AI to do a preliminary evaluation of the mathematics and pointing out possible issues. It seems like plagiarism should be caught rather easily considering the technology to detect it has been used on college papers for a couple decades.

    Manipulated data is probably the largest issue here as without the original, it can only be detected by evaluating the experiment.

  • So, this guy uses two criteria for detecting fake papers: (1) authors who use private, non-institutional email addresses and (2) authors who list an affiliation with a hospital. My field is computer architecture and not medicine, but it seems like the real problem is low quality conferences and journals. Obviously the guy's two criteria are useless for computer architecture conferences. I wonder if these two criteria are even slightly useful for top-level medical journals.

    The other big problem is the de

    • by Uecker ( 1842596 )

      Neither private email nor affiliation with a hospital is a indication for a fake publication. Also in medical journals review is blind (or double blind). In most scientific journals (not just medical) authors can suggest reviewers, but an editor would not select a reviewer which might be biased. There are low-quality journals (in all fields) that published essentially anything, but your reputation as a scientist does not improve if you publish in such journals.

      I also do not believe that most low-quality or

      • by Uecker ( 1842596 )

        Also looking at the original paper it is also not really what they are doing. They look at various indicators for fake papers and try to validate them. They then find that "private email" and "hospital affiliation" has a detection sensitivity was 0.86 and the false-alarm rate
        0.44 when validated on a set of papers known as fakes and presumed not fakes and suggest that this criterion could be used as a red flag.

  • It's not that hard to get published. [isotropic.org] The hard part is delivering it at a conference [youtu.be] with a straight face.

    In the humor section.

  • And then there's this guy - https://scholar.google.com/cit... [google.com]
  • If there was no requirement to publish crap all the time to stay in a field, keep a teaching job, do some basic but not exciting research on the side, then there would be no incentive to cheat the system on massive scales. And 34% is massive. Even the 2% claimed is really bad for the field of "science" which is supposed to be a search for truth. Oh, the irony!

    Without all the faked up and honest but shitty papers getting cranked out there would be lots more resources available to vet the papers that did g

  • Remember this article the next time "researchers" claim China dominates the world in cited scientific papers.

  • Publishers are meant to be the gatekeepers - they are responsible for the integrity of their publications and need to be held to account for their actions.
  • I wonder what is the fraction of bad papers in the global warming topic....
  • I'd be interested to know if my assumptions about where faked or plagiarised papers come from are true. Eg a country with > 1 billion population?
    • That would be my expectation, based on experience. When I found plagiarism in a paper I was reviewing, that's where it was from. The editors let them get away with it after it was brought to the editors' attention.

  • Comment removed based on user account deletion
  • Hopefully he'll have a sense of humor about this a year from now when his perception of "common" has been adjusted by a few orders of magnitude. :)

  • I understand that it is already inevitable and there is so much of it that it's silly to deny it, but I can't call it a problem. The whole education system is long outdated and this is just a new twist on it, just like the use of gadgets used to be. There have long been resources like https://eduzaurus.com/conclusion-generator/ [eduzaurus.com] that just help you write the same essays by picking up the conclusion that makes writing very fast. Now there are neural networks that perform the same function and everyone will get

In the long run, every program becomes rococco, and then rubble. -- Alan Perlis

Working...