Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
Science

Paper Mills Have Flooded Science With 400,000 Fake Studies, Experts Warn (nature.com) 57

A group of scientific integrity experts is calling for urgent action to combat "paper mills" -- companies that sell fraudulent research papers and fake peer reviews. In a Nature comment piece published January 27, the experts warn that at least 400,000 papers published between 2000 and 2022 show signs of being produced by paper mills, while only 55,000 were retracted or corrected during that period.

Paper Mills Have Flooded Science With 400,000 Fake Studies, Experts Warn

Comments Filter:
  • Paper mills? Well, I mean, aside from the cleanup costs, I don't see the problem. The paper that comes out of a paper mill is blank, after all.
  • Goodhart's law (Score:5, Insightful)

    by Anonymous Coward on Wednesday January 29, 2025 @09:55AM (#65127763)

    Goodhart's law (https://en.wikipedia.org/wiki/Goodhart%27s_law) is "When a metric becomes a target, it ceases to function as a metric". People noticed that good academics tended to publish more than bad ones (though this is not a universal rule!) so we got targets for numbers of publications. And what's the result? Massive amounts of useless/marginal papers contributing only noise. Paper mills exist because bad actors want to compete and can win on these metrics. Hell some people are apparently able to produce multiple papers per day!

    Want to fix this? Fix your metrics. Bad metrics destroy progress because they incentivize the wrong things. You ask for lots of papers, you get lots of noise. You ask for lots of citations, you get citation circles, huge self-citations and friend/courtesty citations, and reviewers demanding you cite them. You want "high impact" papers? You get Wakefield's high impact MMR work, researchers writing about hype instead of rigorous science. And they'll do it because we select out those who don't play the game. When your five good papers are outweighed in job and grant applications by twenty poor ones, that's your career ended in favour of the cheaters.

    Anyone involved in coding knows this - you give a "lines of code" metric, you get lots of verbose code. You give a "commits" metric, you'll be innundated with pointless requests.

    So, fix your metrics. Or better yet, get rid of them entirely and do the hard work of actually evaluating output properly.

    • Right. Quality is difficult, perhaps impossible, to quantify, which is of course an endless frustration for scientists in particular, who have been trained to believe in the primacy of quantity.
      • You can get at it to some extent by the journal in which a paper is published. It's not perfect but it's also hard to game but you do need to know about the field to know which are the top journals.
        • True about the journal, up to a point. What's so funny is 'peer reviewed paper.' As a senior engineer I have A LOT of things to review. I do try and review them but if I gave them the attention they deserved - I'd never get my own work done.
          • We need a statistic of how many of the 400,000 estimated research papers with fake data were funded, in whole or part, by government grants, grants from nonprofits, etc..

            Next, for authors found to falsify a paper's data, we'd need statistics on how many other papers they've published along with grant dollars spent.

            Finally, universities and nonprofits should be required to pay back some portion of a grant that resulted in published papers proven to have false data.

            University of California

            Frequently Asked Qu

            • Another good point. The interesting thing is I wonder what the definition for research infrastructure entails?
          • Also no "peer review" can establish a paper's worth. To do that would require doing a good proportion of the work that went into the paper, and having knowledge that perhaps the "expert reviewers" lack.

            • Yes, I should have mentioned that , too. Short of experimental verification, what does 'review' mean anyway?
      • A friend who knew a good deal about science and management told me that most really good scientists make at most three or four major discoveries in their careers. Some fewer. Obviously asking for 30 or more papers per years is absurd.

        Quality should be dealt with at the higher level - the college or institute should set appropriate goals and then create teams of appropriate size and membership to tackle them. Perhaps the institutions should have to compete, rather than the individuals.

    • Re:Goodhart's law (Score:5, Insightful)

      by Archangel Michael ( 180766 ) on Wednesday January 29, 2025 @10:52AM (#65127937) Journal

      The fix is to incentivize critiques and analysis of the papers, and do that several layers deep, including meta moderation, targeting the critiques and analysis for review as well.

      THIS is the basis for science and digging for truth is hard, tedious and takes a lot of time. It isn't perfect.

      • Absolutely no academic or researcher ever got a promotion based on the quality of their reviews.
        • Absolutely no academic or researcher ever got a promotion based on the quality of their reviews.

          Which is the problem that GP's suggesting we fix.

      • The fix is to incentivize...

        Another part of the fix is to stop using self-important, ugly, garbage words like "incentivize," and go back to using words like "encourage," which say what you mean without trying to impress your readers with how smart you are.
      • The fix is to incentivize critiques and analysis of the papers, and do that several layers deep, including meta moderation, targeting the critiques and analysis for review as well.

        THIS is the basis for science and digging for truth is hard, tedious and takes a lot of time. It isn't perfect.

        Also known as the /. Model

    • This.

      You get what you measure.

      I'd start reform by insisting that every paper include a necessary and sufficient falsifiable hypothesis statement. That will filter out 90% of the fraudulent/irreproducible trash. From both paper mills and "respectable" journals.

    • Too many submissions, not enough people to evaluate output properly.

    • by fropenn ( 1116699 ) on Wednesday January 29, 2025 @12:11PM (#65128201)

      fix your metrics

      I hate hate HATE the article count and citation count as metrics of research success. Not only does it incentivize quantity and linking your work to research clickbait, it promotes the myth of the 'independent researcher working alone in vacuum.' In my career I've always been more successful when I work in a department or setting that is active - lots of data to analyze, lots of money flowing around, etc. I'm the same person, but my output is so heavily dependent on those around me and the structures and resources available in that setting. The third problem with these metrics is there is no incentive to actually solve anything, as writing about the same problem endlessly (without making any real progress) is rewarded just the same as making actual progress.

      I propose that the true measure of success should be at the team or department level (rather than individual), and should emphasize progress in addressing or solving an important problem. And rewards and improvements should be focused on the team and department level too (although sometimes you also have individuals in the team that need to develop new skills or change their behaviors). But that kind of analysis of success is much more difficult and time consuming than simply counting papers and citations and stupid impact factors...so no one ever does it.

    • I think the difficulty is genuine to some extent, and arises because there are so many scientists nowadays - most of them just plain honest people trying to make a living and perhaps get promotion or recognition.

      Ironically, I think the system is broken because none of the higher-ups have given any useful thought to how science should be organised. As you point out, generating tens of thousands of more or less meaningless papers is not useful. What's needed is some way of setting important goals and organisi

      • Goodhart's Law really is the heart of the matter.

        Please notice, also, that most of the foundations of almost every serious science were laid by amateurs - in the sense that they were people working for the satisfaction of it, not salaried employees.

        Aristotle, Euclid, Archimedes, Leonardo da Vinci, Roger Bacon, Francis Bacon, Tycho Brahe, Kepler, Galileo, Descartes, Newton, Leibniz, Faraday, Babbage, Darwin, and on and on up to about WW2. None of them had "qualifications" - much. Some of them, being so incli

  • by Baron_Yam ( 643147 ) on Wednesday January 29, 2025 @10:06AM (#65127801)

    All papers from this point forward should be hand-written in the lead's handwriting, accompanied by a video of the writing process. And the pen(s), for forensic ink comparison.

    Also, any respectable journal will not publish without suspending the submitter over a pit of spikes for two weeks while the work is authenticated. A fail gets them a quick drop.

    • by nightflameauto ( 6607976 ) on Wednesday January 29, 2025 @10:22AM (#65127847)

      All papers from this point forward should be hand-written in the lead's handwriting, accompanied by a video of the writing process. And the pen(s), for forensic ink comparison.

      Also, any respectable journal will not publish without suspending the submitter over a pit of spikes for two weeks while the work is authenticated. A fail gets them a quick drop.

      A suspicious rise in the number of scientific studies being submitted from BDSM club members compared to non BDSM club members has appeared since the new rules were invoked. Studies are being conducted by the dom-sub curious to see why this particular club membership has had such a steep rise in scientific studies being produced.

  • by greytree ( 7124971 ) on Wednesday January 29, 2025 @10:11AM (#65127817)
    Or "Social Studies Departments" as we usually call them.
  • It would be interesting if one or more of these fraudulent studies by chance turned out to contain a major breakthrough in the field it was written about.

    • With infinite monkeys on typewriters you get Hamlet sooner or later I suppose. Part of the problem will be that there isn't much financial incentive for anyone to try to confirm the results of basic research.
      • With my infinites, sure. However, a study was done on the Infinite Monkey theory if given practical constraints.

        It turns out the observable universe is too finite to hold enough monkeys to get the job done within the universe's remaining lifetime.

        So that is a bit of a bummer...

      • by Tablizer ( 95088 )

        With infinite monkeys on typewriters you get Hamlet sooner or later

        Or a babbling orangutan who wins the Presidency via mass Gish Galloping.

  • even if the scholar/investigator/researcher doesn't have any thing unique to report. For-profit publishers (like Elsevier for example) with their paywalls add to the problem. The ease of setting up websites that will publish anything are problematic as well. Maybe in time AI can help strain all that output into some sort of useable form.
  • by methano ( 519830 ) on Wednesday January 29, 2025 @11:31AM (#65128059)
    It's not the fake stuff that bothers me so much. It's the quantity of real stuff of such low quality. If you get rid of the fake stuff, you're about 2% of the way toward fixing the problem.
  • Before relying on a research paper for anything important, wouldn't you first check with the source, or at least with a reputable verification agency?

    If they just use Google to find papers without vetting, they are Grade A Morons.

    If they are mostly used by trolls looking to bolster their case with fake studies, then there is probably no profit for the fake study maker. Unless their agenda is propaganda to give trolls scientific "legitimacy", such as "Transgender people are 40% more likely to eat pets".

    In ot

    • by Idzy ( 1549809 )
      did you read the other comments above yours? This is happening because we incentivize quantity over quality. The incentive being: you get more grant money if you publish lots of papers and don't get grants if you only publish a few quality papers that actually contribute to human knowledge
  • Did they already get the 100,000,000,000 by the tobacco industry? Btw I know of 3 separate instances of paper companies having to pay part of the bill to clean up PCBs they dumped in the river. It was pissing off the locals at one city to see non-english speakers fishing there and not knowing the limit was stated as "0-1 per year" for your own exposure risk.
  • Apparently you do not know this, but:
    A Paper Mill is a manufacturing plant that physically makes paper comma you cloying headline-twisting jerk
    A Paper Mill takes tree material and outputs blank physical paper for printing onto
    A Paper Mill smells bad and all the cars in the parking lot are covered in a fine but harmless dust from the process
    A Paper Mill is not a place where papers are written. Please choose less stupid, overlapping and confusing terms to make up and win likes

  • -- Faked data, manipulated data, but officially peer reviewed (even if review process is questionable): NOT FAKE
    -- Real data, faked peer review: FAKE
  • So now it has become democratized and everybody can have fake studies!

It isn't easy being the parent of a six-year-old. However, it's a pretty small price to pay for having somebody around the house who understands computers.

Working...