Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror
Science

Springer Nature Book on Machine Learning is Full of Made-Up Citations (retractionwatch.com) 53

Springer Nature published a $169 machine learning textbook in April containing citations that appear to be largely fabricated, according to an investigation by Retraction Watch. The site checked 18 of the 46 citations in "Mastering Machine Learning: From Basics to Advanced" by Govindakumar Madhavan and found two-thirds either did not exist or contained substantial errors.

Three researchers contacted by Retraction Watch confirmed their supposedly authored works were fake or incorrectly cited. Yehuda Dar of Ben-Gurion University said a paper cited as appearing in IEEE Signal Processing Magazine was actually an unpublished arXiv preprint. Aaron Courville of Universite de Montreal confirmed he was cited for sections of his "Deep Learning" book that "doesn't seem to exist."

The pattern of nonexistent citations matches known hallmarks of large language model-generated text. Madhavan did not answer whether he used AI to generate the book's content. The book contains no AI disclosure despite Springer Nature policies requiring authors to declare AI use beyond basic copy editing.

Springer Nature Book on Machine Learning is Full of Made-Up Citations

Comments Filter:
  • by gweihir ( 88907 ) on Monday July 07, 2025 @10:54AM (#65502952)

    The fanbois are blind to it, the rest are shaking their heads in disgust. And the people like this "author" are essentially scamming their readers.

    • The publisher’s lack of review and missing due diligence for the price is the real news here. Another one using ChatGPT to write book is nothing new.
      • Another one using ChatGPT to write book is nothing new.

        I would say it is "the new normal." Including the bit where AI use is not disclosed even though disclosure is mandatory.

      • by gweihir ( 88907 )

        Most publishers have been doing nothing for the object published and essentially just skimming money for ages.

      • Springer doing jack shit isn't exactly new either. Aside from taking money from other people work (or in this case, nobody's work).
        • by gweihir ( 88907 )

          Yep, same experience here when I published my PhD. These people are utterly disgusting.

      • The Springer name (and price tag) used to carry weight. Not so much any more.

    • by evanh ( 627108 )

      And Springer too of course. They'll turn a blind eye any chance they get.

      • by gweihir ( 88907 )

        When I published my PhD in 2008, Springer was an option. After looking at them, I was so disgusted that I went to considerable effort to find a small publisher that let me keep the online copyright of my thesis and only took the paper-rights for two years. Oh, and Springer and the like do absolutely nothing on the content-side for you. They just print and sell, everything else is on you.

    • ... the media and general public are believing all the marketing puff and outright lies that the AI industry is churning out such as AGI is just around the corner along with full self driving cars etc etc. Anyone with a clue knows this is total BS but the pump and dump continues.

      • by gweihir ( 88907 )

        Indeed. The general public is not smart and easy to manipulate. This is my core learning for this decade, before I was naïve (or maybe sheltered, if you can be sheltered in your 40's) and thought the average person was somewhat smart and had some insight into how things work. Not anymore.

        As to full self-driving, we have SAE 4 now, with geographic limitations. That is after about 60 years or so of steady research. Give it another 20 or 30 years and SAE 5 may become a reality, probably still with some li

        • This is my core learning for this decade [...]

          As an aside, I find it interesting that the noun "learning" is being used a lot these days. I see it all over the place now, substituted where previously the noun "lesson" would normally be used. It would be interesting to trace the etymology in current usage. Is it a reaction to the fake "learning" of AI (iterative convergence claims without real convergence)? Do people want to re-appropriate the word for themselves? Do people want to signify that they can "l

          • by pjt33 ( 739471 )

            People were complaining about it being a corporate usage [upenn.edu] 15 years ago, so it's nothing to do with LLMs or TikTok.

            • Nice link! I think it's reasonable to make a distinction between the pre-modern usage, eg 18th century, and the recently popular revival. At least it's quite clear from the comments that learnings had almost died out: I conclude that I'm not losing my mind :)
          • by gweihir ( 88907 )

            My take is the difference is that lesson is more externally provided/imposed, while learning is more of an internal (and voluntary) process.

    • A LLM stole your woman, too. This is starting to sound like a country & western song.
  • by Pseudonymous Powers ( 4097097 ) on Monday July 07, 2025 @11:03AM (#65502974)
    I don't understand why fans of LLMs don't simply use a second LLM to fact-check the output of the first one. Though I suppose for that to work the second LLM would need to formally recognize that some sources of truth are better than others, which would strike a killing blow to the heart of the LLM ethos. And then the first LLM would need to rewrite its original draft based on the editorial input of the second one, which would undercut its unmerited bloviating confidence, which would strike a second killing blow to the heart of the LLM ethos.
    • by DarkOx ( 621550 ) on Monday July 07, 2025 @11:12AM (#65502996) Journal

      People absolutely are doing stuff like that. They are even putting LLMs in front of other LLMs to act as WAF-like firewall solutions and such.

      The problem is it is all very compute and memory intensive. I do see some people getting good results tying multiple models together via MCP and other interop solutions, in terms of outputs. The problem of course is you get an application that is painfully slow to use and to expensive to run.

      Its funny the big tech people use to talk about doing things more efficiently so they don't 'boil the ocean' now its all what problems can we throw LLMs at, no matter how inefficient that might be from a required compute perspective and now its how fast can we put up datacenters and restart old power plants ... seemingly with little thought at all when it comes to boiling oceans... Until tax season anyway where they invent some scam accounting game to show how carbon neutral they are..

    • I don't understand why fans of LLMs don't simply use a second LLM to fact-check the output of the first one. Though I suppose for that to work the second LLM would need to formally recognize that some sources of truth are better than others, which would strike a killing blow to the heart of the LLM ethos. And then the first LLM would need to rewrite its original draft based on the editorial input of the second one, which would undercut its unmerited bloviating confidence, which would strike a second killing blow to the heart of the LLM ethos.

      It's far simpler than that. Some of these references just don't exist. Just ask a summer intern or high school student to write a simple script to check for the existence of these references. Of course, since this appears to be so challenging to do, maybe someone can form a startup to address this problem and earn billions of dollars.

      In real-life editing, there are human editors that just check for grammar, formatting, etc. Then there are editors that check for consistency, legal issues, being on messag

    • You would need a 3rd LLM to check the work of the second, right? And a 4th to check the 3rd? So, it's LLM's all the way down, right?
      • No, you don't need a 3rd LLM (but if you find some use for it, go ahead). All you need is some test data. The 2nd LLM is quality assurance for the first LLM. Only the first LLM output is published, so that's the only output you need to check.
      • This is the Maxwell Smart approach.

        Chief: I don't know, Larabee. Maybe I should have gone with Max's plan.
        Larabee: What plan was that, Chief?
        Chief: The one where we use 99 Control agents, so that if 98 of them get taken out...

    • Because that obviously does not work.

      You have to fact check yourself or ask one who trustworthy does it

  • by shm ( 235766 ) on Monday July 07, 2025 @11:05AM (#65502976)

    His bio is fascinating.

    BTech in Chemical Engineering.

    Master level practitioner of Neuro Linguistic Programming.

    Never having heard of Neuro Linguistic Programming before, I looked it up. Wikipedia calls it a pseudoscience.

    But the AI Ethics expertise takes the cake.

    • Neuro Linguistic Programming, sounds like a Pascal derivative.
    • I too use my brain and my language skills to program. Master level practitioning!
    • ... Neuro Linguistic Programming ...

      You never heard of it: It was the new, wizz-bang communication tool about 10 years ago. It's a fancy way of saying, word association: That and repetition are how we remember stuff. Whether word association results in a Pavlovian response is, obviously, fuzzy. The fakeness comes from assuming that the word association I make, causes a Pavlovian response in someone else. A bit like assuming my swelling penis causes panties to drop: We all want easy answers and NLP promises them.

  • Fuck academic publishing.
  • Par for the course (Score:4, Insightful)

    by Mr. Dollar Ton ( 5495648 ) on Monday July 07, 2025 @11:21AM (#65503016)

    The whole AI is shit, why shouldn't the "books" on it be shit too.

    • I gave up my fiction writing career about 4 years ago since everyone I met seemed to be so excited they could now 'write a book' using AI. Screw it, I thought. You want to pump out that crap, you can read that crap.
      • Yeah, right around the time I released my first "real" book, it became impossible to get a legitimate review because the AI books were starting to flood the market and reviewers you used to be able to send free copies to for a fair assessment were closing down their intake because they were being absolutely swamped with books so shitty they couldn't bring themselves to finish them. Not sure how one goes about building a base with no review channels available, I've since decided I'll just finish the books I

  • by Viol8 ( 599362 ) on Monday July 07, 2025 @11:55AM (#65503094) Homepage

    ... the trump administration has just stopped all subscriptions to Springer Nature publications citing junk science in their journals (hopefully not Nature itself) which seems to have got a lot of people hot under the collar.

    I'm no trump fan but Trump et al have long been compared to a broken clock - occasionally correct if you wait long enough. Maybe this is one of those times.

  • This is RFK Jr's DREAM! Finally he can use AI as a scapegoat for his false beliefs.
  • by greytree ( 7124971 ) on Monday July 07, 2025 @12:10PM (#65503132)
    I received "Mastering Machine Learning: From Basics to Advanced" by Govindakumar Madhavan for review and I must say I thoroughly enjoyed it and would recommend it to anyone interested in Machine Learning.

      -- Winston Churchill
    • "Aaron Courville of Universite de Montreal confirmed he was cited for sections of his "Deep Learning" book that "doesn't seem to exist.""

      Oh come on Aaron, it's not like someone can possibly be aware of *all* the books they have written.
    • "Rated 5 out of 5 stars by President Abraham Lincoln, King Tut, and Charlamagne! Read the book that helped build the empire of Alexander the Great!"
  • This is kind of funny, but if you assume that this book was written by AI, which is being used to train humans on it's hallucinations to program / improve AI, this almost becomes a link to singularity, with humans brains as part of the neural net.

    • This is kind of funny, but if you assume that this book was written by AI, which is being used to train humans on it's hallucinations to program / improve AI, this almost becomes a link to singularity, with humans brains as part of the neural net.

      That would be a hive mind, not the singularity, because at no point will it become effectively infinitely intelligent. Instead, the stupidities multiply. It's literally the opposite of a singularity.

  • He ordered minions to write it and he probably didn't even bother to read it before sending it to the publisher. My bet is he sketched an outline and set his minions to work. That's the way it was when I was in academia and why many chapters in books like these seem so disconnected -- each chapter was is assigned to a different person -- none of whom show up as authors.
  • LLMs are terrible at math. So when you ask Copilot a math question, it relies on a separate, math-focused AI to solve the problem. https://www.microsoft.com/en-u... [microsoft.com] If it didn't do this, it would have a really hard time solving math problems, as an LLM.

    The same pattern applies to source citations. The LLM generates text that notes citations at plausible locations within a document, but then doesn't link them to anything meaningful. Perhaps a separate AI, such as a web search engine, could locate the actual

  • It's a text book about machine learning and LLM's. it's teaching you about made up shit that has the appearance of being legit.

    It's what LLMs do best. Make random crap look like real crap.

People who go to conferences are the ones who shouldn't.

Working...