Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
AI Google Medicine

Google Is Testing Its Medical AI Chatbot At the Mayo Clinic 38

According to the Wall Street Journal, Google is testing its Med-PaLM 2 AI chat technology at the Mayo Clinic and other hospitals. It's based on the company's PaLM 2 large language model (LLM) that underpins Bard, Google's ChatGPT rival. Engadget reports: Unlike the base model, Med-PaLM-2 has been trained on questions and answer from medical licensing exams, along with a curated set of medical expert demonstrations. That gives it expertise in answering health-related questions, and it can also do labor-intensive tasks like summarizing documents and organizing research data, according to the report. During I/O, Google released (PDF) a paper detailing its work on Med-PaLM2. On the positive side, it demonstrated features like "alignment with medical consensus," reasoning ability, and even the ability to generate answers that were preferred by respondents over physician-generated responses. More negatively, it showed the same accuracy problems we've seen on other Chat AI models.

In an internal email seen by the WSJ, Google said it believes the updated model could "be of tremendous value in countries that have more limited access to doctors." Still, Google has admitted that the technology is still in its early stages. "I don't feel that this kind of technology is yet at a place where I would want it in my family's healthcare journey," said Google senior research director Greg Corrado. However, he added that the tech "takes the places in healthcare where AI can be beneficial and expands them by 10-fold."
This discussion has been archived. No new comments can be posted.

Google Is Testing Its Medical AI Chatbot At the Mayo Clinic

Comments Filter:
  • I can provide numerous examples of AI providing totally incorrect or slightly wrong information -- even in critical areas such as medicine. (In this example chatGPT, but I bet it could be applicable to Google's chatbot.) I'm not gonna list all the examples here, just one randomly. Here's the scenario: Let's assume you have a doctor that has grown to rely on chatGPT. Now what if he gets a patient with syphilus and barely sort of recalls that he can treat syphilus with Penicillin G, but he forgets which type

    • First, they're not using ChatGPT, this is specially-trained AI using data sources much more reliable than "the internet."

      Second, humans also often provide "totally incorrect or slightly wrong" information. My bet is that, with enough iterations of training, AI will do so much less often than humans.

      • Parroting pieces of text that statistically might have come from a medical textbook is a totally shit way of practicing medicine. I'm OK with an expert system that has rules that encode medical knowledge, and can provide probabilities for diagnoses, can explain what and *why* certain things might be, and can suggest tests that would distinguish between them. That does NOT seem to be where these systems are going.
        • by PCM2 ( 4486 )

          Parroting pieces of text that statistically might have come from a medical textbook is a totally shit way of practicing medicine.

          And yet this is pretty much how medicine is practiced. How else? I don't see a whole lot of difference between this and your "alternate" method.

          • Precisely this. Actually, there is one difference. AI can spot trends that humans haven't actually included in their prescriptive texts. To me, this makes AI a better bet than just textbook knowledge.

        • Parroting pieces of text that statistically might have come from a medical textbook is a totally shit way of practicing medicine.

          That's your own statement, not backed up by anything in the article or by any source anywhere.

          And I don't really see a distinction between your "expert systems" that are hand-coded, and predictive models, that use pattern analysis. Well, there is one difference: AI can find patterns that weren't necessarily spotted by humans.

          Whether the doctor consults a textbook or manual, or AI, the doctor needs to know what they are doing. This is true with coding too. I use ChatGPT regularly to help me find ways to code

      • by narcc ( 412956 )

        this is specially-trained AI using data sources much more reliable than "the internet."

        It's based on Google's PaLM 2, which absolutely was trained on massive amounts of data from the internet.

        Second, humans also often provide "totally incorrect or slightly wrong" information.

        Humans are capable of reviewing their work to identify and correct errors. Language models are not.

        My bet is that, with enough iterations of training, AI will do so much less often than humans.

        That is a bet you will lose. Over training will degrade model performance. This isn't a problem that can be fixed. The nature of the technology makes this inevitable.

        • by dvice ( 6309704 )

          > It's based on Google's PaLM 2, which absolutely was trained on massive amounts of data from the internet.

          PALM2 is a combination of multiple AIs like Gato that has been previously been told to be possible solution to AGI. Major difference to for example ChatGPT is that you can train an individual AI to do a specific task and then attach it to the system. Much like you could give a human a general education, then attach a specific education to inspect X-ray images. So it is not an LLM. It is LLM + other

          • by narcc ( 412956 )

            that has been previously been told to be possible solution to AGI.

            "told"? By whom? What a ridiculous thing to say.

            This is why some people already think that AGI is solved

            Some people think the earth is flat. The world is full of stupid people.

            AGI is science fiction. We are so far from a solution that we don't even know what questions to ask. Anyone making any claims about AGI is either a crackpot or a conman.

        • It's based on Google's PaLM 2, which absolutely was trained on massive amounts of data from the internet.

          PaLM 2 is a technology, not a training method. The training is separate. And I challenge you to show me a source that says Google's Med PaLM 2 was trained randomly from "the internet."

          Humans are capable of reviewing their work to identify and correct errors. Language models are not.

          Humans train language models, they don't train themselves.

          If you use AI as a tool, you don't get to turn your brain off. I use ChatGPT for coding ideas. But when I get a suggestion, I still have to know whether it does what I want it to do, or if it's a viable solution. Likewise in medicine, the doctor still has to know what t

          • by narcc ( 412956 )

            PaLM 2 is a technology, not a training method.

            I didn't say it was, troll. Why tell such an obvious lie?

            I challenge you to show me a source that says Google's Med PaLM 2 was trained randomly from "the internet."

            I didn't say "randomly" you lying troll. As for a source than claims they used data from the internet, that's trivial [ai.google] If you don't know anything about the subject, you shouldn't post.

            Humans train language models, they don't train themselves.

            I never claimed that they did, you lying piece of shit troll.

            Don't waste my time with bullshit like this.

            • You are conflating PaLM 2 with PaLM 2 Med. Your link does not mention PaLM 2 Med at all. The two are based on the same technology, but different training sets.

              Clearly, you have run out of actual arguments, because you have started resorting to insults.

      • by nuntius ( 92696 ) on Monday July 10, 2023 @11:20PM (#63676163)

        Many years ago, some classical AI expert systems claimed to outperform doctors in real-world tests. Never gained much traction. Explanations varied.

        Personally, I am all for encoding medical knowledge in a machine-readable, open format knowledge base. I might even support requiring every journal article to provide input to it. Computers are designed for retrieving information and applying rules such as formal logic and Bayesian inference to filter and aggregate results. People are notorious for memory or logic lapses, especially under fatigue and other stress.

        I can imagine an LLM-frontend might be vastly more doctor friendly than writing formal predicates for input or interpreting raw numbers for output. It could also provide a natural prompting system that recommends tests and warns of rare but serious alternative diagnoses.

        Present AI tools can also look at data to suggest correlations that may merit research into improved treatment protocols.

        I am strongly opposed to the idea of present-gen AI being used to supplant a formal knowledge base by "somehow" learning "better" than expert humans through unsupervised processing of existing texts and data. Such hubris seems endemic to AI hype.

        • I am strongly opposed to the idea of present-gen AI being used to supplant a formal knowledge base

          This is like the relationship between Stack Overflow, and ChatGPT. You can't have ChatGPT spit out answers to a lot of technical questions, without it first consulting the Stack Overflow "knowledge base." They work together. ChatGPT doesn't replace the knowledge base, it just digests it.

      • "Second, humans also often provide "totally incorrect or slightly wrong" information" I was really diagnosed with genital herpes, and was prescribed antivirals and spent two weeks depressed as I think I've even been considering I'd have to tell any potential partners that I have an easily transmitted incurable virus, so basically no more sex. I go in to the doctor between a week and two weeks later and complain that the antivirals aren't doing a damn thing. "Let me see again" ... "Oh yeah and the test res
    • Asking "What is the difference between benzathine penicillin g and procaine penicillin g?" was the wrong question in this case. It might have been appropriate to ask "what would be the best implementation of penicillin for treating syphilis". If the answer to that was wrong the machine would more clearly be at fault.

      And a smart answer would have been to ask for more detail about the specific form of syphilis or other relevant details just like a human would. I expect the machines can do that.

  • Future Anomalous & Catastrophic Events (caused by) Physician Assistant Language Model

  • by crunchy_one ( 1047426 ) on Monday July 10, 2023 @09:26PM (#63675967)
    Can we please experiment with something with less risk of lethal outcomes? Grant writing might be more suitable as no one actually reads the damned things.
  • I think human doctors are often arrogant or poorly trained. They work shifts that are too long, and are often overly motivated to opt for more treatments, and more expensive treatments, because they get paid for each procedure.

    Note that this is not ChatGPT, which was trained on "the internet." This is AI trained on specially curated data sets. I would expect much higher quality than ChatGPT, and with enough iterations and training, I'd literally bet my life that it will have a better track record than human doctors.

    • [...]and are often overly motivated to opt for more treatments, and more expensive treatments, because they get paid for each procedure.

      • And what I intended to say was: You must be from a country with a broken healthcare system. Doctors should be paid a salary, not per procedure.
        • The American Healthcare System (if it can be called a system at all) is all about profit, and pushing expensive procedures, or even better, expensive forever treatments like continual, perpetual drugs, carefully balanced with other drugs to knock back the side effects, carefully balanced with still more drugs to knock back those side effects, and so on and so on. We even got clockers watching the doctors and nurse practitioners (since the clockers said doctors spent too much time with patients) to make sure

          • As Winston Churchill famously said, "Democracy is the worst possible form of government that anyone could imagine, except for all the others."

            Healthcare is kind of like that. Yes, American healthcare is expensive. But it's also effective. If you need treatment, you don't have to wait for months, as you do in many countries where the government pays for everything.

            It's also not totally one-sided for the "owner class" as you say. Providers still have to work within the bounds of what insurance companies will

            • As Winston Churchill famously said, "Democracy is the worst possible form of government that anyone could imagine, except for all the others."

              Healthcare is kind of like that. Yes, American healthcare is expensive. But it's also effective. If you need treatment, you don't have to wait for months, as you do in many countries where the government pays for everything.

              It's only effective if you can afford it. And that waiting months thing, if you actually research it, doesn't appear to happen in the upper-tier "government pays for everything" countries any more than it does here in America. Especially these last two to three years, where everything seems to take for freakin' ever, if you can dare to pay.

              It's also not totally one-sided for the "owner class" as you say. Providers still have to work within the bounds of what insurance companies will pay, and that in turn is limited to what employees and employers are willing to pay. It's not perfect, and it is expensive, and there are losers in the system. But in government-run healthcare systems, there are also losers, particularly those who can't afford to pay for private care if they want it. Money pays for nice things, and no amount of regulation can change that.

              Yeah, the insurance companies are part of the owner class in America. They are massively profitable companies in their own right.

              And while you can brag about how great it

        • Yes, this is true. But in countries where doctors are paid only a salary, they are still motivated by factors not related to the quality of patient care. Even government-sponsored healthcare systems want to control costs. So they might be motivated NOT to perform procedures they really should, in the name of cost savings.

          There are certainly problems with the American healthcare system, but Europeans haven't solved all their healthcare problems either.

    • by wrsi ( 10382619 )

      Let me suggest that you perhaps investigate the range of different payment models present in medical settings wherever you live. Your comment strongly suggests that you're only aware of the "Fee For Services" payment model.

      The array of vectors of force acting on medical interactions between patients and health care providers is bewildering. To use a classic example in the US, there are a ton of factors that go into the high rate of cesarean sections that occur here. To attribute huge, systemic problems to t

      • In the government-funded healthcare model, which I presume you live with, there are a different set of problems. Namely, the government wants to control costs, so it is more likely to deny options or treatments that should be considered.

        Yes, the American system has problems, but every system has problems.

  • I'd like to keep my appendix, tonsils, uvula, pancreas and testicles. The last thing I need is an AI making weird predictions. Oh, I'd also like to keep my brain too, although its utility is somewhat dubious these days.
  • Endeavors like this all too often start with "the company" or the technology dudes having a ghee whiz idea divorced from the reality of the primary matter, in this case doctoring and health care. So, of course they are going to talk it up like the greatest thing ever. It's not.

    The old adage - If you think you know the answer, you don't understand the problem.

    Some things that struck me as funny, or too stupid or obvious to be taken seriously :

    Unlike the base model, Med-PaLM-2 has been trained on questions and answer from medical licensing exams, along with a curated set of medical expert demonstrations. That gives it expertise in answering health-related questions

    These AI bots are supposed to be trained on large language model

  • Chatbots are good for answering questions from idiots. Questions like "I have a sore throat, does that mean I have cancer in my foot."

    Nontrivial questions will fail to be answered either correctly or at all, but your access to a human who can actually answer it will be severely restricted.

Think of it! With VLSI we can pack 100 ENIACs in 1 sq. cm.!

Working...