ChatGPT generates cancer treatment plans that are full of errors — Study finds that ChatGPT provided false information when asked to design cancer treatment plans::Researchers at Brigham and Women’s Hospital found that cancer treatment plans generated by OpenAI’s revolutionary chatbot were full of errors.

  • nfsu2@feddit.cl
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 months ago

    true, I tried to explain this to my parents because they were scared of it and they seemed skeptical.

  • EssentialCoffee@midwest.social
    link
    fedilink
    English
    arrow-up
    0
    arrow-down
    1
    ·
    10 months ago

    I use it for D&D. It’s fantastic at coming up with adventures, NPCs, story hooks, taverns, etc.

    All of those things are made up.

  • zeppo@lemmy.world
    link
    fedilink
    English
    arrow-up
    58
    ·
    11 months ago

    I’m still confused that people don’t realize this. It’s not an oracle. It’s a program that generates sentences word by word based on statistical analysis, with no concept of fact checking. It’s even worse that someone actually did a study instead of simply acknowledging or realizing that ChatGPT is happy to just make stuff up.

    • net00@lemm.ee
      link
      fedilink
      English
      arrow-up
      6
      ·
      11 months ago

      Yeah this stuff was always marketed to automate simple and repetitive things we do daily. it’s mostly the media I guess who started misleading everyone into thinking this was AI like skynet. It’s still useful, not just as a all knowing AI god

    • fubo@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      5
      ·
      11 months ago

      It’s even worse that someone actually did a study instead of simply acknowledging or realizing that ChatGPT is happy to just make stuff up.

      Sure, the world should just trust preconceptions instead of doing science to check our beliefs. That worked great for tens of thousands of years of prehistory.

      • zeppo@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        edit-2
        11 months ago

        It’s not merely a preconception. It’s a rather obvious and well-known limitation of these systems. What I am decrying is that some people, from apparent ignorance, think things like “ChatGPT can give a reliable cancer treatment plan!” or “here, I’ll have it write a legal brief and not even check it for accuracy”. But sure, I agree with you, minus the needless sarcasm. It’s useful to prove or disprove even absurd hypotheses. And clearly people need to be definitely told that ChatGPT is not always factual, so hopefully this helps.

      • yiliu@informis.land
        link
        fedilink
        English
        arrow-up
        3
        ·
        11 months ago

        “After an extensive three-year study, I have discovered that touching a hot element with one’s bare hand does, in fact, hurt.”

        “That seems like it was unnecessary…”

        “Do U even science bro?!”

        Not everything automatically deserves a study. Were there any non-rando people out there claiming that ChatGPT could totally generate legit cancer treatment plans that people could then follow?

    • dual_sport_dork@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      11 months ago

      This is why without some hitherto unknown or so far undeveloped capability of these sorts of AI models, they’ll never actually be useful for performing any kind of mission critical work. The catch-22 is this: You can’t trust the AI to produce correct work without some kind of potentially dangerous, showstopping, or embarassing error. This isn’t a problem if you’re just, say, having it paint pictures. Or maybe even helping you twiddle the CSS on your web site. If there is a failure here, no one dies.

      But what if your application is critical to life or safety? Like prescribing medical care, or designing a building that won’t fall down, or deciding which building the drone should bomb. Well, you have to get a trained or accredited professional in whatever field we’re talking about to check all of its work. And how much effort does that entail? As it turns out, pretty much exactly as much as having said trained or accredited professional do the work in the first place.

    • PreviouslyAmused@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      But it’s supposed to be the future! I want the full talking spaceship like in Star Trek, not this … “initial learning steps” BS!

      I was raised on SciFi and am now mad that I don’t have all the cool made up things from those shows/movies!

    • iforgotmyinstance@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      I know university professors struggling with this concept. They are so convinced using an LLM is plagiarism.

      It can lead to plagiarism if you use it poorly, which is why you control the information you feed it. Then proofread and edit.

      • ZodiacSF1969@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        10 months ago

        I can understand the plagiarism argument, though you have to extend the definition of it. If I am expected to write an essay, but I use ChatGPT instead, then I am fraudulently presenting the work as my own. Plagiarism might not be the right word, or maybe it’s a case where language is going to evolve so that plagiarism includes passing off AI generated work as your own. Either way it’s cheating unless I was specifically allowed to use AI.

        • iforgotmyinstance@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          10 months ago

          If the argument and the sources are incongruous, that isn’t the fault of the LLM/AI. That’s the authors fault for not proofreading and editing.

          You assume an inherent morality of LLMs but they are amoral constructs. They are tools, and you limit yourself by not learning them.

          • ZodiacSF1969@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            0
            arrow-down
            1
            ·
            10 months ago

            I didn’t say anything about the sources being incongruent? That’s a completely separate issue. We were talking about plagiarism.

            I don’t understand the morality comment either, I didn’t ascribe any morality to AI, I was talking about whether using them fits the definition of plagiarism or not.

            If you are expected to write it yourself, and you use an LLM to generate it, then that’s cheating in my opinion. Yes, of course we shoukd learn to use AI, but if you are told to do something and you get a person or LLM to do it for you, then you didn’t complete the task as you were told. And at university that can have consequences.

  • Obinice@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    ·
    10 months ago

    Well, it’s a good thing absolutely no clinician is using it to figure out how to treat their patient’s cancer… then?

    I imagine it also struggles when asked to go to the kitchen and make a cup of tea. Thankfully, nobody asks this, because it’s outside of the scope of the application.

    • clutch@lemmy.ml
      link
      fedilink
      English
      arrow-up
      4
      ·
      10 months ago

      The fear is that hospital administrators equipped with their MBA degrees will think about using it to replace expensive, experienced physicians and diagnosticians

      • whoisearth@lemmy.ca
        link
        fedilink
        English
        arrow-up
        5
        ·
        10 months ago

        They’ve been trying this shit for decades already with established AI like Big Blue. This isn’t a new pattern. Those in charge need to keep driving costs down and profit up.

        Race to the bottom.

      • Obinice@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 months ago

        If that were legal, I’d absolutely be worried, you make a good point.

        Even Doctor need special additional qualifications to do things like diagnose illnesses via radiographic imagery, etc. Specialised AI is making good progress in aiding these sorts of things, but a generalised and very poor AI like ChatGPT will never be legally certified to do this sort of thing.

        Once we have a much more effective generalised AI, things will get more interesting. It’ll have to prove itself thoroughly though, before being certified, so it’ll still be a few years after it appears before we see it used in clinical applications.

    • 5BC2E7@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      10 months ago

      Because it’s been hyped. They had announced it could pass the medical licensing exam with good scores. The belief that it can replace a doctor has already been put forward

  • Rexios@lemm.ee
    link
    fedilink
    English
    arrow-up
    9
    ·
    10 months ago

    Okay and? GPT lies how is this news every other day? Lazy ass journalists.

  • NigelFrobisher@aussie.zone
    link
    fedilink
    English
    arrow-up
    8
    ·
    10 months ago

    People really need to understand what LLMs are, and also what they are not. None of the messianic hype or even use of the term “AI” helps with this, and most of the ridiculous claims made in the space make me expect Peter Molyneux to be involved somehow.

  • Sanctus@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    ·
    11 months ago

    These studies are for the people out there who think ChatGPT thinks. Its a really good email assistant, and it can even get basic programming questions right if you are detailed with your prompt. Now everyone stop trying to make this thing like Finn’s mom in adventure time and just use it to helo you write a long email in a few seconds. Jfc.

    • lonke@feddit.nu
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      11 months ago

      I use ChatGPT primarily for programming, and it’s particularly well suited for programming.

      “Even get basic programming questions right if you are detailed with your prompt”

      is underselling its capabilities in that regard. Especially GPT-4 has been able to help me with everything from obscure adobe ExtendScript scripts to infrequently seen ‘unsafe’ C# OpenGL perspective matrix math. All with prompts of a sentence maximum.

    • Corngood@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      11 months ago

      I’m going to need it to turn those emails back into the bullet points used to create them, so I don’t have to read the filler.

  • SirGolan@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    2
    ·
    11 months ago

    What’s with all the hit jobs on ChatGPT?

    Prompts were input to the GPT-3.5-turbo-0301 model via the ChatGPT (OpenAI) interface.

    This is the second paper I’ve seen recently to complain ChatGPT is crap and be using GPT3.5. There is a world of difference between 3.5 and 4. Unfortunately news sites aren’t savvy enough to pick up on that and just run with “ChatGPT sucks!” Also it’s not even ChatGPT if they’re using that model. The paper is wrong (or it’s old) because there’s no way to use that model in the ChatGPT interface. I don’t think there ever was either. It was probably ChatGPT 0301 or something which is (afaik) slightly different.

    Anyway, tldr, paper is similar to “I tried running Diablo 4 on my Windows 95 computer and it didn’t work. Surprised Pikachu!”

    • eggymachus@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      3
      ·
      11 months ago

      And this tech community is being weirdly luddite over it as well, saying stuff like “it’s only a bunch of statistics predicting what’s best to say next”. Guess what, so are you, sunshine.

      • PreviouslyAmused@lemmy.ml
        link
        fedilink
        English
        arrow-up
        3
        ·
        11 months ago

        I mean, people are slightly more complicated than that. But sure, at their most basic, people simply communicate with statistical models.

      • amki@feddit.de
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 months ago

        Might be true for you but most people do have a concept of true and false and don’t just dream up stuff to say.

      • dukk@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        11 months ago

        IMO for AI to reach a useful point it needs to be able to learn. Now I’m no expert on neural networks, but if it can’t learn anything new once it’s been trained, it’s never really going to reach its true potential. It can imitate a human, but that’s about it. Once AI can really learn, it’ll become an order of magnitude more useful. Don’t get me wrong: all this AI work is a step in the right direction, but we’ll only be able to go so far with pre-trained models.

  • •••@lemmy.ml
    link
    fedilink
    English
    arrow-up
    5
    ·
    10 months ago

    I suppose most sensible people already know that ChatGPT is not the answer for medical diagnosis.

    Prompts were input to the GPT-3.5-turbo-0301 model via the ChatGPT (OpenAI) interface.

    If the researcher wanted to investigate whether LLM is helpful, they should develop a model specifically using cancer treatment plans with GPT-4/3.5 before testing it thoroughly, in addition to entering prompts into the model that is available on OpenAI.

    • wewbull@feddit.uk
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      There have been a number of articles about how GPT has been out-diagnosing doctors in various domains. To me, that isn’t that surprising as diagnosis is a pattern matching problem, something a neuralnet will be very good at. Human doctors were seen to be discounting rare conditions just because they were rare and so “it was much more likely to be something else” even if the symptoms backed up the conclusion. A computer can be more objective about such things.

      …but none of that needs AI/ML. We’ve had expert systems since the 60s.

      It’s also very different from constructing a treatment plan, which is what we’re discussing here.

    • ours@lemmy.film
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      Or they could feed the current model with a reputable source of medical information.