Researchers tested different medical scenarios with the chatbot. In more than half of cases in which doctors would send patients to the ER, the chatbot said it was OK to delay care.

ChatGPT Health — OpenAI’s new health-focused chatbot — frequently underestimated the severity of medical emergencies, according to a study published last week in the journal Nature Medicine.

In the study, researchers tested ChatGPT Health’s ability to triage, or assess the severity of, medical cases based on real-life scenarios.

Previous research has shown that ChatGPT can pass medical exams, and nearly two-thirds of physicians reported using some form of AI in 2024. But other research has shown that chatbots, including ChatGPT, don’t provide reliable medical advice.

  • jacksilver@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    ·
    8 days ago

    You’re being down voted because LLMs aren’t deterministic, it’s basically the biggest issue in productizing them. LLMs have a setting called “temperature” that is used to randomize the next token selection process meaning LLMs are inherently not deterministic.

    If you se the temperature to 0, then it will produce consistent results, but the “quality” of output drops significantly.

    • Kairos@lemmy.today
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      5
      ·
      8 days ago

      If you give whatever random data source it uses the same seed, it will output the same thing.