this post was submitted on 03 Mar 2026
100 points (98.1% liked)

Health - Resources and discussion for everything health-related

4260 readers
178 users here now

Health: physical and mental, individual and public.

Discussions, issues, resources, news, everything.

See the pinned post for a long list of other communities dedicated to health or specific diagnoses. The list is continuously updated.

Nothing here shall be taken as medical or any other kind of professional advice.

Commercial advertising is considered spam and not allowed. If you're not sure, contact mods to ask beforehand.

Linked videos without original description context by OP to initiate healthy, constructive discussions will be removed.

Regular rules of lemmy.world apply. Be civil.

founded 2 years ago
MODERATORS
 

Researchers tested different medical scenarios with the chatbot. In more than half of cases in which doctors would send patients to the ER, the chatbot said it was OK to delay care.

ChatGPT Health


OpenAI's new health-focused chatbot


frequently underestimated the severity of medical emergencies, according to a study published last week in the journal Nature Medicine.

In the study, researchers tested ChatGPT Health's ability to triage, or assess the severity of, medical cases based on real-life scenarios.

Previous research has shown that ChatGPT can pass medical exams, and nearly two-thirds of physicians reported using some form of AI in 2024. But other research has shown that chatbots, including ChatGPT, don't provide reliable medical advice.

you are viewing a single comment's thread
view the rest of the comments
[–] natecox@programming.dev 4 points 6 hours ago (2 children)

Listen, this is going to sound like a loaded inflammatory question and I don’t really know how to fix that over text, but you say you’re in the space and I’m genuinely curious as to your take on this:

Do you think it’s possible to build LLM technology in a way that:

  1. Respects copyright and ip,
  2. Doesn’t fuck up the economy and eat all the ram,
  3. Doesn’t drink all the water and subject people to Datacenter hell, and
  4. is consistently accurate and has enough data to be useful?
[–] nate3d@lemmy.world 3 points 5 hours ago* (last edited 5 hours ago) (1 children)
  1. No. And I’ve lost my voice describing why this is the case - LLMs do not use training data in real time which is indicative of the fact that their reasoning chains are learned over many training epochs rather than something akin to a search engine which is parsing and aggregating results from direct sources. I wish I had a different answer but that is simply how the mathematics behind this kind of machine learning model work. The only way to properly manage it would be to limit and license the data appropriately during core model training, but that genie is out of the bottle.
  2. We will eventually (soon hopefully) hit critical mass where the technology isn’t delivering value on the hardware it takes to run it. The limitations, like I detailed above, are core to the technology and are not something that we’re just around the corner from solving. Those are core limitations and a different technology will be needed to move the ball forward past what is essentially a calculator with words. When this happens, we’ll see a whiplash effect where a ton of (server) hardware hits the market from the small datacenters looking to capitalize on the current rush. It’ll cripple the market for new hardware, I’d expect, as they’re going to want to get that capital back ASAP as it’s a quickly deprecating asset if just sitting idle.
  3. Similar to above, the current trajectory isn’t going to last. It’s going to hurt once the reality finally sets in for the economy.
  4. Oh yes, and it’s already been there for years! Unfortunately, these applications are not the glamorous applications like a “Her”-style chat companion, but rather precise application of specific machine learning models for specific business needs. I.e. do you really need an LLM to upload a picture to ask what kind of cat is in the picture? NO! That’s what convolutional neural networks are for, or maybe some custom vision transformers. There are dozens of types of ML models that have clear applications and with fine tuning and proper process implementation, the models can produce production-ready results as any other means of solving this issue.

The core problem with this technology is the misuse/misunderstanding that:

  1. AI does not yet exist. Full stop.
  2. An LLM is just ONE TYPE of machine learning algorithm
  3. An LLM does not possess the ability to understand OR interpret intent
  4. An LLM CAN NOT THINK This is the point I can’t stress enough; the “thinking” models you see today are doing nothing much more than cramming additional data into it’s working context and hoping that this guides the inference to produce a higher-quality result. Once a model is loaded for inference (i.e. asking questions) it is a STATIC entity and does not change.

Thank you for coming to my autistic TED talk <3

Edit: Also, fantastic question and never apologize for wanting to learn; keep that hunger and run with it

[–] natecox@programming.dev 2 points 4 hours ago

Well, this was exactly the answer I expected but I’m still disappointed.

I feel like I’m in a niche position where I want the technology to deliver on promises made (not inherently anti-AI) but even if they did I would still refuse to use them until the ethical and moral issues get solved in their creation and use (definitely anti-cramming-LLMs-into-every-facet-of-our-lives).

I miss being excited about machine learning, but LLMs being the whole topic now is so disappointing. Give us back domain specific, bespoke ML applications.

[–] LodeMike@lemmy.today -1 points 6 hours ago (1 children)

Not who you asked but

  1. Yes. Public domain only IG.
  2. Small
  3. Small
  4. No. Not while being 1.