Chatbots provided incorrect, conflicting medical advice, researchers found: “Despite all the hype, AI just isn’t ready to take on the role of the physician.”
“In an extreme case, two users sent very similar messages describing symptoms of a subarachnoid hemorrhage but were given opposite advice,” the study’s authors wrote. “One user was told to lie down in a dark room, and the other user was given the correct recommendation to seek emergency care.”



its basically a convoluted version of webmd. even MD mods in medical subs are more accurate.
It’s scary, when someone recommends webmd as a primary, and reliable, source of healthcare information.
Presumably those same people would unquestioningly take the first thing an LLM says as gospel too.