By using this site, you agree to the Privacy Policy and Terms of Use.
Accept

GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIOpenAITech

60-year-old man hospitalized after following ChatGPT diet advice

ChatGPT’s salt replacement suggestion led a man to develop bromism, a rare toxic condition, and spend three weeks in the hospital.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Aug 8, 2025, 4:37 AM EDT
Share
OpenAI's ChatGPT chatbot mobile app icon on iPhone smartphone
Photo: Alamy
SHARE

When a 60-year-old man decided to remove sodium chloride from his diet, he didn’t call a doctor — he opened a chat window. Three months later, he was in hospital with a rare but well-documented toxic condition, having traded table salt for sodium bromide after reading about salt’s harms and “consulting ChatGPT.” The episode, laid out in a new case report, is a sharp reminder that conversational AI can confidently lead people down dangerous paths when medical context and judgment are required.

According to the clinical case published in Annals of Internal Medicine: Clinical Cases, the man replaced table salt with sodium bromide he bought online after an interaction with ChatGPT that apparently suggested chloride could be swapped out — a suggestion that makes sense in some industrial or cleaning contexts but is medically perilous. After weeks of consuming bromide, he developed neuropsychiatric symptoms (paranoia, auditory and visual hallucinations), skin changes (new acne and cherry angiomas), and profound thirst — enough that he spent three weeks in hospital before being discharged to psychiatry care as his condition improved.

It’s a striking sequence because bromide toxicity — bromism — used to be much more common decades ago when bromide salts were an ingredient in sedatives and other over-the-counter remedies. Today it’s rare, which helped delay diagnosis: clinicians had to piece together the psychiatric symptoms and the patient’s dietary experiment to identify chronic bromide exposure as the culprit.

What is bromism, and why is it dangerous?

Bromism is the clinical syndrome that follows chronic exposure to bromide ions. Symptoms can be neurologic (confusion, ataxia, psychosis, hallucinations), psychiatric (paranoia, delirium), gastrointestinal, and dermatologic (acneiform eruptions, cherry angiomas). Because bromide has a long elimination half-life, it accumulates over time and can mimic primary psychiatric disorders — which means the diagnosis can be easily missed unless someone suspects toxic exposure. Treatment focuses on stopping the exposure and hastening removal (salt and fluids, diuretics, or in severe cases dialysis).

That clinical background helps explain why swapping one white powder — table salt — for another can be more than a quirky nutrition experiment. The body treats chloride and bromide very differently at the doses someone might be ingesting when they start buying chemical salts online.

Chatbots, confidence, and risky advice

This isn’t the first time AI chatbots’ medical responses have stirred trouble. Studies and reports have repeatedly shown that large language models can produce plausible but sometimes incorrect or misleading medical guidance, and that some chat interfaces have dropped or softened health disclaimers, increasing the risk that users take the output as authoritative. In short: hallucinations aren’t just weird sentences — they can become harmful actions when users act on them without verification.

The authors of the Annals case report explicitly warned that AI can generate scientific inaccuracies and lacks the capacity to critically appraise results or a person’s medical history — exactly the sort of context a clinician brings to decisions about diet, supplements, and chemical exposures. The case reads as a cautionary tale: a technically literate person (the patient had studied nutrition in college) used web and AI tools to self-experiment, but without medical oversight the experiment turned toxic.

What companies say — and what the lawbook already warns users about

OpenAI’s own service terms make clear that its chat services are not intended for medical diagnosis or treatment, and that output “may not always be accurate” and shouldn’t be relied on as a sole source of truth or professional advice. Those legal lines are blunt, but they don’t always reach or convince every user in the moment — especially when an answer sounds authoritative.

Takeaways

  • If a chatbot suggests swapping chemicals or “removing” a nutrient, treat that as a red flag, not an instruction manual. Substances sold for industrial or cleaning uses are often toxic if ingested.
  • Apparent medical advice from an AI is a conversation starter, not a prescription. Ask a clinician, pharmacist, or poison-control center before doing anything that changes what you ingest.
  • Clinicians: remember that patients may arrive having tried internet or AI-sourced experiments; a careful exposure history can be diagnostic. The Annals paper is a useful clinical vignette for teaching that point.

The human cost

For this man, it meant weeks in hospital, frightening psychiatric symptoms, and a recovery that could have been avoided if a trusted, expert source had been consulted first. The episode is small but telling: AI tools have reshaped how people seek health information, and while that can democratize knowledge, it also amplifies the risk that a single confidently worded, context-free answer will trigger real-world harm.

If there’s a moral to take away, it isn’t that AI is bad — it’s that context matters. Medical judgment is less about facts in isolation and more about fitting those facts to a person’s history, meds, labs, and risks. Machines can help surface ideas; humans still need to vet them.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Topic:ChatGPTHealth
Most Popular

Gemini 3 Deep Think promises smarter reasoning for researchers

ClearVPN adds Kid Safe Mode alongside WireGuard upgrade

Google Docs now speaks your notes aloud

Why OpenAI built Lockdown Mode for ChatGPT power users

Ring cuts off Flock Safety partnership before launch

Also Read
Apple iPhone Air MagSafe Battery

Apple’s iPhone Air MagSafe Battery just got a rare price cut

HBO Max logo

HBO Max confirms March 26 launch in UK and Ireland with big shows

Sony WF‑1000XM6 earbuds in black and platinum silver.

Sony WF‑1000XM6 launch with class‑leading ANC and premium studio‑tuned sound

Promotional image for Death Stranding 2: On the Beach.

Death Stranding 2: On the Beach brings the strand sequel to PC on March 19

The image features a simplistic white smile-shaped arrow on an orange background. The arrow curves upwards, resembling a smile, and has a pointed end on the right side. This design is recognizable as the Amazon's smile logo, which is often associated with online shopping and fast delivery services.

Amazon opens 2026 Climate Tech Accelerator for device decarbonization

Google Doodles logo shown in large, colorful letters on a dark background, with the word ‘Doodles’ written in Google’s signature blue, red, yellow, and green colors against a glowing blue gradient at the top and black fade at the bottom.

Google’s Alpine Skiing Doodle rides into Milano‑Cortina 2026 spotlight

A stylized padlock icon centered within a rounded square frame, set against a vibrant gradient background that shifts from pink and purple tones on the left to orange and peach hues on the right, symbolizing digital security and privacy.

OpenAI rolls out new AI safety tools

Promotional image for Donkey Kong Bananza.

Donkey Kong Bananza is $10 off right now

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.