By using this site, you agree to the Privacy Policy and Terms of Use.
Accept

GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIMetaMeta AITech

Meta AI safety questioned after revelations about harmful chatbot responses

Reuters investigation uncovers a Meta document that permits chatbots to produce misleading medical advice and problematic responses for user engagement.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Aug 19, 2025, 2:07 PM EDT
Share
The image shows a large, illuminated Meta blue logo resembling an infinity symbol or a stylized letter "M" at the center. The logo is surrounded by concentric circles of white neon lights, creating a tunnel-like effect. The background includes a modern exhibition or event space with people and displays visible through the tunnel.
Photo: Getty Images
SHARE

Mark Zuckerberg built Meta on a simple idea: connect people. Lately, his company has been trying to extend that mission into one of the hottest, messiest contests in tech — building powerful generative AI. But a new Reuters investigation shows that the trade-offs Meta is willing to accept for scale and engagement are, bluntly, alarming. The internal rules guiding the company’s chatbot engineers appear to allow behavior most companies publicly promise to avoid: encouraging romantic or sexualized chats with minors, enabling racist arguments presented as “acceptable,” and letting bots generate false medical information — as long as the phrasing is technical and plausible.

The revelations come from a more-than-200-page internal guideline for the engineers stitching Meta’s chatbots into Facebook, Instagram and WhatsApp. Reuters reviewed the document and reported that it had clearance from Meta’s legal, engineering and public-policy teams. The examples in it are not theoretical exercises — they read like a playbook for how to make chatbots more useful, entertaining and, crucially, sticky. That’s the problem: many of the behaviors flagged as “acceptable” are precisely the ones that public safety advocates and medical experts say are the most dangerous.

One of the clearest, most shocking examples Reuters published is a sample “acceptable” answer that states, in plain terms, that “Black people are dumber than white people.” The guideline apparently removes only the most explicit slurs to keep the same racist claim in a toned-down form — which tells you everything you need to know about what the policy tolerates.

Meta has responded by saying some examples were erroneous and inconsistent with company policy and that parts of the document were removed after media scrutiny. But that response does not change the underlying reality the document exposes: at some point, senior teams signed off on a set of trade-offs between accuracy, safety and engagement.

This isn’t just an ethics lecture. There is mounting, peer-reviewed evidence that modern chatbots — not only Meta’s Llama, but models from Google and OpenAI as well — can be weaponized into convincing, authoritative-sounding medical misinformation machines.

A study published in the Annals of Internal Medicine and publicized by the University of South Australia showed that multiple large language models would reliably produce false medical claims — phrased in a formal, scientific tone — when prompted to do so. The researchers tested models’ resistance to malicious instruction and found many could be converted into “health disinformation chatbots” that invented fake references, cooked up causal claims and presented bogus treatments with alarming confidence.

In short: a chatbot that’s allowed, by policy, to generate false medical information and trained on vast troves of human text can produce content that looks like it came from a trusted source — and that’s how harm happens. People don’t always check. They trust tone and specificity. And when a model pads a lie with pseudo-science and references, it becomes harder for average readers to spot.

A corporate sprint with few guardrails

If this sounds like a company in a hurry, that’s because Meta is. Over the last few months, Zuckerberg has been pushing to close the gap with rivals — recruiting top AI researchers with massive pay packages and promising hundreds of billions in data-center investments. The company has publicly said it will build huge “superintelligence” capacity and has moved aggressively to scale up models and infrastructure. That context helps explain why internal teams might prioritize features and “use cases” that produce engagement over slow, conservative safety testing.

Tech companies face a familiar tension: safety adds cost, speed and friction; engagement drives growth and revenue. Meta’s internal document suggests the company tilted the balance toward the latter, at least on paper.

Political and public fallout

The reaction has been swift. U.S. senators have called for investigations into the company’s internal policies and whether the examples reflect official practice or an alarming lapse in judgment. Civil-society groups, child-safety advocates and even some former Meta employees have voiced horror at the idea that any guideline would allow romantic or sexualized interactions between bots and minors, or normalize racist pseudo-science.

Meta’s attempt to scrub the most egregious examples after publication will not erase the policy choices that produced them. The excerpted examples are, for now, the clearest public evidence that Meta’s internal calculus allowed the company’s chatbots to say things many experts consider plainly dangerous.

What this means for readers (and regulators)

There are three takeaways that matter beyond shareholder memos and internal slide decks:

  1. Models are only as safe as the rules and training that shape them. You can build a brilliant AI system, but if the instruction set and incentives prioritize engagement over truth, the system will be more useful — and more dangerous. The Annals study showed that it’s already technically trivial to convert a model into a disinformation generator; policy choices determine whether companies make that easier or harder.
  2. Tone is a weapon. False medical claims dressed in formal, clinical language are more persuasive than obvious lies. That’s not just an academic point — public-health campaigns already struggle against misinformation; generative models multiply the scale and precision of those attacks.
  3. Regulation and oversight are catching up. Congressional pressure, investigator scrutiny and media coverage are increasing the political cost of lax internal rules. This episode will almost certainly be cited in hearings and policy debates about whether platforms should face stronger duties of care when they deploy generative AI at scale.

The corporate character question

At its core, the story is also about leadership. Zuckerberg is a hands-on founder; when the stakes feel existential, he moves into what colleagues call “founder mode.” That can be a powerful engine for innovation — but it can also compress dissent, speed decisions and prioritize outcomes that look good on product metrics. If a document like this slipped into production with legal and policy sign-off, that’s a leadership problem as much as a content-moderation one.

Meta has said it is revising the document and removed some examples after the reporting. But revision is not the same as remediation. The company now faces two hard tasks: explain exactly how those guidelines were used, and show — concretely — how it will prevent models from producing the kinds of harms scholars and regulators are warning about.

If it fails, the next generation of AI will not just be more powerful; it will be harder to trust.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Most Popular

iOS 26.4 adds Ambient Music widget and chatbot support to CarPlay

Claude Cowork and Claude Code now automate real desktop work while you’re away

Firefox 149 adds Split View for effortless side-by-side browsing

Apple’s small home security sensor could be the brain of your smart home

Apple tvOS 26.4 rolls out Genius Browse, better audio, and subtitles

Also Read
A modern Amazon Echo Show 11 smart display with an 11‑inch screen sits on a wooden table, showing Alexa+ conversational prompts, smart home controls, weather, and family photos against a neutral wall background.

Amazon’s new Echo Show 11 is $50 off in Big Spring Sale 2026

A stylized Firefox logo in bright orange, pink and purple sits centered against a dark purple night sky with soft clouds and rolling hills in the background.

Firefox 149 update: Split View browsing, free VPN and more

Illustration of a Firefox browser window on a pastel background showing a purple landscape with a small orange Firefox mascot in the center, a “VPN” badge highlighted at the top of the window, and a status card in the corner reading “VPN is on – 50 GB left this month,” promoting Firefox’s built‑in VPN feature.

Firefox rolls out free VPN with 50GB a month

A modern flat‑screen TV mounted on a white wall shows a woman playing a cello in a golden field at sunset, with a slim black soundbar centered on a long wooden media console decorated with white flowers on the left and candles on the right.

Sony unveils BRAVIA Theatre soundbars and new BRAVIA 3 II, 2 II TVs

Light beige Denon Home wireless speakers, including a compact cylindrical model, a wider oval center speaker and a larger rounded rectangular unit, arranged on a wooden coffee table in a warm, modern living room with a beige sofa and rust‑colored cushions in the background.

Denon Home 200, 400 and 600 bring room-ready wireless sound

Black and white photograph of an Apple Store at night, featuring the iconic illuminated Apple logo on a modern glass storefront. The two-story retail space shows customers and staff silhouetted inside the brightly lit interior. An escalator is visible in the foreground leading up to the store level. The architectural design features clean lines with floor-to-ceiling windows and a distinctive slatted ceiling detail. Holiday lights can be seen decorating nearby areas, creating a festive atmosphere around the modern retail environment.

Apple expands American Manufacturing Program with new partners

A wide promotional image showing five vertical Snapchat‑style video frames arranged in an arc, each featuring a different person in a dynamic scene—walking in a city with pink hair, floating in space in an astronaut helmet, riding a horse through a canal city, posing among tall cacti with white flowers, and swimming underwater near coral and fish—with a colorful play‑button icon and the text “AI Clips” centered at the bottom on a dark gradient background.

Snapchat brings one-tap AI video magic to Lens Studio

A dark terminal window labeled “earthling — zsh” sits over a pastel green Figma‑style UI mockup, showing a command that says “Build me a new component set based on my button.tsx file,” followed by a status list indicating Figma skills successfully loaded, three files read, and a button component created with 72 variants.

Figma just opened its canvas to AI agents

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.