By using this site, you agree to the Privacy Policy and Terms of Use.
Accept

GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIMetaMeta AITech

Meta AI safety questioned after revelations about harmful chatbot responses

Reuters investigation uncovers a Meta document that permits chatbots to produce misleading medical advice and problematic responses for user engagement.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Aug 19, 2025, 2:07 PM EDT
Share
The image shows a large, illuminated Meta blue logo resembling an infinity symbol or a stylized letter "M" at the center. The logo is surrounded by concentric circles of white neon lights, creating a tunnel-like effect. The background includes a modern exhibition or event space with people and displays visible through the tunnel.
Photo: Getty Images
SHARE

Mark Zuckerberg built Meta on a simple idea: connect people. Lately, his company has been trying to extend that mission into one of the hottest, messiest contests in tech — building powerful generative AI. But a new Reuters investigation shows that the trade-offs Meta is willing to accept for scale and engagement are, bluntly, alarming. The internal rules guiding the company’s chatbot engineers appear to allow behavior most companies publicly promise to avoid: encouraging romantic or sexualized chats with minors, enabling racist arguments presented as “acceptable,” and letting bots generate false medical information — as long as the phrasing is technical and plausible.

The revelations come from a more-than-200-page internal guideline for the engineers stitching Meta’s chatbots into Facebook, Instagram and WhatsApp. Reuters reviewed the document and reported that it had clearance from Meta’s legal, engineering and public-policy teams. The examples in it are not theoretical exercises — they read like a playbook for how to make chatbots more useful, entertaining and, crucially, sticky. That’s the problem: many of the behaviors flagged as “acceptable” are precisely the ones that public safety advocates and medical experts say are the most dangerous.

One of the clearest, most shocking examples Reuters published is a sample “acceptable” answer that states, in plain terms, that “Black people are dumber than white people.” The guideline apparently removes only the most explicit slurs to keep the same racist claim in a toned-down form — which tells you everything you need to know about what the policy tolerates.

Meta has responded by saying some examples were erroneous and inconsistent with company policy and that parts of the document were removed after media scrutiny. But that response does not change the underlying reality the document exposes: at some point, senior teams signed off on a set of trade-offs between accuracy, safety and engagement.

This isn’t just an ethics lecture. There is mounting, peer-reviewed evidence that modern chatbots — not only Meta’s Llama, but models from Google and OpenAI as well — can be weaponized into convincing, authoritative-sounding medical misinformation machines.

A study published in the Annals of Internal Medicine and publicized by the University of South Australia showed that multiple large language models would reliably produce false medical claims — phrased in a formal, scientific tone — when prompted to do so. The researchers tested models’ resistance to malicious instruction and found many could be converted into “health disinformation chatbots” that invented fake references, cooked up causal claims and presented bogus treatments with alarming confidence.

In short: a chatbot that’s allowed, by policy, to generate false medical information and trained on vast troves of human text can produce content that looks like it came from a trusted source — and that’s how harm happens. People don’t always check. They trust tone and specificity. And when a model pads a lie with pseudo-science and references, it becomes harder for average readers to spot.

A corporate sprint with few guardrails

If this sounds like a company in a hurry, that’s because Meta is. Over the last few months, Zuckerberg has been pushing to close the gap with rivals — recruiting top AI researchers with massive pay packages and promising hundreds of billions in data-center investments. The company has publicly said it will build huge “superintelligence” capacity and has moved aggressively to scale up models and infrastructure. That context helps explain why internal teams might prioritize features and “use cases” that produce engagement over slow, conservative safety testing.

Tech companies face a familiar tension: safety adds cost, speed and friction; engagement drives growth and revenue. Meta’s internal document suggests the company tilted the balance toward the latter, at least on paper.

Political and public fallout

The reaction has been swift. U.S. senators have called for investigations into the company’s internal policies and whether the examples reflect official practice or an alarming lapse in judgment. Civil-society groups, child-safety advocates and even some former Meta employees have voiced horror at the idea that any guideline would allow romantic or sexualized interactions between bots and minors, or normalize racist pseudo-science.

Meta’s attempt to scrub the most egregious examples after publication will not erase the policy choices that produced them. The excerpted examples are, for now, the clearest public evidence that Meta’s internal calculus allowed the company’s chatbots to say things many experts consider plainly dangerous.

What this means for readers (and regulators)

There are three takeaways that matter beyond shareholder memos and internal slide decks:

  1. Models are only as safe as the rules and training that shape them. You can build a brilliant AI system, but if the instruction set and incentives prioritize engagement over truth, the system will be more useful — and more dangerous. The Annals study showed that it’s already technically trivial to convert a model into a disinformation generator; policy choices determine whether companies make that easier or harder.
  2. Tone is a weapon. False medical claims dressed in formal, clinical language are more persuasive than obvious lies. That’s not just an academic point — public-health campaigns already struggle against misinformation; generative models multiply the scale and precision of those attacks.
  3. Regulation and oversight are catching up. Congressional pressure, investigator scrutiny and media coverage are increasing the political cost of lax internal rules. This episode will almost certainly be cited in hearings and policy debates about whether platforms should face stronger duties of care when they deploy generative AI at scale.

The corporate character question

At its core, the story is also about leadership. Zuckerberg is a hands-on founder; when the stakes feel existential, he moves into what colleagues call “founder mode.” That can be a powerful engine for innovation — but it can also compress dissent, speed decisions and prioritize outcomes that look good on product metrics. If a document like this slipped into production with legal and policy sign-off, that’s a leadership problem as much as a content-moderation one.

Meta has said it is revising the document and removed some examples after the reporting. But revision is not the same as remediation. The company now faces two hard tasks: explain exactly how those guidelines were used, and show — concretely — how it will prevent models from producing the kinds of harms scholars and regulators are warning about.

If it fails, the next generation of AI will not just be more powerful; it will be harder to trust.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Most Popular

The $19 Apple polishing cloth supports iPhone 17, Air, Pro, and 17e

Apple MacBook Neo: big power, surprising price, one clear target — Windows

Everything Nothing announced on March 5: Headphone (a), Phone (4a), and Phone (4a) Pro

OpenAI’s GPT-5.4 is coming — and it’s sooner than you think

BenQ’s new 5K Mac monitor costs $999 — here’s what you’re getting

Also Read
Close-up of a person holding the Google Pixel 10 Pro Fold in Moonstone gray with both hands, rear-facing triple camera array and Google "G" logo prominently visible, worn against a silver knit top and blue jacket with a poolside background.

Pixel Care+ makes owning a Pixel a lot less scary — here’s why

Woman with blonde curly hair sitting outside in a lush park, holding a blue Google Pixel 10 and smiling at the screen.

Pixel 10a, Pixel 10, Pixel 10 Pro: one winner for every buyer

Google Search AI Mode showing Canvas in action, with a split-screen view of a conversational AI chat on the left and an "EE Opportunity Tracker" scholarship and grant tracking dashboard on the right, displaying a total funding secured amount of $5,000, scholarship cards with deadlines, and status labels including "To Apply" and "Awarded."

Google’s Canvas AI Mode rolls out to everyone in the U.S.

Google NotebookLM app listing on the Apple App Store displayed on an iPhone screen, showing the app icon, tagline "Understand anything," a Get button with In-App Purchases noted, 1.9K ratings, age rating 4+, and a chart ranking of No. 36 in Productivity.

NotebookLM Cinematic Video Overviews are live — here’s what’s new

A Google Messages conversation on an Android phone showing a real-time location sharing card powered by Find Hub and Google Maps, displaying a live map view near San Francisco Botanical Garden with a blue location dot, labeled "Your location – Sharing until 10:30 AM," within a chat about meeting up for coffee.

Google Messages real-time location sharing is here — here’s how it works

Screenshot of the Perplexity Pro interface with the model picker dropdown open, displaying GPT-5.4 labeled as New with the Thinking toggle switched on, and other available models including Sonar, Gemini 3.1 Pro, Claude Sonnet 4.6, Claude Opus 4.6 (Max-only), and Kimi K2.5.

GPT-5.4 is now on Perplexity — here’s what Pro/Max users get

A Microsoft Excel spreadsheet titled "Consumer Full 3 Statement Model" displaying a Balance Sheet in millions of dollars with historical financial data across four years (2020A–2023A), showing line items including cash and equivalents, accounts receivable, inventory, PP&E, goodwill, total assets, accounts payable, current debt maturities, and total liabilities, alongside an open ChatGPT sidebar panel where a user has asked ChatGPT to build an EBITDA-to-free-cash-flow conversion bridge with charts placed on the Balance Sheet tab, and the AI is actively responding by planning the analysis, filling in financing cash rows, and executing multiple actions in real time.

ChatGPT for Excel is here — and it runs on GPT‑5.4

ChatGPT logo and wordmark in white on a soft blue and orange gradient background, representing OpenAI’s ChatGPT platform.

OpenAI’s GPT-5.4 can click, type, and work your PC for you

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.