GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIAnthropicSecurityTech

Anthropic’s secret Mythos AI just slipped into the wrong hands

Anthropic swore Mythos was too dangerous for the public, but a small group of outsiders still found a way to sneak inside its most powerful AI system.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Apr 22, 2026, 1:35 PM EDT
Share
We may get a commission from retail offers. Learn more
Anthropic logo displayed as bold black uppercase text on a light beige background.
Image: Anthropic
SHARE

Anthropic built Mythos as the kind of AI you never want loose on the open internet – and yet, according to a report, that is now effectively what has happened, at least for a small but determined group of users. Behind the headline about “unauthorized access” is a story about how frontier AI, messy supply chains, and very human curiosity have collided in a way that should worry pretty much anyone who relies on software – which is to say, all of us.

Mythos is not just another chatbot with a friendly name and a quirky personality; it is Anthropic’s internal, unreleased security super-tool, a frontier model trained specifically to hunt down and exploit vulnerabilities across the digital stack. In Anthropic’s own framing, the Claude Mythos Preview can identify and exploit flaws “in every major operating system and every major web browser” when pointed at a target – putting it in a different league from the consumer-facing assistants we’re used to poking for recipe tips and vacation ideas. That capability is exactly why Mythos was being rolled out in a tightly controlled way, through an effort called Project Glasswing that brings in heavyweight partners like AWS, Apple, Google, Microsoft, NVIDIA, JPMorgan Chase, and major security vendors to use the model for defensive cybersecurity, not chaos.

The idea behind Glasswing is simple on paper: if you’ve built an AI that is better than almost any human at finding bugs, you point it at critical infrastructure and let it stress-test the software before attackers do. Anthropic says Mythos has already uncovered thousands of high-severity vulnerabilities across widely used operating systems, browsers, and other software, with those issues being quietly disclosed so vendors can patch them. In theory, this is a huge win for users: the same intelligence that could crack open banks or hospitals can instead help lock them down, reducing the risk of the kind of catastrophic breaches that dominate headlines every few weeks. But that bargain only holds if the model stays where it’s supposed to be.

According to documentation and sources cited by Bloomberg, that assumption broke on April 7th – the very day Anthropic publicly announced it was giving limited Mythos access to a small set of companies under the Project Glasswing banner. A “small group of unauthorized users” in a private online forum reportedly managed to get their hands on the model through a mix of inside access and good old-fashioned internet sleuthing. The Verge, which also reviewed details of the incident, reports that the group congregates on Discord, focused specifically on tracking and poking at unreleased AI systems.

The key link appears to have been a third-party contractor working with Anthropic. One member of the forum, described as a “third-party contractor for Anthropic,” allegedly used their access to help the group reach Mythos, augmented by clues from a separate security disaster: the recent Mercor breach. Mercor is a $10 billion AI training startup that supplies data and services to labs like OpenAI, Anthropic, and Meta, and it suffered a major supply-chain attack that exposed around 4 terabytes of sensitive data, including internal materials tied to frontier models. Among that haul, according to reporting, were details about Anthropic’s model formats and deployment patterns, enough for motivated sleuths to make an “educated guess” about where Mythos might be reachable online.

Put those pieces together – a contractor with access, leaked information about how Anthropic deploys models, and a Discord full of people obsessed with unreleased AI – and you have the rough recipe for what happened next. The group reportedly located an environment hosting Claude Mythos Preview, then began using it regularly over the following two weeks, even providing screenshots and a live demo to Bloomberg as proof. Anthropic, for its part, has acknowledged that it is investigating a report of unauthorized access to Mythos through “one of our third-party vendor environments” and says it has no evidence, so far, that its core systems were compromised or that the incident extended beyond that vendor’s environment.

The most unsettling detail might be what the group decided not to do. According to Bloomberg’s sources and follow-on coverage, the forum members deliberately avoided using Mythos for cybersecurity tasks – the very thing it was built for – in order to avoid tripping any alarms Anthropic might be running on that environment. Instead, they appear to have treated it more like a rare piece of tech memorabilia: something to play with, show off on a private server, and keep under the radar rather than turn into an obvious weapon. That restraint is cold comfort, though, because the incident reveals just how porous the AI ecosystem around these models really is.

Inside Anthropic, Mythos has been framed as “too dangerous for the wild,” a phrase that sounds melodramatic until you understand what the company’s own red-teamers reportedly found. According to feature reporting, internal teams concluded that Mythos could reliably compromise the underlying systems of modern computing – not just individual apps, but the operating systems and core services that everything else runs on. Government agencies and banks have been racing to secure access to the model under Project Glasswing precisely because they see it as both a powerful ally and a potential national security risk. When a tool can out-hack almost anyone, you don’t just worry about script kiddies; you worry about states, criminal syndicates, and anyone with enough money and patience to chase it.

That’s why the Mythos leak lands differently than, say, a misconfigured chatbot endpoint or a leaked model checkpoint for a generic text generator. It suggests that no matter how carefully a frontier lab tries to wrap a powerful system in rules and access controls, the weakest link might be a contractor’s workstation, a compromised vendor, or a misjudged integration. The Mercor breach is a particularly stark reminder that the AI industry relies on a dense supply chain of vendors, contractors, open source components, and cloud services – each of them a potential attack surface. When one of those links snaps, it’s not just personal data on the line; it can be the blueprints and connective tissue around the most advanced AI systems on earth.

Security researchers have been warning for years that “AI safety” can’t just mean making sure chatbots don’t say offensive things or refuse to write malware on request. Incidents like the Anthropic leak and the Mercor attack are now being cited as early examples of a different class of problem: AI security failures that expose internal models, training data, or deployment details in ways that could shape the global balance of power in AI. In one case described by enterprise security analysts, a release-packaging error at Anthropic exposed source code for its Claude Code assistant and other internal files, while in another, a supply-chain attack led to mass exfiltration of proprietary data from a vendor’s environment. None of these are science-fiction scenarios – they look a lot like the messy, human bugs and oversights that drive ordinary data breaches, just with much higher stakes.

For everyday users, it’s tempting to treat Mythos as something distant and abstract, locked inside corporate networks and government labs. But the whole point of tools like Project Glasswing is to defend the software you actually rely on: the banking apps on your phone, the EHR systems at your hospital, the cloud services your employer runs on, the routers and switches keeping your home internet alive. If a model like Mythos can make defenders an order of magnitude more effective at finding and fixing vulnerabilities in those systems, that’s a big net positive – as long as the model doesn’t leak, get cloned, or end up operating at the direction of people who don’t care about collateral damage. The moment unauthorized access becomes normal, the argument for deploying such capabilities at all starts to look a lot shakier.

This also adds fuel to a brewing policy debate about how frontier AI models should be governed. Anthropic is already flagged as a “supply-chain risk” by the Pentagon, reflecting concerns that dependencies on a single private AI vendor could introduce systemic vulnerabilities for the US government and military. At the same time, the company has been trying to rebuild trust in Washington by pitching Mythos and Project Glasswing as proof that high-end AI can serve national security and critical infrastructure, not just consumer products. Now policymakers have to confront the uncomfortable reality that even those security-focused models can slip the leash, not because a lab suddenly decided to open-source its weights, but because a contractor clicked the wrong link or a vendor failed to lock down a dependency.

We’re also seeing the cultural side of this: a growing underground scene of AI hobbyists and professionals who treat unreleased systems the way console modders once treated dev hardware, as forbidden toys that confer status if you can get your hands on them. A private Discord channel trading tips on hidden endpoints or leaked model formats might sound niche, but those communities can quickly become the connective tissue between security lapses and real-world exploitation. The Mythos incident suggests that this subculture doesn’t always see itself as “the bad guys” – they may genuinely be curious, careful, even proud of not going full black hat – but they still normalize a world where wandering into a restricted AI environment is a fun challenge, not a red line.

For the big AI labs, the lesson is brutally straightforward: your threat model has to include your own ecosystem, not just external attackers trying to brute-force the front door. Anthropic’s experience with Mythos and its previous leaks, combined with the Mercor fallout, point to a future where model governance, red-teaming, and safety benchmarks are only one part of the job. The rest looks a lot like old-school, unglamorous enterprise security work: doing vendor due diligence, controlling contractor access, monitoring strange usage patterns, and building incident response plans that assume the worst has already happened somewhere in your stack.

And for the rest of us – users, customers, voters – the Mythos story is a reminder that AI risk isn’t just about whether a model says something harmful on social media or automates away a job. It’s also about whether powerful new tools that can reshape cybersecurity actually stay in the hands of people trying to make the internet safer. When one of those tools quietly slips into a Discord channel, even for a couple of weeks, it shows just how thin the line is between “too dangerous for release” and “accessible if you know the right people and the right tricks.” That thin line is exactly where the next phase of the AI security conversation is going to be fought.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Leave a Comment

Leave a ReplyCancel reply

Most Popular

Claude for Microsoft 365 is now generally available

How to stream all five seasons of The Boys right now

Anthropic launches full Claude Platform on AWS with native integration

OpenAI upgrades its Realtime API with three new voice AI models

AI-powered Google Finance launches across Europe now

Also Read
Person holding a smartphone displaying the Gemini app in dark mode with an AI-generated optics study guide on screen. The document includes explanations of spherical mirror geometry, focal points, and mirror equations, along with mathematical formulas and bullet-point notes for exam preparation. The phone is held in a warmly lit indoor environment with a blurred background, creating a focused study atmosphere.

Turn handwritten notes into a smart Gemini study guide

Screenshot of a dark-themed terminal window running “Claude Code” on a desktop interface. The terminal displays project task management information for a workspace named “acme,” including one task awaiting input and several completed coding tasks such as test coverage improvements, load testing, payment migration, performance auditing, PR reviews, and dark mode implementation. A highlighted task labeled “release-notes” requests guidance on feature priorities. At the bottom, a command prompt invites the user to “describe a task for a new session.” The interface appears on a muted green background with subtle wave patterns.

Anthropic ships agent view to tame your Claude Code chaos

Apple App Store logo

Apple rebalances South Korea App Store pricing to keep global tiers in line

Close-up mockup of an iPhone displaying an RCS text conversation in the Messages app. The chat is with a contact named “Grace,” shown with a profile photo at the top. Below the contact name, the interface displays “Text Message • RCS” and “Encrypted,” indicating secure RCS messaging support. A green message bubble asks, “How are you doing?” and the reply says, “I’m good thanks. Just got back from a camping trip in Yosemite!” The screen uses Apple’s clean light-mode Messages interface with the Dynamic Island visible at the top.

iOS 26.5 update adds secure RCS messaging for iPhone users

Modern kitchen interior featuring a Samsung Bespoke AI Refrigerator Family Hub in a soft green-themed space. The large white refrigerator has a built-in display panel on the upper door showing abstract artwork. Surrounding the refrigerator are matching pastel green cabinets, a kitchen island with open shelving, and a dark countertop with a gold-tone faucet. Natural light enters through a large window beside the minimalist kitchen setup, highlighting the clean and modern design.

Gemini AI comes to Samsung’s Bespoke AI refrigerator Family Hub screen

Screenshot of the Windows 11 touchpad “Scroll & zoom” settings page in dark mode. The panel shows multiple enabled touchpad options with blue checkmarks, including “Drag two fingers to scroll,” “Automatic scrolling at edge,” “Automatic scrolling with pressure,” “Accelerated scrolling,” and “Pinch to zoom.” A “Single-finger scrolling” option is set to “Right Side.” The interface also includes sliders for “Scroll speed” and “Zoom speed,” along with a dropdown menu for “Scrolling direction” set to “Down motion scrolls up.”

Windows 11 adds custom scroll sliders to Settings

Illustration comparing Gmail writing suggestions before and after personalization. On the left, under the heading “Today,” a generic email draft to “Alex Liu” uses formal, template-style language with placeholder text. On the right, under “With personalization,” the same draft is rewritten in a more natural and conversational tone with specific influencer campaign details, highlighted text snippets, and a personalized sign-off. Along the right side are three colored labels reading “Personalized tone and style,” “Based on past emails,” and “Based on Drive files,” emphasizing how Gmail uses user context to improve writing suggestions.

Help me write in Gmail gets smarter with personalization

Three smartphone mockups displaying a ChatGPT trusted contact safety feature. The first screen explains how adding a trusted contact can help someone receive support during serious mental health or safety concerns. The second screen shows a form for inviting a trusted contact with fields for name, phone, email, and consent confirmation. The third screen confirms that the invitation was sent and offers an option to send a personal note.

OpenAI adds an emergency-style Trusted Contact option inside ChatGPT settings

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.