By using this site, you agree to the Privacy Policy and Terms of Use.
Accept

GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIAnthropicTech

Claude Opus 4.6 and Sonnet 4.6 now support 1M tokens at standard pricing

One million tokens is roughly 750,000 words — or about ten full novels — and Claude can now process all of it in a single prompt at standard pricing.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Mar 14, 2026, 5:02 AM EDT
Share
We may get a commission from retail offers. Learn more
Anthropic Claude head illustration
Image: Anthropic
SHARE

Anthropic just made a significant move that’s been quietly anticipated in the developer community for a while — the company officially announced on March 13, 2026, that its 1 million token context window is now generally available for both Claude Opus 4.6 and Claude Sonnet 4.6, effective immediately across the Claude Platform, Amazon Bedrock, Google Cloud’s Vertex AI, and Microsoft Azure Foundry.

To understand why this is a big deal, it helps to know what a “context window” actually means. Think of it as the AI’s working memory — everything it can read and consider at once before giving you an answer. One million tokens is roughly 750,000 words, or the equivalent of about ten full-length novels. Until now, even the smartest AI models would start forgetting what you told them earlier in a conversation once things got too long. That problem — sometimes called “context rot” — has been a real limitation for engineers, lawyers, researchers, and really anyone trying to use AI for complex, sprawling projects.

What’s changed with today’s announcement isn’t just the raw number. It’s the price. Previously, when Anthropic launched Opus 4.6 back in February, that 1M context window was available in beta — but for prompts exceeding 200K tokens, developers were billed at a premium rate of $10 per million input tokens and $37.50 per million output tokens. That was a steep surcharge that many developers simply couldn’t justify at scale. Starting now, those premium rates are completely gone. The standard pricing — $5 per million input tokens and $25 per million output for Opus 4.6, and $3/$15 for Sonnet 4.6 — applies whether you’re sending a 9,000-token message or a 900,000-token one. No multiplier, no fine print.

Beyond pricing, Anthropic has also lifted a few other practical limitations. The media limit per request has jumped from 100 images or PDF pages all the way to 600 — a six-fold increase that makes a meaningful difference for anyone doing document-heavy work. Full rate limits now apply across the entire context window, which means developers aren’t penalized or throttled just because their requests are longer. And for those who were using the beta header in their API calls to unlock long-context access, Anthropic says it’s no longer needed — requests over 200K tokens just work automatically without any code changes.

The other question worth asking is: does the model actually use all that context effectively, or is it just window dressing? This is where Anthropic has put serious effort. On MRCR v2 — an industry benchmark that tests long-context retrieval by hiding multiple pieces of information deep inside a million-token document and asking the model to find them all — Claude Opus 4.6 scores 78.3% at the 1M token length, the highest among frontier models at that context length. For comparison, Sonnet 4.5, the previous default model, managed just 18.5% on the same test. That’s not a minor improvement. That’s a qualitative leap, the kind of difference that changes whether a feature is actually useful in production or just a marketing claim.

The real-world implications are starting to surface in interesting ways. Anthropic shared a number of testimonials from companies already using the expanded context. One AI research lab says it can now synthesize hundreds of scientific papers, proofs, and codebases in a single pass, dramatically accelerating fundamental physics research. A legal tech company notes that lawyers can finally bring multiple rounds of a 100-page contract negotiation into one session without losing track of changes across versions. An incident response platform says it can keep every signal, entity, and working theory in view from the first alert all the way through remediation — without compaction or context clearing.​

One particularly telling data point comes from a company that raised its Opus context window from 200K to 500K and found the agent actually used fewer tokens overall — because with more context available, the model spent less time re-reading and re-processing earlier information. That counterintuitive result speaks to something deeper about how context efficiency works: more isn’t always wasteful; sometimes it’s actually leaner.

For Claude Code users — Anthropic’s AI-powered coding assistant — this update is especially meaningful. Max, Team, and Enterprise users on Opus 4.6 will now default to 1M context automatically, which means fewer “compaction” events where the model is forced to summarize and discard earlier parts of a long coding session. Developers who have worked with Claude Code at scale know exactly how painful those compaction moments are — you lose details, cross-file dependencies get murky, and you end up re-explaining things you’ve already said. With 1M context running by default, that friction is largely eliminated.​

Sonnet 4.6, which Anthropic made the default model for Free and Pro claude.ai users when it launched in February, also benefits from today’s announcement. The model was already praised for approaching Opus-level intelligence at Sonnet-level pricing, and now it carries the same long-context access without surcharge. For developers building on a budget or teams that need high throughput at reasonable cost, Sonnet 4.6 at $3/$15 per million tokens with a full 1M window is a compelling combination.

In the broader AI landscape, this move puts Anthropic squarely in competition with Google’s Gemini 1.5 Pro and Gemini 2.0, both of which have long offered 1M token contexts at competitive prices. What Anthropic is now arguing is that having the context window isn’t enough — what matters is how well the model retrieves and reasons across that context. With Opus 4.6’s benchmark scores and Anthropic’s claim of being the highest-performing frontier model at 1M tokens, the company is making a quality-over-quantity argument.

For anyone building enterprise software, doing large-scale document analysis, or simply tired of their AI assistant losing the thread halfway through a long conversation — this is the kind of infrastructure update that quietly makes a lot of things better. The 1M context window is available right now across all major cloud platforms, with no extra steps required.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Topic:Claude AI
Leave a Comment

Leave a ReplyCancel reply

Most Popular

What is ChatGPT? The AI chatbot that changed everything

Anthropic launches The Anthropic Institute for frontier AI oversight

Samsung’s Galaxy Book6, Pro and Ultra land in the US today

Alexa+ adds new response styles so your smart speaker feels more personal

Nothing’s new headphones cost less and sound better — here’s why

Also Read
Google Pi Day 2026 logo showing the letter G in Google's signature colors — green, red, yellow, and blue — rendered in a hand-cut paper collage style with a textured matte finish, set against a plain white background.

Google’s Pi Day 2026 Doodle honors Archimedes and his 96-sided polygon trick

A promotional graphic from Meta showing the Facebook logo with the text "Rewarding Original Creators," surrounded by creator profile photos and engagement icons.

Facebook’s war on copycats is real — and original creators are winning

A hero banner image for Facebook Marketplace's 2026 Meta AI update, featuring a mock product listing of a blue and white porcelain tea set priced at $100 with a "Draft your listing with AI" prompt at the top. To the right is a circular seller profile photo of a smiling woman with a 4.9-star rating badge showing 23 reviews, alongside a "Meta AI - Summary" card. A blue "Generate shipping label" button appears at the bottom of the listing. The background features a bold blue circle with Facebook's logo, a thumbs-up icon, a heart reaction, and a hugging emoji scattered around.

Meta AI handles the boring parts of selling on Facebook Marketplace

A SpaceX Falcon 9 rocket lifts off from Space Launch Complex 4 East at Vandenberg Space Force Base in California, surrounded by billowing white smoke and bright engine flames, with a clear blue sky and the Pacific Ocean visible in the background.

SpaceX launches 25 Starlink v2 Mini satellites on Falcon 9

Tesla Powerwall 3 WPfRR4Lt

Tesla Powerwall 3 is officially headed to Japan

A screenshot of a Perplexity-branded document titled "Ways to Reclaim Focus," showing the Final Pass document markup feature in action — a yellow tooltip popup highlights a spelling/grammar suggestion labeled "low" severity, pointing out that "less tabs" should be corrected to "fewer tabs," with the suggested fix reading "your research happens in fewer tabs," demonstrating how Final Pass flags actionable edits inline within the document.

Perplexity Computer can now mark up any document with Final Pass

A dark-themed screenshot of the Perplexity Computer interface on a Mac desktop with a floral wallpaper. The left sidebar shows navigation options including Computer, New Task, Tasks, Files, Connectors, Skills, and Gallery. The main panel displays a task titled "Syncing Google Docs and Notion Editor" in progress. Visible steps include building the updated app, running a terminal command (cd unified-editor && npm run build), restarting the production server with NODE_ENV=production node dist/index.js, and a final deployment step labeled "Deploying with new document open by default." At the bottom, a completion message reads "Done — the app now opens straight into the new document editor with the title field and block editor ready to go," accompanied by a preview thumbnail of the generated app called "Perplexity Docs." Two additional browser tabs labeled "Google Docs" and "Media | Articles | Notion" are open at the top of the window.

Perplexity Computer adds embedded connector support for web app builders

Screenshot of the Perplexity Pro interface showing the model selection dropdown menu with "Nemotron 3 Super" selected, labeled as "New," alongside other available models including Best, Sonar, GPT-5.4, Gemini 3.1 Pro, Claude Sonnet 4.6, and Claude Opus 4.6 (Max, locked). The "Computer" mode button and "Thinking" toggle are also visible in the dark-themed UI.

NVIDIA Nemotron 3 Super lands on Perplexity, Agent API, and Computer

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.