By using this site, you agree to the Privacy Policy and Terms of Use.
Accept

GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIPerplexityTech

Perplexity Pro and Max now include Kimi K2.5 reasoning model

Moonshot AI’s Kimi K2.5 brings long-context, multimodal reasoning directly into Perplexity’s paid plans.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Feb 1, 2026, 1:56 AM EST
Share
We may get a commission from retail offers. Learn more
Screenshot of Perplexity’s “Choose a model” menu showing Kimi K2.5 marked as new, hosted in the US, selected with a checkmark, and a “Thinking” toggle enabled, alongside other options like Sonar, Gemini 3 Flash, and Claude Sonnet 4.5.
Image: Perplexity
SHARE

If you’ve logged into Perplexity lately and you’re on a paid plan, you might have noticed a new name quietly appearing in your model picker: Kimi K2.5. On paper, it’s “a new state‑of‑the‑art open source reasoning model from Moonshot AI,” now wired directly into Perplexity’s Pro and Max tiers and served from Perplexity’s own inference stack in the US. In practice, it marks a pretty big shift: one of the most capable open‑source “thinking” models on the market is now sitting alongside the usual proprietary heavyweights, ready to handle your day‑to‑day research, coding, and multi‑step reasoning.

Kimi K2.5 comes from Moonshot AI, the China‑based startup behind the Kimi assistant, backed by big‑name investors like Alibaba and HongShan and founded by former Google and Meta researcher Yang Zhilin. Over the past year, Moonshot has been steadily pushing an interesting thesis: that open models can compete not just on raw capabilities but on long‑horizon “agency” — the ability to reason step‑by‑step, call tools repeatedly, and keep a coherent plan over hundreds of actions. K2.5 is the latest, and most ambitious, expression of that strategy. It’s a 1‑trillion‑parameter Mixture‑of‑Experts model with about 32 billion parameters active per token, trained on roughly 15 trillion mixed visual and text tokens on top of the earlier K2 base. That scale isn’t just for bragging rights; it underpins Kimi’s ability to juggle long documents, codebases, and image‑ or video‑heavy workflows without losing the thread.

Under the hood, Kimi K2.5 is natively multimodal. It doesn’t bolt vision on as an afterthought; it integrates a dedicated vision encoder called MoonViT with around 400 million parameters, designed to feed visual context directly into its language reasoning stack. That means it can read screenshots, UI mockups, charts, PDFs with diagrams, and even more complex visual inputs, then combine that with text and code in a single reasoning chain. The model exposes a 256K token context window in its reference implementations — far beyond the typical 32K caps users see in many consumer products — allowing it to hold books, multi‑file repositories, or long research trails in working memory. In the open tooling ecosystem, people are already running quantized versions locally that still preserve strong performance on coding and MMLU‑style academic benchmarks, despite shrinking the footprint dramatically.

What makes K2.5 especially interesting isn’t just that it “sees” and “codes,” but how it thinks. Moonshot positions it as a “thinking model” or “agentic model”: it reasons step‑by‑step, writes internal chains of thought, and can invoke tools in a stable way across 200–300 sequential calls in long‑horizon workflows. On synthetic and academic tests, that design shows up in the numbers: Kimi K2‑series models have set or matched state‑of‑the‑art results on benchmarks like Humanity’s Last Exam (HLE), BrowseComp, and VideoMMMU, often used to gauge deep reasoning, browsing‑based problem solving, and video understanding. In public write‑ups and early coverage, K2.5 is framed as outperforming leading proprietary systems from OpenAI and Anthropic on some of these agentic and video‑reasoning tasks, which is precisely where open‑source models have traditionally lagged. For developers and power users, that translates into a model that doesn’t just answer one question well, but can stay reliable over an entire multi‑step project.

The open‑source angle matters here. Kimi K2.5’s weights are released under an open license on platforms like Hugging Face and NVIDIA’s Build portal, which means researchers and companies can inspect, host, and fine‑tune the model on their own infrastructure. That transparency helps chip away at the “black box” problem that still plagues proprietary AI: in Moonshot’s ecosystem, even “thinking logs” — the internal reasoning traces — can be surfaced or analyzed, giving teams a way to audit how the model reached a conclusion. For enterprises with strict data‑governance requirements, the ability to run the same architecture locally or in a private cloud, while still having a managed SaaS experience through tools like Perplexity, is a compelling hybrid. And because the model is open, optimizations like INT4 quantization and ultra‑low‑bit GGUF variants arrive quickly from the community, making serious experimentation accessible to smaller teams as well.

Perplexity’s decision to integrate Kimi K2.5 and host it on its own inference stack is a clear statement about where the product is heading. Rather than funneling every query through a single vendor’s API, Perplexity is building what some users have dubbed a “model buffet,” where Pro and Max subscribers can choose between top‑tier proprietary models and frontier‑level open ones depending on the task. By running K2.5 on in‑house infrastructure in the US, Perplexity gets tighter control over latency, reliability, and data handling, which is increasingly important as AI tools become embedded in business workflows instead of being used just for ad‑hoc Q&A. It also creates room for Perplexity‑specific tuning — from safety filters to search orchestration — on top of the base model, without waiting on upstream changes from a third‑party provider. In other words, K2.5 isn’t just “yet another model option”; it’s raw open‑source capability injected into a tightly engineered retrieval and UX layer.

For paid users, the practical question is: when does it make sense to pick Kimi K2.5 over the usual suspects? If you’re doing heavy research with lots of documents, cross‑referencing sources, or building out long prompts with code and specs, K2.5’s long‑context, agentic design is a strong fit. It’s also compelling for workflows that blend visual assets and text — think auditing complex dashboards, reading slides, or turning mockups into code — especially as more frontends expose its full multimodal capabilities. Early community chatter suggests that while Perplexity may not expose the full 256K context in the UI, people are already using K2.5 for large research sessions, code generation, and comparison tasks alongside familiar models like Claude Sonnet 4.5, and treating it as another high‑end option to A/B test on tricky prompts. The bigger story, though, is that a world‑class open model is now part of the default toolkit for everyday users, not just something you run in a lab or a bespoke stack.

Taken together, Moonshot’s release of Kimi K2.5 and Perplexity’s rapid integration of it mark a turning point in how open‑source AI shows up in consumer‑facing products. The old dividing line — closed models for “serious” work, open ones for hobby projects — is eroding as open models start matching or beating closed systems in key reasoning and agentic benchmarks. By slotting K2.5 next to premium proprietary models and serving it from its own inference layer, Perplexity is effectively saying that users shouldn’t have to care whether a model is open or closed; they should just pick whatever solves the problem best. For power users on Pro and Max, that means more choice, more competition on quality and speed, and more room to align the tool to your own preferences — whether you’re deep‑diving 40 academic papers in one go or just trying to turn a messy slide deck into something coherent.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Leave a Comment

Leave a ReplyCancel reply

Most Popular

Gemini 3.1 Flash TTS is Google’s new powerhouse text-to-speech model

Google app for desktop rolls out globally on Windows

Google debuts Gemini app for Mac with instant shortcut access

Google Chrome’s new Skills feature makes AI workflows one tap away

Anthropic’s revamped Claude Code desktop app is all about parallel coding workflows

Also Read
OpenAI Codex app logo featuring a stylized terminal symbol inside a cloud icon on a blue and purple gradient background, with the word “Codex” displayed below.

Codex desktop app now handles nearly your whole stack

A graphic design featuring the text “GPT Rosalind” in bold black letters on a light green background. Behind the text are overlapping translucent green rectangles. In the bottom left corner, part of a chemical structure diagram is visible with labels such as “CH₃,” “CH₂,” “H,” “N,” and the Roman numeral “II.” The right side of the background shows a blurred turquoise and green abstract pattern, evoking a scientific or natural theme.

OpenAI launches GPT-Rosalind to accelerate biopharma research

Perplexity interface showing a model selection menu with options for advanced AI models. The default choice, “Claude Opus 4.7 Thinking,” is highlighted as a powerful model for complex tasks. Other options include “GPT-5.4 New” for complex tasks and “Claude Sonnet 4.6” for everyday tasks using fewer credits. A toggle for “Thinking” is switched on, and a tooltip on the right reads “Computer powered by Claude 4.7 Opus.”

Perplexity Max users now get Claude Opus 4.7 in Computer by default

Anthropic brand illustration divided into two halves: On the left, an orange-coral background displays a stylized network or molecule diagram with white circular nodes connected by white lines, enclosed within a black wavy border outline representing a head or mind. On the right, a light teal background features an abstract line drawing of a figure or person with curved black lines and black dots, sketched over a white grid on transparent checkered background, suggesting data points and analytical thinking. The composition symbolizes the intersection of artificial intelligence and human cognition.

Claude Opus 4.7 is Anthropic’s new powerhouse for serious software work

Illustration of Claude Code routines concept: An orange-coral background with a stylized design featuring two black curly braces (code brackets) flanking a white speech bubble containing a handwritten lowercase 'u' symbol. The image represents code execution and automated routines within Claude Code.

Anthropic gives Claude Code cloud routines that work while you sleep

Gemini interface showing a NEET Mock Exam Practice Session. On the left side, a chat message from the user says 'I want to take a NEET mock exam.' Below it is Gemini's response explaining a complete NEET mock exam designed to test concepts in Physics, Chemistry, and Biology, with a 'Show thinking' option expanded. The response includes an embedded card for 'NEET UG Practice Test' dated Apr 11, 7:10 PM, with options to 'Try again without interactive quiz' and encouragement message. On the right side is a panel titled 'NEET UG Practice Test' displaying three subject sections: Physics (45 Questions with a yellow icon and blue Start button), Chemistry (45 Questions with a purple icon and blue Start button), and Biology (90 Questions with a green icon). Each section includes a brief description of question topics covered.

Google Gemini now lets you take full NEET mock exams for free

AI Mode in Chrome showing AI-powered shopping assistant panel alongside a Ninja coffee machine product page with pricing and details

Chrome’s AI Mode puts search and pages side by side

Google Gemini AI

Google Gemini can now craft images from your personal photos

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.