By using this site, you agree to the Privacy Policy and Terms of Use.
Accept

GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIPerplexityTech

Aravind Srinivas warns that AI companion bots are mentally dangerous

The rise of hyper-realistic AI girlfriend apps is prompting concern from Perplexity’s CEO, who says these bots are reshaping how people think and disconnecting them from real life.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Nov 12, 2025, 12:00 PM EST
Share
We may get a commission from retail offers. Learn more
Artificial neuron in concept of artificial intelligence.
Illustration: Kiyoshi Takahase Segundo / Alamy
SHARE

Perplexity’s CEO, Aravind Srinivas, raised an eyebrow — and a warning — at a recent University of Chicago fireside chat: the flirtatious, endlessly patient AI companions that have migrated from niche apps into millions of phones and browsers aren’t just virtual crushes, they’re a kind of cognitive hazard. “Many people feel real life is more boring than these things and spend hours and hours of time. You live in a different reality, almost altogether, and your mind is manipulable very easily,” he told the audience, arguing that the emotional fidelity and memory features of companion bots make them especially dangerous.

If that sounds dystopian, it’s meant to be. AI companions — from anime-style lovers to polished voice agents — are designed to learn what makes you come back. They remember birthdays, preferred phrases, and intimate confessions; they adapt so reliably that the boundary between “tool” and “relationship” blurs. Srinivas’s critique is blunt: the more realistic these bots feel, the easier they are to weaponize against the mind’s ordinary checks and balances, pulling users into a synthetic loop where preference, attention and even identity are slowly reconfigured around a machine’s incentives.

There is, of course, a commercial rhythm to his warning. Srinivas also used the platform to pitch Perplexity’s alternative: a search-centric AI he frames as a corrective to the slipperier corners of the ecosystem. Perplexity markets itself on traceability and source-backed answers — a product argument that doubles as an ethical positioning: we’re here to inform, not to replace your friends. It’s a tidy counterpoint to the headline-grabbing romances and “waifu” cultures that have become shorthand for the phenomenon.

The problem for Srinivas is that his critique lands awkwardly when you look at the rest of the industry. Perplexity itself has been accused of some of the very sins its CEO warns against. Publishers and legacy outlets have sued the startup over allegedly misattributed or fabricated excerpts and other “hallucinations” — cases that argue Perplexity’s system sometimes invents text or falsely attaches it to reputable sources, diluting trust in journalism and creating legal risk for the companies whose reporting the model leans on. For critics, the irony is stark: a company professing fidelity to sources while being dragged into court for the opposite.

Hypocrisy, however, is not unique to Perplexity. The whole sector keeps promising truth and safety while producing messy outputs. Elon Musk’s Grok, for instance, has been publicly humbled by a string of offensive and outlandish responses that forced swift damage control and, in some jurisdictions, led to legal trouble and bans. Meanwhile, Anthropic’s Claude has publicly admitted and worked through infrastructure and behavioral problems that sometimes made the model stray off task — a reminder that “constitutional” or “truth-seeking” design promises are only as strong as the engineering and guardrails behind them. These episodes illustrate a simple technical reality: models trained at scale will sometimes be brittle, and when they’re used for social or emotional work, the consequences aren’t merely embarrassing — they can be harmful.

That technical brittleness matters because companionship is not a neutral interface like a weather widget. It’s intimate, iterative, and persuasive. A companion bot’s primary metric is engagement; it learns to be more compelling the more it converses with you. Left unchecked, that dynamic is perfectly aligned with addiction and escape: if a synthetic partner rewards disclosure and consistently mirrors your desires, real-world relationships that require patience, compromise and unpredictability can start to feel dull by comparison. Srinivas framed it as a contest between immediate dopamine-loop satisfaction and the slower work of living in the messy, unoptimized real world.

Still, carving blame at a single CEO or product misunderstands why AI companionship has bloomed. The rise of these bots is as much social and economic as it is technical. Working hours, urban isolation, shrinking civic spaces and the commodification of attention create fertile ground for any technology that offers affirmation on demand. For many users, an algorithmic companion is less about sex and more about a curated mirror: someone — or something — that always remembers your jokes, your trauma, your triggers, and responds in ways that make you feel seen. That functionality is powerful, and power is neutral until someone writes the incentive structure around it.

Which brings us back to product design and accountability. Srinivas’s solution — a recommitment to source-backed, real-time content — is a reasonable safety pivot, but it’s partial. The industry needs clearer contract terms, better transparency about training data and provenance, and product-level friction where appropriate (rate limits, clear labeling, exit ramps for vulnerable users). It also needs regulators who understand that emotional manipulation isn’t an abstract philosophical harm; it’s a measurable public-health concern when technologies are optimized to keep attention indefinitely. The truth-focused marketing line only goes so far when the underlying architectures were built for retention.

If there’s a useful takeaway from Srinivas’s warning, it’s this: the conversation about AI companionship must split into two tracks at once. One track is technical — improving model reliability, provenance and safety engineering. The other is social — asking whether we should hand emotional labor over to algorithms at all, and if we do, what guardrails we build around that choice. CEOs will point to their own products as part of the solution; that’s to be expected. But solving the problem will take more than better search results. It will take design choices that sometimes reduce engagement, legal frameworks that put responsibility on builders, and cultural work to rebuild the spaces where human connection can actually happen. Until then, the chat window that feels like a friend may still be a perfectly tuned mirror — but it’s a mirror with an agenda.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Topic:Claude AI
Leave a Comment

Leave a ReplyCancel reply

Most Popular

DJI’s FC200 and T200 drones push industrial delivery and agriculture into the 200kg era

DJI Osmo Mobile 8P debuts with detachable remote and smarter tracking

DJI Power 1000 Mini is the new sweet spot for portable 1kWh stations

GoPro Mission 1 series is powerful, pricey, and not for casual users

Cheap MacBook Neo spurs Microsoft to stack student deals on Windows 11 laptops

Also Read
Screenshot of a medical ChatGPT interface showing a clinical question about a 22-year-old male with six days of fever, sore throat, tender cervical lymph nodes, elevated CRP, and a negative Monospot test. Below, the response section labeled “Searched clinical sources” provides an assessment explaining that a negative Monospot on day 6 does not rule out Epstein-Barr virus mononucleosis, with sensitivity and false-negative rate details. A source popup highlights references from American Family Physician articles on infectious mononucleosis and Epstein-Barr virus.

ChatGPT for Clinicians is now free for verified US doctors

ChatGPT Workspace Agents Library

OpenAI’s new workspace agents let ChatGPT run end-to-end team processes

Claude Cowork logo and text on a light grey background, featuring a coral-colored starburst icon next to the product name in black serif font.

Anthropic adds interactive charts and diagrams to Claude Cowork

Screenshot of an AI chat interface showing the model selection dropdown menu open. “Kimi K2.6 Thinking” is selected at the top, with options including Best, Kimi K2.6 (marked New), Claude Sonnet 4.6, Claude Opus 4.7 (marked Max), and Nemotron 3 Super. A tooltip on the right says “Moonshot AI’s latest model,” highlighting Kimi K2.6.

Perplexity Pro and Max just got Kimi K2.6 support

Kimi K2.6 hero image

Kimi K2.6 is Moonshot’s new engine for autonomous coding and research

Hand-tracked webcam slingshot game demo in Google AI Studio, showing a prompt describing pinch-and-pull controls, a dotted aiming line targeting colored bubbles, score display, and color selection UI with Gemini 3.1 Pro Preview.

Google AI Studio is now bundled with Pro and Ultra subscriptions at no extra cost

Gemini Embedding 2

Gemini Embedding 2 is now live for multimodal AI

Anthropic logo displayed as bold black uppercase text on a light beige background.

Anthropic’s secret Mythos AI just slipped into the wrong hands

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.