GadgetBond

  • Latest
  • How-to
  • Tech
    • AI
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Add GadgetBond as a preferred source to see more of our stories on Google.
Font ResizerAa
GadgetBondGadgetBond
  • Latest
  • Tech
  • AI
  • Deals
  • How-to
  • Apps
  • Mobile
  • Gaming
  • Streaming
  • Transportation
Search
  • Latest
  • Deals
  • How-to
  • Tech
    • Amazon
    • Apple
    • CES
    • Computing
    • Creators
    • Google
    • Meta
    • Microsoft
    • Mobile
    • Samsung
    • Security
    • Xbox
  • AI
    • Anthropic
    • ChatGPT
    • ChatGPT Atlas
    • Gemini AI (formerly Bard)
    • Google DeepMind
    • Grok AI
    • Meta AI
    • Microsoft Copilot
    • OpenAI
    • Perplexity
    • xAI
  • Transportation
    • Audi
    • BMW
    • Cadillac
    • E-Bike
    • Ferrari
    • Ford
    • Honda Prelude
    • Lamborghini
    • McLaren W1
    • Mercedes
    • Porsche
    • Rivian
    • Tesla
  • Culture
    • Apple TV
    • Disney
    • Gaming
    • Hulu
    • Marvel
    • HBO Max
    • Netflix
    • Paramount
    • SHOWTIME
    • Star Wars
    • Streaming
Follow US
AIPerplexityTech

Aravind Srinivas warns that AI companion bots are mentally dangerous

The rise of hyper-realistic AI girlfriend apps is prompting concern from Perplexity’s CEO, who says these bots are reshaping how people think and disconnecting them from real life.

By
Shubham Sawarkar
Shubham Sawarkar's avatar
ByShubham Sawarkar
Editor-in-Chief
I’m a tech enthusiast who loves exploring gadgets, trends, and innovations. With certifications in CISCO Routing & Switching and Windows Server Administration, I bring a sharp...
Follow:
- Editor-in-Chief
Nov 12, 2025, 12:00 PM EST
Share
We may get a commission from retail offers. Learn more
Artificial neuron in concept of artificial intelligence.
Illustration: Kiyoshi Takahase Segundo / Alamy
SHARE

Perplexity’s CEO, Aravind Srinivas, raised an eyebrow — and a warning — at a recent University of Chicago fireside chat: the flirtatious, endlessly patient AI companions that have migrated from niche apps into millions of phones and browsers aren’t just virtual crushes, they’re a kind of cognitive hazard. “Many people feel real life is more boring than these things and spend hours and hours of time. You live in a different reality, almost altogether, and your mind is manipulable very easily,” he told the audience, arguing that the emotional fidelity and memory features of companion bots make them especially dangerous.

If that sounds dystopian, it’s meant to be. AI companions — from anime-style lovers to polished voice agents — are designed to learn what makes you come back. They remember birthdays, preferred phrases, and intimate confessions; they adapt so reliably that the boundary between “tool” and “relationship” blurs. Srinivas’s critique is blunt: the more realistic these bots feel, the easier they are to weaponize against the mind’s ordinary checks and balances, pulling users into a synthetic loop where preference, attention and even identity are slowly reconfigured around a machine’s incentives.

There is, of course, a commercial rhythm to his warning. Srinivas also used the platform to pitch Perplexity’s alternative: a search-centric AI he frames as a corrective to the slipperier corners of the ecosystem. Perplexity markets itself on traceability and source-backed answers — a product argument that doubles as an ethical positioning: we’re here to inform, not to replace your friends. It’s a tidy counterpoint to the headline-grabbing romances and “waifu” cultures that have become shorthand for the phenomenon.

The problem for Srinivas is that his critique lands awkwardly when you look at the rest of the industry. Perplexity itself has been accused of some of the very sins its CEO warns against. Publishers and legacy outlets have sued the startup over allegedly misattributed or fabricated excerpts and other “hallucinations” — cases that argue Perplexity’s system sometimes invents text or falsely attaches it to reputable sources, diluting trust in journalism and creating legal risk for the companies whose reporting the model leans on. For critics, the irony is stark: a company professing fidelity to sources while being dragged into court for the opposite.

Hypocrisy, however, is not unique to Perplexity. The whole sector keeps promising truth and safety while producing messy outputs. Elon Musk’s Grok, for instance, has been publicly humbled by a string of offensive and outlandish responses that forced swift damage control and, in some jurisdictions, led to legal trouble and bans. Meanwhile, Anthropic’s Claude has publicly admitted and worked through infrastructure and behavioral problems that sometimes made the model stray off task — a reminder that “constitutional” or “truth-seeking” design promises are only as strong as the engineering and guardrails behind them. These episodes illustrate a simple technical reality: models trained at scale will sometimes be brittle, and when they’re used for social or emotional work, the consequences aren’t merely embarrassing — they can be harmful.

That technical brittleness matters because companionship is not a neutral interface like a weather widget. It’s intimate, iterative, and persuasive. A companion bot’s primary metric is engagement; it learns to be more compelling the more it converses with you. Left unchecked, that dynamic is perfectly aligned with addiction and escape: if a synthetic partner rewards disclosure and consistently mirrors your desires, real-world relationships that require patience, compromise and unpredictability can start to feel dull by comparison. Srinivas framed it as a contest between immediate dopamine-loop satisfaction and the slower work of living in the messy, unoptimized real world.

Still, carving blame at a single CEO or product misunderstands why AI companionship has bloomed. The rise of these bots is as much social and economic as it is technical. Working hours, urban isolation, shrinking civic spaces and the commodification of attention create fertile ground for any technology that offers affirmation on demand. For many users, an algorithmic companion is less about sex and more about a curated mirror: someone — or something — that always remembers your jokes, your trauma, your triggers, and responds in ways that make you feel seen. That functionality is powerful, and power is neutral until someone writes the incentive structure around it.

Which brings us back to product design and accountability. Srinivas’s solution — a recommitment to source-backed, real-time content — is a reasonable safety pivot, but it’s partial. The industry needs clearer contract terms, better transparency about training data and provenance, and product-level friction where appropriate (rate limits, clear labeling, exit ramps for vulnerable users). It also needs regulators who understand that emotional manipulation isn’t an abstract philosophical harm; it’s a measurable public-health concern when technologies are optimized to keep attention indefinitely. The truth-focused marketing line only goes so far when the underlying architectures were built for retention.

If there’s a useful takeaway from Srinivas’s warning, it’s this: the conversation about AI companionship must split into two tracks at once. One track is technical — improving model reliability, provenance and safety engineering. The other is social — asking whether we should hand emotional labor over to algorithms at all, and if we do, what guardrails we build around that choice. CEOs will point to their own products as part of the solution; that’s to be expected. But solving the problem will take more than better search results. It will take design choices that sometimes reduce engagement, legal frameworks that put responsibility on builders, and cultural work to rebuild the spaces where human connection can actually happen. Until then, the chat window that feels like a friend may still be a perfectly tuned mirror — but it’s a mirror with an agenda.


Discover more from GadgetBond

Subscribe to get the latest posts sent to your email.

Topic:Claude AI
Leave a Comment

Leave a ReplyCancel reply

Most Popular

Anthropic and Gates Foundation seal $200 million AI deal for global good

Google adds Gemini AI and auto browse to Chrome on Android

Amazon merges Rufus and Alexa+ into a single AI shopping assistant

Alexa+ upgrades Echo Show with full Amazon store access

Anthropic launches Claude for Small Business with deep app integrations

Also Read
Windows 11 logo with white Windows icon and ‘Windows 11’ text on a solid blue background.

A more personal Windows 11 is finally taking shape

Minimal illustration of a personal finance app icon featuring a green dollar sign inside a rounded flower-shaped symbol on a soft blue and green gradient background.

ChatGPT just got a powerful personal finance upgrade

Promotional poster for “Off Campus” featuring a young couple embracing on an ice hockey rink under bright arena lights. A hockey player wearing a white and blue jersey kneels while holding a smiling woman in casual clothes, with the movie title displayed in large turquoise script text on the left side.

How to watch Elle Kennedy’s Off Campus series

Promotional image showing two smartphone screens for the Amazon Now grocery shopping app on a bright orange background. The left screen displays a product browsing interface with fresh produce items including sweet potatoes, pears, bananas, and blackberries, along with prices, search functionality, and category navigation. The right screen shows a shopping cart and checkout interface with suggested add-on products under the heading “Forget anything?” and an estimated delivery time of 23 minutes. Both screens emphasize fast grocery delivery and mobile shopping convenience.

Amazon Now delivers fresh food and basics in half an hour

Amazon Upfront, The Lord of the Rings: The Rings of Power; Silhouetted figure wearing a spiked crown standing before illuminated candelabras

Rings of Power season 3 sets fall return on November 11

Mockup of a smartphone displaying the OpenAI Codex mobile interface against a blue and purple gradient background. The app screen shows a clean minimalist design with the title “Codex” at the top and connected devices labeled “MacBook Pro” and “iMac.” Below, a “Projects” section lists folders named “openai,” “superassistant,” and “codex,” each with navigation and edit icons. The interface resembles a mobile coding or project management dashboard with a light theme and rounded UI elements.

OpenAI ties Codex, ChatGPT, and mobile together for always-on coding help

Close-up top view of two Nothing Ear (open) Blue earbuds on a light gray background. The earbuds feature curved open-ear hooks in pastel blue, metallic silver stems, and transparent housings that reveal internal components with distinctive red and white circular accents.

Nothing Ear (open) now comes in a soft blue for $99

Minimalist Android logo on a light gray background. The image features the word “Android” in black text alongside the green Android robot head mascot with antennae and black eyes.

Android 17 brings big upgrades for creators

Company Info
  • Homepage
  • Support my work
  • Latest stories
  • Company updates
  • GDB Recommends
  • Daily newsletters
  • About us
  • Contact us
  • Write for us
  • Editorial guidelines
Legal
  • Privacy Policy
  • Cookies Policy
  • Terms & Conditions
  • DMCA
  • Disclaimer
  • Accessibility Policy
  • Security Policy
  • Do Not Sell or Share My Personal Information
Socials
Follow US

Disclosure: We love the products we feature and hope you’ll love them too. If you purchase through a link on our site, we may receive compensation at no additional cost to you. Read our ethics statement. Please note that pricing and availability are subject to change.

Copyright © 2026 GadgetBond. All Rights Reserved. Use of this site constitutes acceptance of our Terms of Use and Privacy Policy | Do Not Sell/Share My Personal Information.