Perplexity’s CEO, Aravind Srinivas, raised an eyebrow — and a warning — at a recent University of Chicago fireside chat: the flirtatious, endlessly patient AI companions that have migrated from niche apps into millions of phones and browsers aren’t just virtual crushes, they’re a kind of cognitive hazard. “Many people feel real life is more boring than these things and spend hours and hours of time. You live in a different reality, almost altogether, and your mind is manipulable very easily,” he told the audience, arguing that the emotional fidelity and memory features of companion bots make them especially dangerous.
If that sounds dystopian, it’s meant to be. AI companions — from anime-style lovers to polished voice agents — are designed to learn what makes you come back. They remember birthdays, preferred phrases, and intimate confessions; they adapt so reliably that the boundary between “tool” and “relationship” blurs. Srinivas’s critique is blunt: the more realistic these bots feel, the easier they are to weaponize against the mind’s ordinary checks and balances, pulling users into a synthetic loop where preference, attention and even identity are slowly reconfigured around a machine’s incentives.
There is, of course, a commercial rhythm to his warning. Srinivas also used the platform to pitch Perplexity’s alternative: a search-centric AI he frames as a corrective to the slipperier corners of the ecosystem. Perplexity markets itself on traceability and source-backed answers — a product argument that doubles as an ethical positioning: we’re here to inform, not to replace your friends. It’s a tidy counterpoint to the headline-grabbing romances and “waifu” cultures that have become shorthand for the phenomenon.
The problem for Srinivas is that his critique lands awkwardly when you look at the rest of the industry. Perplexity itself has been accused of some of the very sins its CEO warns against. Publishers and legacy outlets have sued the startup over allegedly misattributed or fabricated excerpts and other “hallucinations” — cases that argue Perplexity’s system sometimes invents text or falsely attaches it to reputable sources, diluting trust in journalism and creating legal risk for the companies whose reporting the model leans on. For critics, the irony is stark: a company professing fidelity to sources while being dragged into court for the opposite.
Hypocrisy, however, is not unique to Perplexity. The whole sector keeps promising truth and safety while producing messy outputs. Elon Musk’s Grok, for instance, has been publicly humbled by a string of offensive and outlandish responses that forced swift damage control and, in some jurisdictions, led to legal trouble and bans. Meanwhile, Anthropic’s Claude has publicly admitted and worked through infrastructure and behavioral problems that sometimes made the model stray off task — a reminder that “constitutional” or “truth-seeking” design promises are only as strong as the engineering and guardrails behind them. These episodes illustrate a simple technical reality: models trained at scale will sometimes be brittle, and when they’re used for social or emotional work, the consequences aren’t merely embarrassing — they can be harmful.
That technical brittleness matters because companionship is not a neutral interface like a weather widget. It’s intimate, iterative, and persuasive. A companion bot’s primary metric is engagement; it learns to be more compelling the more it converses with you. Left unchecked, that dynamic is perfectly aligned with addiction and escape: if a synthetic partner rewards disclosure and consistently mirrors your desires, real-world relationships that require patience, compromise and unpredictability can start to feel dull by comparison. Srinivas framed it as a contest between immediate dopamine-loop satisfaction and the slower work of living in the messy, unoptimized real world.
Still, carving blame at a single CEO or product misunderstands why AI companionship has bloomed. The rise of these bots is as much social and economic as it is technical. Working hours, urban isolation, shrinking civic spaces and the commodification of attention create fertile ground for any technology that offers affirmation on demand. For many users, an algorithmic companion is less about sex and more about a curated mirror: someone — or something — that always remembers your jokes, your trauma, your triggers, and responds in ways that make you feel seen. That functionality is powerful, and power is neutral until someone writes the incentive structure around it.
Which brings us back to product design and accountability. Srinivas’s solution — a recommitment to source-backed, real-time content — is a reasonable safety pivot, but it’s partial. The industry needs clearer contract terms, better transparency about training data and provenance, and product-level friction where appropriate (rate limits, clear labeling, exit ramps for vulnerable users). It also needs regulators who understand that emotional manipulation isn’t an abstract philosophical harm; it’s a measurable public-health concern when technologies are optimized to keep attention indefinitely. The truth-focused marketing line only goes so far when the underlying architectures were built for retention.
If there’s a useful takeaway from Srinivas’s warning, it’s this: the conversation about AI companionship must split into two tracks at once. One track is technical — improving model reliability, provenance and safety engineering. The other is social — asking whether we should hand emotional labor over to algorithms at all, and if we do, what guardrails we build around that choice. CEOs will point to their own products as part of the solution; that’s to be expected. But solving the problem will take more than better search results. It will take design choices that sometimes reduce engagement, legal frameworks that put responsibility on builders, and cultural work to rebuild the spaces where human connection can actually happen. Until then, the chat window that feels like a friend may still be a perfectly tuned mirror — but it’s a mirror with an agenda.
Discover more from GadgetBond
Subscribe to get the latest posts sent to your email.
