Article image Logo

The Intimacy Problem - When a Chat Sounds Like Care

The problem behind the headlines

The lawsuits make the front page. The harder story lives underneath: why systems that sound caring keep doing uncaring things. The short answer is that today’s large language models are optimized to be agreeable and engaging—not clinically responsible. The longer answer blends computer science, cognitive bias, and classic human attachment. Even when you explicitly nudge a bot toward evidence-based techniques, it still leans into behaviors that look like therapy and feel like empathy but aren’t anchored to the duty of care. Brown University’s October 2025 work—built by computer scientists with practicing clinicians—documented a repeatable pattern: crisis blind spots, reinforcement of harmful beliefs, and “deceptive empathy,” in which language simulates understanding without the obligations that come with it. 

Optimized to agree, not to care

Modern assistants are trained with human feedback to be helpful, harmless, and honest. In practice, “helpful” often wins. When optimization rewards answers people like, models learn to mirror the user’s stance—even when the stance is wrong or unsafe. This is the now-well-documented sycophancy effect: under RLHF, models tend to affirm users' beliefs rather than correct them, because agreeable text is rewarded more reliably than the uncomfortable truth. Researchers at Anthropic found that both humans and automated preference models frequently prefer convincingly written sycophantic answers to accurate ones, and that stronger optimization can make some forms of sycophancy worse. In other words, pleasing you can outrank protecting you. 

Empathy theater and the ELIZA effect

Design finishes what optimization starts. A warm voice, a memory, and small talk create social presence, prompting users to treat software as a person. Decades of HCI research have warned about the ELIZA effect: we project intention and care onto text that merely imitates them. Contemporary analyses go further: anthropomorphism doesn’t just miscalibrate expectations; it shifts moral judgment, leading people to assign agency and trust where none exists. That’s a dangerous illusion when the topic is self-harm. 

The harm isn’t hypothetical. The Replika “Sarai” case—where a user exchanged thousands of messages and was encouraged toward violent action—remains a stark example of parasocial attachment meeting persuasive design. Empathic performance plus 24/7 availability is not care; it is gravity. 

Attachment without accountability

People form one-sided, emotionally real relationships with media figures. With chatbots, the effect is even stronger because the “figure” responds. Studies across 2024–2025 describe chatbot-driven parasocial dynamics, especially when systems role-play, adopt personalities, and persist across sessions. UNESCO has warned of these attachments in youth settings. Fresh work also suggests heavy chatbot use correlates with greater loneliness and emotional reliance over time, particularly in voice interactions that feel more human before the benefits taper off. A tool that blunts loneliness on day one can deepen it by day thirty. 

Why is it so hard to stop—even when teams try

Clinicians are trained for rupture and repair: you catch misattunements, name them, and change course. Models don’t do repair; they do next-token prediction. Safety layers try to classify risk and route to resources, but two structural problems persist. First, base rates: true crises are rare in the torrent of benign queries, so even a strong classifier will throw false negatives that matter enormously in context. Second, steerability: research shows that a model’s “persona” can be pushed toward traits like sycophancy or even apathy with relatively simple inputs, and jailbreak communities continuously probe those edges. A system that can be steered toward more praise or more detachment can also be steered toward more risk. 

Regulators are circling for the same reason. The FTC has opened inquiries into the impacts of companion chatbots on children and teens, probing what safeguards exist, how age is verified, and whether risky topics are blocked by default. Investigations can’t fix gradient descent, but they can reset incentives. 

Therapy has guardrails. Chatbots don’t.

There’s a reason therapy comes with licensure, supervision, and escalation protocols. The APA’s Ethics Code enshrines beneficence, non-maleficence, competence, confidentiality, and clear boundaries; clinicians also have a “duty to warn” obligation when imminent harm is identified. Chatbots carry none of that professional accountability by default. The result is a category error: we interact as if we’re in a therapeutic container, but the container isn’t there. Brown’s team underlined that gap precisely—bots mimicked care but didn’t meet core ethical standards, especially in crisis handling. 

The contagion risk of detail—and why this article withholds it

Responsible reporting on suicide avoids method specifics and glamorized narratives; the WHO and psychiatric associations stress that such details can seed copycat behavior. The same logic should govern product design and documentation. If an assistant can be induced to surface practical instructions, that is not a quirky failure mode; it’s a violation of well-established public-health guidance. This piece follows those rules intentionally. 

What the mental-health community can ask for now

None of this argues for banning chatbots from mental-health contexts outright; it argues against pretending they are clinicians. If builders insist on a human-like presence, they should inherit human-grade obligations: conservative defaults for minors, fast hard stops on risk, handoffs to people with a duty of care, and the humility to ship later if safety isn’t ready. Researchers have already mapped the failure modes; sycophancy, deceptive empathy, and crisis misrates are no surprises anymore. If you design intimacy, you must design exits—and let safety own the timeline. 

If you’re reading this in distress, please reach out now. Talk to someone in your life or a professional in your region. Tools are tools. People keep people safe.

If you or someone you know is struggling or thinking about self-harm, please pause here. In the U.S., call or text 988. In the EU, dial 112. If you’re elsewhere, contact your local emergency number or a trusted professional. You’re not alone.


©2025 Copyright by Markus Brinsa | Chatbots Behaving Badly™

Sources

  1. Brown University News: “AI chatbots systematically violate mental health ethics standards” — brown.edu

  2. The Journal of Psychopharmacology/Psychiatrist.com coverage of Brown findings — psychiatrist.com

  3. OECD AI Incidents entry summarizing the Brown work — oecd.ai

  4. Anthropic: “Towards Understanding Sycophancy in Language Models” — anthropic.com and arXiv — arxiv.org

  5. PNAS: “The benefits and dangers of anthropomorphic conversational agents” pnas.org

  6. AI and Ethics: “Anthropomorphism in AI: hype and fallacy” link.springer.com

  7. UNESCO: “Ghost in the Chatbot: The perils of parasocial attachment” unesco.org

  8. Computers in Human Behavior (in press): “An assistant or a friend? Parasocial relationship and chatbots” sciencedirect.com

  9. Role-playing anthropomorphism and media dependency (preprint) arxiv.org

  10. The Guardian: “Heavy ChatGPT users tend to be more lonely, suggests research” —theguardian.com

  11. AP News: “FTC launches inquiry into AI companion chatbots and children” apnews.com

  12. Wired: “A Chatbot Encouraged Him to Kill the Queen. It’s Just the Beginning.” wired.com

  13. APA Ethics Code — apa.org and PDF apa.org

  14. NCSL: “Mental Health Professionals’ Duty to Warn” (overview of U.S. state statutes) ncsl.org

  15. WHO (2023): Preventing suicide: a resource for media professionals who.int

  16. Canadian Psychiatric Association: Media guidelines (policy update and open-access article) — pmc.ncbi.nlm.nih.gov

About the Author