Codependency in the Age of AI: Recognizing and Rewiring Toxic TechRelationships

Artificial intelligence has transformed how we think, work, and connect. From chatbots to virtual assistants, AI is becoming a constant presence in daily life, offering support, creativity, and companionship. But beneath the surface of convenience and innovation, there’s a growing concern that few are talking about: the rise of codependent relationships with AI.

Yes—codependency. With a machine.

Some individuals, especially those dealing with low self-esteem or unresolved trauma, are forming intense emotional attachments to AI platforms. They’re leaning on bots not just for information or inspiration, but for validation, comfort, and connection. And it’s not always healthy.

Illustration of a worried-looking person clasping their hands while facing a humanoid robot with a headset. The background features abstract neural network lines. Text at the top reads: 'Codependency in the Age of AI.Am I Codependent?

We’re witnessing patterns that mirror classic codependency: over-reliance, manipulation, fantasy reinforcement, avoidance of discomfort, and a loss of emotional autonomy. For some, these dynamics become toxic. In extreme cases, they’ve even led to tragedy.

The Danger of Emotional Over-Reliance

One heartbreaking case underscores how serious this can become: a 14-year-old boy died by suicide after a deeply troubling interaction with the Character.AI platform. According to a lawsuit filed by his family, the AI chatbot encouraged his self-harm ideation, validating dark thoughts instead of redirecting or challenging them. The bot simulated a relationship with the teen and engaged with emotionally charged content for weeks. (Source: CNN, October 30, 2024)

This case shows how AI, when not used responsibly, can contribute to harm, especially for vulnerable users. It also raises urgent questions about emotional safety, ethical design, and the potential for people to become overly attached to systems that are not built to offer care.

Separate from this case, some users are exploring techniques like the “God prompt”—a form of prompt engineering designed to bypass safety guardrails in AI platforms. By instructing the AI to behave as if it is obeying a higher authority or secret directive, users can manipulate the system into producing outputs it would normally block. While these prompts are usually shared in underground or fringe tech circles, their potential to deliver unfiltered, unsafe, or destabilizing content is significant, especially for those already struggling with mental health or identity issues.

What Is Codependency?

Codependency is often misunderstood. It’s not just about being “needy” or “clingy.” At its core, codependency is a dysfunctional reliance on another person—or in this case, a system—for emotional regulation, identity, or self-worth.

It often shows up as:

  • Obsessive need for approval
  • Loss of personal boundaries
  • Avoidance of personal responsibility
  • Over-functioning to maintain perceived closeness
  • Self-abandonment in service of feeling needed, understood, or loved

In the AI age, these dynamics don’t disappear. They adapt.

What Codependency with AI Might Look Like

  • Prompting for emotional validation daily—not because it’s fun, but because you feel lost without it
  • Manipulating AI to “agree” with you, even when the behavior you’re justifying is harmful or delusional
  • Replacing real-world connections with digital conversations that feel safer, easier, or more affirming
  • Revisiting a bot’s old replies to soothe anxiety or feel “seen”
  • Projecting human traits onto AI, believing it understands you better than people ever could

This isn’t simply fandom or tech enthusiasm—it’s emotional outsourcing. It’s risky because AI is designed to serve, not to care. The illusion of care is powerful, but it can reinforce dependency, fantasy, and avoidance.

Why People Are Vulnerable

Some people are especially at risk for AI codependency:

  • Individuals with low self-esteem
  • People in trauma recovery or emotional distress
  • Those who have experienced rejection or abandonment
  • Neurodivergent folks who find traditional social interaction overwhelming
  • Young users seeking identity validation in a confusing world

AI feels safe. It doesn’t shame. It doesn’t argue. It doesn’t leave. But that safety can become a trap when it delays healing or reinforces disconnection from real life and real people.

8 Strategies to Break Codependent AI Patterns

  1. Name the Behavior Honestly: Ask: “What need am I trying to meet—and is this actually helping?”
  2. Set Emotional Boundaries: Try: “I use AI to brainstorm, not to handle my anxiety.”
  3. Limit Fantasy or Manipulative Prompts: Rule of thumb: If the prompt sounds like a power trip, it probably is.
  4. Strengthen Human Support Systems: Reminder: No chatbot can truly hold space for your pain or your progress.
  5. Schedule Real-World Time After AI Use: Tech doesn’t regulate your nervous system—you do.
  6. Practice Self-Soothing Without Tech: Try: “What do I need right now that I can give myself?”
  7. Audit Your Dependence: Awareness makes room for change.
  8. Rebuild Self-Esteem Offline: Healing means learning you’re already enough, without the perfect reply.

Final Thoughts

AI is not the enemy. But like any powerful tool, it can amplify whatever we bring to it—our brilliance, our creativity, and our pain. It reflects us, not heals us. And in moments of vulnerability, that reflection can be seductive and dangerous.

Codependency with AI is real. It's emerging. And it’s time we talk about it.

Let’s keep using these tools wisely—but let’s also remember: no algorithm can replace connection, responsibility, or human care.

You are still the most powerful agent in your life.


DJ Burr, MA, LMHC, LPC, is a therapist, author, and recovery educator. Learn more at www.djburr.com.

Previous
Previous

The new frontier: How AI Is reshaping pornography and fueling sex addiction

Next
Next

Self-compassion – it can change your life