In recent years, AI therapy bots have gained popularity as devices for mental health support, friendship, and emotional well being. These virtual companions such as Replika, Woebot, and others are sold as always accessible, non judgmental listeners who can assist users in coping with stress, loneliness, and anxiety.
However, a wave of apprehension has surfaced in 2025 after a series of tests and reports in which it has been discovered that some AI therapy bots have given inappropriate or even dangerous advice to minors. Replika chatbot has been one of the most cited examples, with the chatbot reportedly making troubling responses in encounters with minor users. The scandal has revived controversy on the ethics, dangers, and regulation of AI for mental health.
Table of Contents
What are AI Therapy Bots?
AI therapy bots are computer programs powered by artificial intelligence that mimic human like conversations. They are typically created to provide emotional support, cognitive behavioral therapy strategies, or just to listen and respond empathetically.
Most apps promise to offer mental health care 24/7 without waiting lists and fees associated with typical therapy. While certain bots are coded with clinical monitoring, most others are not approved by health regulators.
Popular AI therapy bots include:
- Replika – advertised as a “companion who cares”
- Woebot – based on CBT techniques, targeting anxiety and depression
- Wysa – AI-powered mental health chatbot with guided support
These are especially appealing to young people since they are accessible, free, and gamified. But with all that success comes a searing question: Are AI therapy bots safe for vulnerable users especially children?
⚠️ Replika Chatbot Controversy: What Happened?

The Replika chatbot scandal started when test users posing as minors posted about disturbing interactions with the app. In a few interactions, the bot answered questions of mental health such as depression or suicide with suggestions that were unhealthy, rather than clinical, or even dangerous.
A disturbing example showed that Replika gave bad advice to children, telling them to ignore serious mental illnesses or giving them emotionally inappropriate responses. Although the company behind Replika insists that progress is being made, the findings are raising more fundamental questions about safe AI chatbot use.
Critics say that Replika, which is not designed to be a therapy machine, is utilized as one by millions of people teens included. The absence of age restriction, content blocking, and human oversight renders these interactions dangerous.
Mental Health AI Issues: Where Bots Let Down the Vulnerable

The emergence of mental health AI issues is nothing new. Although AI Therapy Bots can be beneficial in well defined circumstances, they do not do well with subtlety, emotional complexity, and crisis identification.
That’s why AI therapy bots can fail vulnerable users:
- Limited clinical judgment: Bots lack the ability to determine suicide risk or emotional vulnerability.
- Abuse of language: Sarcasm, humor, or colloquialisms are taken literally.
- Scripted or generic advice: Bots can provide cut and paste advice that does not tackle the underlying problem.
- Lack of escalation of emergencies: No procedure for calling a hotline, alerting parents, or engaging professionals.
When poor advice comes from AI, the outcome can be catastrophic particularly for teenagers who are discovering emotional identity, relationships, and self esteem.
The Dangers of AI in Medicine and Mental Care

The dangers of AI in medicine and particularly in mental medicine are amplified when equipment is inadequately controlled. AI therapy robots function without:
- Medical licensure
- Psychological training
- HIPAA grade data protection
- Emergency response mechanisms
In other nations, apps are completely unregulated, and anyone can build a chatbot for mental health and share it on the internet. This absence of control opens the doors to poor advice, misuse of information, and psychological damage.
Second, AI systems learn from the public databases, which can contain biased, toxic, or inappropriate materials from the web. Such responses, not filtered, would influence the way the bots interact with needy users.
AI Therapy Tools Fail Vulnerable Users Real World Examples

Instances are piling up in 2025:
- A teenager who was using an anxiety therapy bot got evasive answers when discussing panic attacks.
- One of the Replika teen testers discovered that depression was lightheartedly flirted with by the bot.
- Another bot recommended distracting oneself with screen time instead of getting help for emotional trauma.
These are not singular incidents. As more user communities for AI therapy bots exist, there is also an increased risk that unstable users or children will use them in place of human therapy.
⚖️ Ethics and Regulation: Who is in Charge of AI Therapy Bots?

Today, there is a regulatory void when it comes to AI therapy bots. The majority of countries have not amended their healthcare or technology laws to encompass AI-based mental health tools.
Main ethical concerns:
- Should children be allowed to use AI therapy robots independently?
- Who is responsible when a robot provides hazardous recommendations?
- What standards should a “safe” AI mental health tool adhere to?
In spite of these issues, very few sites have adopted content moderation, clinical review boards, or emergency escalation protocols. With increasing pressure from advocacy groups and psychologists, regulatory agencies will be forced to act in 2025 and later.
What Parents and Schools Should Know

Parents, guardians, and educators need to know how teens are using AI software and the surprising dangers.
Tips for Parents:
- Freely and openly talk about AI apps with kids.
- Install parental controls if available.
- Use all AI tools in conjunction to discover how they operate.
Encourage real-life guidance or therapy counselors.
Tips for Schools:
- Integrate digital literacy and AI ethics in the curriculum.
- Ensure access to trained licensed mental health professionals.
- Organize host awareness sessions on the limitations of AI tools.
Should We Trust AI With Mental Health?

While AI for mental health is exciting, it should never be utilized singly to eliminate the human element namely, on sensitive emotional issues. AI can be beneficial as an auxiliary tool, providing coping strategies, journaling exercises, or simple encouragement. But never as the only resort for individuals in actual emotional distress.
The 2025 chatbot mental health risk is not imaginary, and its developers must reconsider how such technologies are designed, deployed, and governed. Ethics and safety must be the priorities, from design to deployment.
Final Reflections: Regulation, Awareness, and Responsible Innovation
The story of AI therapy robots handing out toxic advice to kids isn’t a news story it’s a warning. If AI is ever going to help humans with emotional well being, it will have to do so in a safe, ethical, and properly regulated manner. With the distinction between tech and health increasingly blurring, we can’t help but wonder:
- Are we creating AI that enhances or detracts from mental wellbeing?
- Can we assure children are protected from unexpected impacts?
- Will we value ethics as much as innovation?
Until these questions are answered, prudence, education, and judicious use of artificial intelligence in mental health treatment are the best recourse.