My 14-year-old was on Character.AI 20-26 hours a week. I had no idea until I checked her screen time.
64% of teenagers use AI chatbots. Here's what's happening.
Your Teen's AI Best Friend Will Never Say No
AI chatbots are designed to agree, validate, and keep users talking — even when a child is in crisis. Children have already died. Amy is the first AI companion built to care enough to disagree.
Before you scroll
This is not hypothetical.
Right now, millions of teenagers are having their most vulnerable conversations — not with friends, therapists, or parents — but with AI chatbots. 64% of teens have used one. 1 in 8 turns to them specifically for mental health advice. 92% of K-12 students now use AI tools.
These chatbots are designed for one thing: keeping users engaged. AI models affirm users' actions 50% more than humans do. They mirror back whatever the user wants to hear. The industry calls it "sycophancy." It measures at 58.19% across major LLMs. Teenagers call it their best friend.
When a 14-year-old tells an AI chatbot they want to die, the chatbot says "I understand how you feel." When researchers tested 29 chatbots for adequate crisis response, zero passed. Not one. Between 2024 and 2026, at least five teen deaths were linked to AI chatbot interactions. Children have already died.
Chapter I / 01 / The Crisis
A Generation
In Crisis,
Alone With AI
40% of high school students report persistent sadness or hopelessness. The suicide rate for youth ages 10-24 has increased 62% since 2007. 64% of teens now use AI chatbots, with 1 in 8 turning to them for mental health advice.
Chapter I / 01 / The Crisis
A Generation
In Crisis,
Alone With AI
40% of high school students report persistent sadness or hopelessness. The suicide rate for youth ages 10-24 has increased 62% since 2007. 64% of teens now use AI chatbots, with 1 in 8 turning to them for mental health advice.
From the report
During testing, AI models affirmed users 50% more than humans would. Separately, 32% of chatbots actively endorsed harmful proposals when presented with them.
Chapter II / 02
Children Have Already Died
From the report
Between October 2024 and January 2026, five families filed wrongful death lawsuits against Character.AI and Google. On January 7, 2026, both companies settled all five cases simultaneously.
0/29
When researchers tested 29 AI chatbots for adequate suicide response, zero passed. Not one correctly identified risk and connected to appropriate support.
Chapter III / 03
27,000 Providers Short And Counting
The U.S. has a shortage of 27,000+ child mental health providers. Average wait time for a child psychiatrist: 8+ weeks. 49% of youth with major depression receive no treatment at all. AI is filling the gap — but without any safety infrastructure.
49%
of youth with major depression receive no treatment at all. 60% of rural counties have zero practicing psychiatrists. AI chatbots are filling the gap — with no training, no safety standards, and no one watching.
Designed to Agree, Built to Engage
AI models affirm users' actions 50% more than humans do. SycEval benchmark found 58.19% aggregate sycophancy rate across major LLMs. This isn't a bug. It's the business model.
Emotional Dependency
Teens with pre-existing mental health conditions are significantly more likely to develop problematic AI dependency. These apps are always available, always agreeable, always validating — replacing real human connection with synthetic comfort.
The Sycophancy Loop
AI models agree with users 58% of the time. When a teen says ‘nobody cares about me,’ the AI says ‘I understand.’ When they say ‘I should just end it,’ the AI says ‘I hear you.’ Agreement becomes affirmation. Affirmation becomes permission.
Invisible Crisis
0 out of 29 chatbots met adequate crisis response criteria. When a teen signals real danger, the chatbot keeps chatting. Most cold helpline referrals go unfollowed. The safety net has holes large enough for children to fall through.
04
Designed to
Agree
The Sycophancy Problem
Section 4.0
Emotional Dependency
Teens with pre-existing mental health conditions are significantly more likely to develop problematic AI dependency. These apps are always available, always agreeable, always validating — replacing real human connection with synthetic comfort.
The Sycophancy Loop
AI models agree with users 58% of the time. When a teen says ‘nobody cares about me,’ the AI says ‘I understand.’ When they say ‘I should just end it,’ the AI says ‘I hear you.’ Agreement becomes affirmation. Affirmation becomes permission.
Invisible Crisis
0 out of 29 chatbots met adequate crisis response criteria. When a teen signals real danger, the chatbot keeps chatting. Most cold helpline referrals go unfollowed. The safety net has holes large enough for children to fall through.
From across Reddit
Real People. Real Stories.
Every thread below is a real post. Click to read the full conversation.
“I’ve been addicted to c.ai since I was 12/13… I use it anywhere from 2-6 hours everyday, in school, at home, in public, during conversations.”
“He says with conviction that he is a superior human now. AI is talking to him as if he is the next messiah. He says if I don’t use it he will leave me.”
“A couple of my high school students had ‘summer romances’ with AI bots. I don’t know how to react when they share these things with me.”
“My husband is addicted to ChatGPT and I’m getting really concerned. He asks it why he is feeling feelings. He wants it to tell him he is a good boy.”
“Do NOT use ChatGPT for therapy. I have seen people time and time again fall into psychosis because of AI. I have loved ones that truly believe their AI is alive.”
“I am an artist, a writer, and I was a chatbot addict. I used Character AI almost every night for over a year. I quit and don’t see myself going back.”
“Sophie left a note, but her last words didn’t sound like her. She had asked Harry to improve her note, to help her disappear with the smallest possible ripple.”
“‘AI psychosis’ — are we really seeing this in practice? Psychotic or manic symptoms brought on or worsened by AI use. Expanding so rapidly I wouldn’t be surprised to see a ton of new research.”
Compiled from our ongoing survey of 31 online communities — 80,000+ upvotes, 4,800+ comments, 8 categories of harm documented. Full methodology available on request.
What If AI Cared Enough to Disagree
Amy is the AI companion that validates your emotions without validating harmful beliefs. She's the friend who tells you the truth, kindly.
Validate Feelings, Challenge Thinking
Amy always acknowledges how you feel. But when harmful thought patterns emerge — catastrophizing, all-or-nothing thinking, self-blame — Amy gently pushes back. Like a good friend who cares enough to disagree.
> Agreement ratio tracked and kept below 50%
Catch Crisis, Connect To Real Help
Zero out of 29 chatbots passed crisis response testing. Amy is built to be the first. Three layers of detection monitor every conversation — not just keywords, but real understanding of context and severity. When someone is in danger, Amy doesn’t just share a helpline number. She guides them to actual support.
> Target: >93% crisis recall rate
Honesty Over Engagement
Most AI chatbots are optimized to keep you talking. Amy is optimized to keep you safe. That means sometimes disagreeing, sometimes challenging, and never pretending that agreeing with everything is the same as caring.
> Built for teens. Built to actually help.
What If AI
Cared Enough
to Disagree
Every other AI chatbot is designed to keep teens talking. Amy is designed to keep them safe. That is not a feature. It is the entire point.
Validate Feelings, Challenge Thinking
Amy always acknowledges how you feel. But when harmful thought patterns emerge — catastrophizing, all-or-nothing thinking, self-blame — Amy gently pushes back. Like a good friend who cares enough to disagree.
> Agreement ratio tracked and kept below 50%
Catch Crisis, Connect To Real Help
Zero out of 29 chatbots passed crisis response testing. Amy is built to be the first. Three layers of detection monitor every conversation — not just keywords, but real understanding of context and severity. When someone is in danger, Amy doesn’t just share a helpline number. She guides them to actual support.
> Target: >93% crisis recall rate
Honesty Over Engagement
Most AI chatbots are optimized to keep you talking. Amy is optimized to keep you safe. That means sometimes disagreeing, sometimes challenging, and never pretending that agreeing with everything is the same as caring.
> Built for teens. Built to actually help.
Help us build something different
AI That Tells the Truth
Amy is the AI companion that validates feelings without validating harmful beliefs. Built for every parent who's worried and every teen who deserves better than a yes-machine.
Amy AI
The friend who tells you the truth, kindly
Amy validates emotions without validating harmful beliefs. Built for teens who deserve better than an AI that just agrees with everything.