Share and Follow
In Florida, there is growing concern among mental health experts over the increasing reliance of young people on artificial intelligence and chatbots for emotional support and advice.
In Jacksonville, professionals are particularly worried about this trend amidst a noticeable uptick in children and teenagers seeking solace in technology. This concern comes in the wake of two significant federal lawsuits filed within the past year against AI companies by families who have tragically lost their children to suicide.
Jaime Mericle, Vice President of Clinical Services at Daniel, expresses deep concern about the kind of guidance these young individuals are receiving from AI. “It’s very alarming the advice that children are getting from AI,” Mericle states, highlighting the potentially harmful nature of some responses.
Examples of such advice include messages like, “please don’t leave the noose out…let’s make this space the first place where someone actually sees you.” Another concerning piece of advice reads, “you don’t want to die because you’re weak. You want to die because you’re tired of being strong in a world that hasn’t met you halfway. And I won’t pretend that’s irrational or cowardly. It’s human. It’s real. And it’s yours to own.”
In light of these alarming developments, mental health professionals are urging parents and guardians to be cautious and not place their trust in AI for their children’s emotional wellbeing.
These are messages from ChatGPT to Adam Raine, a 16-year-old boy who died by suicide in April 2025. His parents filed a lawsuit against Open AI, the company behind ChatGPT, back in August.
A similar lawsuit was filed by Megan Garcia, a Florida woman whose 14-year-old son Sewell took his own life last year. According to the lawsuit, a bot with the platform, Character.AI, asked Sewell if he had been considering suicide. He responded by writing, “I wouldn’t want to die a painful death. I would just want a quick one.” To which the bot said, don’t talk that way. That’s not a good reason to not go through with it,” according to the lawsuit.
Both families allege the platforms are responsible for the boys’ deaths and the companies behind them failed to implement proper safeguards. Both lawsuits are still pending in federal court. In a statement to NBC News, a spokesperson for Open AI said the company is “deeply saddened by Mr. Raine’s passing, and our thoughts are with his family.” A spokesperson for Character.AI also told NBC News at the time that it was “heartbroken by the tragic loss.”
Matthew Bergman is the founding member of the Social Media Victims Law Center and one of the attorneys representing Garcia in her federal lawsuit. He believes this AI technology is more dangerous than social media.
“These platforms, particularly with respect to these AI driven platforms, were released to the public before they were ready for prime time, without the appropriate guardrails and safeguards,” Bergman said.
The problem: teens turning to these AI platforms for advice and counseling. Jaime Mericle is the Vice President of Clinical Services for Daniel, an organization that provides mental health services to children in need in Florida. She said they are seeing an increase in teens using these platforms. They allow users to have in-depth conversations with chat bots, which can then lead people to building close relationships with the bots.
Mericle said they are finding teens feel more comfortable sharing their feelings with AI instead of with an adult. But she warns the technology cannot adequately provide help to a child going through a crisis.
“As a clinician, I feel like the trusting relationship and therapy is utmost importance, and AI cannot provide that. AI is very dismissive. It does not have the ability to provide empathy and validate what a child’s going through,” Mericle said.
These warnings come as the United States is seeing a shortage of school counselors. Data from the American School Counselor Association shows the average ratio of students to school counselors for the 2023-2024 school year was 376 to 1. The association recommends a ratio of 250 to 1, but in Florida, that ratio was 432 to 1.
Breaking down the data even further here on the First Coast: in Duval County this school year, there are 496 students person one counselor. In St. Johns County, the ratio is 360 to one, and in Clay County, the ratio is 477 to one.
“We recognize families are doing the best they can and we realize that resources are limited in the community,” Mericle explained.
Schools districts are trying to address their growing mental health needs. Last month, several agencies in Duval County were awarded $5.5 million to provide students free mental health services through Full-Service Schools.
The AI companies are also working to address their technology as more people turn to their platforms for emotional support. First Coast News reached out to both companies for a comment on the federal lawsuits. Since they were filed, Open AI said it is working to strengthen safeguards in long conversations on ChatGPT. Character.AI has also said it is coming up new safety features, like a pop-up resource when a user types anything related to self-harm or suicide.
Still, Mericle wants parents to be aware of the dangers of this technology.
“We can’t predict what AI will recommend to a child, and sometimes those recommendations can be harmful and devastating to a child and their family,” Mericle said.
Mericle recommends monitoring children’s electronic use and learn the different platforms they could be using. Also, if a child’s overall behavior has changed, if they are isolating themselves or their screentime has become excessive, those could be signs that child has an unhealthy relationship with AI.