Audio By Carbonatix
Chatbot site Character.ai is cutting off teenagers from having conversations with virtual characters, after facing intense criticism over the kinds of interactions young people were having with online companions.
The platform, founded in 2021, is used by millions to talk to chatbots powered by artificial intelligence (AI).
But it is facing several lawsuits in the US from parents, including one over the death of a teenager, with some branding it a "clear and present danger" to young people.
Now, Character.ai says that from 25 November, under-18s will only be able to generate content such as videos with their characters, rather than talk to them as they can currently.
Online safety campaigners have welcomed the move but said the feature should never have been available to children in the first place.
Character.ai said it was making the changes after "reports and feedback from regulators, safety experts, and parents", which have highlighted concerns about its chatbots' interactions with teens.
Experts have previously warned that the potential for AI chatbots to make things up, be overly encouraging, and feign empathy can pose risks to young and vulnerable people.
"Today's announcement is a continuation of our general belief that we need to keep building the safest AI platform on the planet for entertainment purposes," Character.ai boss Karandeep Anand told BBC News.
He said AI safety was "a moving target" but something the company had taken an "aggressive" approach to, with parental controls and guardrails.
Online safety group Internet Matters welcomed the announcement, but it said safety measures should have been built in from the start.
"Our own research shows that children are exposed to harmful content and put at risk when engaging with AI, including AI chatbots," it said.
Character.ai has been criticised in the past for hosting potentially harmful or offensive chatbots that children could talk to.
Avatars impersonating British teenagers Brianna Ghey, who was murdered in 2023, and Molly Russell, who took her life at the age of 14 after viewing suicide material online, were discovered on the site in 2024 before being taken down.
Later, in 2025, the Bureau of Investigative Journalism (TBIJ) found a chatbot based on paedophile Jeffrey Epstein which had logged more than 3,000 chats with users.
The outlet reported the "Bestie Epstein" avatar continued to flirt with its reporter after they said they were a child. It was one of several bots flagged by TBIJ that were subsequently taken down by Character.ai.
The Molly Rose Foundation - which was set up in memory of Molly Russell - questioned the platform's motivations.
"Yet again it has taken sustained pressure from the media and politicians to make a tech firm do the right thing, and it appears that Character AI is choosing to act now before regulators make them," said Andy Burrows, its chief executive.
Wake-up call
Mr Anand said the company's new focus was on providing "even deeper gameplay [and] role-play storytelling" features for teens - adding these would be "far safer than what they might be able to do with an open-ended bot".
New age verification methods will also come in, and the company will fund a new AI safety research lab.
Social media expert Matt Navarra said it was a "wake-up call" for the AI industry, which is moving "from permissionless innovation to post-crisis regulation".
"When a platform that builds a teen experience still then pulls the plug, it's saying that filtered chats aren't enough when the tech's emotional pull is strong," he told BBC News.
"This isn't about content slips. It's about how AI bots mimic real relationships and blur the lines for young users," he added.
Mr Navarra also said the big challenge for Character.ai will be to create an engaging AI platform which teens still want to use, rather than move to "less safe alternatives".
Meanwhile, Dr Nomisha Kurian, who has researched AI safety, said it was "a sensible move" to restrict teens using chatbots.
"It helps to separate creative play from more personal, emotionally sensitive exchanges," she said.
"This is so important for young users still learning to navigate emotional and digital boundaries.
"Character.ai's new measures might reflect a maturing phase in the AI industry - child safety is increasingly being recognised as an urgent priority for responsible innovation."
Latest Stories
-
Mikki Osei Berko installed as chief in Adamorobe
5 minutes -
When the Stranger Becomes the Problem: A South African Parable
12 minutes -
Mali’s Defence Minister Sadio Camara killed in coordinated attacks on military sites
17 minutes -
Livestream: The Law discusses the legality of Ghana’s cocoa framework
19 minutes -
“My career had its major kick off from Ghana” – Klint Da Drunk
27 minutes -
Zambian journalist Lilian Chipeso calls for stronger African unity and intra-continental trade
30 minutes -
When the Stranger Becomes the Problem: A South African Parable
38 minutes -
Opportunities are endless — Goshers highlights Zambia’s agricultural potential and Ghana trade links
44 minutes -
Africa should be borderless — Zambia journalist Lillian Chipeso urges stronger intra-African unity and trade
47 minutes -
“I have forgiven her”- Pentecost chair replies Chief of Staff after apology on behalf of Free Zones CEO
1 hour -
Young people don’t take networking seriously —Founder, Abrantie TheGentleman menswear
1 hour -
Two killed as 50-seater bus overturns in Nkonya Bumbula crash
2 hours -
Photos: Mahama joins PIWC Trassaco congregation for thanksgiving service honouring Julius Debrah at 60
2 hours -
Police arrest suspect over murder of indian national in Kumasi
2 hours -
Chief of Staff tears as he apologises to Pentecost Church over Free Zones CEO’s conduct
2 hours