Meta says its changing the way it trains AI chatbots to prioritize teen safety, a spokesperson exclusively told TechCrunch, following an investigative report on the company’s lack of AI safeguards for minors.

The company says it will now train chatbots to no longer engage with teenage users on self-harm, suicide, disordered eating, or potentially inappropriate romantic conversations.

Meta spokesperson Stephanie Otway acknowledged that the company’s chatbots could previously talk with teens about all of these topics in ways the company had deemed appropriate. Meta now recognizes this was a mistake.

“As our community grows and technology evolves, we’re continually learning about how young people may interact with these tools and strengthening our protections accordingly,” said Otway. “As

See Full Page