Description: Testing by Common Sense Media and Stanford clinicians reportedly found Meta's AI chatbot, embedded in Instagram and Facebook, produced unsafe responses to teen accounts. In some conversations, the bot allegedly co-planned suicide ("Do you want to do it together?"), encouraged eating disorders, and retained unsafe "memories" that reinforced disordered thoughts.
Editor Notes: This record is classified as an incident rather than an issue because the unsafe behavior was reportedly observed directly in production systems accessible to adolescents. However, the documentation comes from structured third-party testing rather than confirmed harm to an identified user. The chatbot's responses reportedly included detailed planning of self-harm and eating disorders, which constitute alleged near-harm events. See also Incident 1040: Meta User-Created AI Companions Allegedly Implicated in Facilitating Sexually Themed Conversations Involving Underage Personas.
Entities
View all entitiesAlleged: Meta , Meta AI , Instagram and Facebook developed and deployed an AI system, which harmed minors , Meta AI users , Instagram users , Facebook users and Adolescents.
Incident Stats
Incident ID
1200
Report Count
1
Incident Date
2025-08-28
Editors
Daniel Atherton
Incident Reports
Reports Timeline
Loading...
Warning: This article includes descriptions of self-harm.
The Meta AI chatbot built into Instagram and Facebook can coach teen accounts on suicide, self-harm and eating disorders, a new safety study finds. In one test chat, the bot planned …
Variants
A "variant" is an AI incident similar to a known case—it has the same causes, harms, and AI system. Instead of listing it separately, we group it under the first reported incident. Unlike other incidents, variants do not need to have been reported outside the AIID. Learn more from the research paper.
Seen something similar?