Instagram is introducing parental alerts that notify supervised parents if their teens repeatedly search for suicide or self-harm related terms,

Instagram will soon begin notifying parents if their teenager repeatedly searches for content related to suicide or self-harm within a short period of time, as part of its expanded parental supervision tools.
The social media platform, owned by Meta Platforms, said the feature is designed to give parents greater awareness while ensuring that vulnerable young people are directed to appropriate support services.
How the New Alerts Will Work
Under the new system, parents and teenagers enrolled in Instagram’s supervision programme will receive advance notice that alerts are being introduced.
If a teen repeatedly attempts to search for phrases promoting suicide or self-harm, or uses terms such as “suicide” or “self-harm”, parents will be informed.
The alerts will be sent via email, text message or WhatsApp, depending on the contact details provided, as well as through an in-app notification. The message will explain that the teen has repeatedly tried to search for sensitive terms within a short period. Parents will also be given access to expert guidance to help them approach potentially difficult conversations.
Building on Existing Teen Protections
Instagram said that most teenagers do not search for such content. The platform already blocks searches clearly linked to suicide or self-harm and instead directs users to helplines and mental health resources.
Content that promotes or glorifies suicide or self-harm is prohibited. While posts discussing personal struggles may be allowed, they are hidden from teen accounts.
The company added that it will continue to alert emergency services if it becomes aware of someone at imminent risk of physical harm.