MUMBAI: Suicide and self-harm are complex mental health issues that can have devastating consequences. Meta said that it has worked with experts for years – including its suicide and self-harm advisory group and members of Meta’s own safety teams – to develop an informed and thoughtful approach to suicide and self harm content shared on its apps.
Yet, like many other types of potentially problematic content, suicide and self-harm content is not limited to any one platform. To be truly effective in responding to this content, tech companies need to work together it said. That’s why it has worked with the Mental Health Coalition to establish Thrive, the first signal sharing programme to share signals about violating suicide and self-harm content.
Through Thrive, participating tech companies will be able to share signals about violating suicide or self-harm content so that other companies can investigate and take action if the same or similar content is being shared on their platforms. Meta is providing the technical infrastructure that underpins Thrive – the same technology we provide to the Tech Coalition’s Lantern programme – which enables signals to be shared securely.
In a blog post Meta’s head of safety Antigone Davis wrote, “Participating companies will start by sharing hashes – numerical codes that correspond to violating content – of images and videos showing graphic suicide and self-harm, and of content depicting or encouraging viral suicide or self-harm challenges. We’re prioritizing this content because of its propensity to spread across different platforms quickly. These initial signals represent content only, and will not include identifiable information about any accounts or individuals.
“Thrive builds on the work we already do at Meta to remove harmful content that shows graphic imagery or encourages suicide or self-harm, while still giving space for people to talk about their own experiences. We also provide support to those sharing and searching for content related to suicide or self-harm by connecting them to local organizations around the world, including Suicide and Crisis Lifeline and Crisis Text Line in the US.”
“Between April and June this year, we took action on over 12 million pieces of suicide and self-harm content on Facebook and Instagram. While we allow people to discuss their experiences with suicide and self-harm – as long as it’s not graphic or promotional – this year we’ve taken important steps to make this content harder to find in Search and to hide it completely from teens, even if it’s shared by someone they follow.”
“Thrive will help keep people safe not just on Meta’s apps, but across all the apps and services they use. We’re proud to work with the Mental Health Coalition and our industry partners Snap and TikTok on this vital work.”