Instagram Bans Self-Harm Images After Teen’s Suicide
Facebook, which owns Instagram, will use its image recognition technology to help the social media app implement its new self-harm policy.
Instagram has agreed to ban graphic images related to self-harm after a father claims the social media app contributed to his daughter’s suicide.
Instagram chief Adam Mosseri made the announcement Thursday, stating the company will also ban non-graphic, self-harm related content in its search feature and through hashtags, reports CBS News.
“We need to do more to consider the effect of these images on other people who might see them. This is a difficult but important balance to get right,” he said. “We will get better and we are committed to finding and removing this content at scale, and working with experts and the wider industry to find ways to support people when they’re most in need.”
Mosseri added the company will not completely ban non-graphic, self-harm content because “we don’t want to stigmatize or isolate people who may be in distress and posting self-harm related content as a cry for help.”
Facebook, which owns Instagram, said in a statement that independent experts have advised them to “allow people to share admissions of self-harm and suicidal thoughts but should not allow people to share content promoting it.”
Facebook will also use its investment in image recognition technology to help Instagram implement its new policy, reports The Guardian.
“The more people report such images to the platform, the better the algorithm becomes in recognizing such images and becomes quicker in removing them,” said cybersecurity expert Jake Moore. “It is, therefore, a joint effort from both Instagram and its users to remove self-harm images, which will take time.”
The call for change was supported by the British government after the family of 14-year-old Molly Russell found material related to depression and suicide on her Instagram account. She took her own life in 2017.
Her father, Ian Russell, said the content his daughter viewed on Instagram played a role in her suicide.
The changes were announced after the company and other tech firms, including Facebook, Snapchat and Twitter, met with British Health Secretary Matt Hancock and representatives from the Samaritans, a mental health charity that works to prevent suicide.
Leading in Turbulent Times: Effective Campus Public Safety Leadership for the 21st Century
This new webcast will discuss how campus public safety leaders can effectively incorporate Clery Act, Title IX, customer service, “helicopter” parents, emergency notification, town-gown relationships, brand management, Greek Life, student recruitment, faculty, and more into their roles and develop the necessary skills to successfully lead their departments. Register today to attend this free webcast!