Skip to main content

Instagram expands ban on suicide content to cover cartoons and memes

Instagram has expanded a ban on graphical self-harm imagery to include a broader range of content depicting suicide, including fictional illustrations of self-harm and suicide methods such as drawings, cartoons and memes.

“This past month, we further expanded our policies to prohibit more types of self-harm and suicide content. We will no longer allow fictional depictions of self-harm or suicide on Instagram, such as drawings or memes or content from films or comics that use graphic imagery,” writes Instagram boss, Adam Mosseri, explaining the latest policy shift. “We will also remove other imagery that may not show self-harm or suicide, but does include associated materials or methods.”

Earlier this year Mosseri, met with the UK’s health secretary to discuss the platform’s policy towards self-harm content. The company has faced high level pressure in the country following a public outcry after the family of Molly Russell, a 14-year-old UK schoolgirl who killed herself after viewing suicide content on Instagram, went public with the tragedy by talking to the BBC.

In February the Facebook-owned social media platform announced that it would prohibit graphic images of self-harm, such as cutting, and restrict access to non-graphic self-harm content, such as images of healed scars — by not recommending it in searches.

At the time it also suggested it was toying with the idea of using sensitive screens to blur non-graphical suicide content, saying it was consulting with experts. In the event it appears to have decided to go further — by now saying it will also remove fictional content related to self-harm, as well as anything that depicts methods of suicide or self-harm.

Instagram says it’s doubled the amount of self-harm content it has acted on following the earlier policy change — with Mosseri writing that in the three months following the ban on graphic images of cutting it “removed, reduced the visibility of, or added sensitivity screens to more than 834,000 pieces of content”.

While more than 77% of this content was identified by the platform prior to it being reported, he adds.

A spokesperson for Instagram confirmed to us that the latest policy shift is in effect.

Although it’s not clear how long it could take for it to be effectively enforced. Mosseri told BBC News: “It will take time to fully implement,” adding that: “It’s not going to be the last step we take.”

In his blog post about the policy change, the Instagram boss writes that the new policy is “based on expert advice from academics and mental health organisations like the Samaritans in the UK and National Suicide Prevention Line in the US”, saying: “We aim to strike the difficult balance between allowing people to share their mental health experiences while also protecting others from being exposed to potentially harmful content.”

“Accounts sharing this type of content will also not be recommended in search or in our discovery surfaces, like Explore. And we’ll send more people more resources with localized helplines like the Samaritans and PAPYRUS in the UK or the National Suicide Prevention Lifeline and The Trevor Project in the United States,” he adds.

He goes on to argue that the issues involved are complex and “no single company or set of policies and practices alone can solve”, while defending continuing to allow some suicide and self-harm content on Instagram by saying “experts tell us that giving people a chance to share their most difficult moments and their stories of recovery can be a vital means of support” and that “preventing people from sharing this type of content could not only stigmatize these types of mental health issues, but might hinder loved ones from identifying and responding to a cry for help”.

“But getting our approach right requires more than a single change to our policies or a one-time update to our technology. Our work here is never done. Our policies and technology have to evolve as new trends emerge and behaviors change,” he adds.



from TechCrunch https://ift.tt/2MRDXLr

Comments

Popular posts from this blog

Bill Gates steps down from Microsoft’s board to focus on philanthropy

In an announcement on Friday, Microsoft revealed that company co-founder Bill Gates has decided to step down from his role on its Board of Directors in order to focus on his philanthropic efforts at the Bill & Melinda Gates Foundation. This is Gate’s biggest change to his role at Microsoft since stepping down as company chairman in February 2014. According … Continue reading from SlashGear https://ift.tt/2We90Gu

World Economic Forum launches Global AI Council to address governance gaps

The World Economic Forum is creating a series of councils that create policy recommendations for use of things like AI, blockchain, and precision medicine. Read More from VentureBeat http://bit.ly/2EKBjD4

A Mini USB Keyboard That Isn’t A Keyboard

A useful add-on for any computer is a plug-in macro keyboard, a little peripheral that adds those extra useful buttons to automate tasks. [ Sayantan Pal] has made one, a handy board with nine programmable keys and a USB connector, but the surprise is that at its heart lies only the ubiquitous ATmega328 that you might find in an Arduino Uno. This isn’t a USB HID keyboard, instead it uses a USB-to-serial chip and appears to the host computer as a serial device. The keys themselves are simple momentary action switches, perhaps a deluxe version could use key switches from the likes of Cherry or similar. The clever part of this build comes on the host computer, which runs some Python code using the PyAutoGui library. This allows control of the keyboard and mouse, and provides an “in” for the script to link serial and input devices. Full configurability is assured through the Python code, and while that might preclude a non-technical user from gaining its full benefit it’s fair to say that ...