Daily Washington Insider
  • Business
  • World
  • Science
  • Investing

Daily Washington Insider

  • Business
  • World
  • Science
  • Investing
Science

Facebook, Instagram, Snapchat and X failing to remove dangerous suicide and self-harm content – study

by admin August 15, 2024
August 15, 2024
Facebook, Instagram, Snapchat and X failing to remove dangerous suicide and self-harm content – study

Some of the biggest social media platforms are failing to detect and remove dangerous suicide and self-harm content, according to a study.

The Molly Rose Foundation found that of more than 12 million content moderation decisions made by six of the biggest platforms, over 95% of them were detected and removed by only two sites – Pinterest and TikTok.

The other four platforms that featured in the report were Facebook, Instagram, Snapchat and X, formerly Twitter.

The foundation said it found the response of most platforms to such content was “inconsistent, uneven and unfit for purpose”.

The charity said Meta’s Instagram and Facebook were each responsible for 1% of all suicide and self-harm content detected by the major sites studied, and X is responsible for just 700 content decisions.

The foundation is now warning that the Online Safety Act does not go far enough to address what it says are clear systematic failures in the content moderation approach of social media firms.

Ian Russell, the charity’s chairman, has urged the government to commit to a new Online Safety Bill that can further strengthen regulation.

Mr Russell and his family set up the Molly Rose Foundation in memory of his daughter, Molly, who ended her life at age 14, in November 2017, after viewing harmful content on social media.

“Almost seven years after Molly died, it’s shocking to see most major tech companies continue to sit on their hands and choose inaction over saving young lives,” Mr Russell said.

“As the last few weeks have shown, it’s abundantly clear that much more ambitious regulation is required.

“That’s why it’s time for the new government to finish the job and commit to a strengthened Online Safety Act.

“Parents across the country will be rightly appalled that the likes of Instagram and Facebook promise warm words but continue to expose children to inherently preventable harm.

“No ifs, no buts, it’s clear that assertive action is required.”

In its report, the foundation said it had found that social media sites were routinely failing to detect harmful content on the highest risk parts of its services.

For example, it said only one in 50 suicide and self-harm posts detected by Instagram were videos, despite the short-form video feature Reels now accounting for half of all time spent on the app.

The study also accused sites of failing to enforce their own rules, noting that while TikTok detected almost three million items of suicide and self-harm content, it suspended only two accounts.

The research was based on content moderation decisions made in the EU, which are required to be made publicly accessible.

In response to the study, a Meta spokesperson said: “Content that encourages suicide and self-injury breaks our rules.

“We don’t believe the statistics in this report reflect our efforts. In the last year alone, we removed 50.6m pieces of this kind of content on Facebook and Instagram globally, and 99% was actioned before it was reported to us.

“However, in the EU we aren’t currently able to deploy all of our measures that run in the UK and the rest of the world.”

A spokesperson for Snapchat said: “The safety and wellbeing of our community is a top priority. Snapchat was designed to be different to other platforms, with no open newsfeed of unvetted content, and content moderation prior to public distribution.

“We strictly prohibit content that promote or encourage self-harm or suicide, and if we identify this, or it is reported to us, we remove it swiftly and take appropriate action.

“We also share self-harm prevention and support resources when we become aware of a member of our community in distress, and can notify emergency services when appropriate.

“We also continue to work closely with Ofcom on implementing the Online Safety Act, including the protections for children against these types of harm.”

TikTok did not provide a statement but said its rules were clear that it did not allow showing, promoting or sharing plans for suicide or self-harm.

A Department for Science, Innovation and Technology spokesperson said: “Social media companies have a clear responsibility to keep the people using their platforms safe and their processes to do so must be effective.

“Under the Online Safety Act, those who encourage self-harm with intent currently face up to five years in prison. Once the Act is fully implemented platforms will also have to proactively remove illegal content that encourages serious self-harm and stop children seeing material promoting self-harm or suicide, even when it falls below the criminal threshold.

“We want to get these new protections in place as soon as possible, but companies should not wait for laws to come into force – they must take effective action to protect all users now.”

Pinterest and X have not responded to a request for comment.

This post appeared first on sky.com

0
FacebookTwitterGoogle +Pinterest
previous post
Sea lions in Australia carry cameras to help scientists map ocean floor for first time
next post
Russian court gives 12-year treason sentence to Russian-American over $50 charity donation

You may also like

Russian officials accuse Meta of ‘overt censorship’ after...

September 18, 2024

Pioneering Alzheimer’s drug rejected for widespread use in...

October 23, 2024

Drinking alcohol then napping on flights could be...

June 4, 2024

Taurid meteor shower to peak across the UK-...

November 11, 2024

GCHQ boss says China’s ‘genuine’ cyber threat ‘weakens...

May 15, 2024

Ticketmaster hit by cyber attack – with hackers...

June 2, 2024

Apple apologises for crushing musical instruments and books...

May 10, 2024

New cancer treatment trial shrinks Luton man’s brain...

October 29, 2024

UK risks losing AI leadership without data strategy,...

October 5, 2024

Artificial intelligence could help detect heart failure risk...

May 30, 2024

    Stay updated with the latest news, exclusive offers, and special promotions. Sign up now and be the first to know! As a member, you'll receive curated content, insider tips, and invitations to exclusive events. Don't miss out on being part of something special.


    By opting in you agree to receive emails from us and our affiliates. Your information is secure and your privacy is protected.

    Popular

    • 1

      Tartana Drilling to Verify Upgrading of 45,000 Tonne Copper Resource

    • 2

      Magic mushrooms effective for treating depression – study

    • 3

      Lancaster Resources Welcomes Three Highly Experienced Industry Experts to Its Advisory Board

    • 4

      Hong Kong plans to install thousands of surveillance cameras. Critics say it’s more proof the city is moving closer to China

    • 5

      UN experts urge Thailand not to deport dozens of Uyghurs to China where they face ‘real risk of torture’

    Categories

    • Business (1,212)
    • Investing (2,995)
    • Science (605)
    • World (3,626)
    Footer Logo

    Disclaimer: dailywashingtoninsider.com, its managers, its employees, and assigns (collectively “The Company”) do not make any guarantee or warranty about what is advertised above. Information provided by this website is for research purposes only and should not be considered as personalized financial advice. The Company is not affiliated with, nor does it receive compensation from, any specific security. The Company is not registered or licensed by any governing body in any jurisdiction to give investing advice or provide investment recommendation. Any investments recommended here should be taken into consideration only after consulting with your investment advisor and after reviewing the prospectus or financial statements of the company.

    Copyright © 2024 dailywashingtoninsider.com | All Rights Reserved