The news: Facebook will start directing people who have interacted with misinformation on coronavirus to a myth-busting page on the World Health Organization’s website. “We’re going to start showing messages in News Feed to people who have liked, reacted or commented on harmful misinformation about COVID-19 that we have since removed,” Facebook’s VP of integrity, Guy Rosen, wrote in a blog post. These messages will appear in a post the top of people’s news feeds, labelled ““Help friends and family avoid false information about COVID-19.”
The context: There are a lot of harmful myths and hoaxes about covid-19 being promulgated on social media, most notably the idea there’s a link with new 5G networks, which has spread across Europe and led to attacks on phone masts in the UK. People are also sharing dangerous falsehoods about cures and claims that the virus is some sort of manmade weapon. Human rights group Avaaz released a report this week which examined 100 pieces of misinformation on Facebook and found the posts had been shared over 1.7 million times and seen approximately 117 million times. Social media companies have promised to take a more proactive approach to taking down misinformation about coronavirus, but the scale of problem is huge.
What this step will (and won’t) do: This does not mean Facebook is going to alert you if you’ve viewed or shared a lie. It will only show people a WHO link in their news feed, which they can easily scroll past and ignore, or click on and not read properly. Facebook has always been loathe to take an interventionist approach on fact-checking. However, in its report, Avaaz says a new study shows that if Facebook proactively “corrected the record” by providing users with corrections from fact-checkers, it could cut belief in falsehoods by an average of 50%.