Social media-fueled anti-vaxxer propaganda is the latest online damage the U.K. government is targeting.
Speaking on BBC Radio 4’s Today software this morning, health secretary Matt Hancock said he will meet with representatives from social media platforms on Monday to pressure them into doing more to prevent false information about the safety of vaccinations from being amplified by their platforms.
“I’m seeing them on Monday to demand that they do more to take down wrong — well lies essentially — that are promoted on social media about the impact of vaccination,” he said, when asked about a warning by an u.K. public health body about the danger of a public health emergency being caused by an increase in the number of British children who have not received the measles vaccination.
“Vaccination is safe; it’s very, very important for the public health, for everybody’s health and we’re going to tackle it.”
The head of NHS England also warned last month about anti-vaccination messages gaining traction on social media.
“We need to tackle this danger in people not vaccinating,” Hancock added. “One of the things I’m particularly worried about is the spread of anti-vaccination messages online. I’ve named
in the social media companies like we had to for self-harming imagery a couple of months ago.”
That followed a public outcry over suicide content spreading on Instagram after a british schoolgirl was reported to have been encouraged to kill herself by viewing illustration content on the Facebook -owned platform.
Instagram subsequently announced a policy change saying it would remove illustration images of self damage and demote non-illustration self-damage images so they don’t show up in searches, relevant hashtags or the explore tab.
But it remains to be seen whether platforms will be as immediately responsive to amped-up political pressure to scrub anti-vaccination content entirely given the stage of assist this kind of misinformation can attract among social media users.
Earlier this year Facebook said it would downrank anti-vax content in the News Feed and hide it on Instagram in an effort to minimize the spread of vaccination misinformation.
It also said it would point users toward “authoritative” vaccine-related information — i.e. information that’s been corroborated by the health and scientific establishment.
But deleting such content entirely was not part of Facebook’s announced strategy.
We’ve reached out to Facebook for any response to Hancock’s comments.
In the longer term, social media platforms operating in the U.K. could face laws that demand them to remove content deemed to pose a danger to public health if ordered to by a dedicated regulator, as a result of a wide-ranging government plan to tackle a range of online harms.
Earlier this month the U.K. government set out a broad policy plan for regulating online harms.
The Online Harms Whitepaper proposes to put a mandatory duty of care on platforms to take reasonable levels to defend users from a range of harms — including those linked to the spread of disinformation.
It also proposes a dedicated, overarching regulator to supervise internet companies to ensure they meet their responsibilities.
The government is currently running a public consultation on the proposals, which ends July 1, after which it says it will set out any next actions as it works on developing draft legislation.