Social media sites should not ban misleading content, UK scientists say
Calls for social media sites to remove misleading content — for example about vaccines, climate change and 5G technology — should be rejected, according to the UK’s senior scientific academy.
After investigating the sources and impact of online misinformation, the Royal Society concluded that removing false claims and offending accounts would do little to limit their harmful effects. Instead, bans could drive misinformation “to harder-to-address corners of the internet and exacerbate feelings of distrust in authorities,” its report says.
In the UK there have been calls from across the political spectrum for Twitter, Facebook and other platforms to remove antivax posts. However, “clamping down on claims outside the consensus may seem desirable but it can hamper the scientific process and force genuinely malicious content underground”, said Frank Kelly, mathematics professor at the University of Cambridge who chaired the Royal Society inquiry.
He added that removing content and driving users away from mainstream platforms makes it harder for scientists to engage with people such as anti-vaxxers. “A more nuanced, sustainable and focused approach is needed,” he said.
While illegal content that incites violence, racism or child sex abuse must be removed, legal material that runs counter to the scientific consensus should not be banned, the report said. Instead there should be wide-ranging action to “build collective resilience” so that people can detect harmful misinformation and react against it.
“We need new strategies to ensure high quality information can compete in the online attention economy,” said Gina Neff, professor of technology and society at the University of Oxford, and a co-author of the report. “This means investing in life-long information literacy programmes, provenance-enhancing technologies and mechanisms for data sharing between platforms and researchers.”
The well informed majority can act as a “collective intelligence” guarding against misinformation and calling out inaccuracies when they come across them, said Sir Nigel Shadbolt, executive chair of the UK Open Data Institute and another co-author. “Many eyes can provide powerful scrutiny of content, as we see in Wikipedia,” he added.
Some fears about the amplification of misinformation on the internet — such as the existence of “echo chambers” and “filter bubbles”, which lead people only to encounter information that reinforces their own beliefs — have been exaggerated, the report found.
While the internet has led to a vast proliferation of all kinds of information, the vast majority of people in the UK hold views close to those of mainstream science, according to a YouGov poll commissioned for the report. The proportions of the 2,000 participants agreeing that Covid vaccines are not safe were 7 per cent for the BioNTech/Pfizer jab and 11 per cent for Oxford-AstraZeneca jab, while 90 per cent said human activity is changing the climate.
Opponents of vaccination would eventually have to face up to the evidence that their opposition to Covid jabs is wrong, Shadbolt said: “The great natural experiment about the efficacy and safety of being vaccinated is the best evidence we have. For [anti-vaxxers] the evidence is not good.”