For decades, anti-vaccine movements have generated and spread rumors that vaccines cause serious health problems. The rollout of COVID vaccines has provided new opportunities to spread misinformation.
At the start of the pandemic, people were already worried about the virus and the impact of other public health measures, such as lockdowns, on their physical and social well-being. As COVID vaccines were rolled out, concerns mounted about the small but serious risk of blood clots linked to the AstraZeneca vaccine.
Alongside this, there has been a degree of panic around unsubstantiated rumors of adverse events—extremely rare medical problems after being vaccinated—circulating on social media.
But contrary to the popular belief that social media creates these rumors, our new research suggests social media generally only aids the spread of these rumors.
What ‘vaccine harms’ are people sharing on social media?
We have been studying community attitudes to COVID vaccinations, including the flow of information on social media, the kinds of information being shared, and by whom.
In our latest study, we tracked emerging concerns about alleged adverse events globally. We used Google Trends and Crowdtangle—a research platform for studying Facebook’s public-facing data. We focused on the most commonly searched and discussed events to track where they were coming from.
We dug into the five most frequently searched adverse events: clotting, fainting, Bell’s palsy, premature death and infertility.
Clotting was associated with the AstraZeneca vaccine and the rare instances of thrombosis with thrombocytopenia syndrome (TTS). This led to the vaccine’s suspension or authorities placing age restrictions on its recipients in many countries.
News reporting on clotting was generally reasonable and in line with the threat the condition posed. Because the issue was newsworthy on its own, it did not require sensationalist reporting. Social media spread these reports globally, so the first reports of clotting, emerging from Austria, spread on Facebook pages as far as Ghana, the Philippines and Mexico within eight hours.
Fainting, Bell’s palsy and premature death
There was no scientific basis for the other four rumors we investigated. However, three of them drew specifically from “traditional” (television and newspaper) news reporting on specific incidents.
For example, a Tennessee nurse fainted on television shortly after receiving the Pfizer vaccine. Traditional media reports included the nurse’s own disclosure of a history of fainting and cautioned against attributing it to the vaccine.
Likewise, elderly baseball legend Hank Aaron died from natural causes two weeks after receiving a COVID vaccine on camera. He had hoped to inspire other Black Americans to be vaccinated.
These two incidents were widely reported in traditional media and soon flowed into social media posts attributing them to the vaccine.
The Bell’s palsy rumor came out of news reports in Bangladesh, which were then picked up by a U.K. outlet, attributing the rare condition to the Pfizer vaccine.
The rumors of COVID vaccines causing infertility were the only ones that we could not trace to an original “traditional media” source. Instead, two internet stories misrepresenting the work, and words of scientists were shared widely on social media. Traditional media only picked up the story to report on the misinformation occurring.
We describe this as an example of vaccine skeptics “theory crafting” online. This is when a group of people on the internet use their collective resources to analyze information to develop plausible explanations for events.
In the case of infertility, a willing community misused two scientific sources to construct what they represented as compelling evidence of a cover-up. This theory then led to a persistent internet rumor that COVID vaccines caused fertility problems.
In the other four cases above, we found traditional media still played an important role in determining people’s awareness of alleged adverse events.
What did mainstream news outlets do?
Traditional media outlets were important to those sharing the social media posts, as they treated mainstream media reports as markers of credibility.
Vaccine-skeptical communities used international media sources to build “evidence” for adverse events. They then redistributed this “evidence” among their international networks.
Disreputable outlets chased “clickbait,” accelerating the spread of misinformation. For instance, when 86-year-old Aaron died, one site led with the headline “Hank Aaron Death: MLB Legend Shockingly Passes Away Weeks After Taking COVID-19 Vaccine.” This headline spread much faster and further on social media than the majority of reports that explained Aaron’s death was not a result of his vaccination.
Inaccurate and sensationalist headlines in mainstream media went on to drive significant searches and shares. The rumors flowed globally, unfettered by national boundaries.
Despite most of the rumors we investigated gaining traction because of media reporting, journalists also played an important role in suppressing or debunking illegitimate claims.
The disruption of earlier media models clearly poses a challenge for the accuracy of information shared on the internet. The imperative for news sources to generate clicks can outweigh the imperative to provide accurate and reliable information.
So what’s the solution?
We can see no easy answers for resolving the flow of misinformation online.
However, the use of credibility markers for both authors and stories on social media is one possible solution. A system where publicly recognized topic experts can “upvote” and “downvote” news stories to produce a “credibility score” would help readers judge the perceived credibility of particular stories and information.
In the meantime, we recommend scientists and health professionals, where possible, promote their own perspectives when a story about alleged adverse events needs clarifying. Doing so can potentially change the trajectory and spread of a story.
Scientists and health professionals speaking out can’t prevent the stories from being shared within online communities of vaccine-refusers. These people are invested in sharing such information regardless of its veracity. However, professionals can limit the damaging spread of rumors once media outlets begin to report their debunking
Social media spreads rumors about COVID vaccine harms, but it doesn’t always start them (2022, June 3)
retrieved 3 June 2022
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.