YouTube said on Wednesday that it would take down content which contradicts expert consensus about COVID-19 vaccines, updating its policies on misinformation about the pandemic.
The move was the latest by online platforms struggling to contain the spread of hoaxes and false information about the coronavirus and treatments.
YouTube, the Google-owned video-sharing service, said that it was expanding its medical misinformation policy ‘to remove claims about COVID-19 vaccinations that contradict expert consensus from local health authorities or the World Health Organisation.’
YouTube said that it was acting in anticipation of the release of one or more vaccines, and scepticism among many people about their usefulness.
It said that content to be removed would include claims that a vaccine could kill people or cause infertility, or that microchips will be implanted in people who receive a vaccination.
YouTube said that it had removed more than 200,000 videos with ‘dangerous or misleading’ COVID-19 information since February, including unverified claims about transmission or unsubstantiated treatments.
In a related action, Facebook on Tuesday announced a ban on ads that discourage people from getting vaccinated, as part of its efforts to contain misinformation.
The tech giants had regularly been accused of allowing anti-vaccine movements to flourish.
Want stories like this in your inbox?
Sign up to exclusive daily email
More Stories from Internet