Major websites are turning to automatic systems to moderate content as they tell their staff to work from home.
YouTube, Twitter and Facebook are all relying on artificial intelligence and automated tools to find problematic material on their platforms.
The tech giants admit this may lead to some mistakes - but say they still need to remove harmful content.
The coronavirus scare has led to a surge of medical misinformation across the web.
Google, which owns YouTube, said appeals about content wrongfully removed could take longer under the new measures.
Twitter, meanwhile, promised that no accounts suspended by automated software would be permanently banned without a human review.
Computer errors
Content review operations for Facebook, Twitter and Google are spread around the globe, including in the US, India and Spain.
All those countries have said employees should work from home – but switching the content review operation to remote working is complicated.
Facebook has sent home all its content reviewers until further notice, and says it is paying them during this time.
In a blogpost, Facebook said: "With fewer people available for human review we'll continue to prioritise imminent harm and increase our reliance on proactive detections in other areas."
Twitter said it would increase the use of machine-learning and automation but acknowledged they could "sometimes lack the context that our teams bring, and this may result in us making mistakes".
As a result, it said it would not permanently ban any accounts based solely on automated systems.
And nearly all of Google's full-time employees worldwide have been ordered to work from home due to the coronavirus pandemic.
"This means automated systems will start removing some content without human review," YouTube said in a blog.
"As we do this, users and creators may see increased video removals, including some videos that may not violate policies.
"Our workforce precautions will also result in delayed appeal reviews."
It added it would also be more cautious about what content gets promoted, including livestreams.
It comes at a time when the tech giants are being asked to ramp up their removal of coronavirus misinformation on their platforms.
The UK's Digital, Culture, Media and Sport committee has asked the government to explain why it has taken two months to set up a unit to counter the spread of disinformation about the virus.
MPs expressed concern that false narratives about coronavirus could undermine efforts to deal with the crisis.
Latest Stories
-
Scottish Premiership: Mathew Anim Cudjoe and Dundee United secure promotion after Championship win
39 mins -
Leicester City trigger buy option for Fatawu Issahaku
1 hour -
We will ensure recent poor run of results is reversed – Hearts of Oak management to fans
1 hour -
GRA terminates 2 SML contracts, suspends upstream and minerals audit deal
1 hour -
Sister Derby joins Medikal on stage at 02 concert
2 hours -
Medikal’s 02 Concert: Shatta Wale, Sarkodie bring fireworks on stage
2 hours -
Livestream: Newsfile discusses Cecilia Dapaah’s case, vote-buying allegations in Ejisu
2 hours -
NACOC embarks on a search at Nsawam Prison
2 hours -
Ban on noise-making commences May 6
2 hours -
Foden wins Football Writers’ Association Footballer of the Year award
3 hours -
FDA discovers expired food items at Zuarungu SHS storeroom
3 hours -
Bahamas World Relays: Ghana paired with world and Olympic champions in Heat 1
11 hours -
Bahamas 2024 Relays: Star-studded cast ready to put on pre-Paris show
11 hours -
Bahamas Relays preview: clash of global title-winning teams in men’s 4x100m
11 hours -
Arsenal keeper Raya wins Premier League Golden Glove
12 hours