
“Some of them may not be able to work on certain things that they would work on in the office,” Kevin Chan, Facebook Canada’s head of public policy, told The Canadian Press.
“They’re looking at potentially private, and sensitive things that have been reported to them and we need to make sure….that these things can be treated in the secure and private manner that they deserve.”
Full-time Facebook employees have stepped up and are taking on some of the moderating work, including from contractors who can’t have proprietary and sensitive content at home. These workers are dealing with content related to “real-world harm” like child safety and suicide and self-injury.
“There is no question this is going to pose challenges to the degree to which we can be as responsive,’ Chan said.
To deal with the situation, Facebook has rolled out measures meant to curb the flow of COVID-19 misinformation and is focused on weeding out and removing content around terrorism and anything inciting violence or linking to “dangerous” individuals and organizations.
At Twitter, machine learning and automation is being used to help the company review reports most likely to cause harm first and to help rank content or “challenge” accounts automatically.
“While we work to ensure our systems are consistent, they can sometimes lack the context that our teams bring, and this may result in us making mistakes,” Twitter said in a blog. “As a result, we will not permanently suspend any accounts based solely on our automated enforcement systems.”
Google has also upped its reliance on machine-based systems to reduce the need for people to work from the office and said the increase in automation has many downsides, including a potential increase in content classified for removal and slower turnaround times for appeals.
“They are not always as accurate or granular in their analysis of content as human reviewers,” added a Google blog released in March.
This is a sentiment Priebe has encountered many times, but he has a counter-argument: “AI is not perfect but…humans are also not perfect.”
He gives the example of a child playing a game at home during the pandemic, when pedophiles might be more active online and trying to contact young people.
“You have three different humans look at the same conversation and they’re not going to give you the same answer. Some of them are going to call it grooming and some of them aren’t,” said Priebe.
Priebe believes an ideal system blends humans and AI because the latter is good at knowing what to do with obvious cases like when a user’s content is flagged almost a dozen times in a short period of time or when someone gets a message that only reads hello and hits report just to see what the button does.
“You don’t need a human to have to be looking at their screen and looking at this absolutely sexual content in front of potentially their children who snuck up behind them because artificial intelligence is going to win every time on that,” he said.
“Let humans do what humans do well, which is deal with that middle category of stuff that is subjective, difficult or hard to understand, that the AI is not confident about.”
Regardless of how the moderation gets done, some things will always slip through the cracks, especially in a pandemic, said Dunn.
“No system is perfect.”
This report by The Canadian Press was first published June 7, 2020.
By Tara Deschamps, The Canadian Press












