AI moderation threatens uploaded video evidence of war crimes


Student journalist Baraa Razzouk has been documenting daily life under Syria’s civil war for years, from bombings and protests to the opening of a new health centre. Razzouk has been sharing his video for the whole world to see on YouTube.

But just recently, he started getting automated emails from YouTube alerting him that his videos violated its policy, and that they would be deleted. As of this month, more than a dozen of his videos had been removed, Razzouk said.

“Documenting the (Syrian) protests in videos is really important. Also, documenting attacks by regime forces,” he told the Thomson Reuters Foundation in a phone interview. “This is something I had documented for the world and now it’s deleted.”

YouTube, Facebook, and Twitter warned in March that videos and other content may be erroneously removed for policy violations, as the coronavirus pandemic forced them to empty offices and rely on automated takedown software.

But those AI-enabled tools risk confusing human rights and historical documentation like Razzouk’s videos with problematic material like terrorist content – particularly in war-torn countries like Syria and Yemen, digital rights activists warned.

“AI is notoriously context-blind,” said Jeff Deutch, a researcher for Syrian Archive, a nonprofit which archives video from conflict zones in the Middle East.

“It is often unable to gauge the historical, political or linguistic settings of posts … human rights documentation and violent extremist proposals are too often indistinguishable,” he said in a phone interview.

Erroneous takedowns threaten content like videos that are used as formal evidence of rights violations by international bodies such as the International Criminal Court and the United Nations, said Dia Kayyali of digital rights group Witness.

After the Thomson Reuters Foundation flagged Razzouk’s account to YouTube, a spokesman said the company had deleted the videos in error, although the removal was not appealed through their internal process. They have now restored 17 of Razzouk’s videos.

“With the massive volume of videos on our site, sometimes we make the wrong call,” the spokesman said in emailed comments. “When it’s brought to our attention that a video has been removed mistakenly, we act quickly to reinstate it.”

In recent years social media platforms have come under increased pressure from governments to quickly remove violent content and disinformation from their platforms – increasing their reliance on AI systems.

With the help of automated software, YouTube removes millions of videos a year, and Facebook deleted more than 1 billion accounts last year for violating rules like posting terrorist content.

Last year social media companies pledged to block extremist content following a livestreamed terror attack on Facebook of a gunman killing 51 people at two mosques in Christchurch, New Zealand.

Companies like Facebook have also pledged to remove misinformation about the coronavirus outbreak that could contribute to imminent physical harm.

These pressures, combined with an increased reliance on AI during the pandemic, puts human rights content in particular jeopardy, said Kayyali.

Social media firms typically do not disclose how frequently their AI tools mistakenly take down content.

So, the Syrian Archive group has been using its own data to approximate change over time in the rate of deletions of human rights documentation on crimes committed in Syria, which has been battered by nearly a decade of war.

“Our research suggests that since the beginning of the year, the rate of content takedowns of Syrian human rights documentations on YouTube roughly doubled (from 13% to 20%)” said Deutch, calling the increase “unprecedented”.

In May, Syrian Archive detected more than 350,000 videos on YouTube had disappeared – up from 200,000 in May 2019, including videos of aerial attacks, protests, and destruction of civilians homes in Syria.

Deutch said he had seen content takedowns in other war-torn countries in the region, including Yemen and Sudan. “Users in conflict zones are more vulnerable,” he said.

Other groups, including Amnesty International and Witness, have warned of the trend elsewhere, including in sub-Saharan Africa.

Last month Yahya Daoud, a Syrian humanitarian worker with the White Helmets emergency response group, shared a post and a photo showing a woman who died in a 2012 massacre by the forces of Syrian President Bashar al-Assad in the Houla region.

By the end of the month Daoud said his account – which he had used since 2011 to document his life in Syria – was automatically deleted without explanation. “I was depending on Facebook to be an archive for me,” he said.

“So many memories have been lost: the death of my friends, the day I became displaced, the death of my mother,” he said, adding that he had unsuccessfully tried to appeal the decision through Facebook’s automated complaints system.

Facebook did not respond to requests for comment. (Source: Thomson Reuters Foundation)