Advertisement

Tech execs tell lawmakers they’re acting faster on extremist content

Monika Bickert, Nick Pickles and Derek Slater
Facebook’s Monika Bickert, left, Twitter’s Nick Pickles and Google’s Derek Slater testify Wednesday before the Senate Commerce Committee.
(J. Scott Applewhite / Associated Press)
Share via

Facebook, Google and Twitter executives told members of Congress on Wednesday that they’ve gotten better and faster at detecting and removing violent extremist content on their social media platforms in the face of hatred-fueled mass shootings.

Questioned at a hearing by the Senate Commerce Committee, the executives said they are spending money on technology to improve their ability to flag extremist content and taking the initiative to reach out to law enforcement authorities to try to head off potential violent incidents.

“We will continue to invest in the people and technology to meet the challenge,” said Derek Slater, Google’s director of information policy.

Advertisement

The lawmakers want to know what the companies are doing to remove hate speech from their platforms and how they are coordinating with law enforcement.

“We are experiencing a surge of hate. Social media is used to amplify that hate,” said Sen. Maria Cantwell of Washington state, the panel’s senior Democrat.

The company executives testified that their technology is improving for identifying and taking down suspect content faster.

Advertisement

Of the 9 million videos removed from Google’s YouTube in the second quarter of the year, 87% were flagged by a machine using artificial intelligence, and many of them were taken down before they got a single view, Slater said.

After the February 2018 shooting that killed 17 people at a high school in Florida, Google began to proactively reach out to law enforcement authorities to see how they could better coordinate, Slater said. Before that shooting, the suspect posted on a YouTube page, “I’m going to be a professional school shooter,” authorities said.

Word came this week from Facebook that it will work with law enforcement organizations to train its AI systems to recognize videos of violent events as part of a broader effort to crack down on extremism. Facebook’s AI systems were unable to detect live-streamed video of the mosque shootings in New Zealand in March that killed 50 people. The self-professed white supremacist accused of the shootings had live-streamed the attack.

Advertisement

The effort will use body camera footage of firearms training provided by U.S. and British government and law enforcement agencies.

Facebook also is expanding its definition of terrorism to include not just acts of violence intended to achieve a political or ideological aim, but also attempts at violence, especially when aimed at civilians with the intent to coerce and intimidate. The company has had mixed success in its efforts to limit the spread of extremist material on its service.

Facebook appears to have made little progress, for example, on its automated systems for removing prohibited content glorifying groups such as Islamic State in the four months since the Associated Press detailed how Facebook pages auto-generated for businesses are aiding Middle East extremists and white supremacists in the United States. The new details come from an update of a complaint to the Securities and Exchange Commission that the National Whistleblower Center plans to file this week.

Facebook said in response that it removes any auto-generated pages “that violate our policies. While we cannot catch every one, we remain vigilant in this effort.”

Monika Bickert, Facebook’s head of global policy management, said at the Senate hearing that the company has increased its ability to detect terror, violence and hate speech much sooner. “We know that people need to be safe,” she said. Bickert said Facebook removes any content that promotes violence, white supremacy or nationalism as well as indicating suicide, and disables accounts when threats are detected.

Twitter’s director of public policy strategy, Nick Pickles, said his company suspended more than 1.5 million accounts for promoting terrorism from Aug. 1, 2015, to Dec. 31, 2018. More than 90% of the accounts are suspended through Twitter’s proactive measures, he said, not waiting for reports from government and law enforcement.

Advertisement

Sen. Rick Scott (R-Fla.) asked Pickles why Twitter hadn’t suspended the account of Venezuelan socialist leader Nicolas Maduro, who has presided over a deepening economic and political crisis and has threatened opposition politicians with criminal prosecution.

If Twitter removed Maduro’s account, “it would not change facts on the ground,” Pickles said.

Scott said he disagreed, saying that Maduro’s account, with some 3.7 million followers, provides him with legitimacy as a world leader.

Advertisement