Illustration: Rebecca Zisser/Axios

The online spread of the Christchurch mosque killer's sickening first-person video divided experts, industry insiders and the broader public into two opposite camps: Some saw the debacle as proof that Facebook and YouTube can't police their platforms. Others saw it as evidence that they won't.

Why it matters: How we define the platforms' struggle to block the New Zealand shooter's video will shape how we respond to the problem. Either way, Facebook and YouTube don't come off well.

Driving the news:

  • Late Monday night, Facebook posted new details about the video, reporting that the shooter's original live stream was viewed fewer than 200 times in total. None of those viewers flagged it to moderators. The first report about the video came in 12 minutes after the live stream ended.
  • "Before we were alerted... a user on [troll site] 8chan posted a link to a copy of the video on a file-sharing site," Facebook says. Once Facebook took down the original, users began reposting copies.
  • Facebook previously reported that in the 24 hours after the shooting, it removed 1.5 million copies of the video — 1.2 million of which were blocked as they were uploaded.
  • YouTube didn't report numbers, but a Washington Post account said that despite the video platform's efforts to expand human moderation and automated systems, "humans determined to beat the company’s detection tools won the day."

One widely held view is that Facebook and YouTube are simply too big to monitor and control, even with the legions of human moderators they employ and the AI-driven recognition tools they are beginning to deploy.

  • "Social-media platforms were eager to embrace live streaming because it promised growth. Now scale has become a burden," Neima Jahromi wrote in The New Yorker.
  • In this picture, outnumbered moderators will always be a step behind masses of determined users, and the whack-a-mole game will never end.
  • As platforms get better at identifying and blocking particular classes and instances of undesirable content, the content's proponents will find new tactics for modifying, hiding and redistributing the material.

Another view holds that Facebook and YouTube have both repeatedly shown their ability to police their vast online estates when given no alternative:

  • They've taken strong measures against child pornography, and indeed kept most kinds of porn at bay.
  • They've moved forcefully to keep ISIS recruiting videos from publicly circulating.
  • They've worked to eliminate access to Nazi propaganda in Germany, where it's outlawed.
  • They've cracked down effectively on distribution of copyrighted materials via their services.
  • If strong enough legal, financial and socio-political incentives have done the trick in these areas, the argument goes, surely Facebook and YouTube can also take effective action against violent right-wing extremists.

These two scenarios paint two very different pictures of what's going on.

  • In one, platform managers are playing a Sisyphean delete-and-block game against persistent and inventive opponents.
  • In the other, companies that prioritize engagement metrics are protecting their business interests by failing to limit offensive content — except when media coverage and ad boycotts make action unavoidable.

Be smart: Hard as the problem is, people are going to keep pushing the platforms to solve it. And there are plenty of other steps Facebook and YouTube could take.

  • For instance, during a crisis they could suspend real-time uploads, or temporarily block those coming from new or unverified accounts.
  • In a Twitter thread, Homebrew's Hunter Walk (a former YouTube exec) proposed methods for YouTube to protect freedom of speech while curtailing "freedom of reach."

Go deeper: The real tech regulators

Go deeper

Updated 22 mins ago - Politics & Policy

Coronavirus dashboard

Illustration: Sarah Grillo/Axios

  1. Politics: Obama: Trump is "jealous of COVID's media coverage" Axios-Ipsos poll: Federal response has only gotten worse.
  2. Health: Hospitals face a crush — 13 states set single-day case records last week.
  3. Business: Winter threat spurs new surge of startup activity.
  4. Media: Pandemic causes TV providers to lose the most subscribers ever.
  5. States: Nearly two dozen Minnesota cases traced to three Trump campaign events.
  6. World: Putin mandates face masks.

Louisiana braces for 3rd hurricane in 2 months as Tropical Storm Zeta nears

Municipality workers clean the streets of garbage in Playa del Carmen, Mexico, on Tuesday that was left by Zeta, which struck the Yucatan Peninsula as a Category 1 Hurricane a day earlier — causing no major damage to infrastructure. Photo: Medios y Media/Getty Images

Tropical Storm Zeta is expected to strengthen back into a hurricane and bring dangerous storm surge conditions to parts of the northern Gulf Coast on Wednesday, per the National Hurricane Center.

The state of play: Louisiana Gov. John Bel Edwards (D) requested a pre-landfall Federal Declaration of Emergency in a letter to President Trump on Tuesday, ahead of the storm's expected arrival south of New Orleans.

2 hours ago - Technology

Trump's campaign website hacked

A screenshot of the Trump campaign website after it was hacked.

The Trump campaign website briefly went down and its "About" page was modified after hackers attacked the site Tuesday evening.

The big picture: With just seven days before the election, the hackers emulated the FBI and declared on the "About" page that: "this was seized. the world has had enough of the fake-news spreaded [sic] daily by president donald j trump. it is time to allow the world to know truth." Two addresses linked to the cryptocurrency Monero appeared on the site. Trump campaign spokesperson Tim Murtaugh in a statement said no sensitive data had been exposed in the attack.

Go deeper: Twitter hack raises fears of an unstable election