๐ Stopping viral toxicity before it starts
Can we use machine learning to circuit-break misinformation?
This week at a glance:
๐ข Circuit breakers suggest a way algorithms can keep toxicity off platforms
๐ฉ๐ฝโ๐ป The tech may not be there yet and human moderation is still vital
๐ค Even if they work as intended, circuit breakers are not a panacea
This week weโre exploring circuit breakers:1 an alluring use of machine learning thatโs supposed to be able to detoxify the big social platforms, assist human moderators, and stop lies and propaganda in their tracks before they spread, if โ and itโs a big โifโ โ the technology can work as intended.
In August 2020 the Center for American Progress proposed a system for social platforms modeled on Wall Streetโs trading curbs, often called circuit breakers. A sudden drop in a stock or market index can trigger a trading halt as short as 15 minutes or as long as the rest of the day, depending on how steep the drop is.
When trading halted in March 2020, New York Stock Exchange President Stacey Cunningham described the circuit breakers as โa precautionary measure that we put in place so that the market can slow down for a minute.โ One study found that in 2020, โwithout the circuit breaker rule, the market would continue to slump by up to -19.98% on the days following the previous crashes. In reality, the rule has successfully prevented the predicted plummet and further stimulated the market to recover for up to 9.31% percent.โย
According to CAP, the key takeaway from Wall Street circuit breakers is the introduction of friction by putting in a kind of speed bump. They imagined something similar in spirit could be used by platforms to prevent the spread of Covid misinformation:ย
Platforms should detect, label, suspend algorithmic amplification, and prioritize rapid review and fact-checking of trending coronavirus content that displays reliable misinformation markers, which can be drawn from the existing body of coronavirus mis/disinformation.
And according to Politico, Big Tech was already deep into an experiment in bringing far more machine learning into identifying harmful and illegal content.ย
In October 2020, the New York Post published a story about a laptop supposedly belonging to Hunter Biden. But before the link could go viral on Facebook, it was caught by their internal โviral content review system,โ a system to prevent viral mis- and disinformation from spreading. The tech giant touted it as a test case that showed how machine learning could react faster than human moderators.
But last August, a man livestreamed his drive to D.C., where he then threatened to set off a bomb outside the U.S. Capitol. He was live on Facebook for five hours until moderators disabled the broadcast. Why didnโt the automation kick in this time?
The appeal of less harmful and illegal content on social platforms is probably self-evident, especially to readers of this newsletter. But thereโs also a near-constant crisis on most platforms around the significant toll that reviewing sensitive content takes on workers in trust and safety, sometimes called integrity. Facebook has agreed to pay $52 million in a settlement with moderators who say they developed PTSD on the job, and TikTok was recently sued by a moderator alleging โsevere psychological trauma including depression.โ It makes sense why companies would want to shift more of this burden to their algorithms.
Since the pandemic, with tens of thousands of content moderators sent home, Politico reported that Facebook, Google, and Twitter have all announced increased automation in content moderation, but they have offered few specifics about its inner workings.2
However, TikTok has offered more detail. Last July, in announcing that TikTok would begin to introduce automation into their moderation process, Eric Han, the companyโs Head of US Safety, wrote: โAutomation will be reserved for content categories where our technology has the highest degree of accuracy, starting with violations of our policies on minor safety, adult nudity and sexual activities, violent and graphic content, and illegal activities and regulated goods.โ
It makes sense that TikTok would use computers to identify relatively straightforward violations, because context can be extremely tricky. Facebook cites bullying as an example of something particularly nuanced and contextual. This may also explain why the livestreamed bomb threat (as well as many other examples) failed to trigger a rapid response from Facebook.ย
More fundamentally: using previous violations to predict potential violations is not a perfect system. According to Politico, after Facebook, Google, and Twitter increased automated moderation in 2020, there were fewer post removals, more successful content appeals โ and more hate speech. Similarly, TikTok saw takedown disputes go through the roof after their change last summer.3
But perhaps the best indication that this technology might not be ready for primetime comes from the hundreds of Facebook moderators who wrote this open letter in late 2020: โManagement told moderators that we should no longer see certain varieties of toxic content coming up in the review tool from which we workโ such as graphic violence or child abuse, for example.โ However, โthe AI wasnโt up to the job,โ they write. โImportant speech got swept into the maw of the Facebook filterโand risky content, like self-harm, stayed up.โ (Facebook responded, but not about the automation, and they have been pretty quiet about their โviral content review systemโ since touting the Hunter Biden story.)ย
What if we could make these automated moderation systems more accountable to the public? Returning to the Wall Street analogy, thereโs another important detail: circuit breakers were not just โput in placeโ by stock exchanges, but were actually mandated by the Securities and Exchange Commission after the Dow Jones fell 22.6% in the 1987 Black Monday crash. Assuming this technology is refined and improved, could platforms be similarly regulated? Might OSHA set circuit breaker-based labor standards for trust and safety workers?4 And could Congress compel platforms to use automated moderation or lose their Section 230 protection?
Over email, I reached out to Daphne Keller, Director of Program on Platform Regulation at the Cyber Policy Center at Stanford. Keller is an expert on platform regulation and userโs rights. As she notes in a paper on amplification, thereโs an important distinction between harmful, yet constitutionally protected speech, and illegal speech, such as terrorism.
Using a circuit breaker to automate the moderation of harmful speech โwould be a very complicated thing for the law to mandate, though, since it essentially suppresses distribution of potentially lawful speech at a time when no one knows yet whether it is illegal,โ Keller told me. She added that the European Union is โcurrently contemplating mandating the opposite, in order to avoid the harm from โheckler's vetoesโ targeting lawful speech.โ
As we often say, there are no panaceas.5 It appears that we are a long way away from platforms employing machine learning so sophisticated that it keeps our human mods healthy and our feeds free of mis- and disinformation. But thereโs reason to think we might get closer: Big Tech has been developing these technologies for years, and they have previously successfully applied them to recognizing content as complex as copyrighted songs and child pornography.ย AI is rapidly evolving, and who knows what it will be capable of in a few years.
In the meantime, perhaps we should turn back to the idea of deliberately introducing friction. Some of our favorite examples of digital public spaces employ a lot of it, from Front Porch Forumโs once-a-day delivery, to some of the best โslow internetโ projects. But a lot of friction isnโt ideal for every situation; sometimes itโs important to learn things right away.
The appropriate amount of friction, and the human care required to facilitate it, is worth deeply considering when designing a digital space. Iโd love to know what you think: what are great examples of friction in your online life?ย
More moderation
If youโre interested in this topic, consider reading the Everything in Moderation newsletter, which was part of our recent #Lookoutfor2022 project. Itโs a pithy, weekly dive into moderation, written by Friend of New_ Public Ben Whitelaw. Ben is in London and heโs been covering these issues for years as a freelance journalist, so he has a great sense of the long-term and more of a global perspective than we often consider here in the US.
Seeking a Head of Product
In addition to looking to hire a Community Architect and Design Facilitator, weโre now interested in hiring a full-time Head of Product to help us develop the New_ Public Incubator. (!) More details on that to come soon. If you, or anyone you know, specializes in product management and design, as well as growth marketing and people management, please see the job description and apply here.
Getting carried away with the footnotes these days,
Josh
Photos by Markus Spiske on Unsplash
New_ Public is a partnership between the Center for Media Engagement at the University of Texas, Austin, and the National Conference on Citizenship, and was incubated by New America.
Daphne Keller told me in an email: โI'm not sure there is consensus on what counts as a โcircuit breaker.โ I've mostly heard it used to describe purely quantitative caps in virality... That would limit the spread of dangerous lies, but would also limit the spread of the next Lil Nas X, or activist campaign, etc.โ This is obviously a much bigger kind of limit, that would pose an existential threat to nearly all existing social platforms, for better or for worse. In this newsletter, Iโm instead using โcircuit breakerโ as it has appeared in the press.
A serious caveat: platforms are not required to share data. Much of what we know about algorithms and moderation has been volunteered by the companies, discovered by investigative journalists, or leaked by ex-employees. Lawmakers have recently proposed laws to require platforms to make more data available, but so far our understanding is necessarily incomplete.
Itโs worth noting that correlation is not causation: we donโt know for sure that increasing automation resulted in these changes, or even whatโs a reasonable baseline, because, say it with me: we donโt have the data.
If you have expertise in labor law Iโd love to hear what you think, just reply to this email.
This was a rallying cry of Dr. Elinor Ostrom, whose work on โthe commonsโ we look to for instruction and inspiration.ย