Earlier this month, I chose to leave my position leading trust and safety at Elon Musk’s Twitter.
As the company’s head of policy, my teams were responsible for drafting Twitter’s rules and figuring out how to apply them consistently to hundreds of millions of tweets per day. In my more than seven years at the company, we exposed government-backed troll farms meddling in elections, introduced new tools for contextualizing dangerous misinformation and, yes, banned President Trump from the service. The Cornell professor Tarleton Gillespie called teams like mine the “custodians of the internet.” The work of online sanitation is unrelenting and contentious.
Enter Mr. Musk.
In a news release announcing his agreement to acquire the company, Mr. Musk laid out a simple thesis: “Free speech is the bedrock of a functioning democracy, and Twitter is the digital town square where matters vital to the future of humanity are debated.” He said he planned to revitalize Twitter by eliminating spam and drastically altering its policies to remove only illegal speech.
Since closing the deal on Oct. 27, many of the changes implemented by Mr. Musk and his team have been sudden and alarming for employees and users alike, including rapid-fire layoffs and an ill-fated foray into reinventing Twitter’s verification system. A wave of employee resignations caused the hashtag “#RIPTwitter” to trend on the site on Thursday — not for the first time — alongside questions about whether a skeleton crew of remaining staff can keep the 16-year-old service afloat.
And yet when it comes to content moderation, much has stayed the same since Mr. Musk’s acquisition. Twitter’s rules continue to ban a wide range of “lawful but awful” speech. Mr. Musk has insisted publicly that the company’s practices and policies are unchanged. Are we just in the early days — or has the self-declared free speech absolutist had a change of heart?
The truth is that even Elon Musk’s brand of radical transformation has unavoidable limits.
Advertisers have played the most direct role thus far in moderating Mr. Musk’s free speech ambitions. As long as 90 percent of the company’s revenue comes from ads (as was the case when Mr. Musk bought the company), Twitter has little choice but to operate in a way that won’t imperil the revenue streams that keep the lights on. This has already proved to be challenging.
Almost immediately upon the acquisition’s close, a wave of racist and antisemitic trolling emerged on Twitter. Wary marketers, including those at General Mills, Audi and Pfizer, slowed down or paused ad spending on the platform, kicking off a crisis within the company to protect precious ad revenue.
In response, Mr. Musk empowered my team to move more aggressively to remove hate speech across the platform — censoring more content, not less. Our actions worked: Before my departure, I shared data about Twitter’s enforcement of hateful conduct showing that, by some measures, Twitter was actually safer under Mr. Musk than it had been before.
Marketers have not shied away from using the power of the purse: In the days following Mr. Musk’s acquisition, the Global Alliance for Responsible Media, a key ad industry trade group, published an open call to Twitter to adhere to existing commitments to “brand safety.” It’s perhaps for this reason that Mr. Musk has said he wants to move away from ads as Twitter’s primary revenue source: His ability to make decisions unilaterally about the site’s future is constrained by a marketing industry he neither controls nor has managed to win over.
But even if Mr. Musk is able to free Twitter from the influence of powerful advertisers, his path to unfettered speech is still not clear. Twitter remains bound by the laws and regulations of the countries in which it operates. Amid the spike in racial slurs on Twitter in the days after the acquisition, the European Union’s chief platform regulator took to the site to remind Mr. Musk that, in Europe, an unmoderated free-for-all won’t fly. In the United States, members of Congress and the Federal Trade Commission have raised concerns about the company’s recent actions. And outside of the United States and the European Union, the situation becomes even more complex: Mr. Musk’s principle of keying Twitter’s policies on local laws could push the company to censor speech it has been loath to restrict in the past, including political dissent.
Regulators have significant tools at their disposal to enforce their will on Twitter and on Mr. Musk. Penalties for noncompliance with Europe’s Digital Services Act could total as much as 6 percent of the company’s annual revenue. In the United States, the F.T.C. has shown an increasing willingness to exact significant fines for noncompliance with their orders (like a blockbuster $5 billion fine imposed on Facebook in 2019). In other key markets for Twitter, such as India, in-country staff work with the looming threat of personal intimidation and arrest if their employers fail to comply with local directives. Even a Musk-led Twitter will struggle to shrug off these constraints.
There is one more source of power on the web — one that most people don’t think much about, but which may be the most significant check on unrestrained speech on the mainstream internet: the app stores operated by Google and Apple.
While Twitter has been publicly tight-lipped about how many people use the company’s mobile apps (rather than visiting Twitter.com on a browser), the company’s 2021 annual report didn’t mince words: “Our release of new products … is dependent upon and can be impacted by digital storefront operators” that decide the guidelines and enforce them, it reads in part. “Such review processes can be difficult to predict and certain decisions may harm our business.”
“May harm our business” is an understatement. Failure to adhere to Apple and Google’s guidelines would be catastrophic, risking Twitter’s expulsion from their app stores and making it more difficult for billions of potential users to access Twitter’s services. This gives Apple and Google enormous power to shape the decisions Twitter makes.
Apple’s guidelines for developers are reasonable and plainly stated: They emphasize creating “a safe experience for users” and stress the importance of protecting children. The guidelines quote the Supreme Court justice Potter Stewart’s “I know it when I see it” quip, saying the company will ban apps that are “over the line.”
In practice, the enforcement of these rules is fraught.
In my time at Twitter, representatives of the app stores regularly raised concerns about content available on our platform. On one occasion, a member of an app review team contacted Twitter, saying with consternation that he had searched for “#boobs” in the Twitter app and was presented with … exactly what you’d expect. Another time, on the eve of a major feature release, a reviewer sent screenshots of several days-old tweets containing an English-language racial slur, asking Twitter representatives whether they should be permitted to appear on the service.
Reviewers hint that app approval could be delayed or perhaps even withheld entirely if issues are not resolved to their satisfaction — although the standards for resolution are often inferred. Even as they appear to be driven largely by manual checks and anecdotes, these review procedures have the power to derail company road maps and trigger all-hands-on-deck crises for weeks or months at a time.
Whose values are these companies defending when they enforce their policies? While the wide array of often conflicting global laws no doubt play a part, the most direct explanation is that platform policies represent the preferences of a small group of predominantly American tech executives. Steve Jobs didn’t believe porn should be allowed in the App Store, and so it isn’t allowed. Stripped bare, there’s a dismaying lack of legitimacy to the decisions here.
It’s this very lack of legitimacy that Mr. Musk, correctly, points to when he calls for greater free speech, and for the establishment of a “content moderation council” to guide the company’s policies — an idea Google and Apple would be right to borrow for the governance of their app stores. But even as he criticizes the capriciousness of platform policies, he perpetuates this same lack of legitimacy through his impulsive changes and tweet-length pronouncements about Twitter’s rules. In appointing himself “Chief Twit,” Mr. Musk has made clear that at the end of the day, he’ll be the one calling the shots.
It was for this reason that I ultimately chose to leave the company: A Twitter whose policies are defined by unilateral edict has little need for a trust and safety function dedicated to its principled development.
So where will Twitter go from here? Some of the company’s decisions in the weeks and months to come, like the near-certainty of allowing Donald Trump’s account back on the service, will have an immediate, perceptible impact. But to truly understand the shape of Twitter going forward, I’d encourage looking not just at the choices the company makes but at how Mr. Musk makes them. Should it materialize, will the moderation council represent more than just the loudest, predominantly American voices complaining about censorship — including, critically, the approximately 80 percent of Twitter users who reside outside of the United States? Will the company continue to invest in features like Community Notes, which bring Twitter users into the work of platform governance? Will Mr. Musk’s tweets announcing policy changes become less frequent and abrupt?
Longer term, the moderating influences of advertisers, regulators and — most critically of all — app stores may be welcome for those of us hoping to avoid an escalation in the volume of dangerous speech online. Twitter will have to balance its new owner’s goals against the practical realities of life on Apple and Google’s internet — no easy task for the employees who have chosen to remain. And as I departed the company, the calls from the app review teams had already begun.
Yoel Roth (@yoyoel) was the head of trust and safety at Twitter, where he spent seven years directing the company’s policy and enforcement work on abuse, election security and anti-spam issues.
The Times is committed to publishing a diversity of letters to the editor. We’d like to hear what you think about this or any of our articles. Here are some tips. And here’s our email: email@example.com.
Follow The New York Times Opinion section on Facebook, Twitter (@NYTopinion) and Instagram.