Last August, Twitter’s top executives gathered at the company’s headquarters to discuss how to make the site safer for its users. Two attendees proposed banning all speech that could be considered “dehumanizing.” For an example of what they meant, they showed a sample post that featured the words President Trump used to compare certain nations to excrement.
By January, Twitter had backed off from deeming that sample tweet dehumanizing. Instead, the post was included in an internal company slide show, which helps train Twitter moderators, as the kind of message that should be allowed on the platform.
And on Tuesday, when Twitter rolled out its first official guidelines around what constitutes dehumanizing speech on its service, the sample post was nowhere in sight. The company had narrowed its policymaking to focus only on banning speech that is insulting and unacceptable if directed at religious groups.
“While we have started with religion, our intention has always been and continues to be an expansion to all protected categories,” Jerrel Peterson, Twitter’s head of safety policy, said in an interview. “We just want to be methodical.”
The scaling back of Twitter’s efforts to define dehumanizing speech illustrates the company’s challenges as it sorts through what to allow on its platform. While the new guidelines help it draw starker lines around what it will and will not tolerate, it took Twitter nearly a year to put together the rules — and even then they are just a fraction of the policy that it originally said it intended to create.
Twitter said it had ratcheted down the policy’s scope partly because it kept running into obstacles. When the company sought users’ feedback last year on what it thought such speech might include, people pushed back on the proposed definitions. Over months of discussions late last year and early this year, Twitter employees also worried that such a policy might be too sweeping, potentially resulting in the removal of benign messages and in haphazard enforcement.
“We get one shot to write a policy that has to work for 350 million people who speak 43-plus languages while respecting cultural norms and local laws,” Mr. Peterson said. “It’s incredibly difficult, and we can’t do it by ourselves. We realized we need to be really small and specific.”
Twitter unveiled its new policy ahead of a social media summit at the White House on Thursday that is likely to thrust it and other Silicon Valley companies under the spotlight for what they will and won’t allow.
For the event, Mr. Trump has invited conservative activists who have thrived on social media, such as Charlie Kirk, president of Turning Point USA, which advocates limited government and other issues. Many of the attendees have accused social media companies of anti-conservative bias.
Twitter declined to comment on the meeting.
In the past, Twitter has focused its removal policies on posts that may directly harm an individual, such as threats of violence or messages that contain personal information or nonconsensual nudity. Under the new rules, the company is adding a sentence that says users “may not dehumanize groups based on their religion, as these remarks can lead to offline harm.”
Twitter said that included any tweets that might compare people in religious groups to animals, insects, bacteria and other categories.
The company quickly put the change into effect on Tuesday. Twitter said it had removed a tweet in which Louis Farrakhan, the outspoken black nationalist minister, compared Jewish people to termites because it violated the dehumanization policy.
Rashad Robinson, the president of Color of Change, a civil rights group, said Twitter’s new policy fell short of where it should go.
“Dehumanization is a great start, but if dehumanization starts and stops at religious categories alone, that does not encapsulate all the ways people have been dehumanized,” he said.
Twitter’s work around a dehumanization policy began in August after the company faced a firestorm for not immediately barring Alex Jones, the right-wing conspiracy theorist, when Apple, Facebook and others did. Twitter eventually did bar Mr. Jones, and its chief executive, Jack Dorsey, said at the time that the incident had forced the company to consider “that safety should come first.”
“That’s a conversation we need to have,” he added.
Mr. Dorsey delegated the task of figuring out what makes up dehumanizing speech on Twitter to the company’s legal, policy and safety teams, which are led by Vijaya Gadde. Mr. Dorsey took a hands-off approach because he wanted to empower Ms. Gadde to make the decisions, a Twitter spokeswoman said.
The discussions began with the meeting at Twitter’s headquarters, which included the sample tweet featuring Mr. Trump’s unflattering description of nations such as Haiti. At the end of that meeting, executives agreed to draft a policy about dehumanizing speech and open it to the public for comments.
In September, Twitter published the draft policy of what dehumanizing speech would be forbidden. It included posts likening people to animals or suggesting that certain groups serve a single, mechanistic purpose.
“I like to think of this as us trying to be experimental, the way that our colleagues in product and engineering are very experimental and they’re trying new things,” Ms. Gadde said in an interview at the time.
The response from users was swift — and critical. Twitter received more than 8,000 pieces of feedback from people in more than 30 countries. Many said the draft made no sense, pointing out cases in which the policy would lead to takedowns of posts that lacked any negative intent.
In one example, fans of Lady Gaga, who call themselves “Little Monsters” as a term of endearment, worried that they would no longer be able to use the phrase. Some gamers complained that they would be unable to discuss killing a character in a video game.
©2019 The New York Times News Service