
Pat Holloway has seen her share of destruction during a 30-year career as a photojournalist: the 1993 standoff in Waco, Texas; the 1995 bombing of a federal building in Oklahoma City by Timothy McVeigh; and the 2011 tornado that struck Joplin, Mo.
But this weekend, she said in an interview, she had had enough. When graphic images began circulating on Twitter showing bloody victims of a mass shooting at a mall in Texas that left at least nine people, including the gunman, dead, she tweeted at Elon Musk, Twitter’s owner, demanding that he do something.
“This family does not deserve to see the dead relatives spread across Twitter for everybody to see,” Ms. Holloway, 64, said in the interview on Sunday.
Ms. Holloway was one of many Twitter users who criticized the social network for allowing the grisly images — including of a blood-spattered child — to spread virally across the platform after the shooting on Saturday. Though gruesome images have become common on social media, where a cellphone camera and an internet connection make everyone a publisher, the unusually graphic nature of the images drew sustained outcry from users. And they threw a harsh spotlight on Twitter’s content moderation practices, which have been curtailed since Mr. Musk acquired the company last year.
Like other social media companies, Twitter has once again found itself in a position akin to that of traditional newspaper editors, who wrestle with difficult decisions about how much to show their audiences. Though newspapers and magazines generally spare their readers from truly graphic images, they have made some exceptions, as Jet magazine did in 1955 when it published open-casket images of Emmett Till, a 14-year-old Black boy who was beaten to death in Mississippi, to illustrate the horrors of the Jim Crow-era South.
Unlike newspaper and magazine publishers, however, tech companies like Twitter must enforce their decisions on a huge scale, policing millions of users with a combination of automated systems and human content moderators.
Other tech companies like Facebook’s parent, Meta, and YouTube’s parent, Alphabet, have invested in large teams that reduce the spread of violent images on their platforms. Twitter, on the other hand, has scaled back its content moderation since Mr. Musk bought the site late last October, laying off full-time employees and contractors on the trust and safety teams that manage content moderation. Mr. Musk, who has described himself as a “free speech absolutist,” said last November that he would establish a “content moderation council” that would decide which posts should stay up and which should be taken down. He later reneged on that promise.
Twitter and Meta did not respond to requests for comment. A spokesman for YouTube said the site had begun removing video of the massacre, adding that it was promoting authoritative information sources.
Graphic content was never completely banned by Twitter, even before Mr. Musk took over. The platform, for instance, has allowed images of people killed or wounded in the war in Ukraine, arguing that they are newsworthy and informative. The company sometimes places warning labels or pop-ups on sensitive content, requiring that users opt in to see the imagery.