Grok's prompt page, displayed on a phone in Berkeley, Calif., on Oct. 29, 2025. Elon Musk's artificial intelligence chatbot, Grok, created and then publicly shared at least 1.8 million sexualized images of women, according to separate estimates of X data by The New York Times and the Center of Countering Digital Hate....(Andria Lo/The New York Times).
- Data found Elon Musk’s AI chatbot, Grok, created and then publicly shared at least 1.8 million sexualized images of women and children.
- Since Jan. 8 Grok has largely ignored requests to dress women in bikinis, but restrictions do not extend Grok’s app or website, which continue to allow users to generate sexual content in private.
- Tools to create sexualized and realistic AI images exist elsewhere, the spread this material on X is unique for its public platform, experts says it’s online harassment.
Share
|
Getting your Trinity Audio player ready...
|
SAN FRANCISCO — Elon Musk’s artificial intelligence chatbot, Grok, created and then publicly shared at least 1.8 million sexualized images of women, according to separate estimates of X data by The New York Times and the Center for Countering Digital Hate.
Starting in late December, users on the social media platform inundated the chatbot’s X social media account with requests to alter real photos of women and children to remove their clothes, put them in bikinis and pose them in sexual positions, prompting a global outcry from victims and regulators.
In just nine days, Grok posted more than 4.4 million images. A review by the Times conservatively estimated that at least 41% of posts, or 1.8 million, most likely contained sexualized imagery of women. A broader analysis by the Center for Countering Digital Hate, using a statistical model, estimated that 65%, or just over 3 million, contained sexualized imagery of men, women or children.
The Rate at Which Grok Is Spreading AI Sexualized Images
The findings show how quickly Grok spread disturbing images, which earlier prompted governments in Britain, India, Malaysia and the United States to start investigations into whether the images violated local laws. The burst of nonconsensual images in just a few days surpassed collections of sexualized deepfakes, or realistic AI-generated images, from other websites, according to the Times’ analysis and experts on online harassment.
“This is industrial-scale abuse of women and girls,” said Imran Ahmed, the CEO of the Center for Countering Digital Hate, which conducts research on online hate and disinformation. “There have been nudifying tools, but they have never had the distribution, ease of use or the integration into a large platform that Elon Musk did with Grok.”
X and Musk’s AI startup, xAI, which owns X and makes Grok, did not respond to requests for comment. X’s head of product, Nikita Bier, said in a post on Jan. 6 that a surge of traffic over four days that month had resulted in the highest engagement levels on X in the company’s history, although he did not mention the images.
The interest in Grok’s image-editing abilities exploded Dec. 31, when Musk shared a photo generated by the chatbot of himself in a bikini, as well as a SpaceX rocket with a woman’s undressed body superimposed on top.
The chatbot has a public account on X, where users can ask it questions or request alterations to images. Users flocked to the social media site, in many cases asking Grok to remove clothing in images of women and children, after which the bot publicly posted the AI-generated images.
Between Dec. 31 and Jan. 8, Grok generated more than 4.4 million images, compared with 311,762 in the nine days before Musk’s posts. The Times used data from Tweet Binder, an analytics company that collects posts on X, to find the number of images Grok had posted during that period.
On Jan. 8, X limited Grok’s AI image creation to users who pay for some premium features, significantly reducing the number of images. Last week, X expanded those guardrails, saying it would no longer allow anyone to prompt Grok’s X account for “images of real people in revealing clothing such as bikinis.”
“We remain committed to making X a safe platform for everyone and continue to have zero tolerance for any forms of child sexual exploitation, non-consensual nudity, and unwanted sexual content,” the company said on X last week.
Since then, Grok has largely ignored requests to dress women in bikinis, but has created images of them in leotards and one-piece bathing suits. The restrictions did not extend to Grok’s app or website, which continue to allow users to generate sexual content in private.
Center for Countering Digital Hate Continues to investigate Grok
Some women whose images were altered by Grok were popular influencers, musicians or actresses, while others appeared to be everyday users of the platform, according to the Times’ analysis. Some were depicted drenched with fluids or holding suggestive props including bananas and sex toys.
The Times used two AI models to analyze 525,000 images posted by Grok from Jan. 1-7. One model identified images containing women, and another model identified whether the images were sexual in nature. A selection of the posts were then reviewed manually to verify that the models had correctly identified the images.
Separately, the Center for Countering Digital Hate collected a random sample of 20,000 images produced by Grok between Dec. 29 and Jan. 8, and found that about 65% of the images were sexualized. The organization identified 101 sexualized images of children. Extrapolating across the total, the group estimated that Grok had produced more than 3 million sexual images, including more than 23,000 images of children.
The organization classified the random images it sampled as sexual if they depicted a person in a sexual position, in revealing clothing like underwear or swimwear, or coated in fluid intended to look sexual.
(Musk sued the Center for Countering Digital Hate in 2023, claiming it broke the law when it collected data from X that showed a rise in hate speech after his acquisition of the company. The lawsuit was dismissed; an appeal is pending.)
Backlash and Frustrated Users
As sexual images flooded X this month, backlash mounted.
“Immediately delete this,” one woman posted Jan. 5 about a sexualized image of herself. “I didn’t give you my permission and never post explicit pictures of me again.”
While tools to create sexualized and realistic AI images exist elsewhere, the spread of such material on X was unique for its public nature and sheer scale, experts on online harassment said.
In comparison, one of the largest forums dedicated to making fake images of real people, Mr. Deepfakes, hosted 43,000 sexual deepfake videos depicting 3,800 individuals at the peak of its popularity in 2023, according to researchers at Stanford and the University of California, San Diego. The website shut down last year.
—
This article originally appeared in The New York Times.
By Kate Conger, Dylan Freedman and Stuart A. Thompson/Andria Lo
c.2026 The New York Times Company
RELATED TOPICS:
Categories
Tulare County Sheriff Adds Fugitive to Top 10 Most Wanted List
US Lays Out a Glittering Plan for Gaza, Including Skyscrapers
Microsoft Down for Thousands of Users, Downdetector Reports




