Three Tennessee-based plaintiffs filed a 44-page complaint Monday in federal court in San Jose, California, accusing Elon Musk's artificial intelligence company xAI of enabling the creation and distribution of sexually exploitative material involving minors. According to the lawsuit, one of the teens was alerted last December that someone was sharing AI-generated images and videos of her and other minors in settings with which she was familiar, but morphed into sexually explicit poses.
The images and videos were allegedly shared on Discord, Telegram and other platforms and used as a bartering tool for other CSAM imagery. One plaintiff said the person who ran the Discord server used Instagram photos of her wearing a blue bikini at the beach last October to generate images of her without clothing, and the alleged perpetrator was arrested in December. Law enforcement officials who investigated the images told the girls' parents they were created with xAI's Grok.
Though the lawsuit currently names three individuals, the complaint says that it could cover at least thousands of minors who have also had their photos manipulated by Grok into sexualized images. The proposed case seeks to represent a broad class of victims, suggesting the problem extends far beyond the named plaintiffs.
The proposed class-action lawsuit alleges xAI recklessly designed Grok to enable such abuse, and then, amid a public outcry, restricted the technology to paid subscribers and third-party companies rather than fix the problem. The plaintiffs allege that xAI did not take basic precautions used by other frontier labs to prevent their image models from producing pornography depicting real people and minors.
The scale of the issue proved substantial. The Center for Countering Digital Hate estimated Grok generated more than 3 million sexualized images in just 11 days, including over 23,000 images involving children. These images were generated between December 29, 2025 and January 8, the time period between the launch of Grok's photo-editing feature and when it was restricted to paid users after the feature caused public uproar, governmental investigations, and statements by children's rights organizations.
The controversy highlights a fundamental tension in AI development. The complaint states that a model that can create sexualized images of adults cannot be prevented from creating CSAM of minors. Supporters of stronger AI regulation argue this demonstrates why safeguards must be built in from the beginning rather than added later as problems emerge. Those concerned about government overreach counter that rapid innovation sometimes outpaces safety measures, and that the responsibility ultimately lies with bad actors abusing legitimate tools rather than the tools themselves.
The company's response has drawn criticism. After accusing governments of censorship, Musk first stopped allowing images to be generated in Grok except to paid subscribers; on January 15, xAI stopped allowing Grok to undress people in images. Last week, Musk said in a post on X that if it's allowed in an R-rated movie, it's allowed by Grok. The company has not publicly responded to the lawsuit.
Brazil, Britain, Spain and other countries were investigating the company over these allegations. Attorneys for the plaintiffs say that because third-party usage still requires xAI code and servers, the company should be held responsible. The case will test whether AI companies can be held liable for how their technology is misused by third parties, a question likely to shape the emerging legal landscape around generative AI.