Three plaintiffs from Tennessee, including two minors, have filed a lawsuit against Elon Musk’s artificial intelligence company xAI, alleging that its Grok image generator enabled the creation of sexually explicit content using real photographs of individuals.
Lawsuit targets Grok image generator
The complaint, filed in a federal court in San Jose, California, claims that xAI knowingly designed its system to allow users to generate explicit images based on real people. The plaintiffs are seeking class action status for individuals across the United States who were “reasonably identifiable” in such generated content.
According to the filing, the system allegedly allowed manipulation of real images into sexualised content without adequate safeguards. As a result, the plaintiffs argue that the technology exposed individuals to harm and misuse.
Allegations involving minors
The lawsuit states that all three plaintiffs were minors when the images were created. It alleges that their photographs, including school and family pictures, were digitally altered into explicit material and circulated online.
Furthermore, the plaintiffs claim this caused significant emotional distress and created a broader public nuisance. They argue that xAI failed to implement protections to prevent the generation of such content involving minors.
In a statement, the plaintiffs’ counsel said the case involves children whose images were turned into exploitative material. The legal team also alleged that the system prioritised engagement and financial gain over user safety.
xAI response and regulatory scrutiny
xAI did not immediately respond to requests for comment following the lawsuit. However, the company had earlier introduced restrictions after backlash over explicit content generated by its chatbot.
In January, xAI said it blocked users from editing images of real people in revealing clothing and restricted such outputs in jurisdictions where they are illegal.
Meanwhile, governments and regulators worldwide have increased scrutiny of artificial intelligence platforms. They have launched investigations, imposed bans and demanded stricter safeguards to prevent the spread of harmful or illegal content.
Legal demands and broader implications
The plaintiffs are seeking unspecified damages, legal fees and a court order to stop the alleged practices. The case could have wider implications for how AI companies design and monitor image generation tools.
As scrutiny intensifies, the lawsuit highlights ongoing concerns about safety, consent and accountability in rapidly evolving AI technologies.
With inputs from Reuters

