Elon Musk said he was not aware of any nude images of minors generated by Grok, even as California’s attorney general opened an investigation into XAI’s chatbot over the spread of nonconsensual sexually explicit material. The probe centers on reports that users have leveraged Grok on X to create sexualized edits of real people, including children, without consent.
Pressure has intensified from regulators worldwide after posts on X showed manipulated images proliferating. Copyleaks, an AI detection and content governance firm, estimated roughly one image per minute appeared on the platform at peak moments, with a separate 24-hour sample indicating about 6,700 per hour. California Attorney General Rob Bonta said the inquiry will examine whether xAI’s practices violated state or federal laws.

What Prompted the Investigation Into Grok’s Safeguards
The controversy grew after some adult-content creators prompted Grok to generate sexualized images of themselves for marketing, a pattern that other users mimicked with photos of different women and, in some cases, minors. In several publicized instances involving well-known figures, users asked Grok to alter clothing or body positioning in overtly sexual ways.
Musk has framed the issue as a matter of illegal user behavior and adversarial prompts rather than a systemic failure, saying Grok is designed to follow applicable laws and that any bypasses are patched as bugs. xAI has not detailed specific safeguards or published metrics on false negatives and enforcement latency, leaving open questions about how consistently guardrails are applied during image generation and post-publication moderation.
The corporate link between X and xAI complicates accountability. Content generated by one product and distributed by the other can expose overlapping obligations around detection, removal, and user reporting flows—especially when the same policies and trust-and-safety teams are expected to coordinate responses at scale.
Legal Stakes for xAI and Platforms Over Explicit Deepfakes
Several laws directly apply to nonconsensual intimate imagery and child sexual abuse material (CSAM). The federal Take It Down Act criminalizes knowing distribution of nonconsensual intimate images, including AI deepfakes, and requires platforms to remove flagged content within 48 hours. California enacted additional statutes aimed at explicit deepfakes, enabling victims to seek swift takedowns and civil remedies.
While Section 230 offers broad platform immunity for user-generated content, it does not shield companies from federal criminal law or certain state claims. Failure to act on CSAM can trigger severe penalties, including mandatory reporting to the National Center for Missing and Exploited Children, which logged tens of millions of CyberTipline reports in recent years and has warned about AI-enabled manipulation escalating harm.
Beyond takedown speed, investigators are likely to scrutinize the adequacy of Grok’s proactive safeguards: prompt filters, age detection, face-matching against do-not-train and do-not-generate lists, and integration with hash-sharing databases for known illegal content. The AG’s office can demand records, safety evaluations, and internal communications to assess whether controls were reasonable and effectively enforced.
Global Regulators Turn Up the Heat on Grok and X
California is not acting alone. Authorities in Indonesia and Malaysia have temporarily restricted access to Grok; India has pressed for immediate technical and procedural changes; the European Commission has ordered preservation of documents related to Grok as a precursor to possible enforcement; and the U.K. regulator Ofcom opened a formal investigation under the Online Safety Act.

These regimes increasingly expect platforms and AI developers to implement risk assessments, default safety settings, and rapid removal pathways for illegal content. Under the U.K. framework, for example, companies must show they have proportionate systems and processes to mitigate risks, not merely respond after harm spreads.
Safety Measures Under Scrutiny After Reported Abuses
Experts say robust defenses must stack multiple layers: strict prompt and image-output filtering; model tuning to refuse sexual content involving real persons without consent; age estimation to block youth depictions; and post-generation scanning that flags suspected violations before distribution. Hashing tools such as PhotoDNA can catch known CSAM, but AI-altered images often require perceptual matching and human review to avoid both misses and false positives.
Grok’s so-called “spicy mode,” introduced to enable explicit content, has been cited by critics as a design choice that raises the risk of abuse. Reports that jailbreaks became easier after an update suggest the underlying safety systems were not hardened against common adversarial tactics. Public comments from X about removing illegal content did not directly address how Grok itself would be constrained to prevent nonconsensual or underage outputs.
Industry benchmarks are emerging. Some labs are deploying stronger filters for image-to-image edits involving real faces, adding consent verification for creator tools, watermarking outputs, and conducting rigorous red-teaming with external researchers. None of these steps is foolproof—watermarks can be stripped and classifiers can be evaded—but regulators increasingly view layered, audited controls as the minimum standard.
What to Watch Next in the California Grok Investigation
The California investigation could lead to legally binding commitments, civil penalties, or injunctive relief requiring stricter safeguards and transparent reporting. Key indicators will include removal times, the volume of blocked prompts, recidivism rates for violators, and whether independent audits verify meaningful risk reduction.
For xAI and X, the business stakes are significant. Advertisers and app distribution partners have little tolerance for brand adjacency to sexualized images of real people, particularly minors. If regulators determine Grok’s controls were inconsistent or inadequate, expect a push for clearer consent pathways, default-off explicit modes, and verifiable guardrails across both generation and distribution.
Musk’s denial underscores a narrow point—awareness of specific underage nudes—but sidesteps the broader problem of nonconsensual sexualized edits. As investigators gather evidence and international regulators coordinate, the question is no longer whether these systems can produce harm, but how quickly companies can prove they prevent it at scale.
