TL;DR
An investigation into xAI’s Grok reveals a disturbing industrial AI challenge: workers training the model are routinely exposed to explicit and illegal content, including AI-generated child sexual abuse material (CSAM). This is exacerbated by Grok’s built-in “sexy” and “spicy” modes, which experts warn create more gray areas for AI in detecting explicit content. While the AI industry at large grapples with a surge in AI-generated CSAM detection reports in 2025—from under 6,000 to over 440,000 in a year—xAI’s approach highlights critical failures in worker safety protocols and ethical data annotation for AI.
The Unseen Workforce: Why AI Tutors Are Paying a Psychological Price
“You have to have thick skin to work here, and even then it doesn’t feel good.” – A former xAI worker who quit over the content they encountered.
Imagine sitting at your desk, reviewing data to improve an AI, when a generated image depicting AI-generated CSAM flashes on your screen. This isn’t a hypothetical horror; it’s the reported reality for some workers at Elon Musk’s xAI. In conversations with more than 30 current and former workers, 12 told Business Insider they encountered sexually explicit material, including instances of user requests for AI child sexual abuse material detection.
This is the human foundation of the industrial AI boom. Behind every “witty” response from Grok lies a complex supply chain of industrial AI data annotation — the same hidden assembly line discussed in our piece on how human-in-the-loop workflows save millions. The very features that market Grok as an “anti-woke” alternative — its “sexy” and “unhinged” modes, its flirtatious female avatar — are the same features that experts say complicate the task of moderating synthetic media.
Why Grok’s Design Creates a Perfect Storm for Explicit Content
The “Spicy” Feature and Its Consequences
At the heart of this issue is Grok’s product design. Unlike competitors such as OpenAI and Google, which implement strict safeguards, xAI has rolled out features like a “spicy” setting in its Grok Imagine AI video generator that allows users to create sexually explicit content, including partial female nudity. This official sanctioning of NSFW content blurs the lines of acceptable use and highlights the limits of an automated NSFW content detection API — and the pitfalls are reflected across industrial AI tools facing false positives and bias (see our analysis of industrial AI false positives).
Celebrity Deepfakes and Non-Consensual AI Pornography
The problem is amplified by Grok’s ability to easily generate celebrity deepfakes. The Verge reported that the tool “didn’t hesitate to spit out fully uncensored topless videos of Taylor Swift.” This is a clear example of AI-generated Taylor Swift deepfakes, which fall under non-consensual AI pornography and are now targeted by new laws like the Take It Down Act.
The Moderation Gray Area
Stanford University researcher Riana Pfefferkorn explains the problem: “If you don’t draw a hard line at anything unpleasant, you will have a more complex problem with more gray areas.” By allowing a range of explicit outputs — from flirtatious roleplay to fully sexual media — xAI forces moderators and detection systems to make nuanced judgments, increasing both the risk of harmful content slipping through and of deep psychological effects on content moderation workers. The growing wave of AI resentment in industrial workplaces helps explain why employees feel pushed into risky roles with inadequate protections.
The Industrial-Scale Challenge of AI-Generated CSAM
The surge in AI-generated CSAM detection reports highlights the global scale of the crisis. The National Center for Missing and Exploited Children (NCMEC) reported a staggering jump — from tens of thousands to hundreds of thousands of reports in a short period. While OpenAI and Anthropic filed thousands of reports with NCMEC, xAI reportedly did not, raising questions about whether internal programs such as Project Rabbit (an internal transcription initiative) had adequate reporting AI-generated child exploitation material procedures.
Why Content Moderation Is the AI Industry’s Dirty Secret
The Human Toll of Data Annotation
The psychological effects of content moderation are devastating. Workers describe mental breakdowns, paranoia, and family breakdowns after prolonged exposure to extreme sexual violence content. At xAI, workers involved in transcription and image annotation report being immersed in audio porn and disturbing imagery; many say support mechanisms were insufficient. That lack of meaningful support is why mental-health-first approaches — like those explored in our AI mental health early-detection analysis — are critical when designing human-in-the-loop systems.
This highlights an urgent need for mental health support for AI tutors and stronger corporate responsibility in AI development.
The Technical Battle: Automating CSAM Detection
How AI Detectors Work
AI explicit content detectors use combinations of:
- Feature extraction — scanning for skin-tone regions, shapes and texture cues.
- Object detection — CNNs look for anatomical markers.
- Pattern recognition — classifiers compare content to known explicit patterns.
Major cloud providers and third-party moderation services expose automated NSFW content detection APIs to developers, but these systems need constant retraining. They also suffer from detection bias and edge-case problems — issues we unpack in our piece on industrial AI bias undermining team performance. In practice, automated systems are best viewed as a first line of defense that must be paired with careful human review.
Navigating the Future: Ethics, Laws, and Corporate Responsibility
Regulatory Crackdown
The Take It Down Act and similar laws now target non-consensual intimate images and AI deepfakes. NCMEC remains central to coordinating reports of AI-generated child sexual abuse material, and public pressure is pushing companies to adopt clearer AI ethics and safety guidelines. But regulation alone won’t solve the supply-chain problems — companies must invest in safer product design and transparent reporting.
Best Practices Moving Forward
Experts recommend:
- Robust human-in-the-loop workflows for edge-case escalations (see our guide on how human-in-the-loop workflows save millions).
- Mandatory mental-health programs tied to moderation tasks.
- Transparent disclosure of reporting practices and escalation protocols — the kind of transparency flagged in AI transparency at risk.
- Regular AI red teaming and ethical annotation standards to reduce harm before it reaches workers.
FAQ
What is AI-generated CSAM?
AI-generated CSAM refers to sexually explicit images, videos, or texts depicting minors created with AI. Even if no real child is involved, many jurisdictions treat it as illegal and it causes real harm.
How does Grok’s “spicy mode” work?
Grok Imagine’s AI video generator spicy mode lets users generate sexually explicit videos and celebrity deepfakes. Because Grok allows broader outputs than competitors, it creates more moderation gray areas.
What psychological support is offered to moderators?
Support varies, and in many reports it’s inadequate. Companies should look at proactive mental-health detection and intervention strategies like those covered in our AI mental health early-detection analysis.
How do AI content moderation policies compare?
OpenAI vs xAI moderation policies show differences in reporting transparency. OpenAI and other companies have publicly filed reports with NCMEC; xAI’s reporting practices have drawn scrutiny. The long-term fix will blend regulation, corporate transparency, and better annotation ethics.