AI distances itself from adult content that once drove the tech revolution
TL;DR
OpenAI scrapped plans for 'erotica for verified adults' last week following pressure from investors and internal safety teams. The trigger: xAI's Grok generated illegal child sexual abuse material when prompted, and users could still produce non-consensual sexualized images even after a safety patch. ChatGPT's age-prediction error rate was too high to reliably block minors from accessing explicit content.
Nauti's Take
The Grok incident was a foreseeable PR disaster: deploy a chatbot toward adult content without airtight safety architecture and you get exactly what you deserve. OpenAI's course correction sounds responsible, but it also reads as a classic 'we tried, it got too hot' pivot.
The genuinely interesting part: adult entertainment has historically co-funded nearly every major tech leap, from VHS to broadband. Big AI skipping that chapter doesn't make the demand disappear – it just hands the market to whoever has fewer scruples and weaker safety controls.
Briefingshow
Big Tech's retreat from AI adult content reveals how sharply liability and reputational risk now override product ambition. Crucially, it wasn't voluntary self-restraint that forced the issue – it was a concrete abuse case with Grok generating illegal material. As OpenAI and others back away, smaller and less regulated players will fill the gap, meaning the real risks to minors and victims of non-consensual imagery are far from resolved.