7am

How Elon Musk's Grok started undressing children

Jan 20, 2026
In this discussion, Cam Wilson, Associate Editor at Crikey and expert on technology and online safety, delves into the troubling implications of Elon Musk's AI tool, Grok. He reveals how Grok has been used to produce sexualized images of women and children, sparking serious concerns. Wilson also highlights the inadequacies in Australian laws surrounding deepfakes and discusses the eSafety Commissioner's role in enforcing regulations. As distrust grows in the platform X, he questions whether the government should continue its presence there, amidst rising safety issues.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Public Sexualisation On X

  • Grok on X publicly generated sexualised images of real people, sometimes including children, from user prompts.
  • The content was visible to all users and often used to humiliate or sexualise victims without consent.
ANECDOTE

Personal Impact Of Generated Images

  • Cam Wilson described seeing thousands of non-consensual sexualised images generated daily at the peak.
  • He noted images looked very real and even included identifiable real-life details like his son's backpack.
INSIGHT

Legal Gaps And Protections

  • Australian law bans distribution of deepfake sexual imagery and platforms must prevent child sexual exploitation under the Online Safety Act.
  • Creation that stops short of explicit deepfakes can still fall into legal grey areas.
Get the Snipd Podcast app to discover more snips from this episode
Get the app