Weekend Times


Google Workspace

Business News

Sexualised deepfakes on X are a sign of things to come. NZ law is already way behind

  • Written by Cassandra Mudgway, Senior Lecturer in Law, University of Canterbury
Sexualised deepfakes on X are a sign of things to come. NZ law is already way behind

Elon Musk finally responded last week to widespread outrage[1] about his social media platform X letting users create sexualised deepfakes with Grok, the platform’s artificial intelligence (AI) chatbot.

Musk has now assured the United Kingdom government he will block Grok from making deepfakes[2] in order to comply with the law. But the change will likely only apply to users in the UK.

These latest complaints were hardly new, however. Last year, Grok users were able to “undress” posted pictures[3] to produce images of women in underwear, swimwear or sexually suggestive positions. X’s “spicy” option[4] let them to create topless images without any detailed prompting at all.

And such cases may be signs of things to come if governments aren’t more assertive about regulating AI.

Despite public outcry and growing scrutiny from regulatory bodies[5], X initially made little effort to address the issue and simply limited access[6] to Grok on X to paying subscribers.

Various governments took action, with the UK announcing plans[7] to legislate against deepfake tools, joining Denmark[8] and Australia[9] in seeking to criminalise such sexual material. UK regulator Ofcom launched an investigation[10] of X, seemingly prompting Musk’s about-turn.

So far, the New Zealand government has been silent on the issue, even though domestic law is doing a poor job of preventing or criminalising non-consensual sexualised deepfakes.

Holding platforms accountable

The Harmful Digital Communications Act 2015[11] does offer some pathways to justice, but is far from perfect. Victims are required to show they’ve suffered “serious emotional distress”, which shifts focus to their response rather than the inherent wrong of non-consensual sexualisation.

Where images are entirely synthetic rather than “real” (generated without a reference photo, for example), legal protection becomes even less certain.

A members’ bill[12] is expected to be introduced later this year that would criminalise the creation, possession and distribution of sexualised deepfakes without consent.

This reform is both necessary and welcome. But it only tackles part of the problem.

Criminalisation holds individuals accountable after harm has already occurred. It does not hold companies accountable for designing and deploying the AI tools that produce these images in the first place.

We expect social media providers to take down child sexual abuse material[13], so why not deepfakes of women? While users are responsible for their actions, platforms such as X provide an ease of access that removes the technical barrier to deepfake creation.

The Grok case has been in the news for many months, so the resulting harm is easily foreseeable. Treating such incidents as isolated misuse distracts from the platform’s responsibility.

Light-touch regulation is not working

Social media companies (including X) have signed the voluntary Aotearoa New Zealand Code of Practice for Online Safety and Harms[14], but this is already out of date.

The code does not set standards for generative AI, nor does it require risk assessments prior to implementing an AI tool, or set meaningful consequences for failing to prevent predictable forms of abuse.

This means X can get away with allowing Grok to produce deepfakes while still technically complying with the code.

Victims could also hold X responsible by complaining to the Privacy Commissioner[15] under the Privacy Act 2020[16].

The commissioner’s guidance on AI[17] suggests that both the use of someone’s image as a prompt and the generated deepfake could count as personal information.

However, these investigations can take years, and any compensation is usually small. Responsibility is often split among the user, the platform and the AI developer. This does little to make platforms or AI tools such as Grok safer in the first place.

New Zealand’s approach reflects a broader political preference[18] for light-touch AI regulation that assumes technological development will be accompanied by adequate self-restraint and good-faith governance.

Clearly, this isn’t working. Competitive pressures to release new features quickly prioritise novelty and engagement over safety, with gendered harm often treated as an acceptable byproduct.

A sign of things to come

Technologies are shaped by the social conditions in which they are developed and deployed. Generative AI systems trained on masses of human data inevitably absorb misogynistic norms.

Integrating these systems into platforms without robust safeguards allows sexualised deepfakes that reinforce existing patterns of gender-based violence.

These harms extend beyond individual humiliation. The knowledge that a convincing sexualised image can be generated at any time – by anyone – creates an ongoing threat[19] that alters how women engage online.

For politicians and other public figures, that threat can deter participation in public debate[20] altogether. The cumulative effect is a narrowing of digital public space.

Criminalising deepfakes alone won’t fix this. New Zealand deserves a regulatory framework that recognises AI-enabled, gendered harm as foreseeable and systemic.

That means imposing clear obligations on companies that deploy these AI tools, including duties to assess risk, implement effective guardrails, and prevent predictable misuse before it occurs.

Grok offers an early signal of the challenges ahead. As AI becomes embedded across digital platforms, the gap between technological capabilities and legislation will continue to widen unless those in power take action.

At the same time, Elon Musk’s response to legislative action in the UK demonstrates how effective political will and robust regulation can be.

The authors acknowledge the contribution of Chris McGavin to the preparation of this article.

References

  1. ^ widespread outrage (www.rnz.co.nz)
  2. ^ block Grok from making deepfakes (www.theguardian.com)
  3. ^ “undress” posted pictures (www.404media.co)
  4. ^ “spicy” option (www.theverge.com)
  5. ^ regulatory bodies (www.ofcom.org.uk)
  6. ^ limited access (www.rnz.co.nz)
  7. ^ announcing plans (www.bbc.com)
  8. ^ Denmark (www.theguardian.com)
  9. ^ Australia (www.legislation.gov.au)
  10. ^ launched an investigation (www.ofcom.org.uk)
  11. ^ Harmful Digital Communications Act 2015 (www.legislation.govt.nz)
  12. ^ members’ bill (www.legislation.govt.nz)
  13. ^ take down child sexual abuse material (www.legislation.govt.nz)
  14. ^ Aotearoa New Zealand Code of Practice for Online Safety and Harms (thecode.org.nz)
  15. ^ Privacy Commissioner (www.privacy.org.nz)
  16. ^ Privacy Act 2020 (www.legislation.govt.nz)
  17. ^ guidance on AI (www.privacy.org.nz)
  18. ^ political preference (www.beehive.govt.nz)
  19. ^ creates an ongoing threat (onlinelibrary.wiley.com)
  20. ^ deter participation in public debate (www.ipu.org)

Authors: Cassandra Mudgway, Senior Lecturer in Law, University of Canterbury

Read more https://theconversation.com/sexualised-deepfakes-on-x-are-a-sign-of-things-to-come-nz-law-is-already-way-behind-273562

The Weekend Times Magazine

Why You Should Hire an Agent When Shopping For a Luxury Home

Many home buyers find themselves in a conundrum when they think about buying a luxury property. They're excited to shop for such an amazing home, but overwhelmed by the amount...

How Homeowners Can Prepare for Asbestos Removal in Melbourne

If you own an older home in Melbourne, there’s a chance asbestos-containing materials may be present somewhere on your property. That’s why professional asbestos removal Melbourne services are necessary, but as...

Prime Minister interview with Karl Stefanovic and Alison Langdon, Today

KARL STEFANOVIC: Joining us now from Kirribilli House in Sydney. PM, good morning to you. Thank you for your time.    PRIME MINISTER: G’day Karl.   STEFANOVIC: The Premiers don't seem to be listening...

Make Your Holiday Merry with Christmas Inflatables

The holiday season is all about bringing joy and festivity to your home or event. One of the most fun and visually captivating ways to do this is by incorporating...

Car subscription offers part-time workers access to a car during COVID-19

New research commissioned by Carly, Australia’s first flexible car subscription provider, surveyed more than 1200 Australians and found that 48% of part time workers would consider car subscription instead of...

How Offroad Caravans For Sale Offer Freedom, Durability And Adventure-Ready Performance

Australians who love exploring rugged landscapes and remote travel destinations often look for reliable offroad caravans for sale that can withstand tough terrain while providing a comfortable home on wheels...

When AEC IT breaks, It Rarely Looks Like IT

AEC businesses rarely lose time to one dramatic outage. What hurts more are the small, repeat delays that show up right when pressure is highest. The drawing set stalls, the...

Essential Packaging Materials That Support Business Efficiency in Melbourne

Reliable access to packaging materials is a key factor in how smoothly businesses operate across storage, shipping, and distribution processes. Companies that depend on packaging supplies Melbourne understand that packaging is...

Effective Pest Control Solutions in Ballarat: What You Need to Know

Living in Ballarat means enjoying a beautiful regional lifestyle, but it also comes with its share of challenges—one of which is managing pests. From termites threatening your home’s structure to...

hacklink hack forum hacklink film izle hacklink testjetbahistipobetslogan bahis girişirocket play casino australiapadişahbetgalabetcasinolevantcasibommarsbahiscasibomjojobet