Content
summary Summary

While AI image generators are mainly used for creative purposes, they also open the doors for fake nude photos. AI artist Denis Shiryaev wants to educate the public about this.

With great power comes great responsibility – a phrase that also applies to AI image generators. Because closely linked to them are deepfakes, which we define as AI-generated synthetic visual media for ethically questionable or criminal purposes.

The history of deepfakes shows that public AI image generators have their origins in pornographic images and videos without consent. Especially celebrities, for whom there is plenty of AI training material on the Internet, were the target of early deepfake creators.

But with the latest tools, anyone can become a victim – all it takes is a single image as a template, as AI artist Denis Shiryaev shows.

Ad
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer

Undressed by AI

Technological achievements like img2img (“picture to picture”) or inpainting, i.e. replacing certain image content with your own visuals, make it even easier to create deepfakes.  That’s why we need to talk about what people with bad intentions can do with the technology that is currently available.

Anyone who hears about deepfakes probably first thinks of entertaining videos like that of the fake Margot Robbie or Tom Cruise on TikTok, or the possible spread of misinformation. Things get intimate when people are virtually stripped and their underwear is replaced with realistic-looking sexual characteristics.

Example of DeepNude’s capabilities from 2019. | Image: DeepNude via Motherboard / The Verge

A software called DeepNude was capable of doing this three years ago. The program made headlines and spread quickly, but was just as quickly discontinued due to heavy criticism. The reason given by the operators at the time: “The world is not ready yet.”

Whether the world is ready in 2022 remains to be seen, but technology is advancing unabated.

Image AI providers ban NSFW content, but those bans have limited effect

The companies behind DALL-E 2, Midjourney, or Stable Diffusion know the risks and are taking precautions, at least on their platforms, such as word filters and permanent blocks for repeated violations of content guidelines.

Recommendation

No adult content or gore. Please avoid making visually shocking or disturbing content. We will block some text inputs automatically,” for example, Midjourney’s terms of use state.

Stability AI, the company behind Stable Diffusion, also prohibits the generation of “NSFW, lewd, or sexual material” for its DreamStudio web service. The restrictions of the open-source model, however, are somewhat different: only sexual content without the consent of the people “who might see it” is prohibited.

Furthermore, compliance with rules for open-source models cannot be controlled. This is also not possible for other freely available generative image AIs, which do not need to be as powerful as the known models, but could probably be fine-tuned with manageable effort, especially for special use cases like nudity. DeepNude already achieved this three years ago.

Shiryaev sees another DeepNude scandal brewing

Denis Shiryaev, CEO of AI toolbox “neural.love” and known as an AI artist, thinks a new DeepNude scandal is possible once the public learns how easy it is to use Stable Diffusion to turn a photo with clothes into a nude photo.

Ad
Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.

He demonstrates AI-generated nudity using Stable Diffusion as an example. To do so, he commits a violation of the open-source model’s guidelines.

For a test, AI artist Denis Shiryaev violated Stable Diffusion’s guidelines to fake a nude image via inpainting. It worked easily, according to Shiryaev. | Image: Denis Shiryaev via Twitter

Today’s software is also much easier to use than the GAN-based DeepNude, Shiryaev says. Basically, any person who has even a single photo available on the Internet or on a hard drive is at risk of being defamed by fake nude photos.

“I believe it is better to prepare the general audience (outside of the AI industry bubble) that not all that they see in photos is real anymore, since almost anyone can reproduce those results quite easily and fast,” Shiryaev writes on Twitter.

Support our independent, free-access reporting. Any contribution helps and secures our future. Support now:
Bank transfer
Summary
  • In 2019, an AI-based app made the rounds that allowed clothed people to be unclothed. The clothes were replaced with fantasized naked body parts.
  • AI artist Denis Shiryaev now shows that this approach is even easier with today's AI image generators.
  • A critical mass of, for example, training images is no longer needed; a single image as a template is sufficient, and the quality is higher.
Jonathan is a freelance tech journalist who focuses primarily on consumer electronics. He educates his fellow human beings about how easily AI can already be used today and how it can support everyday life.
Join our community
Join the DECODER community on Discord, Reddit or Twitter - we can't wait to meet you.