Page Loader
Summarize
Instagram is profiting from ads promoting nonconsensual nude image creation
The usage of such image generators is spreading

Instagram is profiting from ads promoting nonconsensual nude image creation

Apr 23, 2024
04:35 pm

What's the story

Instagram, a Meta subsidiary, is profiting from advertisements that encourage the creation of nonconsensual nude images through AI-powered apps. These ads are not hidden but openly promoted on social media platforms. Despite several such ads being removed as indicated by Meta's Ad Library, many continue to lure users into generating explicit content. One disturbing ad featured a photo of Kim Kardashian with the caption "Undress any girl for free. Try It."

Functionality

Clicking on ad redirects users to a nude image generator

Clicking on the aforementioned ad redirects users to a popular web-based app that generates nonconsensual nude images. Users can upload a photo of a real person and pay for the app to create an image depicting the person as nude. Multiple versions of this ad were active on Instagram and Facebook between March 13 and 19, according to Meta's Ad Library.

Loophole

The apps were labeled as 'art generators' on app stores

Another ad showcased an AI-created split image of a woman, clothed on one side and nude on the other, with the text "Try it now. Any clothing DELETE. 1. Erase anywhere. 2. Output image! This APP can delete..." These ads and their variations were active across Facebook, Instagram, and Facebook Messenger, from April 10-20. Clicking these ads would lead users to the Apple App Store where these apps are labeled as "art generators," cleverly bypassing policies against promoting adult content.

Enforcement

Meta has only removed some of these ads

Despite these apps advertising their nonconsensual deepfake porn creation capabilities on various sites, Meta has only removed some of these ads after they were reported by journalists. The company continues to struggle with enforcing its policies regarding who can purchase ads on its platforms. "Meta does not allow ads that contain adult content and when we identify violating ads we work quickly to remove them, as we're doing here," stated Meta Spokesperson Daniel Roberts.

App proliferation

The usage of these apps is spreading

Despite negative reviews and some apps failing to generate nude images as advertised, the usage of these "undress" apps has spread. This trend has even led to the arrest of two middle school students in Florida. These apps are readily available on Google and Apple app stores, making nonconsensual AI-generated porn easier to create. Other social media platforms like TikTok and X have also been reported to promote these types of apps.