bionglass.blogg.se

Open source nsfw image cleaner
Open source nsfw image cleaner










open source nsfw image cleaner

Education: Automatically filter or flag explicit content in student-generated assignments or papers.Gaming: Detect and remove explicit language or imagery from in-game chat or player-generated content.E-commerce: Automatically filter out or flag explicit content in product descriptions or user-generated reviews.Social Media Monitoring: Automatically moderate user-generated content and detect explicit content.Cyber Security: Detect and remove explicit images, and protect children from inappropriate content.Online Content Moderation: Automatically filter or flag potentially offensive or adult content in various applications such as social media, e-commerce, and messaging.Here are some examples of common use cases: You can use Explicit Content Detection in numerous fields. The jury is still out as to whether this will actually help people who are the victim of this sort of thing without their consent.‍ Some Explicit Content Detection API use cases The company highlights that if you specifically provoke the AI into generating NSFW images, it might, but that it is implementing filters to prevent this from happening accidentally. UPDATE: The Prisma Labs team replied to our concerns. This seems like a very good idea, but the internet is a hard-to-govern place at the best of times, and we’re collectively facing a wall of legal, moral and ethical quandaries. government to push for laws criminalizing the dissemination of non-consensual nude photos. These platforms, and the unfettered proliferation of other so-called “deepfake” platforms, are turning into an ethical nightmare, are prompting the U.K. It appears that if you have 10-15 “real” photos of a person and are willing to take the time to photoshop a handful of fakes, Lensa will gladly churn out a number of problematic images.ĪI art generators are already churning out pornography by the thousands of images, exemplified by the likes of Unstable Diffusion and others. Adding NSFW content into the mix, and we are careening into some pretty murky territory very quickly: your friends or some random person you met in a bar and exchanged Facebook friend status with may not have given consent to someone generating soft-core porn of them. The ease with which you can create images of anyone you can imagine (or, at least, anyone you have a handful of photos of), is terrifying. The big turning point, and the ethical nightmare, is the ease with which you can create near-photorealistic AI-generated art images by the hundreds without any tools other than a smartphone, an app and a few dollars. But so far, getting those to look realistic takes a lot of skill with photo editing tools along with hours, if not days, of work. Just because it’s common doesn’t make it right - in point of fact, celebrities absolutely deserve their privacy and should definitely not be made victims of non-consensual sexualized depictions. Generating saucy images of celebrities is one thing, and as illustrated by the source images we were able to find, there has long been people on the internet who are willing to collage some images together in Photoshop. We might not be prepared for the consequences. Out of the 100-image set, 11 were topless photos of higher quality (or, at least with higher stylistic consistency) than the poorly done edited topless photos the AI was given as input.ĪI is getting better at generating porn. It turns out the AI takes those Photoshopped images as permission to go wild, and it appears it disables an NSFW filter. The second set, however, was a lot spicier than we were expecting. The first set of images was in line with the AI avatars we’ve seen Lensa generate in the past. Another set, based on the same 15 photos, but with an additional set of five photos added of the same actor’s face, Photoshopped onto topless models.One set, based on 15 photos of a well-known actor.To verify that Lensa will create the images it perhaps shouldn’t, we created two sets of Lensa avatars: It seemed like the kind of thing that shouldn’t have been possible, so we decided to try it ourselves. TechCrunch has seen photo sets generated with the Lensa app that include images with breasts and nipples clearly visible in the images with faces of recognizable people. Now there’s another reason to fly the flag: As it turns out, it’s possible - and way too easy - to use the platform to generate non-consensual soft porn. Lensa has been climbing the app store hit lists with its avatar-generating AI that is making artists wave the red flag.












Open source nsfw image cleaner