Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
AI Government The Internet

White House Gets Voluntary Commitments From AI Companies To Curb Deepfake Porn (engadget.com) 50

In a statement today, the White House said it has received commitments from several AI companies to curb the creation and distribution of deepfake porn, also known as image-based sexual abuse material. Engadget reports: The participating businesses have laid out the steps they are taking to prevent their platforms from being used to generate non-consensual intimate images (NCII) of adults and child sexual abuse material (CSAM). Specifically, Adobe, Anthropic, Cohere, Common Crawl, Microsoft and OpenAI said they'll be: "responsibly sourcing their datasets and safeguarding them from image-based sexual abuse."

All of the aforementioned except Common Crawl also agreed they'd be: "incorporating feedback loops and iterative stress-testing strategies in their development processes, to guard against AI models outputting image-based sexual abuse" and "removing nude images from AI training datasets" when appropriate. [...] The notable absences from today's White House release are Apple, Amazon, Google and Meta.

This discussion has been archived. No new comments can be posted.

White House Gets Voluntary Commitments From AI Companies To Curb Deepfake Porn

Comments Filter:
  • by Iamthecheese ( 1264298 ) on Thursday September 12, 2024 @07:42PM (#64784207)
    So now I'm abusing someone if I so much as have my own computer make a fake with their face on it? This has gone several steps too far. How to reacquire liberty?
    • by gweihir ( 88907 ) on Thursday September 12, 2024 @08:01PM (#64784251)

      Too many people are profiting from these claims. Hence you would have to reform a major part of law-enforcement, politics and some part of the IT security industry.

      As to whether that is doable, refer to the "War on drugs".

      • Too many people are profiting from these claims.

        Profit? Man the models required to do face swap / body impersonation are openly published and there's plenty of open source projects that will let you play around with the concept and I doubt any of them have a magic - "fail if I see a titty" component in their models.

        • by gweihir ( 88907 )

          And if you could read, you would have understood that I was referring to something else.

    • by 1nt3lx ( 124618 )

      Your freedom to swing your arms ends at my nose.

      • by Baron_Yam ( 643147 ) on Thursday September 12, 2024 @08:52PM (#64784323)

        Which means distributing a malicious deepfake should be a crime, which actually is probably already covered under numerous other laws. Actually creating it for personal use if nobody ever finds out? Banning that is getting creepily close to thoughtcrime laws.

        • by AmiMoJo ( 196126 )

          All they are doing is asking them not to offer this as a service. Not even banning offering that service.

          You made a huge leap to it being *illegal* to even create them for "personal use".

          Clearly AI deepfake nudes are not the same as just drawing them or photoshopping them yourself. We didn't use to see an epidemic of children using those things to abuse classmates, or people advertising to photoshop photos for the purpose of sexual blackmail. Asking companies to not help criminals doesn't seem like anything

    • by AmiMoJo ( 196126 )

      "So now"? It's always been that way. It's just that it was harder to do and the results often were not very good, and many times the victim never found out anyway.

      Now there are companies offering this as a service using AI, no skill required. Children are using it to bully each other, it is so accessible.

      It doesn't seem at all unreasonable to ask AI companies to try to limit this kind of abuse. Even Photoshop has restrictions on things like editing and printing images of dollar bills. This doesn't affect yo

      • Even Photoshop has restrictions on things like editing and printing images of dollar bills.

        I know Xerox type machines have prevention measures for money...but this Photoshop thing must be relatively new.

        There were no such boundaries I knew of when I used PS...that was before they went to the "rental" model.

        But you could manipulate any images you pleased as far as I know prior to that. I don't believe any other of the competitive tools prevent this, like Affinity Photo or GIMP....

        • by AmiMoJo ( 196126 )

          Bank note detected was added to the first version of Photoshop CS, which was released in 2003.

          Colour photocopiers had it even before that.

    • Many people look similar, near doubles can always be found. Same for speech and mannerisms. See SNL - parody has been going on for ages, legally. Fewer can do spoonerisms, but I digress. Leftists are distressed that AI will create provably different faces, with mathematical proof that are not the same. Courts should not rubber stamp or entertain such cases. HOWEVER all AI images should state they are AI images and have the same in EXIF metadata. No new laws are needed. Existing misrepresentation laws cover
    • So now I'm abusing someone if I so much as have my own computer make a fake with their face on it? This has gone several steps too far. How to reacquire liberty?

      Good question. The answers might not be simple.

      A "White House"* that gets "voluntary" agreements to censor everything just might not be the answer?

      • A "White House"* that gets "voluntary" agreements to censor everything just might not be the answer?

        It's nothing new for them to try...ie: Trying to stop "misinformation" at Facebook and Twitter a few years ago for instance.

        We need to fight this tooth and nail...this is NOT something the govt. is there for....

  • by algaeman ( 600564 ) on Thursday September 12, 2024 @07:47PM (#64784215)
    These sorts of agreements have worked out so well for the public in recent history. I'm sure there's no way they would renege on such a binding commitment, once the government commits to not legislating.
    • There's a reason strip clubs are called gentleman's clubs.

    • These sorts of agreements have worked out so well for the public in recent history. I'm sure there's no way they would renege on such a binding commitment, once the government commits to not legislating.

      Those commitments don't work when there's a strong motivation to break them, that's not that case here. This is more a "we commit not to do something incredibly controversial for virtually zero benefit".

      Even Grok isn't interested in pron, despite letting through virtually everything else go [acs.org.au].

      I'm not sure why Apple, Amazon, Google and Meta weren't in the list, though I'm guessing it's more to do with internal red tape than wanting to make Deepfake porn.

      • This is more a "we commit not to do something incredibly controversial for virtually zero benefit".

        This is pointless. The underlying models don't make decisions about how they are used. You can't train a simple face swap model without someone throwing it in an open source wrapper innocently, and then some idiot on 4chan using it to copy taylor swift's face on the body of a porn star.

        That's not how this works. AI models can't be restricted from doing things without damage of the underlying model. At best companies can lock down their models, but there's already plenty of publicly published models out ther

    • What agreements could possibly work here? The models for detecting and manipulating faces are openly published and are just models, they can't make decisions about the content they are being used for. The tools required to use these models exist in the open source world.

      Closing the barn door doesn't help when you have no horses left.

  • by Anonymous Coward

    Won't this make it easier, if not less expensive, to use real people? Does this mean that the government thinks that only real people should be used to make pornography?

  • by account_deleted ( 4530225 ) on Thursday September 12, 2024 @09:06PM (#64784337)
    Comment removed based on user account deletion
  • by Powercntrl ( 458442 ) on Thursday September 12, 2024 @09:07PM (#64784339) Homepage

    Granted, I'll be the first to admit I've never tried to find an AI porn generator, but it certainly doesn't seem like it's as easy as running a Google search. I've already run headfirst into the "responsible AI guidelines" on one of the major ones trying to make some silly cartoon artwork for AI sung songs, so I doubt any of the mainstream AI stuff will let you make porn. Heck, they don't even let you make political cartoons of well-known celebs without complaining that you're being too naughty.

    By comparison, doing the ol' "head switcheroo" on Photoshop is brain dead easy these days.

    • by AmiMoJo ( 196126 )

      Headshot swapping in Photoshop is not trivial. You need two images where the heads are in mostly the same position, and things like skin tones mostly match. Hair has to be right, or at least easy to remove. Even then, it takes some skill to do convincingly.

      Harassment on social media is now pretty widespread. Someone posts a normal photo of themselves, and others respond with deepfakes of them. They are trivial to generate, it takes seconds and costs almost nothing. The sites are not difficult to find, and s

      • Headshot swapping in Photoshop is not trivial. You need two images where the heads are in mostly the same position, and things like skin tones mostly match. Hair has to be right, or at least easy to remove. Even then, it takes some skill to do convincingly.

        Anyone with a modicum of experience can headswap in PS quite easily and do a good job of it.

        Blending tones and colors is basic stuff for editing photos....

    • The last time we talked about this, I used a computer at the library to do a Google search. Let me just say that it was definitely as easy as running a Google search. Feel free to repeat my experiment, but I suggest not using your own device.

      How legitimate are the services? I don't know. I uploaded a photo of myself and about half of the services errored out. The other half stated that they had made something great and that I could enter my credit card number to see it. Needless to say I did not pr

  • I always tell the truth and I never lie. Honest.
  • Don't think i will trust any of this to not be abused to claim something is fake but turns out to be true and something that is true turns out to be fake in which interestingly all goes 1 way like how media always gets it wrong against 1 direction.
    • We already have politicians insisting real things are fake and fake things are real. How do AI generators change anything?
  • Last time that someone (Stability AI) tried a little too hard to make its new model (Stable Diffusion 3) as SFW as possible we got unfixable anatomic monstrosities, you couldn't even prompt for an human lying on grass without getting a nightmarish mess. That wasn't even nerfing, it was straight up downgrading something that used to work just fine to something broken and unusable, as a result they fell into irrelevance (as it should be) and got superspeed by Flux (from Black Forest Labs). The true point here
    • by hughJ ( 1343331 )
      I think if aliens were to arrive on Earth to assess our current state they'd probably have things like "catastrophic climate change" and "nuclear weapons" on their bingo cards, but I doubt they'd have anticipated "developmentally stunted AI due to sex organ taboo." I'm trying to imagine how children would come to understand the human body if they didn't have their own to look at as a baseline, and only had the sanitized public education curriculum a source.
      • I'm trying to imagine how children would come to understand the human body if they didn't have their own to look at as a baseline, and only had the sanitized public education curriculum a source.

        Simple: They wait for the stork to show up with their kid after coming into direct contact with someone else. While feeling really bad about it and fearing that they will soon be in hell. Thankfully, they all died out due to not knowing how to reproduce.

  • this is kind of pointless, since open source models can do it easily.
  • Abscences included google and meta? In other words, they got non binding non enforceable voluntary commitments from 10 percent of the important players, to take unspecified action on a problem. Eyeroll.
    • I believe that Google and Meta already have robust protections in place for many types of abuse, this included. What would they gain by entering such an agreement?
  • Once you make so much money or are too politically connected in the United States, the law no longer applies to you.
  • What is stopping me from creating a Common Crawl - Nudity Addon pack that can be used by models? One that doesn't contain porn, but just nudity.

    Deep fakes aren't going anywhere and being able to generate AI fake nudity is an artistic expression that should be covered by the 1st Amendment.

    Trading an image that is of someone else, like Taylor Swift, should be considered harassment (or whatever) if it's portrayed as real, otherwise it's a parody.

    I get going after the websites that do this for a living

You are in a maze of little twisting passages, all alike.

Working...