this post was submitted on 18 Jan 2024
420 points (94.9% liked)

Technology

58303 readers
29 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Rep. Joe Morelle, D.-N.Y., appeared with a New Jersey high school victim of nonconsensual sexually explicit deepfakes to discuss a bill stalled in the House.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 14 points 9 months ago (2 children)

We're not talking about whether we should make fakes. We're talking about whether people who do, should be prosecuted - IE physically overpowered by police officers, restrained with handcuffs, and locked up in a prison cell. Some empathy?

If some classmate of your little cousin makes a fake, should the police come and drag them out of school and throw them in prison? You think that would help?

Realistically, it's as likely to happen as prosecution of kids who "get into fights" for assault. Kids tell mean lies about each other but that is not resolved in civil suits over defamation. Even between adults, that's not the usual thing.

Civil suits under this bill would be mainly targeted against internet services, because they have the money. And it would largely be used over celebrity fakes. That's the overwhelming part of fakes out there and they have the money to splurge on suing people who can't pay. It would be wealthy, powerful people using it against horny teens.

Also, this bill is so ripe for industrial abuse. Insert a risqué scene in a movie, and suddenly "pirates" can be prosecuted under this.

[–] [email protected] 18 points 9 months ago

You do have a point about the excesses of police work, but if you want to talk about empathy you should also consider the position of the kid who is harassed and traumatized over something they didn't even have any say over. There is some discussion to be had over what degree of punishment ought to be appropriate, and the need to limit police brutality, well beyond this particular matter.

But as far as demanding that every such work is taken down, and giving vulnerable people the means to demand so without exposing themselves further, it is perfectly reasonable.

Realistically, it’s as likely to happen as prosecution of kids who “get into fights” for assault. Kids tell mean lies about each other but that is not resolved in civil suits over defamation. Even between adults, that’s not the usual thing.

Except that in the case of deepfake porn it's not a matter of fuzzy two-sided conflicts. One side is creating the whole problem, and one side is just the victim of it despite not being involved in any way. That's the whole point of deepfake. The most that lies might play into it is in finding out that the porn is real, and in such case there is even more reason to take it down.

Civil suits under this bill would be mainly targeted against internet services, because they have the money. And it would largely be used over celebrity fakes. That’s the overwhelming part of fakes out there and they have the money to splurge on suing people who can’t pay. It would be wealthy, powerful people using it against horny teens.

Gotta say I have a hard time feeling sorry for the people who can't be satisfied by the frankly immense amount of porn we have and decided that they absolutely must have porn from that one specific person who never consented to it. Maybe they are wealthy and powerful, sure. Does that mean it's a free pass to fabricate deepfake porn with their likenesses? I don't think so. Nobody is owed that. As much as you insist that it will be used by the powerful against the poor masses, it still seems to me that whatever regular dude decides to do it is crossing serious boundaries. This is not brave freedom fighter, it's just an asshole.

I think most likely what will happen is that these internet services will just take those down. As they should.

[–] [email protected] 8 points 9 months ago (1 children)

If my little cousin makes AI child porn, of anyone at all let alone a classmate he knows physically in real life, I dont think he should be allowed to kick his feet and go about his day.

Like..... Making kiddie porn of your classmates is not excusable because youre a horny teen. Sorry, bud, its fucking not

[–] [email protected] -2 points 9 months ago (1 children)

If two 14-year-olds get it on, they should both be prosecuted for child abuse? That is what you are actually saying?

[–] [email protected] 0 points 9 months ago* (last edited 9 months ago)

You can only fuck by creating AI porn of the person you are trying to have sex with against their will? Are you a robot?

The people who think creating non consentual AI child porn is equivalent to sex need to spend time outside