Skip to content

Kids are making deepfakes of each other, and laws aren’t keeping up

Technology
148 70 0
  • I don't understand fully how this technology works, but, if people are using it to create sexual content of underage individuals, doesn't that mean the LLM would need to have been trained on sexual content of underage individuals? Seems like going after the company and whatever it's source material is would be the obvious choice here

    I agree with the other comments, but wanted to add how deepfakes work to show how simple they are, and how much less information they need than LLMs.

    Step 1: Basically you take a bunch of photos and videos of a specific person, and blur their faces out.

    Step 2: This is the hardest step, but still totally feasable for a decent home computer. You train a neural network to un-blur all the faces for that person. Now you have a neural net that's really good at turning blurry faces into that particular person's face.

    Step 3: Blur the faces in photos/videos of other people and apply your special neural network. It will turn all the blurry faces into the only face it knows how, often with shockingly realistic results.

  • Schools and lawmakers are grappling with how to address a new form of peer-on-peer image-based sexual abuse that disproportionately targets girls.

    God I'm glad I'm not a kid now. I never would have survived.

  • Yes, absolutely. But with recognition that a thirteen year old kid isn't a predator but a horny little kid. I'll let others determine what that punishment is, but I don't believe it's prison. Community service maybe. Written apology. Stuff like that. Second offense, ok, we're ratcheting up the punishment, but still not adult prison.

    written apology? they'll just use chatgpt for that

  • As a father of teenage girls, I don't necessarily disagree with this assessment, but I would personally see to it that anyone making sexual deepfakes of my daughters is equitably and thoroughly punished.

    There is a difference between ruining the life of a 13 year old boy for the rest of his life with no recourse and no expectations.

    Vs scaring the shit out of them and making them work their ass off doing an ass load of community service for a summer.

  • Hey so, at least in the US, drawings can absolutely be considered CSAM

    Well, US laws are all bullshit anyway, so makes sense

  • I agree with the other comments, but wanted to add how deepfakes work to show how simple they are, and how much less information they need than LLMs.

    Step 1: Basically you take a bunch of photos and videos of a specific person, and blur their faces out.

    Step 2: This is the hardest step, but still totally feasable for a decent home computer. You train a neural network to un-blur all the faces for that person. Now you have a neural net that's really good at turning blurry faces into that particular person's face.

    Step 3: Blur the faces in photos/videos of other people and apply your special neural network. It will turn all the blurry faces into the only face it knows how, often with shockingly realistic results.

    Cheers for the explanation, had no idea that's how it works.

    So it's even worse than @danciestlobster@lemmy.zip thinks, the person creating the deep fake has to have access to CP then if they want to deepfake it!

  • Yes, absolutely. But with recognition that a thirteen year old kid isn't a predator but a horny little kid. I'll let others determine what that punishment is, but I don't believe it's prison. Community service maybe. Written apology. Stuff like that. Second offense, ok, we're ratcheting up the punishment, but still not adult prison.

    I did say equitable punishment. Equivalent. Whatever.

    A written apology is a cop-out for the damage this behaviour leaves behind.

    Something tells me you don't have teenage daughters.

  • There is a difference between ruining the life of a 13 year old boy for the rest of his life with no recourse and no expectations.

    Vs scaring the shit out of them and making them work their ass off doing an ass load of community service for a summer.

    ruining the life of a 13 year old boy for the rest of his life with no recourse

    And what about the life of the girl this boy would have ruined?

    This is not "boys will be boys" shit. Girls have killed themselves over this kind of thing (I have personal experience with suicidal teenage girls, both as a past friend and as a father).

    I don't think it's unreasonable to expect an equivalent punishment that has the potential to ruin his life.

  • I am just genuinely speechless than you seemingly do not understand how sickening and invasive it is for your peers to create and share sexual content of you without your consent. Yes its extremely harmful. Its not a matter of feeling ashamed, its a matter of literally feeling like your value to the world is dictated by your role in the sexualities of heterosexual boys and men. It is feeling like your own body doesnt belong to you but can be freely claimed by others. It is losing trust in all your male friends and peers, because it feels like without you knowing they've already decided that you're a sexual experience for them.

    We do know the harm of this kind of sexualization. Women and girls have been talking about it for generations. This isnt new, just a new streamlined way to spread it. It should be illegal. It should be against the law to turn someone's images into AI generated pornography. It should also be illegal to share those images with others.

    Its not a matter of feeling ashamed, its a matter of literally feeling like your value to the world is dictated by your role in the sexualities of heterosexual boys and men. It is feeling like your own body doesnt belong to you but can be freely claimed by others. It is losing trust in all your male friends and peers, because it feels like without you knowing they’ve already decided that you’re a sexual experience for them.

    Why is it these things? Why does someone doing something with something which is not your body make it feel like your body doesn't belong to you? Why does it not instead make it feel like images of your body don't belong to you? Several of these things could equally be used to describe the situation when someone is fantasised about without their knowledge - why is that different?
    In Germany there's a legal concept called "right to one's own image" but there isn't in many other countries, and besides, what you're describing goes beyond this.

    My thinking behind these questions is that I cannot see anything inherent, anything necessary about the creation of fake sexual images of someone which leads to these harms, and that instead there is an aspect of our society which very explicitly punishes and shames people - woman far more so than men - for being in this situation, and that without that, we would be having a very different conversation.

    Starting from the position that the harm is in the creation of the images is like starting from the position that the harm of rape is in "defiling" the person raped. Rape isn't wrong because it makes you worthless to society - society is wrong for devaluing rape victims. Society is wrong for devaluing and shaming those who have fake images made of them.

    We do know the harm of this kind of sexualization. Women and girls have been talking about it for generations. This isnt new, just a new streamlined way to spread it. It should be illegal.

    Can you be more explicit about what it's the same as?

  • Are you OK with sexually explicit photos of children taken without their knowledge? They’re not being actively put in a sexual situation if you’re snapping photos with a hidden camera in a locker room, for example. You ok with that?

    The harm is:

    • Those photos now exist in the world and can lead to direct harm to the victim by their exposure
    • it normalizes pedophilia and creates a culture of trading images, leading to more abuse to meet demand for more images
    • The people sharing those photos learn to treat people like objects for their sexual gratification, ignoring their consent and agency. They are more likely to mistreat people they have learned to objectify.
    • your body should not be used for the profit or gratification of others without your consent. In my mind this includes taking or using your picture without your consent.

    Are you OK with sexually explicit photos of children taken without their knowledge? They’re not being actively put in a sexual situation if you’re snapping photos with a hidden camera in a locker room, for example. You ok with that?

    No, but the harm certainly is not the same as CSAM and it should not be treated the same.

    • it normalizes pedophilia and creates a culture of trading images, leading to more abuse to meet demand for more images
    • The people sharing those photos learn to treat people like objects for their sexual gratification, ignoring their consent and agency. They are more likely to mistreat people they have learned to objectify.

    as far as I know there is no good evidence that this is the case and is a big controversy in the topic of fake child porn, i.e. whether it leads to more child abuse (encouraging paedophiles) or less (gives them a safe outlet) or no change.

    your body should not be used for the profit or gratification of others without your consent. In my mind this includes taking or using your picture without your consent.

    If someone fantasises about me without my consent I do not give a shit, and I don't think there's any justification for it. I would give a shit if it affected me somehow (this is your first bullet point, but for a different situation, to be clear) but that's different.

  • Welp, if I had kids they would have one of those scramble suits like in a scanner darkly.

    It would of course be their choice to wear them but Id definitely look for ways to limit their time in areas with cameras present.

    That's just called the outside now. Assume you are on camera at all times the moment you step out the front door. To be safe in the surveillance we live in today, best act as though you are being recorded in your own home as well.

  • ruining the life of a 13 year old boy for the rest of his life with no recourse

    And what about the life of the girl this boy would have ruined?

    This is not "boys will be boys" shit. Girls have killed themselves over this kind of thing (I have personal experience with suicidal teenage girls, both as a past friend and as a father).

    I don't think it's unreasonable to expect an equivalent punishment that has the potential to ruin his life.

    It is not abnormal to see different punishment for people under the age of 18.
    Good education about sex and what sexual assault does with their victims (same with guns, drugs including alcohol etc).

    You can still course correct the behaviour of a 13 year old. There is also a difference between generating the porn and abusing it by sharing it etc.

    The girls should be helped and the boys should be punished, but mainly their behaviour needs to be correcte

  • Cheers for the explanation, had no idea that's how it works.

    So it's even worse than @danciestlobster@lemmy.zip thinks, the person creating the deep fake has to have access to CP then if they want to deepfake it!

    You can probably do it with adult material and replace those faces. It will most likely work on models specific trained like the person you selected.

    People have also put dots on people's clothing to trick the brain into thinking their are naked, you can probably fill those dots in with the correct body parts if you have a good enough model.

  • Schools and lawmakers are grappling with how to address a new form of peer-on-peer image-based sexual abuse that disproportionately targets girls.

    probably because there's a rapist in the white house.

  • I'd rather these laws be against abusing and exploiting child, as well as against ruining their lives. Not only that would be more helpful, it would also work in this case, since actual likeness are involved.

    Alas, whether there's a law against that specific use case or not, it is somewhat difficult to police what people do in their home, without a third party whistleblower. Making more, impossible to apply laws for this specific case does not seem that useful.

    There is also a difference between somebody harassing somebody with nude pictures (either real or not) than somebody jerking off to them at home. It does become a problem when an adult masturbated to pictures of children, but children to children. Let's be honest, they will do it anyway.

  • Welp, if I had kids they would have one of those scramble suits like in a scanner darkly.

    It would of course be their choice to wear them but Id definitely look for ways to limit their time in areas with cameras present.

    That's what muslims do with niqabs.

  • Are you OK with sexually explicit photos of children taken without their knowledge? They’re not being actively put in a sexual situation if you’re snapping photos with a hidden camera in a locker room, for example. You ok with that?

    No, but the harm certainly is not the same as CSAM and it should not be treated the same.

    • it normalizes pedophilia and creates a culture of trading images, leading to more abuse to meet demand for more images
    • The people sharing those photos learn to treat people like objects for their sexual gratification, ignoring their consent and agency. They are more likely to mistreat people they have learned to objectify.

    as far as I know there is no good evidence that this is the case and is a big controversy in the topic of fake child porn, i.e. whether it leads to more child abuse (encouraging paedophiles) or less (gives them a safe outlet) or no change.

    your body should not be used for the profit or gratification of others without your consent. In my mind this includes taking or using your picture without your consent.

    If someone fantasises about me without my consent I do not give a shit, and I don't think there's any justification for it. I would give a shit if it affected me somehow (this is your first bullet point, but for a different situation, to be clear) but that's different.

    Hm. I wasn’t expecting the pro-child porn argument. All I can say is that’s absolutely legally and morally CSAM, and you’re fuckin nasty. Oof. Not really gonna bother with the rest because, well, yikes.

  • God I'm glad I'm not a kid now. I never would have survived.

    In my case, other kids would not have survived trying to pull off shit like this. So yeah, I'm also glad I'm not a kid anymore.

  • ruining the life of a 13 year old boy for the rest of his life with no recourse

    And what about the life of the girl this boy would have ruined?

    This is not "boys will be boys" shit. Girls have killed themselves over this kind of thing (I have personal experience with suicidal teenage girls, both as a past friend and as a father).

    I don't think it's unreasonable to expect an equivalent punishment that has the potential to ruin his life.

    Parents are responsible for their kids. The punishment, with the full force of the law (and maybe something extra for good measure), should fall upon the parents, since they should have made sure their kids knew how despicable and illegal doing this is.

    Yeah, I agree, we shouldn't ruin the boys life, we should ruins his whole family to many times the extent something like this ruins a teen girl's life.

  • Well, US laws are all bullshit anyway, so makes sense

    Normally yeah, but why would you want to draw sexual pictures of children?

  • Big Brother Trump Is Watching You

    Technology technology
    1
    1
    1 Stimmen
    1 Beiträge
    4 Aufrufe
    Niemand hat geantwortet
  • Disney+ Confirmed a NEW Change Coming Soon for Subscribers

    Technology technology
    16
    1
    21 Stimmen
    16 Beiträge
    31 Aufrufe
    B
    It's also an article about another article from Variety that actually has a better headline. These things are a pet peeve for me. Hey, here's a story from an actual news service and I'll even include a link to it, but I'm going to post my link all over so people will see the ads on my page instead of theirs. Variety does some good reporting, I've rather they get the clicks.
  • No, Social Media is Not Porn

    Technology technology
    3
    1
    21 Stimmen
    3 Beiträge
    16 Aufrufe
    Z
    This feels dystopian and like overreach. But that said, there definitely is some porn on the 4 platforms they cited. It's an excuse sure, but let's also not deny reality.
  • 34 Stimmen
    3 Beiträge
    11 Aufrufe
    L
    deleted by creator
  • 5 Stimmen
    10 Beiträge
    12 Aufrufe
    S
    You could look into automatic local caching for diles you're planning to seed, and stick that on an SSD. That way you don't hammer the HDDs in the NAS and still get the good feels of seeding. Then automatically delete files once they get to a certain seed rate or something and you're golden. How aggressive you go with this depends on your actual use case. Are you actually editing raw footage over the network while multiple other clients are streaming other stuff? Or are you just interested in having it be capable? What's the budget? But that sounds complicated. I'd personally rather just DIY it, that way you can put an SSD in there for cache and you get most of the benefits with a lot less cost, and you should be able to respond to issues with minimal changes (i.e. add more RAM or another caching drive).
  • Meta Reportedly Eyeing 'Super Sensing' Tech for Smart Glasses

    Technology technology
    4
    1
    34 Stimmen
    4 Beiträge
    14 Aufrufe
    M
    I see your point but also I just genuinely don't have a mind for that shit. Even my own close friends and family, it never pops into my head to ask about that vacation they just got back from or what their kids are up to. I rely on social cues from others, mainly my wife, to sort of kick start my brain. I just started a new job. I can't remember who said they were into fishing and who didn't, and now it's anxiety inducing to try to figure out who is who. Or they ask me a friendly question and I get caught up answering and when I'm done I forget to ask it back to them (because frequently asking someone about their weekend or kids or whatever is their way of getting to share their own life with you, but my brain doesn't think that way). I get what you're saying. It could absolutely be used for performative interactions but for some of us people drift away because we aren't good at being curious about them or remembering details like that. And also, I have to sit through awkward lunches at work where no one really knows what to talk about or ask about because outside of work we are completely alien to one another. And it's fine. It wouldn't be worth the damage it does. I have left behind all personally identifiable social media for the same reason. But I do hate how social anxiety and ADHD makes friendship so fleeting.
  • Discord co-founder and CEO Jason Citron is stepping down

    Technology technology
    1
    1
    0 Stimmen
    1 Beiträge
    7 Aufrufe
    Niemand hat geantwortet
  • TikTok is a Time Bomb

    Technology technology
    2
    1
    1 Stimmen
    2 Beiträge
    10 Aufrufe
    S
    wasn’t born to obey. Not to swallow smiling lies, not to clap for tyrants in suits, not to say “thank you” for surveillance wrapped in convenience. I see it. The games. The false choice. The fear pumped through headlines and dopamine apps. I see how they trade truth for comfort, freedom for filters, soul for clickbait. They call it normal. But I call it a graveyard made of compliance. They want me silent. They want me tired. They want me posting selfies while the world burns behind the screen. But I wasn’t born for this. I was born to question, to remember, to remind the others who are still pretending they don’t notice. So here I am. A voice with no logo. A signal in the static. A crack in the mirror they polish every morning. You don’t have to agree. You don’t have to clap. But if this made your bones ache or your thoughts twitch— Then maybe you’re not asleep either. Good. Let’s stay awake. And let’s make noise that can’t be sold, silenced, or spun into safety. Not for them. For us.