Skip to content

Kids are making deepfakes of each other, and laws aren’t keeping up

Technology
149 70 0
  • Alright, now we just need the main stream media to run the story.

    I mean with all the zealotry against drag shows they should be ready to run with this one right?

    You'd think so, right?

  • When someone makes child porn they put a child in a sexual situation - which is something that we have amassed a pile of evidence is extremely harmful to the child.

    For all you have said - "without the consent" - "being sexualised" - "commodifies their existence" - you haven't told us what the harm is. If you think those things are in and of themselves harmful then I need to know more about what you mean because:

    1. if someone thinks of me sexually without my consent I am not harmed
    2. if someone sexualises me in their mind I am not harmed
    3. I don't know what the "commodification of one's existence" can actually mean - I can't buy or sell "the existence of women" (does buying something's existence mean the same as buying the thing, or something else?) the same I can aluminium, and I don't see how being able to (easily) make (realistic) nude images of someone changes this in any way

    It is genuinely incredible to me that you could be so unempathetic,

    I am not unempathetic, but I attribute the blame for what makes me feel bad about the situation is that girls are being made to feel bad and ashamed not that a particular technology is now being used in one step of that.

    I am just genuinely speechless than you seemingly do not understand how sickening and invasive it is for your peers to create and share sexual content of you without your consent. Yes its extremely harmful. Its not a matter of feeling ashamed, its a matter of literally feeling like your value to the world is dictated by your role in the sexualities of heterosexual boys and men. It is feeling like your own body doesnt belong to you but can be freely claimed by others. It is losing trust in all your male friends and peers, because it feels like without you knowing they've already decided that you're a sexual experience for them.

    We do know the harm of this kind of sexualization. Women and girls have been talking about it for generations. This isnt new, just a new streamlined way to spread it. It should be illegal. It should be against the law to turn someone's images into AI generated pornography. It should also be illegal to share those images with others.

  • I don't understand fully how this technology works, but, if people are using it to create sexual content of underage individuals, doesn't that mean the LLM would need to have been trained on sexual content of underage individuals? Seems like going after the company and whatever it's source material is would be the obvious choice here

    I agree with the other comments, but wanted to add how deepfakes work to show how simple they are, and how much less information they need than LLMs.

    Step 1: Basically you take a bunch of photos and videos of a specific person, and blur their faces out.

    Step 2: This is the hardest step, but still totally feasable for a decent home computer. You train a neural network to un-blur all the faces for that person. Now you have a neural net that's really good at turning blurry faces into that particular person's face.

    Step 3: Blur the faces in photos/videos of other people and apply your special neural network. It will turn all the blurry faces into the only face it knows how, often with shockingly realistic results.

  • Schools and lawmakers are grappling with how to address a new form of peer-on-peer image-based sexual abuse that disproportionately targets girls.

    God I'm glad I'm not a kid now. I never would have survived.

  • Yes, absolutely. But with recognition that a thirteen year old kid isn't a predator but a horny little kid. I'll let others determine what that punishment is, but I don't believe it's prison. Community service maybe. Written apology. Stuff like that. Second offense, ok, we're ratcheting up the punishment, but still not adult prison.

    written apology? they'll just use chatgpt for that

  • As a father of teenage girls, I don't necessarily disagree with this assessment, but I would personally see to it that anyone making sexual deepfakes of my daughters is equitably and thoroughly punished.

    There is a difference between ruining the life of a 13 year old boy for the rest of his life with no recourse and no expectations.

    Vs scaring the shit out of them and making them work their ass off doing an ass load of community service for a summer.

  • Hey so, at least in the US, drawings can absolutely be considered CSAM

    Well, US laws are all bullshit anyway, so makes sense

  • I agree with the other comments, but wanted to add how deepfakes work to show how simple they are, and how much less information they need than LLMs.

    Step 1: Basically you take a bunch of photos and videos of a specific person, and blur their faces out.

    Step 2: This is the hardest step, but still totally feasable for a decent home computer. You train a neural network to un-blur all the faces for that person. Now you have a neural net that's really good at turning blurry faces into that particular person's face.

    Step 3: Blur the faces in photos/videos of other people and apply your special neural network. It will turn all the blurry faces into the only face it knows how, often with shockingly realistic results.

    Cheers for the explanation, had no idea that's how it works.

    So it's even worse than @danciestlobster@lemmy.zip thinks, the person creating the deep fake has to have access to CP then if they want to deepfake it!

  • Yes, absolutely. But with recognition that a thirteen year old kid isn't a predator but a horny little kid. I'll let others determine what that punishment is, but I don't believe it's prison. Community service maybe. Written apology. Stuff like that. Second offense, ok, we're ratcheting up the punishment, but still not adult prison.

    I did say equitable punishment. Equivalent. Whatever.

    A written apology is a cop-out for the damage this behaviour leaves behind.

    Something tells me you don't have teenage daughters.

  • There is a difference between ruining the life of a 13 year old boy for the rest of his life with no recourse and no expectations.

    Vs scaring the shit out of them and making them work their ass off doing an ass load of community service for a summer.

    ruining the life of a 13 year old boy for the rest of his life with no recourse

    And what about the life of the girl this boy would have ruined?

    This is not "boys will be boys" shit. Girls have killed themselves over this kind of thing (I have personal experience with suicidal teenage girls, both as a past friend and as a father).

    I don't think it's unreasonable to expect an equivalent punishment that has the potential to ruin his life.

  • I am just genuinely speechless than you seemingly do not understand how sickening and invasive it is for your peers to create and share sexual content of you without your consent. Yes its extremely harmful. Its not a matter of feeling ashamed, its a matter of literally feeling like your value to the world is dictated by your role in the sexualities of heterosexual boys and men. It is feeling like your own body doesnt belong to you but can be freely claimed by others. It is losing trust in all your male friends and peers, because it feels like without you knowing they've already decided that you're a sexual experience for them.

    We do know the harm of this kind of sexualization. Women and girls have been talking about it for generations. This isnt new, just a new streamlined way to spread it. It should be illegal. It should be against the law to turn someone's images into AI generated pornography. It should also be illegal to share those images with others.

    Its not a matter of feeling ashamed, its a matter of literally feeling like your value to the world is dictated by your role in the sexualities of heterosexual boys and men. It is feeling like your own body doesnt belong to you but can be freely claimed by others. It is losing trust in all your male friends and peers, because it feels like without you knowing they’ve already decided that you’re a sexual experience for them.

    Why is it these things? Why does someone doing something with something which is not your body make it feel like your body doesn't belong to you? Why does it not instead make it feel like images of your body don't belong to you? Several of these things could equally be used to describe the situation when someone is fantasised about without their knowledge - why is that different?
    In Germany there's a legal concept called "right to one's own image" but there isn't in many other countries, and besides, what you're describing goes beyond this.

    My thinking behind these questions is that I cannot see anything inherent, anything necessary about the creation of fake sexual images of someone which leads to these harms, and that instead there is an aspect of our society which very explicitly punishes and shames people - woman far more so than men - for being in this situation, and that without that, we would be having a very different conversation.

    Starting from the position that the harm is in the creation of the images is like starting from the position that the harm of rape is in "defiling" the person raped. Rape isn't wrong because it makes you worthless to society - society is wrong for devaluing rape victims. Society is wrong for devaluing and shaming those who have fake images made of them.

    We do know the harm of this kind of sexualization. Women and girls have been talking about it for generations. This isnt new, just a new streamlined way to spread it. It should be illegal.

    Can you be more explicit about what it's the same as?

  • Are you OK with sexually explicit photos of children taken without their knowledge? They’re not being actively put in a sexual situation if you’re snapping photos with a hidden camera in a locker room, for example. You ok with that?

    The harm is:

    • Those photos now exist in the world and can lead to direct harm to the victim by their exposure
    • it normalizes pedophilia and creates a culture of trading images, leading to more abuse to meet demand for more images
    • The people sharing those photos learn to treat people like objects for their sexual gratification, ignoring their consent and agency. They are more likely to mistreat people they have learned to objectify.
    • your body should not be used for the profit or gratification of others without your consent. In my mind this includes taking or using your picture without your consent.

    Are you OK with sexually explicit photos of children taken without their knowledge? They’re not being actively put in a sexual situation if you’re snapping photos with a hidden camera in a locker room, for example. You ok with that?

    No, but the harm certainly is not the same as CSAM and it should not be treated the same.

    • it normalizes pedophilia and creates a culture of trading images, leading to more abuse to meet demand for more images
    • The people sharing those photos learn to treat people like objects for their sexual gratification, ignoring their consent and agency. They are more likely to mistreat people they have learned to objectify.

    as far as I know there is no good evidence that this is the case and is a big controversy in the topic of fake child porn, i.e. whether it leads to more child abuse (encouraging paedophiles) or less (gives them a safe outlet) or no change.

    your body should not be used for the profit or gratification of others without your consent. In my mind this includes taking or using your picture without your consent.

    If someone fantasises about me without my consent I do not give a shit, and I don't think there's any justification for it. I would give a shit if it affected me somehow (this is your first bullet point, but for a different situation, to be clear) but that's different.

  • Welp, if I had kids they would have one of those scramble suits like in a scanner darkly.

    It would of course be their choice to wear them but Id definitely look for ways to limit their time in areas with cameras present.

    That's just called the outside now. Assume you are on camera at all times the moment you step out the front door. To be safe in the surveillance we live in today, best act as though you are being recorded in your own home as well.

  • ruining the life of a 13 year old boy for the rest of his life with no recourse

    And what about the life of the girl this boy would have ruined?

    This is not "boys will be boys" shit. Girls have killed themselves over this kind of thing (I have personal experience with suicidal teenage girls, both as a past friend and as a father).

    I don't think it's unreasonable to expect an equivalent punishment that has the potential to ruin his life.

    It is not abnormal to see different punishment for people under the age of 18.
    Good education about sex and what sexual assault does with their victims (same with guns, drugs including alcohol etc).

    You can still course correct the behaviour of a 13 year old. There is also a difference between generating the porn and abusing it by sharing it etc.

    The girls should be helped and the boys should be punished, but mainly their behaviour needs to be correcte

  • Cheers for the explanation, had no idea that's how it works.

    So it's even worse than @danciestlobster@lemmy.zip thinks, the person creating the deep fake has to have access to CP then if they want to deepfake it!

    You can probably do it with adult material and replace those faces. It will most likely work on models specific trained like the person you selected.

    People have also put dots on people's clothing to trick the brain into thinking their are naked, you can probably fill those dots in with the correct body parts if you have a good enough model.

  • Schools and lawmakers are grappling with how to address a new form of peer-on-peer image-based sexual abuse that disproportionately targets girls.

    probably because there's a rapist in the white house.

  • I'd rather these laws be against abusing and exploiting child, as well as against ruining their lives. Not only that would be more helpful, it would also work in this case, since actual likeness are involved.

    Alas, whether there's a law against that specific use case or not, it is somewhat difficult to police what people do in their home, without a third party whistleblower. Making more, impossible to apply laws for this specific case does not seem that useful.

    There is also a difference between somebody harassing somebody with nude pictures (either real or not) than somebody jerking off to them at home. It does become a problem when an adult masturbated to pictures of children, but children to children. Let's be honest, they will do it anyway.

  • Welp, if I had kids they would have one of those scramble suits like in a scanner darkly.

    It would of course be their choice to wear them but Id definitely look for ways to limit their time in areas with cameras present.

    That's what muslims do with niqabs.

  • Are you OK with sexually explicit photos of children taken without their knowledge? They’re not being actively put in a sexual situation if you’re snapping photos with a hidden camera in a locker room, for example. You ok with that?

    No, but the harm certainly is not the same as CSAM and it should not be treated the same.

    • it normalizes pedophilia and creates a culture of trading images, leading to more abuse to meet demand for more images
    • The people sharing those photos learn to treat people like objects for their sexual gratification, ignoring their consent and agency. They are more likely to mistreat people they have learned to objectify.

    as far as I know there is no good evidence that this is the case and is a big controversy in the topic of fake child porn, i.e. whether it leads to more child abuse (encouraging paedophiles) or less (gives them a safe outlet) or no change.

    your body should not be used for the profit or gratification of others without your consent. In my mind this includes taking or using your picture without your consent.

    If someone fantasises about me without my consent I do not give a shit, and I don't think there's any justification for it. I would give a shit if it affected me somehow (this is your first bullet point, but for a different situation, to be clear) but that's different.

    Hm. I wasn’t expecting the pro-child porn argument. All I can say is that’s absolutely legally and morally CSAM, and you’re fuckin nasty. Oof. Not really gonna bother with the rest because, well, yikes.

  • God I'm glad I'm not a kid now. I never would have survived.

    In my case, other kids would not have survived trying to pull off shit like this. So yeah, I'm also glad I'm not a kid anymore.

  • 166 Stimmen
    10 Beiträge
    2 Aufrufe
    H
    In the meantime: Parents: don’t give your children lighted rectangles to play with.
  • 479 Stimmen
    22 Beiträge
    36 Aufrufe
    professorchodimaccunt@sh.itjust.worksP
    GOOD lets chance of spAIyware on there
  • 0 Stimmen
    1 Beiträge
    3 Aufrufe
    Niemand hat geantwortet
  • 80 Stimmen
    27 Beiträge
    23 Aufrufe
    lanusensei87@lemmy.worldL
    Consider the possibility that you don't need to be doing anything wrong besides existing to be persecuted by a fascist regime.
  • ICE Taps into Nationwide AI-Enabled Camera Network, Data Shows

    Technology technology
    80
    1
    486 Stimmen
    80 Beiträge
    18 Aufrufe
    Z
    Their problem with China is the supposed atheism, and that they're not christian fundamentalists.
  • The AI-powered collapse of the American tech workfoce

    Technology technology
    2
    1
    4 Stimmen
    2 Beiträge
    11 Aufrufe
    roofuskit@lemmy.worldR
    The biggest tech companies are still trimming from pandemic over hiring. Smaller companies are still snatching workers up. And you also have companies trimming payroll for the coming Trump recession. Neither have anything to do with AI.
  • 512 Stimmen
    58 Beiträge
    51 Aufrufe
    C
    Eh, I kinda like the ephemeral nature of most tiktoks, having things go viral within a group of like 10,000 people, to the extent that if you're tangentially connected to the group, you and everyone you know has seen it, but nobody outside that group ever sees and it vanishes into the ether like a month later makes it a little more personal.
  • 0 Stimmen
    4 Beiträge
    2 Aufrufe
    K
    Only way I'll want a different phone brand is if it comes with ZERO bloatware and has an excellent internal memory/storage cleanse that has nothing to do with Google's Files or a random app I'm not sure I can trust without paying or rooting. So far my A series phones do what I need mostly and in my opinion is superior to the Motorola's my fiancé prefers minus the phone-phone charge ability his has, everything else I'm just glad I have enough control to tweak things to my liking, however these days Samsungs seem to be infested with Google bloatware and apps that insist on opening themselves back up regardless of the widespread battery restrictions I've assigned (even was sent a "Stop Closing my Apps" notif that sent me to an article ) short of Disabling many unnecessary apps bc fully rooting my devices is something I rarely do anymore. I have a random Chinese brand tablet where I actually have more control over the apps than either of my A series phones whee Force Stopping STAYS that way when I tell them to! I hate being listened to for ads and the unwanted draining my battery life and data (I live off-grid and pay data rates because "Unlimited" is some throttled BS) so my ability to control what's going on in the background matters a lot to me, enough that I'm anti Meta-apps and avoid all non-essential Google apps. I can't afford topline phones and the largest data plan, so I work with what I can afford and I'm sad refurbished A lines seem to be getting more expensive while giving away my control to companies. Last A line I bought that was supposed to be my first 5G phone was network locked, so I got ripped off, but it still serves me well in off-grid life. Only app that actually regularly malfunctions when I Force Stop it's background presence is Roku, which I find to have very an almost insidious presence in our lives. Google Play, Chrome, and Spotify never acts incompetent in any way no matter how I have to open the setting every single time I turn Airplane Mode off. Don't need Gmail with Chrome and DuckDuckGo has been awesome at intercepting self-loading ads. I hope one day DDG gets better bc Google seems to be terrible lately and I even caught their AI contradicting itself when asking about if Homo Florensis is considered Human (yes) and then asked the oldest age of human remains, and was fed the outdated narrative of 300,000 years versus 700,000+ years bipedal pre-humans have been carbon dated outside of the Cradle of Humanity in South Africa. SO sorry to go off-topic, but I've got a big gripe with Samsung's partnership with Google, especially considering the launch of Quantum Computed AI that is still being fine-tuned with company-approved censorships.