634
submitted 11 months ago by 0x815@feddit.de to c/europe@feddit.de

Police investigation remains open. The photo of one of the minors included a fly; that is the logo of Clothoff, the application that is presumably being used to create the images, which promotes its services with the slogan: “Undress anybody with our free service!”

you are viewing a single comment's thread
view the rest of the comments
[-] alvvayson@lemmy.world 127 points 11 months ago

There's one blessing coming out of that mess, though: For girls who did take pictures, and had them leaked, saying "they're AI generated" is becoming a plausible way out.

Indeed, once the AI gets good enough, the value of pictures and videos will plummet to zero.

Ironically, in a sense we will revert back to the era before photography existed. To verify if something is real, we might have to rely on witness testimony.

[-] Blapoo@lemmy.ml 57 points 11 months ago

Politics is about to get WILD

[-] jet@hackertalks.com 19 points 11 months ago* (last edited 11 months ago)

Dwayne Elizondo Mountain Dew Herbert Camacho approves!

Shit's going to get real emotional

[-] Blapoo@lemmy.ml 4 points 11 months ago

I'd vote for Terry Crews. No lie.

[-] taladar@feddit.de 37 points 11 months ago

To verify if something is real, we might have to rely on witness testimony.

This is not going to work. Just because images and videos become less reliable that doesn't mean we will forget about the fact that eyewitness testimony is very unreliable.

[-] Khanzarate@lemmy.world 26 points 11 months ago

You say "forget" like it's not still incredibly common as evidence.

There's lots of data showing that eyewitnesses aren't reliable but that doesn't mean courts actually stopped relying on it. Ai making another form of evidence untrustworthy will result in eyewitnesses taking its place.

[-] JoBo@feddit.uk 30 points 11 months ago

Indeed, once the AI gets good enough, the value of pictures and videos will plummet to zero.

This just isn't true. They will still be used to sexualise people, mostly girls and women, against their consent. It's no different from AI-generated child pornography. It does harm even if no 'real' people appear in the images.

Fucking horrible world we're forced to live in. Where's the fucking exit?

[-] GreatGrapeApe@reddthat.com 13 points 11 months ago

It is different than AI-generated CSAM because real people are actually being harmed by these deepfake images.

[-] JoBo@feddit.uk 10 points 11 months ago* (last edited 11 months ago)

I was replying to someone who was claiming they aren't harmful as long as everyone knows they're fake. Maybe nitpick them, not me?

Reak kids are harmed by AI CSAM normalising a problem they should be seeking help for, not getting off on.

load more comments (4 replies)
load more comments (3 replies)
[-] hansl@lemmy.world 18 points 11 months ago

A bit off topic, but I wonder if the entertainment industry as a whole is going to be completely destroyed by AI when it gets good enough.

I can totally see myself prompting “a movie about love in the style of Star Wars, with Ryan Gosling and Audrey Hepburn as the leads, directed by Alfred Hitchcock, written by Vincent Hugo.” And then what? It’s game over for any content creation.

Curious if I’ll see that kind of power at home (using open source tools) in my lifetime.

[-] Benj1B@sh.itjust.works 9 points 11 months ago

I envisage a world where your browsing Netflix, and based on past preferences some of the title cards are generated on the fly for you. Then based on what you click, the AI engine warms us and generates the film for you in real-time. Essentially indistinguishable from the majority of Hollywood regurgitation.

And because the script is just a series of autogenerated prompts, its like a choose your own adventure book, you can steer the narrative the way you want if you elect to. Otherwise it'll be good enough to keep most monkey brains happy and you won't even be able to tell the difference most of the time.

[-] Rootiest@lemmy.world 8 points 11 months ago

Then the real money will be in hipster retro human-generated movies

[-] Gsus4@feddit.nl 4 points 11 months ago

And it will work, because we've grown used to Hollywood being so repetitive.

[-] Zoomboingding@lemmy.world 6 points 11 months ago

I know it's impossible to perfectly predict future technology, but I believe AI will exist alongside traditional filmmaking. You'll NEVER get something with the emotional impact of Up or Schindler's List from an AI. You'll be able to make fun action or fantasy movies though, and like you said, fully customized for the viewer. I imagine it'll be like CGI vs traditional animation now - you only see the latter for passion projects, but for most uses, CGI works well enough.

[-] ParsnipWitch@feddit.de 2 points 11 months ago

This is already starting to happen for digital illustration. With better models and enough images saved, you can already train a model to replicate the art created by an artist.

load more comments (3 replies)
[-] Cruxifux@lemmy.world 10 points 11 months ago

Holy shit, I never thought of the whole witness testimony aspect. For some reason my mind was just like “well, nothing we see in videos or pictures is real anymore, guess everyone is just gonna devolve into believing whatever confirms their bias and argue endlessly about which pictures are fake and which are real.”

Witness testimony and live political interactions are going to become incredibly important for how our society views “the truth” in world events in the near future. I don’t know if I love or hate that.

[-] hardware26@discuss.tchncs.de 6 points 11 months ago

Not necessarily, solutions can implemented. For example, footage from private security cameras can be sent to trusted establishment (trusted by the court at least) in real time which can be timestamped and stored (maybe not necessarily even stored there, encryption with timestamp may be enough). If source private camera and the network is secure, footage is also secure.

[-] AeroLemming@lemm.ee 8 points 11 months ago* (last edited 11 months ago)

All they'd need to send to trusted servers is some basic metadata and hashes of segments of video. That's enough to verify that the video hasn't been tampered with after it was filmed as long as you trust the hashing function.

[-] taladar@feddit.de 1 points 11 months ago

I don't think that will matter very much considering how many real time video modifications we can do already today. Not to mention synthesizing video before the time it is supposed to take place.

[-] Benj1B@sh.itjust.works 3 points 11 months ago

Network security is a pretty big ask though - just look at how many unsecured cameras are around now. And once an attacker is in anything generated on that network becomes suspect - how do you know the security camera feed wasn't intercepted, manipulated, or replaced altogether?

[-] lambalicious 2 points 11 months ago

To verify if something is real, we might have to rely on ~~witness testimony~~ flagrancy.

FTFY. Witness has never been that good a means to verify something is real.

[-] Mattol@discuss.tchncs.de 2 points 11 months ago

Maybe there will be cameras as well that sign the pictures they take?

load more comments (30 replies)
this post was submitted on 19 Sep 2023
634 points (98.0% liked)

Europe

8324 readers
1 users here now

News/Interesting Stories/Beautiful Pictures from Europe 🇪🇺

(Current banner: Thunder mountain, Germany, 🇩🇪 ) Feel free to post submissions for banner pictures

Rules

(This list is obviously incomplete, but it will get expanded when necessary)

  1. Be nice to each other (e.g. No direct insults against each other);
  2. No racism, antisemitism, dehumanisation of minorities or glorification of National Socialism allowed;
  3. No posts linking to mis-information funded by foreign states or billionaires.

Also check out !yurop@lemm.ee

founded 1 year ago
MODERATORS