166
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 16 Feb 2024
166 points (94.6% liked)
Technology
59066 readers
4390 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
Digitally watermark the image for identification purposes. Hash the hardware, MAC ID, IP address, of the creator and have that inserted via steganography or similar means. Just like printers use a MIC for documents printed on that printer, AI generated imagery should do the same at this point. It’s not perfect and probably has some undesirable consequences, but it’s better than nothing when trying to track down deepfake creators.
The difference is it costs billions of dollars to run a company manufacturing printers and it’s easy for law enforcement to pressure them into not printing money.
It costs nothing to produce an AI image, you can run this stuff on a cheap gaming PC or laptop.
And you can do it with open source software. If the software has restrictions on creating abusive material, you can find a fork with that feature disabled. If it has stenography, you can find one with that disabled too.
You can tag an image to prove a certain person (or camera) took a photo. You can’t stop people from removing that.
Honestly, I can't tell if this is sarcasm or not.
Samsung's AI does watermark their image at the exif, yes it is trivial for us to remove the exif, but it's enough to catch these low effort bad actors.
I think all the main AI services watermark their images (invisibly, not in the metadata). A nudify service might not, I imagine.
I was rather wondering about the support for extensive surveillance.