56
submitted 8 months ago by xptiger@lemmy.world to c/fediverse@lemmy.world

Though Lemmy and Mastodon are public sites, and their structures are open-source I guess? (I'm not a programmer/coder), can they really dodge the ability of AI s to collect/track any data everytime they search everywhere on Internet?

you are viewing a single comment's thread
view the rest of the comments
[-] Jeremyward@lemmy.world 9 points 8 months ago

They can put a robots.txt file in their root structure which can tell robots (AI scrapers) to ignore that website. However that only works on robots which follow that rule, it's self enforced so it's a crap shoot of it'll be followed. Otherwise to be honest there isn't a lot a public facing website can do to avoid being scraped. Maybe put up a captcha on every page?

this post was submitted on 13 Jan 2024
56 points (87.8% liked)

Fediverse

27735 readers
333 users here now

A community to talk about the Fediverse and all it's related services using ActivityPub (Mastodon, Lemmy, KBin, etc).

If you wanted to get help with moderating your own community then head over to !moderators@lemmy.world!

Rules

Learn more at these websites: Join The Fediverse Wiki, Fediverse.info, Wikipedia Page, The Federation Info (Stats), FediDB (Stats), Sub Rehab (Reddit Migration), Search Lemmy

founded 1 year ago
MODERATORS