554
submitted 1 month ago* (last edited 1 month ago) by ForgottenFlux@lemmy.world to c/technology@lemmy.world

Reddit CEO Steve Huffman is standing by Reddit’s decision to block companies from scraping the site without an AI agreement.

Last week, 404 Media noticed that search engines that weren't Google were no longer listing recent Reddit posts in results. This was because Reddit updated its Robots Exclusion Protocol (txt file) to block bots from scraping the site. The file reads: "Reddit believes in an open Internet, but not the misuse of public content." Since the news broke, OpenAI announced SearchGPT, which can show recent Reddit results.

The change came a year after Reddit began its efforts to stop free scraping, which Huffman initially framed as an attempt to stop AI companies from making money off of Reddit content for free. This endeavor also led Reddit to begin charging for API access (the high pricing led to many third-party Reddit apps closing).

In an interview with The Verge today, Huffman stood by the changes that led to Google temporarily being the only search engine able to show recent discussions from Reddit. Reddit and Google signed an AI training deal in February said to be worth $60 million a year. It's unclear how much Reddit's OpenAI deal is worth.

Huffman said:

Without these agreements, we don’t have any say or knowledge of how our data is displayed and what it’s used for, which has put us in a position now of blocking folks who haven’t been willing to come to terms with how we’d like our data to be used or not used.

“[It’s been] a real pain in the ass to block these companies,” Huffman told The Verge.

you are viewing a single comment's thread
view the rest of the comments
[-] morgunkorn@discuss.tchncs.de 426 points 1 month ago* (last edited 1 month ago)

Honestly, any platforms hosting user-generated content who use the legal argument that they only provide hosting and aren't responsible for what their user post shouldn't also be able to sell the same data and claim owning any of it.

Otherwise, take away their legal immunity. Nazis or pedophiles post something awful? You get in front of the judge.

edit: typo

[-] jlh@lemmy.jlh.name 181 points 1 month ago

Exactly this. You can claim that their scraping is abusing your servers, but the moment you claim copyright for the content of the site, then you give up your Section 230 rights.

You'd also probably lose a whole lot more processing power trying to stop the crawlers vs just letting them have API access with some sort of limit to queries.

[-] ptz@dubvee.org 29 points 1 month ago* (last edited 1 month ago)

Eh, not really.

I block bot user agents to my Lemmy instance, and the overhead is pretty negligible for that (it's all handled in my web firewall/load balancer).

Granted, those are bots that correctly identify themselves via user agent and don't spoof a browser's.

It's also cheaper and easier to add another load balancer than size up or scale out my DB server to handle the bot traffic.

[-] AlexanderESmith@social.alexanderesmith.com 10 points 1 month ago* (last edited 1 month ago)

Systems admin here: You'll lose almost the same amount of processing time looking up limits and providing a "you're over your credits / rate limit" as you would by just providing the data.

Also, everyone will game the system with multiple accounts (cost be damned, the entities who want the data have cash to burn).

[-] rbits@lemm.ee 6 points 1 month ago

I don't think they actually block malicious bots, the change they've made is just to the robots.txt, they don't have to do anything.

[-] tb_@lemmy.world 9 points 1 month ago

Robots.txt does literally nothing. It's a piece of courtesy that's easily ignored if you don't care.

[-] rbits@lemm.ee 4 points 1 month ago* (last edited 1 month ago)

Yeah but it stops bing and a bunch of AI scrapers that want to act like they're following the rules

[-] echodot@feddit.uk 0 points 1 month ago

How do we know it stops bing? As far as anyone knows they could have instructed their programmers to alter the crawlers so that it ignores robots.txt when on Reddit - that should have taken them a whole 2 minutes.

Reddit blocking any search crawler via robots.txt is such a non thing that it shouldn't even be reported.

[-] givesomefucks@lemmy.world 38 points 1 month ago

Can't sell something you don't own.

So if they're selling the parts people want, they need to own the parts no one wants.

[-] jlh@lemmy.jlh.name 3 points 1 month ago* (last edited 1 month ago)

Well, you can give money to Reddit for a piece of paper, but unless Reddit is claiming copyright to the content posted there, then they can't sue anyone for not paying. It would be very interesting to see the text of these "licensing agreements".

[-] lemmyvore@feddit.nl 4 points 1 month ago

They're not claiming copyright. They have a perpetual, non-revokable license to the content, granted by the people who use their site when they post the content.

[-] aramis87@fedia.io 4 points 1 month ago
this post was submitted on 01 Aug 2024
554 points (97.8% liked)

Technology

57944 readers
2909 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS