110
NYT Disables Open AI bot (www.theverge.com)

NYT looks like it's updated it's robots.txt file to disallow the Open AI bot from scraping it's data. Pretty interested to see if they just update their user agent string or if they'll respect it

all 8 comments
sorted by: hot top controversial new old
[-] plz1@lemmy.world 36 points 10 months ago

Updating user agent doesn't natter unless NYT is actively blocking that, too. Updating robots.txt is purely a "gentleman's agreement" that OpenAI will respect it. OpenAI would be dumb to ignore it, hat all said, because it'd trigger the lawyer shenanigans to ensue.

[-] DocMcStuffin@lemmy.world 12 points 10 months ago

NYT is already considering a lawsuit against OpenAI. So, not just dumb but arrogantly stupid when the lawyers are already in the room.

[-] iforgotmyinstance@lemmy.world 9 points 10 months ago

The burden of proof will fall upon the NYT and it will be extremely difficult to prove OpenAI is culpable for any infringement that it's end users perform.

It's new territory and will be expensive, but NYT is old money and has the liquidity to burn cash all day.

[-] autotldr@lemmings.world 8 points 10 months ago

This is the best summary I could come up with:


Based on the Internet Archive’s Wayback Machine, it appears NYT blocked the crawler as early as August 17th.

The change comes after the NYT updated its terms of service at the beginning of this month to prohibit the use of its content to train AI models.

OpenAI didn’t immediately reply to a request for comment.

The NYT is also considering legal action against OpenAI for intellectual property rights violations, NPR reported last week.

If it did sue, the Times would be joining others like Sarah Silverman and two other authors who sued the company in July over its use of Books3, a dataset used to train ChatGPT that may have thousands of copyrighted works, as well as Matthew Butterick, a programmer and lawyer who alleges the company’s data scraping practices amount to software piracy.

Update August 21st, 7:55PM ET: The New York Times declined to comment.


The original article contains 202 words, the summary contains 146 words. Saved 28%. I'm a bot and I'm open source!

[-] cbarrick@lemmy.world 6 points 10 months ago

But all those reposts on Reddit and Lemmy are still fair game...

[-] simonced@lemmy.one 11 points 10 months ago

shared by humans is not the same as crawled by bots...

[-] WarmSoda@lemm.ee 4 points 10 months ago

I wonder how much of a boost sites get from Reddit and lemmy, etc. Even with posts that have the text copy/pasted I imagine it has to give them traffic.

this post was submitted on 23 Aug 2023
110 points (94.4% liked)

Technology

55919 readers
2765 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS