←back to thread

211 points CrankyBear | 2 comments | | HN request time: 0.411s | source
Show context
giancarlostoro ◴[] No.45106227[source]
I'm not sure why they don't just cache the websites and avoid going back for at least 24 hours, especially in the case of most sites. I swear its like we're re-learning software engineering basics with LLMs / AI and it kills me.
replies(8): >>45106404 #>>45106430 #>>45106554 #>>45107000 #>>45107104 #>>45107170 #>>45107187 #>>45112971 #
1. immibis ◴[] No.45107187[source]
It's because they don't give a shit whether the product works properly or not. By blocking AI scraping, sites are forcing AI companies to scrape faster before they're blocked. And faster means sloppier.
replies(1): >>45107489 #
2. lovich ◴[] No.45107489[source]
There’s also the point that if the web site is down after you scraped it, then that’s 1 more sites data you’ve scraped that your competition now cant