←back to thread

646 points blendergeek | 1 comments | | HN request time: 0s | source
Show context
observationist[dead post] ◴[] No.42726591[source]
[flagged]
jsheard ◴[] No.42726726[source]
This is a really bad take, it's not like this server is hacking clients which connect to it. It's providing perfectly valid HTTP responses that just happen to be slow and full of markov gibberish, any harm which comes of that is self inflicted by assuming that websites must provide valuable data as a matter of course.

If AI companies want to sue webmasters for that then by all means, they can waste their money and get laughed out of court.

replies(3): >>42726813 #>>42726898 #>>42729375 #
observationist[dead post] ◴[] No.42726813[source]
[flagged]
blibble ◴[] No.42726981[source]
> If you want to protect your content, use the technical mechanisms that are available,

> You can choose to gatekeep your content, and by doing so, make it unscrapeable, and legally protected.

so... robots.txt, which the AI parasites ignore?

> Also, consider that relatively small, cheap llms are able to parse the difference between meaningful content and Markovian jabber such as this software produces.

okay, so it's not damaging, and there you've refuted your entire argument

replies(1): >>42727385 #
observationist[dead post] ◴[] No.42727385{3}[source]
[flagged]
jsheard ◴[] No.42727632{4}[source]
> No, put up a loginwall or paywall, authenticate users, and go private.

We know for a fact that AI companies don't respect that, if they want data that's behind a paywall then they'll jump through hoops to take it anyway.

https://www.theguardian.com/technology/2025/jan/10/mark-zuck...

If they don't have to abide by "norms" then we don't have to for their sake. Fuck 'em.

replies(1): >>42727731 #
observationist[dead post] ◴[] No.42727731{5}[source]
[flagged]
1. tir ◴[] No.42728192{6}[source]
>the law explicitly allows scraping and crawling.

Nepenthes also allows scraping and crawling, for as long as you like.