←back to thread

257 points ColinWright | 1 comments | | HN request time: 0.208s | source
Show context
bakql ◴[] No.45775259[source]
>These were scrapers, and they were most likely trying to non-consensually collect content for training LLMs.

"Non-consensually", as if you had to ask for permission to perform a GET request to an open HTTP server.

Yes, I know about weev. That was a travesty.

replies(15): >>45775283 #>>45775392 #>>45775754 #>>45775912 #>>45775998 #>>45776008 #>>45776055 #>>45776210 #>>45776222 #>>45776270 #>>45776765 #>>45776932 #>>45777727 #>>45777934 #>>45778166 #
XenophileJKO ◴[] No.45775283[source]
What about people using an LLM as their web client? Are you now saying the website owner should be able to dictate what client I use and how it must behave?
replies(2): >>45776757 #>>45777923 #
aDyslecticCrow ◴[] No.45776757[source]
> Are you now saying the website owner should be able to dictate what client I use and how it must behave?

Already pretty well established with Ad-block actually. It's a pretty similar case even. AI's don't click ads, so why should we accept their traffic? If it's un-proportionally loading the server without contributing to the funding of the site, get blocked.

The server can set whatever rules it wants. If the maintainer hates google and wants to block all chrome users, it can do so.

replies(1): >>45777624 #
XenophileJKO ◴[] No.45777624[source]
That was kind of what I was really hinting at, as the HN community tends to embrace things like ad blockers and archive links on stories, but god forbid someone read a site using an LLM.
replies(2): >>45778268 #>>45778307 #
1. 1gn15 ◴[] No.45778268[source]
Humans are usually hypocritical. They support whatever they personally use while opposing whatever inconveniences them, even though they're basically the same thing.

This whole thing has made me hate humans, so so much. Robots are much better.