←back to thread

770 points ta988 | 4 comments | | HN request time: 0.894s | source
Show context
Ukv ◴[] No.42550989[source]
Are these IPs actually from OpenAI/etc. (https://openai.com/gptbot.json), or is it possibly something else masquerading as these bots? The real GPTBot/Amazonbot/etc. claim to obey robots.txt, and switching to a non-bot UA string seems extra questionable behaviour.
replies(2): >>42551196 #>>42563566 #
equestria ◴[] No.42551196[source]
I exclude all the published LLM User-Agents and have a content honeypot on my website. Google obeys, but ChatGPT and Bing still clearly know the content of the honeypot.
replies(3): >>42551318 #>>42551321 #>>42551783 #
1. jonnycomputer ◴[] No.42551321[source]
how do you determine that they know the content of the honeypot?
replies(1): >>42551553 #
2. arrowsmith ◴[] No.42551553[source]
Presumably the "honeypot" is an obscured link that humans won't click (e.g. tiny white text on a white background in a forgotten corner of the page) but scrapers will. Then you can determine whether a given IP visited the link.
replies(2): >>42551731 #>>42551733 #
3. 55555 ◴[] No.42551731[source]
I interpreted it to mean that a hidden page (linked as u describe) is indexed in Bing or that some "facts" written on a hidden page are regurgitated by ChatGPT.
4. jonnycomputer ◴[] No.42551733[source]
I know what a honeypot is, but the question is how the know the scraped data was actually used to train llms. I wondered whether they discovered or verified that by getting the llm to regurgitate content from the honeypot.