Most active commenters
  • jowea(3)

←back to thread

511 points andy99 | 19 comments | | HN request time: 0.476s | source | bottom
1. k__ ◴[] No.44536047[source]
"respecting web crawling opt-outs during data acquisition produces virtually no performance degradation"

Great to read that!

replies(3): >>44536377 #>>44538760 #>>44539036 #
2. Onavo ◴[] No.44536377[source]
No performance degradation on training metrics except for the end user. At the end of the day users and website owners have completely orthogonal interests. Users want answers and content, website owners want attention so they can upsell/push ads. You can only serve one master.
replies(1): >>44536456 #
3. esafak ◴[] No.44536456[source]
> Users want answers and content, website owners want attention so they can upsell/push ads. You can only serve one master

How are you going to serve users if web site owners decide to wall their content? You can't ignore one side of the market.

replies(1): >>44536786 #
4. Onavo ◴[] No.44536786{3}[source]
You don't. You bypass them with crawlers and don't reveal your training data. And this is exactly why open source models can't surpass open weight models.
replies(1): >>44537425 #
5. diggan ◴[] No.44537425{4}[source]
> And this is exactly why open source models can't surpass open weight models.

It is a fair point, but how strong of a point it is remains to be seen, some architectures are better than others, even with the same training data, so not impossible we could at one point see some innovative architectures beating current proprietary ones. It would probably be short-lived though, as the proprietary ones would obviously improve in their next release after that.

replies(1): >>44537860 #
6. jowea ◴[] No.44537860{5}[source]
How can open source models respectful of robots.txt possibly perform equally if they are missing information that the other models have access to?
replies(3): >>44538826 #>>44539844 #>>44540399 #
7. JKCalhoun ◴[] No.44538760[source]
Is there not yet a Source where the web has already been scraped and souped down to just the text? It would seem someone would have created such a thing in order to save LLM training from having to reinvent the wheel.

I understand the web is a dynamic thing but still it would seem to be useful on some level.

replies(1): >>44540972 #
8. datameta ◴[] No.44538826{6}[source]
How can we possibly find out without trying?
replies(1): >>44538893 #
9. jowea ◴[] No.44538893{7}[source]
It is logically impossible for a LLM to, for example, to know that fooExecute() takes two int arguments if the documentation is blocked by robots.txt and there are no examples of fooExecute() usage in the wild, don't you agree?
replies(2): >>44539576 #>>44540377 #
10. stephen_cagle ◴[] No.44539036[source]
I wonder if the reason for these results is that any data on the internet is already copied to other locations by actors who ignore crawling opt-outs. So, even if they respect all web crawling opt-outs, they are still effectively copying the data because someone else did not respect it who does not include an opt-out.
replies(2): >>44539370 #>>44539981 #
11. conradkay ◴[] No.44539370[source]
My guess is that it doesn't remove that much of the data, and the post-training data (not just randomly scraped from the web) probably matters more
12. tharant ◴[] No.44539576{8}[source]
Sure, the model would not “know” about your example, but that’s not the point; the penultimate[0] goal is for the model to figure out the method signature on its own just like a human dev might leverage her own knowledge and experience to infer that method signature. Intelligence isn’t just rote memorization.

[0] the ultimate, of course, being profit.

replies(1): >>44539831 #
13. jowea ◴[] No.44539831{9}[source]
I don't think a human dev can divine a method signature and effects in the general case either. Sure the add() function probably takes 2 numbers, but maybe it takes a list? Or a two-tuple? How would we or the LLM know without having the documentation? And yeah sure the LLM can look at the documentation while being used instead of it being part of the training dataset, but that's strictly inferior for practical uses, no?

I'm not sure if we're thinking of the same field of AI development. I think I'm talking about the super-autocomplete with integrated copy of all of digitalized human knowledge, while you're talking about trying to do (proto-)AGI. Is that it?

replies(1): >>44540274 #
14. Dylan16807 ◴[] No.44539844{6}[source]
Maybe the missing data makes it 3% worse but the architecture is 5% better. Or your respect for robots.txt gets you more funding and you gain a 4% advantage by training longer.

Don't focus too much on a single variable, especially when all the variables have diminishing returns.

15. lllllm ◴[] No.44539981[source]
Yes this is an interesting question. In our arxiv paper [1] we did study this for news articles, and also removed duplicates of articles (decontamination). We did not observe an impact on the downstream accuracy of the LLM, in the case of news data.

[1] https://arxiv.org/abs/2504.06219

16. heavenlyblue ◴[] No.44540274{10}[source]
> Sure the add() function probably takes 2 numbers, but maybe it takes a list? Or a two-tuple? How would we or the LLM know without having the documentation?

You just listed possible options in the order of their relative probability. Human would attempt to use them in exactly that order

17. diggan ◴[] No.44540377{8}[source]
I agree, but also think it's less important. I don't want a big fat LLM that memorized every API out there, and as soon as the API changed, the weights have to updated. I like the current approach of Codex (and similar) where they can look up the APIs they need to use as they're doing the work instead, so same weights will continue to work no matter how much the APIs change.
18. lllllm ◴[] No.44540399{6}[source]
this is what this paper tries to answer: https://arxiv.org/abs/2504.06219 the quality gap is surprisingly small between compliant and not
19. CaptainFever ◴[] No.44540972[source]
Common Crawl, maybe?