←back to thread

505 points andy99 | 8 comments | | HN request time: 0.611s | source | bottom
1. WeirderScience ◴[] No.44536327[source]
The open training data is a huge differentiator. Is this the first truly open dataset of this scale? Prior efforts like The Pile were valuable, but had limitations. Curious to see how reproducible the training is.
replies(2): >>44536400 #>>44537249 #
2. layer8 ◴[] No.44536400[source]
> The model will be fully open: source code and weights will be publicly available, and the training data will be transparent and reproducible

This leads me to believe that the training data won’t be made publicly available in full, but merely be “reproducible”. This might mean that they’ll provide references like a list of URLs of the pages they trained on, but not their contents.

replies(3): >>44536448 #>>44536623 #>>44536818 #
3. WeirderScience ◴[] No.44536448[source]
Yeah, I suspect you're right. Still, even a list of URLs for a frontier model (assuming it does turn out to be of that level) would be welcome over the current situation.
4. glhaynes ◴[] No.44536623[source]
That wouldn't seem reproducible if the content at those URLs changes. (Er, unless it was all web.archive.org URLs or something.)
replies(1): >>44536997 #
5. TobTobXX ◴[] No.44536818[source]
Well, when the actual content is 100s of terabytes big, providing URLs may be more practical for them and for others.
replies(1): >>44537342 #
6. dietr1ch ◴[] No.44536997{3}[source]
This is a problem with the Web. It should be easier to download content like it was updating a git Repo.
7. evolvedlight ◴[] No.44537249[source]
Yup, it’s not a dataset packaged like you hope for here, as it still contains traditionally copyrighted material
8. layer8 ◴[] No.44537342{3}[source]
The difference between content they are allowed to train on vs. being allowed to distribute copies of is likely at least as relevant.