←back to thread

524 points andy99 | 6 comments | | HN request time: 1.625s | source | bottom
1. oytis ◴[] No.44536348[source]
The press release talks a lot about how it was done, but very little about how capabilities compare to other open models.
replies(2): >>44536398 #>>44536523 #
2. pantalaimon ◴[] No.44536398[source]
It's a university, teaching the 'how it's done' is kind of the point
replies(1): >>44536825 #
3. joot82 ◴[] No.44536523[source]
The model will be released in two sizes — 8 billion and 70 billion parameters [...]. The 70B version will rank among the most powerful fully open models worldwide. [...] In late summer, the LLM will be released under the Apache 2.0 License.

We'll find out in September if it's true?

replies(2): >>44536885 #>>44536967 #
4. EA-3167 ◴[] No.44536825[source]
Sure, but usually you teach something that is inherently useful, or can be applied to some sort of useful endeavor. In this case I think it's fair to ask what the collision of two bubbles really achieves, or if it's just a useful teaching model, what it can be applied to.
5. k__ ◴[] No.44536885[source]
I hope DeepSeek R2, but I fear Llama 4.
6. oytis ◴[] No.44536967[source]
Yeah, I was thinking more of a table with benchmark results