←back to thread

32 points oldfuture | 1 comments | | HN request time: 0.41s | source
Show context
tripplyons ◴[] No.44505634[source]
Nice to see how open it is! However, if you are just looking for the best model, Mistral Small 3.2 appears to be a stronger model with fewer parameters compared to OLMo 2 32B. It would be interesting to see how far these "fully open" models can get to their "open weight" counterparts.
replies(1): >>44505916 #
real0mar ◴[] No.44505916[source]
The inconvenient truth might be that the other models score higher than OLMO because they aren't restricted to purely "open and accessible" training data. Who knows what private or ethically dubious data went into training Mistral or llama, for example.
replies(1): >>44506776 #
1. erlend_sh ◴[] No.44506776[source]
Exactly. If we really wanted to benchmark the various models on the merits of their individual implementations, we should be comparing them all on the same open dataset.