←back to thread

GPT-5.2

(openai.com)
1019 points atgctg | 4 comments | | HN request time: 0.001s | source
1. mattas ◴[] No.46235111[source]
Are benchmarks the right way to measure LLMs? Not because benchmarks can be gamed, but because the most useful outputs of models aren't things that can be bucketed into "right" and "wrong." Tough problem!
replies(2): >>46235164 #>>46235214 #
2. Sir_Twist ◴[] No.46235164[source]
Not an expert in LLM benchmarks, but I generally I think of benchmarks as being good particularly for measuring usefulness for certain usecases. Even if measuring LLMs is not as straightforward as, say, read/write speeds when comparing different SSDs, if a certain model's responses are consistently measured as being higher quality / more useful, surely that means something, right?
3. olliepro ◴[] No.46235214[source]
Do you have a better way to measure LLMs? Measurement implies quantitative evaluation... which is the same as benchmarks.
replies(1): >>46236704 #
4. Wowfunhappy ◴[] No.46236704[source]
I don’t have a good way to measure them, but I think they should be evaluated more like how we evaluate movies, or restaurants. Namely, experienced critics try them and write reviews.