←back to thread

DeepSeek-v3.1

(api-docs.deepseek.com)
776 points wertyk | 6 comments | | HN request time: 0.473s | source | bottom
Show context
hodgehog11 ◴[] No.44977357[source]
For reference, here is the terminal-bench leaderboard:

https://www.tbench.ai/leaderboard

Looks like it doesn't get close to GPT-5, Claude 4, or GLM-4.5, but still does reasonably well compared to other open weight models. Benchmarks are rarely the full story though, so time will tell how good it is in practice.

replies(6): >>44977423 #>>44977655 #>>44977754 #>>44977946 #>>44978395 #>>44978560 #
guluarte ◴[] No.44977946[source]
tbh companies like anthopic, openai, create custom agents for specific benchmarks
replies(2): >>44978101 #>>44979380 #
amelius ◴[] No.44979380[source]
Aren't good benchmarks supposed to be secret?
replies(3): >>44979634 #>>44982470 #>>45056160 #
noodletheworld ◴[] No.44982470[source]
How can a benchmark be secret if you post it to an API to test a model on it?

"We totally promise that when we run your benchmark against our API we won't take the data from it and use to be better at your benchmark next time"

:P

If you want to do it properly you have to avoid any 3rd party hosted model when you test your benchmark, which means you can't have GPT5, claude, etc. on it; and none of the benchmarks want to be 'that guy' who doesn't have all the best models on it.

So no.

They're not secret.

replies(1): >>44982928 #
dmos62 ◴[] No.44982928[source]
How do you propose that would work? A pipeline that goes through query-response pairs to deduce response quality and then uses the low-quality responses for further training? Wouldn't you need a model that's already smart enough to tell that previous model's responses weren't smart enough? Sounds like a chicken and egg problem.
replies(1): >>44983190 #
1. irthomasthomas ◴[] No.44983190[source]
It just means that once you send your test questions to a model API, that company now has your test. So 'private' benchmarks take it on faith that the companies won't look at those requests and tune their models or prompts to beat them.
replies(2): >>44983732 #>>44985081 #
2. dmos62 ◴[] No.44983732[source]
Sounds a bit presumptious to me. Sure, they have your needle, but they also need a cost-efficient way to find it in their hay stack.
replies(2): >>44984722 #>>44986900 #
3. noodletheworld ◴[] No.44984722[source]
Security through obscurity is not security.

Your api key is linked to your credit card, which is linked to your identity.

…but hey, youre right.

Lets just trust them not to be cheating. Cool.

4. merelysounds ◴[] No.44985081[source]
Would the model owners be able to identify the benchmarking session among many other similar requests?
replies(1): >>44985331 #
5. irthomasthomas ◴[] No.44985331[source]
Depends. Something like arc-agi might be easy as it follows a defined format. I would also guess that the usage pattern for someone running a benchmark will be quite distinct from that of a normal user, unless they take specific measures to try to blend in.
6. lucianbr ◴[] No.44986900[source]
They have quite large amounts of money. I don't think they need to be very cost-efficient. And they also have very smart people, so likely they can figure out a somewhat cost-efficient way. The stakes are high, for them.