←back to thread

DeepSeek-v3.1

(api-docs.deepseek.com)
776 points wertyk | 2 comments | | HN request time: 0s | source
Show context
rsanek ◴[] No.44980753[source]
Looks to be the ~same intelligence as gpt-oss-120B, but about 10x slower and 3x more expensive?

https://artificialanalysis.ai/models/deepseek-v3-1-reasoning

replies(5): >>44981187 #>>44981737 #>>44981789 #>>44982171 #>>44982769 #
okasaki ◴[] No.44981737[source]
My experience is that gpt-oss doesn't know much about obscure topics, so if you're using it for anything except puzzles or coding in popular languages, it won't do well as the bigger models.

It's knowledge seems to be lacking even compared to gpt3.

No idea how you'd benchmark this though.

replies(2): >>44981768 #>>44984205 #
1. xadhominemx ◴[] No.44984205[source]
> My experience is that gpt-oss doesn't know much about obscure topics

That is the point of these small models. Remove the bloat of obscure information (address that with RAG), leaving behind a core “reasoning” skeleton.

replies(1): >>44984641 #
2. okasaki ◴[] No.44984641[source]
Yeah I guess. Just wanted to say the size difference might be accounted for by the model knowing more.

Seems more user-friendly to bake it in.