←back to thread

507 points martinald | 1 comments | | HN request time: 0s | source
Show context
sc68cal ◴[] No.45053212[source]
This whole article is built off using DeepSeek R1, which is a huge premise that I don't think is correct. DeepSeek is much more efficient and I don't think it's a valid way to estimate what OpenAI and Anthropic's costs are.

https://www.wheresyoured.at/deep-impact/

Basically, DeepSeek is _very_ efficient at inference, and that was the whole reason why it shook the industry when it was released.

replies(7): >>45053283 #>>45053303 #>>45053401 #>>45053455 #>>45053507 #>>45053923 #>>45054034 #
1. boroboro4 ◴[] No.45053283[source]
DeepSeek inference efficiency comes from two things: MoE and MLA attention. OpenAI was rumored to use MoE around GPT4 moment, I.e loooong time ago.

Given Gemini efficiency with long context I would bet their attention is very efficient too.

GPT OSS uses fp4, which DeepSeek doesn’t use yet btw.

So no, big labs aren’t behind DeepSeek in efficiency. Not by much at least.