←back to thread

507 points martinald | 9 comments | | HN request time: 0.506s | source | bottom
Show context
sc68cal ◴[] No.45053212[source]
This whole article is built off using DeepSeek R1, which is a huge premise that I don't think is correct. DeepSeek is much more efficient and I don't think it's a valid way to estimate what OpenAI and Anthropic's costs are.

https://www.wheresyoured.at/deep-impact/

Basically, DeepSeek is _very_ efficient at inference, and that was the whole reason why it shook the industry when it was released.

replies(7): >>45053283 #>>45053303 #>>45053401 #>>45053455 #>>45053507 #>>45053923 #>>45054034 #
1. phillipcarter ◴[] No.45053455[source]
Uhhh, I'm pretty sure DeepSeek shook the industry because of a 14x reduction in training cost, not inference cost.

We also don't know the per-token cost for OpenAI and Anthropic models, but I would be highly surprised if it was significantly more expensive than open models anyone can use and run themselves. It's not like they're also not investing in inference research.

replies(3): >>45053857 #>>45053879 #>>45053974 #
2. andai ◴[] No.45053857[source]
Isn't training cost a function of inference cost? From what I gathered, they reduced both.

I remember seeing lots of videos at the time explaining the details, but basically it came down to the kind of hardware-aware programming that used to be very common. (Although they took it to the next level by using undocumented behavior to their advantage.)

replies(1): >>45053897 #
3. baxtr ◴[] No.45053879[source]
Because of the alleged reduction in training costs.
replies(1): >>45053970 #
4. booi ◴[] No.45053897[source]
They're typically somewhat related but the difference between training and inference can vary greatly so, i guess the answer is no.

they did reduce both though and mostly due to reduced precision

5. basilgohar ◴[] No.45053970[source]
All reports by companies are alleged until verified by other, more trustworthy sources. I don't think it's especially notable that it's alleged because it's DeepSeek vs. the alleged numbers from other companies.
6. gmd63 ◴[] No.45053974[source]
DeepSeek was trained with distillation. Any accurate estimate of training costs should include the training costs of the model that it was distilling.
replies(1): >>45054081 #
7. ffsm8 ◴[] No.45054081[source]
That makes the calculation nonsensical, because if you go there... you'd also have to include all energy used in producing the content the other model providers used. So now suddenly everyones devices on which they wrote comments on social media, pretty much all servers to have ever served a request to open AI/Google/anthropics bots etc pp

Seriously, that claim was always completely disingenuous

replies(2): >>45055659 #>>45062012 #
8. gmd63 ◴[] No.45055659{3}[source]
I don't think it's that nonsensical to realize that in order to have AI, you need generations of artists, journalists, scientists, and librarians to produce materials to learn from.

And when you're using an actual AI model to "train" (copy), it's not even a shred of nonsense to realize the prior model is a core component of the training.

9. jaakl ◴[] No.45062012{3}[source]
Not just energy cost, but also licensing cost of all this content…