←back to thread

507 points martinald | 2 comments | | HN request time: 0.007s | source
Show context
ankit219 ◴[] No.45053523[source]
This seems very very far off. From the latest reports, anthropic has a gross margin of 60%. It came out in their latest fundraising story. From that one The Information report, it estimated OpenAI's GM to be 50% including free users. These are gross margins so any amortization or model training cost would likely come after this.

Then, today almost every lab uses methods like speculative decoding and caching which reduce the cost and speed up things significantly.

The input numbers are far off. The assumption is 37B of active parameters. Sonnet 4 is supposedly a 100B-200B param model. Opus is about 2T params. Both of them (even if we assume MoE) wont have exactly these number of output params. Then there is a cost to hosting and activating params at inference time. (the article kind of assumes it would be the same constant 37B params).

replies(2): >>45053768 #>>45054031 #
thegeomaster ◴[] No.45054031[source]
Are you saying that you think Sonnet 4 has 100B-200B _active_ params? And that Opus has 2T active? What data are you basing these outlandish assumptions on?
replies(2): >>45054387 #>>45055219 #
Der_Einzige ◴[] No.45055219[source]
Not everyone uses MoE architectures. It's not outlandish at all...
replies(1): >>45058203 #
thegeomaster ◴[] No.45058203[source]
There's no way Sonnet 4 or Opus 4 are dense models.
replies(1): >>45058292 #
1. Der_Einzige ◴[] No.45058292[source]
Citation needed
replies(1): >>45058625 #
2. thegeomaster ◴[] No.45058625[source]
Common sense:

- The compute requirements would be massive compared to the rest of the industry

- Not a single large open source lab has trained anything over 32B dense in the recent past

- There is considerable crosstalk between researchers at large labs; notice how all of them seem to be going in similar directions all the time. If dense models of this size actually provided benefit compared to MoE, the info would've spread like wildfire.