Most active commenters

    ←back to thread

    Anthropic raises $13B Series F

    (www.anthropic.com)
    585 points meetpateltech | 23 comments | | HN request time: 0s | source | bottom
    Show context
    llamasushi ◴[] No.45105325[source]
    The compute moat is getting absolutely insane. We're basically at the point where you need a small country's GDP just to stay in the game for one more generation of models.

    What gets me is that this isn't even a software moat anymore - it's literally just whoever can get their hands on enough GPUs and power infrastructure. TSMC and the power companies are the real kingmakers here. You can have all the talent in the world but if you can't get 100k H100s and a dedicated power plant, you're out.

    Wonder how much of this $13B is just prepaying for compute vs actual opex. If it's mostly compute, we're watching something weird happen - like the privatization of Manhattan Project-scale infrastructure. Except instead of enriching uranium we're computing gradient descents lol

    The wildest part is we might look back at this as cheap. GPT-4 training was what, $100M? GPT-5/Opus-4 class probably $1B+? At this rate GPT-7 will need its own sovereign wealth fund

    replies(48): >>45105396 #>>45105412 #>>45105420 #>>45105480 #>>45105535 #>>45105549 #>>45105604 #>>45105619 #>>45105641 #>>45105679 #>>45105738 #>>45105766 #>>45105797 #>>45105848 #>>45105855 #>>45105915 #>>45105960 #>>45105963 #>>45105985 #>>45106070 #>>45106096 #>>45106150 #>>45106272 #>>45106285 #>>45106679 #>>45106851 #>>45106897 #>>45106940 #>>45107085 #>>45107239 #>>45107242 #>>45107347 #>>45107622 #>>45107915 #>>45108298 #>>45108477 #>>45109495 #>>45110545 #>>45110824 #>>45110882 #>>45111336 #>>45111695 #>>45111885 #>>45111904 #>>45111971 #>>45112441 #>>45112552 #>>45113827 #
    duxup ◴[] No.45105396[source]
    It's not clear to me that each new generation of models is going to be "that" much better vs cost.

    Anecdotally moving from model to model I'm not seeing huge changes in many use cases. I can just pick an older model and often I can't tell the difference...

    Video seems to be moving forward fast from what I can tell, but it sounds like the back end cost of compute there is skyrocketing with it raising other questions.

    replies(9): >>45105636 #>>45105699 #>>45105746 #>>45105777 #>>45105835 #>>45106211 #>>45106364 #>>45106367 #>>45106463 #
    renegade-otter ◴[] No.45105699[source]
    We do seem to be hitting the top of the curve of diminishing returns. Forget AGI - they need a performance breakthrough in order to stop shoveling money into this cash furnace.
    replies(6): >>45105775 #>>45105790 #>>45105830 #>>45105936 #>>45105998 #>>45106035 #
    1. reissbaker ◴[] No.45106035[source]
    According to Dario, each model line has generally been profitable: i.e. $200MM to train a model that makes $1B in profit over its lifetime. But, since each model has been more and more expensive to train, they keep needing to raise more money to train the next generation of model, and the company balance sheet looks negative: i.e. they spent more this year than last (since the training cost for model N+1 is higher), and the model this year made less money this year than they spent (even if the model generation itself was profitable, model N isn't profitable enough to train model N+1 without raising — and spending — more money).

    That's still a pretty good deal for an investor: if I give you $15B, you will probably make a lot more than $15B with it. But it does raise questions about when it will simply become infeasible to train the subsequent model generation due to the costs going up so much (even if, in all likelihood, that model would eventually turn a profit).

    replies(8): >>45106645 #>>45106689 #>>45106988 #>>45107665 #>>45108456 #>>45110567 #>>45112144 #>>45112270 #
    2. viscanti ◴[] No.45106645[source]
    Well how much of it is correlation vs causation. Does the next generation of model unlock another 10x usage? Or was Claude 3 "good enough" that it got traction from early adopters and Claude 4 is "good enough" that it's getting a lot of mid/late adopters using it for this generation? Presumably competitors get better and at cheaper prices (Anthropic charges a premium per token currently) as well.
    3. dom96 ◴[] No.45106689[source]
    > if I give you $15B, you will probably make a lot more than $15B with it

    "probably" is the key word here, this feels like a ponzi scheme to me. What happens when the next model isn't a big enough jump over the last one to repay the investment?

    It seems like this already happened with GPT-5. They've hit a wall, so how can they be confident enough to invest ever more money into this?

    replies(1): >>45107077 #
    4. mandevil ◴[] No.45106988[source]
    I mean, this is how semiconductors have worked forever. Every new generation of fab costs ~2x what the previous generation did, and you need to build a new fab ever couple of years. But (if you could keep the order book full for the fab) it would make a lot of money over its lifetime, and you still needed to borrow/raise even more to build the next generation of fab. And if you were wrong about demand .... you got into a really big bust, which is also characteristic of the semiconductor industry.

    This was the power of Moore's Law, it gave the semiconductor engineers an argument they could use to convince the money-guys to let them raise the capital to build the next fab- see, it's right here in this chart, it says that if we don't do it our competitors will, because this chart shows that it is inevitable. Moore's Law had more of a financial impact than a technological one.

    And now we're down to a point where only TSMC is for sure going through with the next fab (as a rough estimate of cost, think 40 billion dollars)- Samsung and Intel are both hemming and hawing and trying to get others to go in with them, because that is an awful lot of money to get the next frontier node. Is Apple (and Nvidia, AMZ, Google, etc.) willing to pay the costs (in delivery delays, higher costs, etc.) to continue to have a second potential supplier around or just bite the bullet and commit to TSMC being the only company that can build a frontier node?

    And even if they can make it to the next node (1.4nm/14A), can they get to the one after that?

    The implication for AI models is that they can end up like Intel (or AMD, selling off their fab) if they misstep badly enough on one or two nodes in a row. This was the real threat of Deepseek: if they could get frontier models for an order of magnitude cheaper, then the entire economics of this doesn't work. If they can't keep up, then the economics of it might, so long as people are willing to pay more for the value produced by the new models.

    replies(1): >>45108069 #
    5. bcrosby95 ◴[] No.45107077[source]
    I think you're really bending over backwards to make this company seem non viable.

    If model training has truly turned out to be profitable at the end of each cycle, then this company is going to make money hand over fist, and investing money to out compete the competition is the right thing to do.

    Most mega corps started out wildly unprofitable due to investing into the core business... until they aren't. It's almost as if people forget the days of Facebook being seen as continually unprofitable. This is how basically all huge tech companies you know today started.

    replies(3): >>45107188 #>>45109776 #>>45113886 #
    6. serf ◴[] No.45107188{3}[source]
    >I think you're really bending over backwards to make this company seem non viable.

    Having experienced Anthropic as a customer, I have a hard time thinking that their inevitable failure (something i'd bet on) will be model/capability-based, that's how bad they suck at every other customer-facing metric.

    You think Amazon is frustrating to deal with? Get into a CSR-chat-loop with an uncaring LLM followed up on by an uncaring CSR.

    My minimum response time with their customer service is 14 days -- 2 weeks -- while paying 200usd a month.

    An LLM could be 'The Great Kreskin' and I would still try to avoid paying for that level of abuse.

    replies(2): >>45107371 #>>45111602 #
    7. sbarre ◴[] No.45107371{4}[source]
    Maybe you don't want to share, but I'm scratching my head trying to think of something I would need to talk to Anthropic's customer service about that would be urgent and un-straightfoward enough to frustrate me to the point of using the term "abuse"..
    replies(1): >>45108118 #
    8. yahoozoo ◴[] No.45107665[source]
    What about inference costs?
    9. m101 ◴[] No.45108069[source]
    Except it's like second tier semi manufacturer spending 10x less on the same fab in one years time. Here it might make sense to wait a bit. There will be customers, especially considering the diminishing returns these models seem to have come across. If performance was improving I'd agree with you, but it's not.
    10. babelfish ◴[] No.45108118{5}[source]
    Particularly since they seem to be complaining about service as a consumer, rather than an enterprise...
    replies(1): >>45110256 #
    11. Avshalom ◴[] No.45108456[source]
    if you're referring to https://youtu.be/GcqQ1ebBqkc?t=1027 he doesn't actually say that each model has been profitable.

    He says "You paid $100 million and then it made $200 million of revenue. There's some cost to inference with the model, but let's just assume in this cartoonish cartoon example that even if you add those two up, you're kind of in a good state. So, if every model was a company, the model is actually, in this example is actually profitable. What's going on is that at the same time"

    notice those are hypothetical numbers and he just asks you to assume that inference is (sufficiently) profitable.

    He doesn't actually say they made money by the EoL of some model.

    12. Barbing ◴[] No.45109776{3}[source]
    Thoughts on Ed Zitron’s pessimism?

    “There Is No AI Revolution” - Feb ‘25:

    https://www.wheresyoured.at/wheres-the-money/

    replies(1): >>45125350 #
    13. ◴[] No.45110256{6}[source]
    14. 9cb14c1ec0 ◴[] No.45110567[source]
    That can only be true if someone else is subsidizing Anthropic's compute. The calculation is simple: Annualized depreciation costs on the AI buildout (hundreds of billions, possibly a trillion invested) are more that the combined total annualized revenue of the inference industry. A more realistic computation of expenses would show the each model line very deeply in the red.
    15. StephenHerlihyy ◴[] No.45111602{4}[source]
    What's fun is that I have had Anthropic's AI support give me blatantly false information. It tried to tell me that I could get a full year's worth of Claude Max for only $200 dollars. When I asked if that was true it quickly backtracked and acknowledged it's mistake. I figure someone more litigious will eventually try to capitalize.
    replies(1): >>45112519 #
    16. majormajor ◴[] No.45112144[source]
    Do they have a function to predict in advance if the next model is going to be profitable?

    If not, this seems like a recipe for bankruptcy. You are always investing more than you're making, right up until the day you don't make it back. Whether that's next year or in ten or twenty years. It's basically impossible to do it forever - there simply isn't enough profit to be had in the world if you go forward enough orders of magnitude. How will they know when to hop off the train?

    replies(1): >>45113228 #
    17. oblio ◴[] No.45112270[source]
    > According to Dario, each model line has generally been profitable: i.e. $200MM to train a model that makes $1B in profit over its lifetime.

    Surely the Anthropic CEO will have no incentive to lie.

    replies(1): >>45112528 #
    18. nielsbot ◴[] No.45112519{5}[source]
    "Air Canada must honor refund policy invented by airline’s chatbot"

    https://arstechnica.com/tech-policy/2024/02/air-canada-must-...

    19. nielsbot ◴[] No.45112528[source]
    Not saying he's above lying, but I do believe there are potential legal ramifications from a CEO lying. (Assuming they get caught)
    20. ikr678 ◴[] No.45113228[source]
    Back in my day, we called this a pyramid scheme.
    21. ricardobayes ◴[] No.45113886{3}[source]
    It's an interesting case. IMO LLMs are not a product in the classical sense, companies like Anthropic are basically doing "basic research" so others can build products on top of it. Perhaps Anthropic will charge a royalty on the API usage. I personally don't think you can earn billions selling $500 subscriptions. This has been shown by the SaaS industry. But it is yet to be seen whether the wider industry will accept such royalty model. It would be akin to Kodak charging filmmakers based on the success of the movie. Somehow AI companies will need to build a monetization pipeline that will earn them a small amount of money "with every gulp", if we are using a soft drink analogy.
    22. reissbaker ◴[] No.45125350{4}[source]
    Ed Zitron plainly has no idea what he's talking about. For example:

    Putting aside the hype and bluster, OpenAI — as with all generative AI model developers — loses money on every single prompt and output. Its products do not scale like traditional software, in that the more users it gets, the more expensive its services are to run because its models are so compute-intensive.

    While OpenAI's numbers aren't public, this seems very unlikely. Given open-source models can be profitably run for cents per million input tokens at FP8 — and OpenAI is already training (and thus certainly running) in FP4 — even if the closed-source models are many times bigger than the largest open-source models, OpenAI is still making money hand over fist on inference. The GPT-5 API costs $1.25/million input tokens: that's a lot more than it takes in compute to run it. And unless you're using the API, it's incredibly unlikely you're burning through millions of tokens in a week... And yet, subscribers to the chat UI are paying $20/month (at minimum!), which is much higher than a few million tokens a week cost.

    Ed Zitron repeats his claim many, many, excruciatingly many times throughout the article, and it seems quite central to the point he's trying to make. But he's wrong, and wrong enough that I think you should doubt that he knows much about what he's talking about.

    (His entire blog seems to be a series of anti-tech screeds, so in general I'm pretty dubious he has deep insight into much of anything in the industry. But he quite obviously doesn't know about the economics of LLM inference.)

    replies(1): >>45128836 #
    23. Barbing ◴[] No.45128836{5}[source]
    Thank you for your analysis!