Most active commenters

    ←back to thread

    507 points martinald | 11 comments | | HN request time: 0s | source | bottom
    Show context
    _sword ◴[] No.45055003[source]
    I've done the modeling on this a few times and I always get to a place where inference can run at 50%+ gross margins, depending mostly on GPU depreciation and how good the host is at optimizing utilization. The challenge for the margins is whether or not you consider model training costs as part of the calculation. If model training isn't capitalized + amortized, margins are great. If they are amortized and need to be considered... yikes
    replies(7): >>45055030 #>>45055275 #>>45055536 #>>45055820 #>>45055835 #>>45056242 #>>45056523 #
    BlindEyeHalo ◴[] No.45055275[source]
    Why wouldn't you factor in training? It is not like you can train once and then have the model run for years. You need to constantly improve to keep up with the competition. The lifespan of a model is just a few months at this point.
    replies(7): >>45055303 #>>45055495 #>>45055624 #>>45055631 #>>45056110 #>>45056973 #>>45057517 #
    1. jacurtis ◴[] No.45057517[source]
    In a recent episode of Hard Fork podcast, the hosts discussed an on-the-record conversation they had with Sam Altman from OpenAI. They asked him about profitability and he claimed that they are losing money mostly because of the cost of training. But as the model advances, they will train less and less. Once you take training out of the equation he claimed they were profitable based on the cost of serving the trained foundation models to users at current prices.

    Now, when he said that, his CFO corrected him and said they aren't profitable, but said "it's close".

    Take that with a grain of salt, but thats a conversation from one of the big AI companies that is only a few weeks old. I suspect that it is pretty accurate that pricing is currently reasonable if you ignore training. But training is very expensive and the reason most AI companies are losing money right now.

    replies(4): >>45057639 #>>45057962 #>>45060581 #>>45061058 #
    2. pas ◴[] No.45057639[source]
    > most AI companies are losing money right now

    which is completely "normal" at this point, """right"""? if you have billions of VC money chasing returns there's no time to sit around, it's all in, the hype train doesn't wait for bootstrapping profitability. and of course with these gargantuan valuations and mandatory YoY growth numbers, there is no way they are not fucking with the unit economy numbers too. (biases are hard to beat, especially if there's not much conscious effort to do so.)

    replies(1): >>45058223 #
    3. dgfitz ◴[] No.45057962[source]
    > But as the model advances, they will train less and less.

    They sure have a lot of training to do between now and whenever that happens. Rolling back from 5 to whatever was before it is their own admission of this fact.

    replies(1): >>45058471 #
    4. brianwawok ◴[] No.45058223[source]
    Does the cost of good come down 10x or not? For say Uber it didn’t, so we went from great $6 VC funded product to mediocre $24 ride product we have today. I’m not sure I’m going to use Copilot at $1 per request. Or even $0.25. Starts to approach overseas consultant in price and ability.
    5. mindwok ◴[] No.45058471[source]
    I think that actually proves the opposite. People wanted an old model, not a new one, indicating that for that user base they could have just... not trained a new model.
    replies(3): >>45058933 #>>45060288 #>>45060618 #
    6. jazzyjackson ◴[] No.45058933{3}[source]
    for their user base, sure

    for their investors, however, they are promising a revolution

    7. hnfsfdsd ◴[] No.45060288{3}[source]
    If people want old models, they can go to any of the competitor's , deepseek, claud, opensources, etc... That's not good news for OpenAI.
    8. anothernewdude ◴[] No.45060581[source]
    Unfortunately for those companies, their APIs are a commodity, and are very fungible. So they'll need to keep training or be replaced with whichever competitor will. This is an exercise in attrition.
    replies(1): >>45062777 #
    9. PeterStuer ◴[] No.45060618{3}[source]
    That is for a very specific class of usecases. If they would turn up the sycophancy on the new model, those people would not call for the old onee.

    The reasoning here is off. It is like saying new game development is nearly over as some people keep playing old games.

    My feeling: we've yet barely scrarched the surface on the milage we can get out of even today's frontier models, but we are just at the beginning of a huge runway for improved models and architectures. Watch this space.

    10. diamond559 ◴[] No.45061058[source]
    You lost me at "Sam Altman says".
    11. LadyCailin ◴[] No.45062777[source]
    I wonder if we’re reaching a point of diminishing returns with training, at least, just by scaling the data set. I mean, there’s a finite amount of information (that can be obtained reasonably) to be trained on. I think we’re already at a sizable chunk of that, not to mention the cost of naively scaling up. My guess is that the ultimate winner will be the one that figures out how to improve without massive training costs, through better algorithms, or maybe even just better hardware (i.e. neuristors). I mean, we know that at worst case, we should be able to build something with human level intelligence that takes about 20 watts to run, and is about the size of a human head, and you only need to ingest a small slice of all available information to do that. And training should only use about 3.5 MWh, total, and can be done with the same hardware that runs the model.