Most active commenters
  • frozenseven(10)
  • johnnyanmac(10)
  • eric-burel(7)
  • dvfjsdhgfv(6)
  • amanaplanacanal(6)
  • SJC_Hacker(6)
  • thewebguyd(5)
  • jsnell(5)
  • __loam(4)
  • Earw0rm(4)

←back to thread

LLM Inevitabilism

(tomrenner.com)
1655 points SwoopsFromAbove | 409 comments | | HN request time: 3.338s | source | bottom
1. lsy ◴[] No.44568114[source]
I think two things can be true simultaneously:

1. LLMs are a new technology and it's hard to put the genie back in the bottle with that. It's difficult to imagine a future where they don't continue to exist in some form, with all the timesaving benefits and social issues that come with them.

2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them, the majority of consumer usage is at the free tier, the industry is seeing the first signs of pulling back investments, and model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return (the supersonic jetliner), and several that seemed poised to displace both old tech and labor but have settled into specific use cases (the microwave oven). Given the lack of a sufficiently profitable business model, it feels as likely as not that LLMs settle somewhere a little less remarkable, and hopefully less annoying, than today's almost universally disliked attempts to cram it everywhere.

replies(26): >>44568145 #>>44568416 #>>44568799 #>>44569151 #>>44569734 #>>44570520 #>>44570663 #>>44570711 #>>44570870 #>>44571050 #>>44571189 #>>44571513 #>>44571570 #>>44572142 #>>44572326 #>>44572360 #>>44572627 #>>44572898 #>>44573137 #>>44573370 #>>44573406 #>>44574774 #>>44575820 #>>44577486 #>>44577751 #>>44577911 #
2. fendy3002 ◴[] No.44568145[source]
LLMs need significant optimization or we get significant improvement on computing power while keeping the energy cost the same. It's similar with smartphone, when at the start it's not feasible because of computing power, and now we have one that can rival 2000s notebooks.

LLMs is too trivial to be expensive

EDIT: I presented the statement wrongly. What I mean is the use case for LLM are trivial things, it shouldn't be expensive to operate

replies(4): >>44568305 #>>44568319 #>>44568610 #>>44570702 #
3. trashchomper ◴[] No.44568305[source]
Calling LLMs trivial is a new one. Yea just consume all of the information on the internet and encode it into a statistical model, trivial, child could do it /s
replies(2): >>44568630 #>>44568725 #
4. lblume ◴[] No.44568319[source]
Imagine telling a person from five years ago that the programs that would basically solve NLP, perform better than experts at many tasks and are hard not to anthropomorphize accidentally are actually "trivial". Good luck with that.
replies(4): >>44568547 #>>44568655 #>>44569792 #>>44575415 #
5. eric-burel ◴[] No.44568416[source]
Developers haven't even started extracting the value of LLMs with agent architectures yet. Using an LLM UI like open ai is like we just figured fire and you use it to warm you hands (still impressive when you think about it, but not worth the burns), while LLM development is about building car engines (here is you return on investment).
replies(8): >>44568647 #>>44568953 #>>44568969 #>>44569090 #>>44569602 #>>44569667 #>>44570029 #>>44570985 #
6. jrflowers ◴[] No.44568547{3}[source]
>programs that would basically solve NLP

There is a load-bearing “basically” in this statement about the chat bots that just told me that the number of dogs granted forklift certification in 2023 is 8,472.

replies(1): >>44568800 #
7. killerstorm ◴[] No.44568610[source]
LLM can give you thousands of lines of perfectly working code for less than 1 dollar. How is that trivial or expensive?
replies(3): >>44568731 #>>44568734 #>>44569075 #
8. hammyhavoc ◴[] No.44568630{3}[source]
> all of the information on the internet

Total exaggeration—especially given Cloudflare providing free tools to block AI and now tools to charge bots for access to information.

9. clarinificator ◴[] No.44568647[source]
Every booster argument is like this one. $trite_analogy triumphant smile
replies(1): >>44568826 #
10. clarinificator ◴[] No.44568655{3}[source]
Yeah it solved NLP about 50% of the time, and also mangles data badly and in often hard-to-detect ways.
11. fendy3002 ◴[] No.44568725{3}[source]
well I presented the statement wrongly. What I mean is the use case for LLM are trivial things, it shouldn't be expensive to operate
12. zwnow ◴[] No.44568731{3}[source]
Thousands of lines of perfectly working code? Did you verify that yourself? Last time I tried it produced slop, and I've been extremely detailed in my prompt.
replies(2): >>44573478 #>>44575728 #
13. fendy3002 ◴[] No.44568734{3}[source]
well I presented the statement wrongly. What I mean is the use case for LLM are trivial things, it shouldn't be expensive to operate

and the 1 dollar cost for your case is heavily subsidized, that price won't hold up long assuming the computing power stays the same.

replies(1): >>44570666 #
14. Msurrow ◴[] No.44568799[source]
> first signs of pulling back investments

I agree with you, but I’m curious; do you have link to one or two concrete examples of companies pulling back investments, or rolling back an AI push?

(Yes it’s just to fuel my confirmation bias, but it’s still feels nice:-) )

replies(1): >>44569183 #
15. lblume ◴[] No.44568800{4}[source]
Sure, maybe solving NLP is too great a claim to make. It is still not at all ordinary that beforehand we could not solve referential questions algorithmically, that we could not extract information from plain text into custom schemas of structured data, and context-aware mechanical translation was really unheard of. Nowadays LLMs can do most of these tasks better than most humans in most scenarios. Many NLP questions at least I find interesting reduce to questions of the explanability of LLMs.
16. Jensson ◴[] No.44568953[source]
> Developers haven't even started extracting the value of LLMs with agent architectures yet

There are thousands of startups doing exactly that right now, why do you think this will work when all evidence points towards it not working? Or why else would it not already have revolutionized everything a year or two ago when everyone started doing this?

replies(2): >>44569274 #>>44571260 #
17. pydry ◴[] No.44568969[source]
Theyre doing it so much it's practically a cliche.

There are underserved areas of the economy but agentic startups is not one.

18. sgt101 ◴[] No.44569075{3}[source]
Looking up a project on github, downloading it and using it can give you 10000 lines of perfectly working code for free.

Also, when I use Cursor I have to watch it like a hawk or it deletes random bits of code that are needed or adds in extra code to repair imaginary issues. A good example was that I used it to write a function that inverted the axis on some data that I wanted to present differently, and then added that call into one of the functions generating the data I needed.

Of course, somewhere in the pipeline it added the call into every data generating function. Cue a very confused 20 minutes a week later when I was re-running some experiments.

replies(1): >>44569459 #
19. mns ◴[] No.44569090[source]
>evelopers haven't even started extracting the value of LLMs with agent architectures yet.

Which is basically what? The infinite monkey theorem? Brute forcing solutions for problems at huge costs? Somehow people have been tricked to actually embrace and accept that now they have to pay subscriptions from 20$ to 300$ to freaking code? How insane is that, something that was a very low entry point and something that anyone could do, is now being turned into some sort of classist system where the future of code is subscriptions you pay for companies ran by sociopaths who don't care that the world burns around them, as long as their pockets are full.

replies(3): >>44569263 #>>44569305 #>>44570800 #
20. moffkalast ◴[] No.44569151[source]
ML models have the good property of only requiring investment once and can then be used till the end of history or until something better replaces them.

Granted the initial investment is immense, and the results are not guaranteed which makes it risky, but it's like building a dam or a bridge. Being in the age where bridge technology evolves massively on a weekly basis is a recipe for being wasteful if you keep starting a new megaproject every other month though. The R&D phase for just about anything always results in a lot of waste. The Apollo programme wasn't profitable either, but without it we wouldn't have the knowledge for modern launch vehicles to be either. Or to even exist.

I'm pretty sure one day we'll have an LLM/LMM/VLA/etc. that's so good that pretraining a new one will seem pointless, and that'll finally be the time we get to (as a society) reap the benefits of our collective investment in the tech. The profitability of a single technology demonstrator model (which is what all current models are) is immaterial from that standpoint.

replies(1): >>44569360 #
21. 0xAFFFF ◴[] No.44569183[source]
Most prominent example was this one: https://www.reuters.com/technology/microsoft-pulls-back-more...
replies(1): >>44570468 #
22. rrr_oh_man ◴[] No.44569198{4}[source]
Throwing genetic fallacies around instead of engaging with the comment at hand... :)
23. frizlab ◴[] No.44569263{3}[source]
I cannot emphasize how much I agree with this comment. Thank you for writing it, I would never have had written it as well.
24. eric-burel ◴[] No.44569274{3}[source]
Most of them are a bunch of prompts and don't even have actual developers. For the good reason that there is no training system yet and the wording of how you call the people that build these system isn't even there or clearly defined. Local companies haven't even setup a proper internal LLM or at least a contract with a provider. I am in France so probably lagging behind USA a bit especially NY/SF but the word "LLM developer" is just arriving now and mostly under the pressure of isolated developers and companies like me. This feel really really early stage.
replies(3): >>44569675 #>>44569759 #>>44570203 #
25. eric-burel ◴[] No.44569305{3}[source]
I don't have a subscription not even an Open AI account (mostly cause they messed up their google account system). You can't extract value of an LLM by just using the official UI, you just scratch the surface of how they work. And yet there aren't much developers able to actually build an actual agent architecture that does deliver some value. I don't include the "thousands" of startups that are clearly suffer from a signaling bias: they don't exist in the economy and I don't care about them like at all in my reasonning. I am talking about actual LLM developers that you can recruit locally the same way you recruit a web developer today, and that can make sense out of "frontier" LLM garbage talk by using proper architectures. These devs are not there yet.
26. wincy ◴[] No.44569360[source]
Nah, if TSMC got exploded and there was a world war, in 20 years all the LLMs would bit rot.
replies(1): >>44569507 #
27. brulard ◴[] No.44569459{4}[source]
Are you seriously comparing downloading static code from github with bespoke code generated for your specific problem? LLMs don't keep you from coding, they assist it. Sometimes the output works, sometimes it doesn't (on first or multiple tries). Dismissing the entire approach because it's not perfect yet is shortsighted.
replies(1): >>44569848 #
28. moffkalast ◴[] No.44569507{3}[source]
Eh, I doubt it, tech only got massively better in each world war so far, through unlimited reckless strategic spending. We'd probably get a TSMC-like fab on every continent by the end of it. Maybe even optical computers. Quadrotor UAV are the future of warfare after all, and they require lots of compute.

Adjusted for inflation it took over 120 billion to build the fleet of liberty ships during WW2, that's like at least 10 TSMC fabs.

replies(1): >>44572240 #
29. camillomiller ◴[] No.44569602[source]
>> Developers haven't even started extracting the value of LLMs with agent architectures yet.

What does this EVEN mean? Do words have any value still, or are we all just starting to treat them as the byproduct of probabilistic tokens?

"Agent architectures". Last time I checked an architecture needs predictability and constraints. Even in software engineering, a field for which the word "engineering" is already quite a stretch in comparison to construction, electronics, mechanics.

Yet we just spew the non-speak "Agentic architectures" as if the innate inability of LLMs in managing predictable quantitative operations is not an unsolved issue. As if putting more and more of these things together automagically will solves their fundamental and existential issue (hallucinations) and suddenly makes them viable for unchecked and automated integration.

replies(2): >>44571385 #>>44578138 #
30. __loam ◴[] No.44569667[source]
3 years into automating all white collar labor in 6 months.
replies(2): >>44571334 #>>44581011 #
31. __loam ◴[] No.44569675{4}[source]
The smartest and most well funded people on the planet have been trying and failing to get value out of this technology for years and the best we've come up with so far is some statistically unreliable coding assistants. Hardly the revolution its proponents keep eagerly insisting we're seeing.
replies(3): >>44570282 #>>44571073 #>>44571482 #
32. erlend_sh ◴[] No.44569734[source]
Exactly. This is basically the argument of “AI as Normal Technology”.

https://knightcolumbia.org/content/ai-as-normal-technology

https://news.ycombinator.com/item?id=43697717

replies(2): >>44569958 #>>44572137 #
33. aquariusDue ◴[] No.44569759{4}[source]
Between the ridiculously optimistic and the cynically nihilistic I personally believe there is some value that extremely talented people at huge companies can't really provide because they're not in the right environment (too big a scale) but neither can grifters packaging a prompt in a vibecoded app.

In the last few months the building blocks for something useful for small companies (think less than 100 employees) have appeared, now it's time for developers or catch-all IT at those companies and freelancers serving small local companies to "up-skill".

Why do I believe this? Well for a start OCR became much more accessible this year cutting down on manual data entry compared to tesseract of yesteryear.

34. Applejinx ◴[] No.44569792{3}[source]
"hard not to anthropomorphize accidentally' is a you problem.

I'm unhappy every time I look in my inbox, as it's a constant reminder there are people (increasingly, scripts and LLMs!) prepared to straight-up lie to me if it means they can take my money or get me to click on a link that's a trap.

Are you anthropomorphizing that, too? You're not gonna last a day.

replies(1): >>44570331 #
35. ozgrakkurt ◴[] No.44569848{5}[source]
They didn’t dismiss it, they just said it is not really that useful which is correct?
replies(2): >>44572257 #>>44575667 #
36. highfrequency ◴[] No.44569958[source]
Thanks for the link. The comparison to electricity is a good one, and this is a nice reflection on why it took time for electricity’s usefulness to show up in productivity stats:

> What eventually allowed gains to be realized was redesigning the entire layout of factories around the logic of production lines. In addition to changes to factory architecture, diffusion also required changes to workplace organization and process control, which could only be developed through experimentation across industries.

37. ◴[] No.44570029[source]
38. liveoneggs ◴[] No.44570203{4}[source]
is there a non-prompt way to interact with LLMs?
replies(1): >>44570271 #
39. eric-burel ◴[] No.44570271{5}[source]
In an agentic setup the value is half the prompts half how you plug them together. I am opposing for instance a big prompt that is supposed to write a dissertation vs a smart web scraper that builds a knowledge graph out of sources and outputs a specialized search engine for your task. The former is a free funny intern, the latter is growth percentage visible in the economy.
40. eric-burel ◴[] No.44570282{5}[source]
They try to get value at their scale, which is tough. Your local SME definitely sees value in an embedding-based semantic search engine over their 20 years of weird unstructured data.
41. lblume ◴[] No.44570331{4}[source]
I didn't mean typical chatbot output, these are luckily still fairly recognizable due to stylistic preferences learned during fine-tuning. I mean actual base model output. Take a SOTA base model and give it the first two paragraphs of some longer text you wrote, and I would bet on many people being unable to distinguish your continuation from the model's autoregressive guesses.
42. durumu ◴[] No.44570468{3}[source]
I think that's more reflective of the deteriorating relationship between OpenAI and Microsoft than an true lack of demand for datacenters. If a major model provider (OpenAI, Anthropic, Google, xAI) were to see a dip in available funding or stop focusing on training more powerful models, that would convince me we may be in a bubble about to pop, but there are no signs of that as far as I can see.
43. ludicrousdispla ◴[] No.44570520[source]
>> There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return

120+ Cable TV channels must have seemed like a good idea at the time, but like LLMs the vast majority of the content was not something people were interested in.

44. strangescript ◴[] No.44570663[source]
I think the difference between all previous technologies is scope. If you make a super sonic jet that gets people from place A to place B faster for more money, but the target consumer is like "yeah, I don't care that much about that at that price point", then your tech sort is of dead. You are also fully innovated on that product, like maybe you can make it more fuel efficient, sure, but your scope is narrow.

AI is the opposite. There are numerous things it can do and numerous ways to improve it (currently). There is lower upfront investment than say a supersonic jet and many more ways it can pivot if something doesn't work out.

replies(3): >>44570929 #>>44571877 #>>44572175 #
45. killerstorm ◴[] No.44570666{4}[source]
Cheaper models might be around $0.01 per request, and it's not subsidized: we see a lot of different providers offering open source models, which offer quality similar to proprietary ones. On-device generation is also an option now.

For $1 I'm talking about Claude Opus 4. I doubt it's subsidized - it's already much more expensive than the open models.

46. jsnell ◴[] No.44570702[source]
But the thing is, LLMs are already incredibly cheap to operate compared to the alternatives. Both for trivial things and for complex things.
replies(1): >>44570871 #
47. alonsonic ◴[] No.44570711[source]
I'm confused with your second point. LLM companies are not making any money from current models? Openai generates 10b USD ARR and has 100M MAUs. Yes they are running at a loss right now but that's because they are racing to improve models. If they stopped today to focus on optimization of their current models to minimize operating cost and monetizing their massive user base you think they don't have a successful business model? People use this tools daily, this is inevitable.
replies(11): >>44570725 #>>44570756 #>>44570760 #>>44570772 #>>44570780 #>>44570853 #>>44570896 #>>44570964 #>>44571007 #>>44571541 #>>44571655 #
48. airstrike ◴[] No.44570725[source]
No, because if they stop to focus on optimizing and minimizing operating costs, the next competitor over will leapfrog them with a better model in 6-12 months, making all those margin improvements an NPV negative endeavor.
49. bbor ◴[] No.44570756[source]
It’s just the natural counterpart to dogmatic inevitabilism — dogmatic denialism. One denies the present, the other the (recent) past. It’s honestly an understandable PoV though when you consider A) most people understand “AI” and “chatbot” to be synonyms, and B) the blockchain hype cycle(s) bred some deep cynicism about software innovation.

Funny seeing that comment on this post in particular, tho. When OP says “I’m not sure it’s a world I want”, I really don’t think they’re thinking about corporate revenue opportunities… More like Rehoboam, if not Skynet.

replies(1): >>44570924 #
50. mc32 ◴[] No.44570760[source]
Making money and operating at a loss contradict each other. Maybe someday they’ll make money —but not just yet. As many have said they’re hoping capturing market will position them nicely once things settle. Obviously we’re not there yet.
replies(1): >>44570844 #
51. BolexNOLA ◴[] No.44570772[source]
> that's because they are racing improve models. If they stopped today to focus on optimization of their current models to minimize operating cost and monetizing their user base you think they don't have a successful business model?

I imagine they would’ve flicked that switch if they thought it would generate a profit, but as it is it seems like all AI companies are still happy to burn investor money trying to improve their models while I guess waiting for everyone else to stop first.

I also imagine it’s hard to go to investors with “while all of our competitors are improving their models and either closing the gap or surpassing us, we’re just going to stabilize and see if people will pay for our current product.”

replies(1): >>44574022 #
52. ◴[] No.44570780[source]
53. pj_mukh ◴[] No.44570800{3}[source]
I pay $300 to fly from SF to LA when I could've just walked for free. Its true. How classist!
54. colinmorelli ◴[] No.44570844{3}[source]
It is absolutely possible for the unit economics of a product to be profitable and for the parent company to be losing money. In fact, it's extremely common when the company is bullish on their own future and thus they invest heavily in marketing and R&D to continue their growth. This is what I understood GP to mean.

Whether it's true for any of the mainstream LLM companies or not is anyone's guess, since their financials are either private or don't separate out LLM inference as a line item.

55. lordnacho ◴[] No.44570853[source]
Are you saying they'd be profitable if they didn't pour all the winnings into research?

From where I'm standing, the models are useful as is. If Claude stopped improving today, I would still find use for it. Well worth 4 figures a year IMO.

replies(5): >>44570918 #>>44570925 #>>44570962 #>>44571742 #>>44572421 #
56. nyarlathotep_ ◴[] No.44570870[source]
I do wonder where in the cycle this all is given that we've now seen yet another LLM/"Agentic" VSCode fork.

I'm genuinely surprised that Code forks and LLM cli things are seemingly the only use case that's approached viability. Even a year ago, I figured there'd be something else that's emerged by now.

replies(1): >>44570990 #
57. fendy3002 ◴[] No.44570871{3}[source]
Well recently cursor got a heat for rising price and having opaque usage, while anthropic's claude reported to be worse due to optimization. IMO the current LLMs are not sustainable, and prices are expected to increase sooner or later.

Personally, until models comparable with sonnet 3.5 can be run locally on mid range setup, people need to wary that the price of LLM can skyrocket

replies(1): >>44583007 #
58. dvfjsdhgfv ◴[] No.44570896[source]
> If they stopped today to focus on optimization of their current models to minimize operating cost and monetizing their user base you think they don't have a successful business model?

Actually, I'd be very curious to know this. Because we already have a few relatively capable models that I can run on my MBP with 128 GB of RAM (and a few less capable models I can run much faster on my 5090).

In order to break even they would have to minimize the operating costs (by throttling, maiming models etc.) and/or increase prices. This would be the reality check.

But the cynic in me feels they prefer to avoid this reality check and use the tried and tested Uber model of permanent money influx with the "profitability is just around the corner" justification but at an even bigger scale.

replies(1): >>44570940 #
59. dvfjsdhgfv ◴[] No.44570918{3}[source]
For me, if Anthropic stopped now, and given access to all alternative models, they still would be worth exactly $240 which is the amount I'm paying now. I guess Anthropic and OpenAI can see the real demand by clearly seeing what are their free:basic:expensive plan ratios.
replies(1): >>44574494 #
60. dvfjsdhgfv ◴[] No.44570924{3}[source]
> most people understand “AI” and “chatbot” to be synonyms

This might be true (or not), but for sure not on this site.

replies(1): >>44571283 #
61. apwell23 ◴[] No.44570925{3}[source]
> Well worth 4 figures a year IMO

only because software engineering pay hasn't adjusted down for the new reality . You don't know what its worth yet.

replies(2): >>44571084 #>>44574128 #
62. digianarchist ◴[] No.44570929[source]
It's not a great analogy. The only parallel with Concorde is energy consumption. I think a better analogy would have been VR.
replies(1): >>44571217 #
63. ghc ◴[] No.44570940{3}[source]
> In order to break even they would have to minimize the operating costs (by throttling, maiming models etc.) and/or increase prices. This would be the reality check.

Is that true? Are they operating inference at a loss or are they incurring losses entirely on R&D? I guess we'll probably never know, but I wouldn't take as a given that inference is operating at a loss.

I found this: https://semianalysis.com/2023/02/09/the-inference-cost-of-se...

which estimates that it costs $250M/year to operate ChatGPT. If even remotely true $10B in revenue on $250M of COGS would be a great business.

replies(1): >>44571028 #
64. jsnell ◴[] No.44570962{3}[source]
They'd be profitable if they showed ads to their free tier users. They wouldn't even need to be particularly competent at targeting or aggressive with the amount of ads they show, they'd be profitable with 1/10th the ARPU of Meta or Google.

And they would not be incompetent at targeting. If they were to use the chat history for targeting, they might have the most valuable ad targeting data sets ever built.

replies(5): >>44571061 #>>44571136 #>>44572280 #>>44572443 #>>44573390 #
65. dbalatero ◴[] No.44570964[source]
They might generate 10b ARR, but they lose a lot more than that. Their paid users are a fraction of the free riders.

https://www.wheresyoured.at/openai-is-a-systemic-risk-to-the...

replies(3): >>44571830 #>>44572286 #>>44573506 #
66. dvfjsdhgfv ◴[] No.44570985[source]
> Developers haven't even started extracting the value of LLMs with agent architectures yet.

For sure there is a portion of developers who don't care about the future, are not interested in current developements and just live as before hoping nothing will change. But the rest already gave it a try and realized tools like Claude Code can give excellent results for small codebases to fail miserably at more complex tasks with the net result being negative as you get a codebase you don't understand, with many subtle bugs and inconsistencies created over a few days you will need weeks to discover and fix.

replies(1): >>44571363 #
67. alonsonic ◴[] No.44570990[source]
But there are a ton of LLM powered products in the market.

I have a friend in finance that uses LLM powered products for financial analysis, he works in a big bank. Just now anthropic released a product to compete in this space.

Another friend in real estate uses LLM powered lead qualifications products, he runs marketing campaigns and the AI handles the initial interaction via email or phone and then ranks the lead in their crm.

I have a few friends that run small businesses and use LLM powered assistants to manage all their email comms and agendas.

I've also talked with startups in legal and marketing doing very well.

Coding is the theme that's talked about the most in HN but there are a ton of startups and big companies creating value with LLMs

replies(3): >>44572626 #>>44572675 #>>44574169 #
68. ehutch79 ◴[] No.44571007[source]
Revenue is _NOT_ Profit
replies(2): >>44571163 #>>44571412 #
69. dvfjsdhgfv ◴[] No.44571028{4}[source]
As you say, we will never know, but this article[0] claims:

> The cost of the compute to train models alone ($3 billion) obliterates the entirety of its subscription revenue, and the compute from running models ($2 billion) takes the rest, and then some. It doesn’t just cost more to run OpenAI than it makes — it costs the company a billion dollars more than the entirety of its revenue to run the software it sells before any other costs.

[0] https://www.lesswrong.com/posts/CCQsQnCMWhJcCFY9x/openai-los...

replies(2): >>44571100 #>>44571236 #
70. dcow ◴[] No.44571050[source]
The difference is that the future is now with LLMs. There is a microwave (some multiple) in almost every kitchen in the world. The Concord served a few hundred people a day. LLMs are already ingrained into hundreds of millions if not billions of people’s lives, directly and indirectly. My dad directly uses LLMs multiple times a week if not daily in an industry that still makes you rotate your password every 3 months. It’s not a question of whether the future will have them, it’s a question of whether the future will get tired of them.
replies(1): >>44572292 #
71. bugbuddy ◴[] No.44571061{4}[source]
I heard majority of the users are techies asking coding questions. What do you sell to someone asking how to fix a nested for loop in C++? I am genuinely curious. Programmers are known to be the stingiest consumers out there.
replies(9): >>44571134 #>>44571182 #>>44571264 #>>44571269 #>>44572071 #>>44572254 #>>44572375 #>>44572688 #>>44573270 #
72. liveoneggs ◴[] No.44571073{5}[source]
my company has already fired a bunch of people in favor of LLMs so they are realizing all kinds of value
replies(5): >>44571223 #>>44571241 #>>44571416 #>>44572040 #>>44574184 #
73. fkyoureadthedoc ◴[] No.44571084{4}[source]
Can you explain this in more detail? The idiot bottom rate contractors that come through my team on the regular have not been helped at all by LLMs. The competent people do get a productivity boost though.

The only way I see compensation "adjusting" because of LLMs would need them to become significantly more competent and autonomous.

replies(2): >>44571579 #>>44573279 #
74. ghc ◴[] No.44571100{5}[source]
Obviously you don't need to train new models to operate existing ones.

I think I trust the semianalysis estimate ($250M) more than this estimate ($2B), but who knows? I do see my revenue estimate was for this year, though. However, $4B revenue on $250M COGS...is still staggeringly good. No wonder amazon, google, and Microsoft are tripping over themselves to offer these models for a fee.

replies(3): >>44571326 #>>44572365 #>>44575101 #
75. LtWorf ◴[] No.44571134{5}[source]
According to fb's aggressively targeted marketing, you sell them donald trump propaganda.
replies(1): >>44571275 #
76. lxgr ◴[] No.44571136{4}[source]
Bolting banner ads onto a technology that can organically weave any concept into a trusted conversation would be incredibly crude.
replies(4): >>44571218 #>>44571487 #>>44572061 #>>44572225 #
77. throwawayoldie ◴[] No.44571163{3}[source]
And ARR is not revenue. It's "annualized recurring revenue": take one month's worth of revenue, multiply it by 12--and you get to pick which month makes the figures look most impressive.
replies(4): >>44571287 #>>44571311 #>>44571351 #>>44572679 #
78. cuchoi ◴[] No.44571182{5}[source]
I'm not sure that stereotype holds up. Developers spend a lot: courses, cloud services, APIs, plugins, even fancy keyboards.

A quick search shows that click on ads targeting developers are expensive.

Also there is a ton of users asking to rewrite emails, create business plans, translate, etc.

79. smrtinsert ◴[] No.44571189[source]
They didn't really need the cloud either and yet...
80. strangescript ◴[] No.44571217{3}[source]
I mean, thats the point, they aren't the same. Concorde was one dimensional, AI is not.
81. nacnud ◴[] No.44571218{5}[source]
True - but if you erode that trust then your users may go elsewhere. If you keep the ads visually separated, there's a respected boundary & users may accept it.
replies(2): >>44571708 #>>44572379 #
82. dasil003 ◴[] No.44571223{6}[source]
I don’t know your company but this thinking doesn’t necessarily follow logically. In a large company the value of developers is not distributed evenly across people and time, and also has a strong dependency on market realities in front of them.

While it’s true that lots of companies are getting some value out of LLMs, a much larger number are using them as an excuse for layoffs they would have wanted to do anyway—LLMs are just a golden opportunity to tie in an unmitigated success narrative.

replies(1): >>44574544 #
83. matwood ◴[] No.44571236{5}[source]
CapEx vs. OpEx.

If they stop training today what happens? Does training always have to be at these same levels or will it level off? Is training fixed? IE, you can add 10x the subs and training costs stay static.

IMO, there is a great business in there, but the market will likely shrink to ~2 players. ChatGPT has a huge lead and is already Kleenex/Google of the LLMs. I think the battle is really for second place and that is likely dictated by who runs out of runway first. I would say that Google has the inside track, but they are so bad at product they may fumble. Makes me wonder sometimes how Google ever became a product and verb.

replies(1): >>44572608 #
84. Capricorn2481 ◴[] No.44571241{6}[source]
Only as much as replacing all your devs with a frog is "realizing value"
85. ReptileMan ◴[] No.44571260{3}[source]
>Or why else would it not already have revolutionized everything a year or two ago when everyone started doing this?

The internet needed 20 years to take over the world. All of the companies of the first dot com bust are in the past. The tech is solid.

86. disgruntledphd2 ◴[] No.44571264{5}[source]
You'd probably do brand marketing for Stripe, Datadog, Kafka, Elastic Search etc.

You could even loudly proclaim that the are ads are not targeted by users which HN would love (but really it would just be old school brand marketing).

87. Lewton ◴[] No.44571269{5}[source]
> I heard majority of the users are techies asking coding questions.

Citation needed? I can't sit on a bus without spotting some young person using ChatGPT

88. disgruntledphd2 ◴[] No.44571275{6}[source]
It's very important to note that advertisers set the parameters in which FB/Google's algorithms and systems operate. If you're 25-55 in a red state, it seems likely that you'll see a bunch of that information (even if FB are well aware you won't click).
replies(1): >>44573190 #
89. bbor ◴[] No.44571283{4}[source]
I mean...

  LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them,
The only way one could say such a thing is if they think chatbots are the only real application.
90. UK-Al05 ◴[] No.44571287{4}[source]
That's still not profit.
replies(1): >>44571443 #
91. jdiff ◴[] No.44571311{4}[source]
Astonishing that that concept survived getting laughed out of the room long enough to actually become established as a term and an acronym.
replies(3): >>44571567 #>>44571756 #>>44572575 #
92. hamburga ◴[] No.44571326{6}[source]
But assuming no new models are trained, this competitive effect drives down the profit margin on the current SOTA models to zero.
replies(1): >>44571604 #
93. erfgh ◴[] No.44571334{3}[source]
It's 3 years away.
94. airstrike ◴[] No.44571351{4}[source]
You don't get to pick the month. At least not with any half-serious audience.
replies(2): >>44571452 #>>44572405 #
95. eric-burel ◴[] No.44571363{3}[source]
This is a bit developer centric, I am much more impressed by the opportunities I see in consulting rather than applying LLMs to dev tasks. And I am still impressed by the code it can output eventhough we are still in the funny intern stage in this area.
replies(1): >>44573813 #
96. eric-burel ◴[] No.44571385{3}[source]
This means I believe we currently underuse LLM capabilities and their empirical nature makes it difficult to assess their limitations without trying. I've been studying LLMs from various angles during a few months before coming to this conclusion, as an experienced software engineer and consultant. I must admit it is however biased towards my experience as an SME and in my local ecosystem.
97. vuggamie ◴[] No.44571412{3}[source]
It's a good point. Any business can get revenue by selling Twenty dollar bills for $19. But in the history of tech, many winners have been dismissed for lack of an apparent business model. Amazon went years losing money, and when the business stabilized, went years re-investing and never showed a profit. Analysts complained as Amazon expanded into non-retail activities. And then there's Uber.

The money is there. Investors believe this is the next big thing, and is a once in a lifetime opportunity. Bigger than the social media boom which made a bunch of billionaires, bigger than the dot com boom, bigger maybe than the invention of the microchip itself.

It's going to be years before any of these companies care about profit. Ad revenue is unlikely to fund the engineering and research they need. So the only question is, does the investor money dry up? I don't think so. Investor money will be chasing AGI until we get it or there's another AI winter.

98. SketchySeaBeast ◴[] No.44571416{6}[source]
I imagine they HOPE they'll realize value. A lot of people are acting on what might be, rather than what is, which makes sense given that the AI "thought leaders" (CEOs with billions invested that need to start turning a profit) are all promising great things soon™.
99. throwawayoldie ◴[] No.44571443{5}[source]
I know. It's a doubly-dubious figure.
100. throwawayoldie ◴[] No.44571452{5}[source]
We're not talking about a half-serious audience: we're talking about the collection of reposters of press releases we call "the media".
101. SpicyLemonZest ◴[] No.44571482{5}[source]
The best they've come up with is the LLM chatbot, which both OpenAI and Anthropic have as their flagship product because many people find it extremely valuable. Many people I know routinely use ChatGPT to help them write things, even those who were already good at writing, and if you don't think that's true at your workplace I strongly suspect it's because people aren't telling you about it.
replies(1): >>44574220 #
102. Analemma_ ◴[] No.44571487{5}[source]
Like that’s ever stopped the adtech industry before.

It would be a hilarious outcome though, “we built machine gods, and the main thing we use them for is to make people click ads.” What a perfect Silicon Valley apotheosis.

103. JimmaDaRustla ◴[] No.44571513[source]
Investments are mostly in model training. We have trained models now, we'll see a pullback in that regard as businesses will need to optimize to get the best model without spending billions in order to compete on price, but LLMs are here to stay.
104. 827a ◴[] No.44571541[source]
One thing we're seeing in the software engineering agent space right now is how many people are angry with Cursor [1], and now Claude Code [2] (just picked a couple examples; you can browse around these subreddits and see tons of complaints).

What's happening here is pretty clear to me: Its a form of enshittification. These companies are struggling to find a price point that supports both broad market adoption ($20? $30?) and the intelligence/scale to deliver good results ($200? $300?). So, they're nerfing cheap plans, prioritizing expensive ones, and pissing off customers in the process. Cursor even had to apologize for it [3].

There's a broad sense in the LLM industry right now that if we can't get to "it" (AGI, etc) by the end of this decade, it won't happen during this "AI Summer". The reason for that is two-fold: Intelligence scaling is logarithmic w.r.t compute. We simply cannot scale compute quick enough. And, interest in funding to pay for that exponential compute need will dry up, and previous super-cycles tell us that will happen on the order of ~5 years.

So here's my thesis: We have a deadline that even evangelists agree is a deadline. I would argue that we're further along in this supercycle than many people realize, because these companies have already reached the early enshitification phase for some niche use-cases (software development). We're also seeing Grok 4 Heavy release with a 50% price increase ($300/mo) yet offer single-digit percent improvement in capability. This is hallmark enshitification.

Enshitification is the final, terminal phase of hyperscale technology companies. Companies remain in that phase potentially forever, but its not a phase where significant research, innovation, and optimization can happen; instead, it is a phase of extraction. AI hyperscalers genuinely speedran this cycle thanks to their incredible funding and costs; but they're now showcasing very early signals of enshitifications.

(Google might actually escape this enshitification supercycle, to be clear, and that's why I'm so bullish on them and them alone. Their deep, multi-decade investment into TPUs, Cloud Infra, and high margin product deployments of AI might help them escape it).

[1] https://www.reddit.com/r/cursor/comments/1m0i6o3/cursor_qual...

[2] https://www.reddit.com/r/ClaudeAI/comments/1lzuy0j/claude_co...

[3] https://techcrunch.com/2025/07/07/cursor-apologizes-for-uncl...

105. eddythompson80 ◴[] No.44571567{5}[source]
It’s a KPI just like any KPI and it’s gamed. A lot of random financial metrics are like that. They were invented or coined as a short hand for something.

Different investors use different ratios and numbers (ARR, P/E, EV/EBITDA, etc) as a quick initial smoke screen. They mean different things in different industries during different times of a business’ lifecycle. BUT they are supposed to help you get a starting point to reduce noise. Not as a the 1 metric you base your investing strategy on.

replies(1): >>44572237 #
106. dmix ◴[] No.44571570[source]
> model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

What are you basing this on? Personal feelings?

107. lelanthran ◴[] No.44571579{5}[source]
> Can you explain this in more detail?

Not sure what GP meant specifically, but to me, if $200/m gets you a decent programmer, then $200/m is the new going rate for a programmer.

Sure, now it's all fun and games as the market hasn't adjusted yet, but if it really is true that for $200/m you can 10x your revenue, it's still only going to be true until the market adjusts!

> The competent people do get a productivity boost though.

And they are not likely to remain competent if they are all doing 80% review, 15% prompting and 5% coding. If they keep the ratios at, for example, 25% review, 5% prompting and the rest coding, then sure, they'll remain productive.

OTOH, the pipeline for juniors now seems to be irrevocably broken: the only way forward is to improve the LLM coding capabilities to the point that, when the current crop of knowledgeable people have retired, programmers are not required.

Otherwise, when the current crop of coders who have the experience retires, there'll be no experience in the pipeline to take their place.

If the new norm is "$200/m gets you a programmer", then that is exactly the labour rate for programming: $200/m. These were previously (at least) $5k/m jobs. They are now $200/m jobs.

replies(2): >>44571897 #>>44573056 #
108. ghc ◴[] No.44571604{7}[source]
Even if the profit margin is driven to zero, that does not mean competitors will cease to offer the models. It just means the models will be bundled with other services. Case in point: Subversion & Git drove VCS margin to zero (remember BitKeeper?), but Bitbucket and Github wound up becoming good businesses. I think Claude Code might be the start of how companies evolve here.
109. dkdbejwi383 ◴[] No.44571655[source]
How many of those MAUs are crappy startups building a janky layer on top of the OpenAI API which will cease to exist in 2 years?
replies(1): >>44575036 #
110. calvinmorrison ◴[] No.44571708{6}[source]
google did it. LLms are the new google search. It'll happen sooner or later.
replies(1): >>44572195 #
111. vikramkr ◴[] No.44571742{3}[source]
That's calculating value against not having LLMs and current competitors. If they stopped improving but their competitors didn't, then the question would be the incremental cost of Claude (financial, adjusted for switching costs, etc) against the incremental advantage against the next best competitor that did continue improving. Lock in is going to be hard to accomplish around a product that has success defined by its generalizability and adaptability.

Basically, they can stop investing in research either when 1) the tech matures and everyone is out of ideas or 2) they have monopoly power from either market power or oracle style enterprise lock in or something. Otherwise they'll fall behind and you won't have any reason to pay for it anymore. Fun thing about "perfect" competition is that everyone competes their profits to zero

112. singron ◴[] No.44571756{5}[source]
So the "multiply by 12" thing is a slight corruption of ARR, which should be based on recurring revenue (i.e. subscriptions). Subscriptions are harder to game by e.g. channel-stuffing and should be much more stable than non-recurring revenue.

To steelman the original concept, annual revenue isn't a great measure for a young fast-growing company since you are averaging all the months of the last year, many of which aren't indicative of the trajectory of the company. E.g. if a company only had revenue the last 3 months, annual revenue is a bad measure. So you use MRR to get a better notion of instantaneous revenue, but you need to annualize it to make it a useful comparison (e.g. to compute a P/E ratio), so you use ARR.

Private investors will of course demand more detailed numbers like churn and an exact breakdown of "recurring" revenue. The real issue is that these aren't public companies, and so they have no obligation to report anything to the public, and their PR team carefully selects a couple nice sounding numbers.

113. Cthulhu_ ◴[] No.44571830{3}[source]
That's fixable, a gradual adjusting of the free tier will happen soon enough once they stop pumping money into it. Part of this is also a war of attrition though, who has the most money to keep a free tier the longest and attract the most people. Very familiar strategy for companies trying to gain market share.
replies(4): >>44572182 #>>44572199 #>>44572277 #>>44572372 #
114. davidcbc ◴[] No.44571877[source]
The number of things it can actually do is significantly lower than the number of things the hype men are claiming it can do.
115. fkyoureadthedoc ◴[] No.44571897{6}[source]
$200 does not get you a decent programmer though. It needs constant prompting, babysitting, feedback, iteration. It's just a tool. It massively boosts productivity in many cases, yes. But it doesn't do your job for you. And I'm very bullish on LLM assisted coding when compared to most of HN.

High level languages also massively boosted productivity, but we didn't see salaries collapse from that.

> And they are not likely to remain competent if they are all doing 80% review, 15% prompting and 5% coding.

I've been doing 80% review and design for years, it's called not being a mid or junior level developer.

> OTOH, the pipeline for juniors now seems to be irrevocably broken

I constantly get junior developers handed to me from "strategic partners", they are just disguised as senior developers. I'm telling you brother, the LLMs aren't helping these guys do the job. I've let go 3 of them in July alone.

replies(3): >>44572544 #>>44572766 #>>44574269 #
116. sensanaty ◴[] No.44572040{6}[source]
So has mine, and quite predictably our product has gone into the shitter and breaks constantly, requiring reverts almost daily. They've armed a couple of Juniors with Cursor and given them the workload of all those people they fired / have quit since the firings, some of which have been at the company for years and held a lot of institutional knowledge that is now biting them in the ass.

Now sure, "Just don't fire the useful people and get rid of the juniors and supercharge the good devs with AI tooling" or whatever, except the whole reason the C-level is obsessed with this AI shit is because they're sold on the idea of replacing their most expensive asset, devs, because they've been told by people who sell AI as a job that it can replace those pesky expensive devs and be replaced by any random person in the company prompting up a storm and vibecoding it all.

Churn rates are up, we're burning unfathomable amounts of money on the shitty AI tooling and the project has somehow regressed after we've finally managed to get a good foothold on it and start making real progress for once. Oh and the real funny part is they're starting to backpedal a bit and have tried to get some people back in.

I expect to hear a LOT more of this type of thing happening in the near future. As the idiots in charge start slowly realizing all the marketing sold to them on LinkedIn or wherever the fuck it is they get these moronic ideas from are literal, actual literal lies.

117. evilfred ◴[] No.44572061{5}[source]
how is it "trusted" when it just makes things up
replies(3): >>44572191 #>>44572215 #>>44572641 #
118. jsnell ◴[] No.44572071{5}[source]
OpenAI has half a billion active users.

You don't need every individual request to be profitable, just the aggregate. If you're doing a Google search for, like, the std::vector API reference you won't see ads. And that's probably true for something like 90% of the searches. Those searches have no commercial value, and serving results is just a cost of doing business.

By serving those unmonetizable queries the search engine is making a bet that when you need to buy a new washing machine, need a personal injury lawyer, or are researching that holiday trip to Istanbul, you'll also do those highly commercial and monetizable searches with the same search engine.

Chatbots should have exactly the same dynamics as search engines.

119. SirHumphrey ◴[] No.44572137[source]
This seems like one the only sane arguments in this whole sea of articles.
120. api ◴[] No.44572142[source]
My take since day one:

(1) Model capabilities will plateau as training data is exhausted. Some additional gains will be possible by better training, better architectures, more compute, longer context windows or "infinite" context architectures, etc., but there are limits here.

(2) Training on synthetic data beyond a very limited amount will result in overfitting because there is no new information. To some extent you could train models on each other, but that's just an indirect way to consolidate models. Beyond consolidation you'll plateau.

(3) There will be no "takeoff" scenario -- this is sci-fi (in the pejorative sense) because you can't exceed available information. There is no magic way that a brain in a vat can innovate beyond available training data. This includes for humans -- a brain in a vat would quickly go mad and then spiral into a coma-like state. The idea of AI running away is the information-theoretic equivalent of a perpetual motion machine and is impossible. Yudkowski and the rest of the people afraid of this are crackpots, and so are the hype-mongers betting on it.

So I agree that LLMs are real and useful, but the hype and bubble are starting to plateau. The bubble is predicated on the idea that you can just keep going forever.

replies(1): >>44578020 #
121. peder ◴[] No.44572175[source]
Most of the comments here feel like cope about AI TBH. There's never been an innovation like this ever, and it makes sense to get on board rather than be left behind.
replies(1): >>44572535 #
122. sc68cal ◴[] No.44572182{4}[source]
That assumes that everyone is willing to pay for it. I don't think that's an assumption that will be true.
replies(3): >>44572633 #>>44572986 #>>44573012 #
123. andrewflnr ◴[] No.44572191{6}[source]
That's a great question to ask the people who seem to trust them implicitly.
replies(1): >>44572373 #
124. ptero ◴[] No.44572195{7}[source]
Yes, but for a while google was head and shoulders above the competition. It also poured a ton of money into building non-search functionality (email, maps, etc.). And had a highly visible and, for a while, internally respected "don't be evil" corporate motto.

All of which made it much less likely that users would bolt in response to each real monetization step. This is very different to the current situation, where we have a shifting landscape with several AI companies, each with its strengths. Things can change, but it takes time for 1-2 leaders to consolidate and for the competition to die off. My 2c.

125. gmerc ◴[] No.44572199{4}[source]
Competition is almost guaranteed to drive price close to cost of delivery especially if they can't pay trump to ban open source, particularly chinese. With no ability to play the thiel monopoly playbook, their investors would never make their money back if not for government capture and sweet sweet taxpayer military contracts.
replies(1): >>44573094 #
126. dingnuts ◴[] No.44572215{6}[source]
15% of people aren't smart enough to read and follow directions explaining how to fold a trifold brochure, place it in an envelope, seal it, and address it

you think those people don't believe the magic computer when it talks?

127. ModernMech ◴[] No.44572225{5}[source]
I imagine they would be more like product placements in film and TV than banner ads. Just casually dropping a recommendation and link to Brand (TM) in a query. Like those Cerveza Cristal ads in star wars. They'll make it seem completely seamless to the original query.
replies(2): >>44573078 #>>44573705 #
128. jdiff ◴[] No.44572237{6}[source]
I understand the importance of having data, and that any measurement can be gamed, but this one seems so tailored for tailoring that I struggle to understand how it was ever a good metric.

Even being generous it seems like it'd be too noisy to even assist in informing a good decision. Don't the overwhelmingly vast majority of businesses see periodic ebbs and flows over the course of a year?

replies(1): >>44577153 #
129. aydyn ◴[] No.44572240{4}[source]
Technology is an exponential process, and the thing about exponentials is that they are chaotic. You cant use inductive reasoning vis a vis war and technology. The next big one could truly reset us to zero or worse.
replies(1): >>44572604 #
130. yamazakiwi ◴[] No.44572254{5}[source]
A lot of people use it for cooking and other categories as well.

Techies are also great for network growth and verification for other users, and act as community managers indirectly.

131. Matticus_Rex ◴[] No.44572257{6}[source]
Many obviously disagree that it's correct
132. kelseyfrog ◴[] No.44572277{4}[source]
Absolutely, free-tier AI won’t stay "free" forever. It’s only a matter of time before advertisers start paying to have their products woven into your AI conversations. It’ll creep in quietly—maybe a helpful brand suggestion, a recommended product "just for you," or a well-timed promo in a tangential conversation. Soon enough though, you’ll wonder if your LLM genuinely likes that brand of shoes, or if it's just doing its job.

But hey, why not get ahead of the curve? With BrightlyAI™, you get powerful conversational intelligence - always on, always free. Whether you're searching for new gear, planning your next trip, or just craving dinner ideas, BrightlyAI™ brings you personalized suggestions from our curated partners—so you save time, money, and effort.

Enjoy smarter conversations, seamless offers, and a world of possibilities—powered by BrightlyAI™: "Illuminate your day. Conversation, curated."

133. naravara ◴[] No.44572280{4}[source]
If interactions with your AI start sounding like your conversation partner shilling hot cocoa powder at nobody in particular those conversations are going to stop being trusted real quick. (Pop culture reference: https://youtu.be/MzKSQrhX7BM?si=piAkfkwuorldn3sb)

Which may be for the best, because people shouldn’t be implicitly trusting the bullshit engine.

134. Centigonal ◴[] No.44572286{3}[source]
This echoes a lot of the rhetoric around "but how will facebook/twitter/etc make money?" back in the mid 2000s. LLMs might shake out differently from the social web, but I don't think that speculating about the flexibility of demand curves is a particularly useful exercise in an industry where the marginal cost of inference capacity is measured in microcents per token. Plus, the question at hand is "will LLMs be relevant?" and not "will LLMs be massively profitable to model providers?"
replies(13): >>44572513 #>>44572558 #>>44572586 #>>44572813 #>>44573104 #>>44573394 #>>44573558 #>>44573961 #>>44575180 #>>44575826 #>>44577467 #>>44577474 #>>44582494 #
135. jayd16 ◴[] No.44572292[source]
The huge leap that is getting pushback is the sentiment that LLMs will consume every use case and replace human labor. I don't think many are arguing LLMs will die off entirely.
136. giancarlostoro ◴[] No.44572326[source]
> 2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them, the majority of consumer usage is at the free tier, the industry is seeing the first signs of pulling back investments, and model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

You hit the nail on why I say to much hatred from "AI Bros" as I call them, when I say it will not take off truly until it runs on your phone effortlessly, because nobody wants to foot a trillion dollar cloud bill.

Give me a fully offline LLM that fits in 2GB of VRAM and lets refine that so it can plug into external APIs and see how much farther we can take things without resorting to burning billions of dollars' worth of GPU compute. I don't care that my answer arrives instantly, if I'm doing the research myself, I want to take my time to get the correct answer anyway.

replies(2): >>44573467 #>>44573735 #
137. Jach ◴[] No.44572360[source]
I don't really buy your point 2. Just the other day Meta announced hundreds of billions of dollars investment into more AI datacenters. Companies are bringing back nuclear power plants to support this stuff. Earlier this year OpenAI and Oracle announced their $500bn AI datacenter project, but admittedly in favor of your point have run into funding snags, though that's supposedly from tariff fears with foreign investors, not lack of confidence in AI. Meta can just finance everything from their own capital and Zuck's decree, like they did with VR (and it may very well turn out similarly).

Since you brought up supersonic jetliners you're probably aware of the startup Boom in Colorado trying to bring it back. We'll see if they succeed. But yes, it would be a strange path, but a possible one, that LLMs kind of go away for a while and try to come back later.

You're going to have to cite some surveys for the "most people agree that the output is trite and unpleasant" and "almost universally disliked attempts to cram it everywhere" claims. There are some very vocal people against LLM flavors of AI, but I don't think they even represent the biggest minority, let alone a majority or near universal opinions. (I personally was bugged by earlier attempts at cramming non-LLM AI into a lot of places, e.g. Salesforce Einstein appeared I think in 2016, and that was mostly just being put off by the cutesy Einstein characterization. I generally don't have the same feelings with LLMs in particular, in some cases they're small improvements to an already annoying process, e.g. non-human customer support that was previously done by a crude chatbot front-end to an expert system or knowledge base, the LLM version of that tends to be slightly less annoying.)

138. singron ◴[] No.44572365{6}[source]
You need to train new models to advance the knowledge cutoff. You don't necessarily need to R&D new architectures, and maybe you can infuse a model with new knowledge without completely training from scratch, but if you do nothing the model will become obsolete.

Also the semianalysis estimate is from Feb 2023, which is before the release of gpt4, and it assumes 13 million DAU. ChatGPT has 800 million WAU, so that's somewhere between 115 million and 800 million DAU. E.g. if we prorate the cogs estimate for 200 DAU, then that's 15x higher or $3.75B.

replies(1): >>44573670 #
139. SJC_Hacker ◴[] No.44572372{4}[source]
I agree, its easily fixable by injecting ads into the responses for the free tier and probably eventually even the lower paid tiers to some extent
replies(1): >>44572524 #
140. handfuloflight ◴[] No.44572373{7}[source]
They aren't trusted in a vacuum. They're trusted when grounded in sources and their claims can be traced to sources. And more specifically, they're trusted to accurately represent the sources.
replies(3): >>44572554 #>>44572818 #>>44574260 #
141. naravara ◴[] No.44572375{5}[source]
The existence of the LLMs will themselves change the profile and proclivities of people we consider “programmers” in the same way the app-driven tech boom did. Programmers who came up in the early days are different from ones who came up in the days of the web are different from ones who came up in the app era.
142. SJC_Hacker ◴[] No.44572379{6}[source]
There will be a respected boundary for a time, then as advertisers find its more effective the boundaries will start to disappear
143. SJC_Hacker ◴[] No.44572405{5}[source]
> At least not with any half-serious audience.

So I guess this rules out most SV venture capital

144. miki123211 ◴[] No.44572421{3}[source]
But if Claude stopped pouring their money into research and others didn't, Claude wouldn't be useful a year from now, as you could get a better model for the same price.

This is why AI companies must lose money short term. The moment improvements plateau or the economic environment changes, everyone will cut back on research.

145. miki123211 ◴[] No.44572443{4}[source]
and they wouldn't even have to make the model say the ads. I think that's a terrible idea which would drive model performance down.

Traditional banner ads, inserted inline into the conversation based on some classifier seem a far better idea.

146. amrocha ◴[] No.44572513{4}[source]
The point is that if they’re not profitable they won’t be relevant since they’re so expensive to run.

And there was never any question as to how social media would make money, everyone knew it would be ads. LLMs can’t do ads without compromising the product.

replies(9): >>44572606 #>>44572617 #>>44572620 #>>44572951 #>>44573061 #>>44573125 #>>44575104 #>>44575452 #>>44576838 #
147. amrocha ◴[] No.44572524{5}[source]
Literally nobody would talk to a robot that spits back ads at them
replies(4): >>44572645 #>>44573292 #>>44574101 #>>44575780 #
148. Gormo ◴[] No.44572535{3}[source]
> There's never been an innovation like this ever

There have been plenty of innovations like this. In fact, much of the hype around LLMs is a rehash of the hype around "expert systems" back in the '80s. LLMs are marginally more effective than those systems, but only marginally.

replies(1): >>44581530 #
149. handfuloflight ◴[] No.44572544{7}[source]
> It needs constant prompting, babysitting, feedback, iteration.

What do you think a product manager is doing?

replies(1): >>44572886 #
150. andrewflnr ◴[] No.44572554{8}[source]
Nope, lots of idiots just take them at face value. You're still describing what rational people do, not what all actual people do.
replies(1): >>44572590 #
151. ◴[] No.44572558{4}[source]
152. marcosdumay ◴[] No.44572575{5}[source]
Just wait until companies start calculating it on future revenue from people on the trial period of subscriptions... I mean, if we aren't there already.

Any number that there isn't a law telling companies how to calculate it will always be a joke.

153. magicalist ◴[] No.44572586{4}[source]
> LLMs might shake out differently from the social web, but I don't think that speculating about the flexibility of demand curves is a particularly useful exercise in an industry where the marginal cost of inference capacity is measured in microcents per token

That we might come to companies saying "it's not worth continuing research or training new models" seems to reinforce the OP's point, not contradict it.

replies(1): >>44572756 #
154. handfuloflight ◴[] No.44572590{9}[source]
Fair enough.
155. moffkalast ◴[] No.44572604{5}[source]
Sure you can't plan for black swan events, so the only choice you have is to plan for their absence. If we all nuke ourselves tomorrow well at least we don't have to worry about anything anymore. But in case we don't, those plans will be useful.
156. tsukikage ◴[] No.44572606{5}[source]
You’re not thinking evil enough. LLMs have the potential to be much more insidious about whatever it is they are shilling. Our dystopian future will feature plausibly deniable priming.
157. marcosdumay ◴[] No.44572608{6}[source]
That paragraph is quite clear.

OpEx is larger than revenue. CapEx is also larger than the total revenue on the lifetime of a model.

158. kridsdale3 ◴[] No.44572617{5}[source]
Well, they haven't really tried yet.

The Meta app Threads had no ads for the first year, and it was wonderful. Now it does, and its attractiveness was only reduced by 1% at most. Meta is really good at knowing the balance for how much to degrade UX by having monetization. And the amount they put in is hyper profitable.

So let's see Gemini and GPT with 1% of response content being sponsored. I doubt we'll see a user exodus and if that's enough to sustain the business, we're all good.

159. Centigonal ◴[] No.44572620{5}[source]
I can run an LLM on my RTX3090 that is at least as useful to me in my daily life as an AAA game that would otherwise justify the cost of the hardware. This is today, which I suspect is in the upper part of the Kuznets curve for AI inference tech. I don't see a future where LLMs are too expensive to run (at least for some subset of valuable use cases) as likely.
replies(1): >>44573316 #
160. brokencode ◴[] No.44572627[source]
> “most people agree that the output is trite and unpleasant to consume”

That is a such a wild claim. People like the output of LLMs so much that ChatGPT is the fastest growing app ever. It and other AI apps like Perplexity are now beginning to challenge Google’s search dominance.

Sure, probably not a lot of people would go out and buy a novel or collection of poetry written by ChatGPT. But that doesn’t mean the output is unpleasant to consume. It pretty undeniably produces clear and readable summaries and explanations.

replies(13): >>44572785 #>>44572809 #>>44572887 #>>44573175 #>>44573204 #>>44573208 #>>44573215 #>>44573682 #>>44573868 #>>44574638 #>>44574872 #>>44575425 #>>44578731 #
161. Jach ◴[] No.44572626{3}[source]
Yup. Lots of products in the education space. Even doctors are using LLMs, while talking with patients. All sorts of teams are using the adjacent products for image and (increasingly) video generation. Translation freelancers have been hit somewhat hard because LLMs do "good enough" quite a bit better than old google translate.

Coding is relevant to the HN bubble, and as tech is the biggest driver of the economy it's no surprise that tech-related AI usages will also be the biggest causes of investment, but it really is used in quite a lot of places out there already that aren't coding related at all.

162. mike-cardwell ◴[] No.44572633{5}[source]
Those that aren't willing to pay for it directly, can still use it for free, but will just have to tolerate product placement.
163. tsukikage ◴[] No.44572641{6}[source]
“trusted” in computer science does not mean what it means in ordinary speech. It is what you call things you have no choice but to trust, regardless of whether that trust is deserved or not.
replies(2): >>44573318 #>>44573568 #
164. kridsdale3 ◴[] No.44572645{6}[source]
Hundreds of millions of people watch TV and listen to Radio that is at least 30% ad content per hour.
165. materiallie ◴[] No.44572675{3}[source]
It feels like there's a lot of shifting goalposts. A year ago, the hype was that knowledge work would cease to exist by 2027.

Now we are trying to hype up enhanced email autocomplete and data analysis as revolutionary?

I agree that those things are useful. But it's not really addressing the criticism. I would have zero criticisms of AI marketing if it was "hey, look at this new technology that can assist your employees and make them 20% more productive".

I think there's also a healthy dose of skepticism after the internet and social media age. Those were also society altering technologies that purported to democratize the political and economic system. I don't think those goals were accomplished, although without a doubt many workers and industries were made more productive. That effect is definitely real and I'm not denying that.

But in other areas, the last 3 decades of technological advancement have been a resounding failure. We haven't made a dent in educational outcomes or intergenerational poverty, for instance.

166. hobofan ◴[] No.44572679{4}[source]
ARR traditionally is _annual_ recurring revenue. The notion that it may be interpreted as _annualized_ and extrapolatable from MRR is a very recent development, and I doubt that most people interpret it as that.
replies(1): >>44573459 #
167. tsukikage ◴[] No.44572688{5}[source]
…for starters, you can sell them the ability to integrate your AI platform into whatever it is they are building, so you can then sell your stuff to their customers.
168. Centigonal ◴[] No.44572756{5}[source]
The point I'm making is that, even in the extreme case where we cease all additional R&D on LLMs, what has been developed up until now has a great deal of utility and transformative power, and that utility can be delivered at scale for cheap. So, even if LLMs don't become an economic boon for the companies that enable them, the transformative effect they have and will continue to have on society is inevitable.

Edit: I believe that "LLMs transforming society is inevitable" is a much more defensible assertion than any assertion about the nature of that transformation and the resulting economic winners and losers.

replies(1): >>44574055 #
169. lelanthran ◴[] No.44572766{7}[source]
> It needs constant prompting, babysitting, feedback, iteration. It's just a tool. It massively boosts productivity in many cases, yes.

It doesn't sound like you are disagreeing with me: that role you described is one of manager, not of programmer.

> High level languages also massively boosted productivity, but we didn't see salaries collapse from that.

Those high level languages still needed actual programmers. If the LLM is able to 10x the output of a single programmer because that programmer is spending all their time managing, you don't really need a programmer anymore, do you?

> I've been doing 80% review and design for years, it's called not being a mid or junior level developer.

Maybe it differs from place to place. I was a senior and a staff engineer, at various places including a FAANG. My observations were that even staff engineer level was still spending around 2 - 3 hours a day writing code. If you're 10x'ing your productivity, you almost certainly aren't spending 2 - 3 hours a day writing code.

> I constantly get junior developers handed to me from "strategic partners", they are just disguised as senior developers. I'm telling you brother, the LLMs aren't helping these guys do the job. I've let go 3 of them in July alone.

This is a bit of a non-sequitor; what does that have to do with breaking the pipeline for actual juniors?

Without juniors, we don't get seniors. Without seniors and above, who will double-check the output of the LLM?[1]

If no one is hiring juniors anymore, then the pipeline is broken. And since the market price of a programmer is going to be set at $200/m, where will you find new entrants for this market?

Hell, even mid-level programmers will exit, because when a 10-programmer team can be replaced by a 1-person manager and a $200/m coding agent, those 9 people aren't quietly going to starve while the industry needs them again. They're going to go off and find something else to do, and their skills will atrophy (just like the 1-person LLM manager skills will atrophy eventually as well).

----------------------------

[1] Recall that my first post in this thread was to say that the LLM coding agents have to get so good that programmers aren't needed anymore because we won't have programmers anymore. If they aren't that good when the current crop starts retiring then we're in for some trouble, aren't we?

replies(1): >>44573132 #
170. sejje ◴[] No.44572785[source]
Maybe he's referencing how people don't like when other humans post LLM responses in the comments.

"Here's what chatGPT said about..."

I don't like that, either.

I love the LLM for answering my own questions, though.

replies(1): >>44573615 #
171. hattmall ◴[] No.44572809[source]
I'm not really countering that ChatGPT is popular, it certainly is, but it's also sort of like "fastest growing tire brand" that came along with the adoption of vehicles. The amount of smartphone users is also growing at the fastest rate ever so whatever the new most popular app is has a good chance of being the fastest growing app ever.
replies(1): >>44574360 #
172. overfeed ◴[] No.44572813{4}[source]
> This echoes a lot of the rhetoric around "but how will facebook/twitter/etc make money?"

The answer was, and will be ads (talk about inevitability!)

Can you imagine how miserable interacting with ad-funded models will be? Not just because of the ads they spew, but also the penny-pinching on training and inference budgets, with an eye focused solely on profitability. That is what the the future holds: consolidations, little competition, and models that do the bare-minimum, trained and operated by profit-maximizing misers, and not the unlimited intelligence AGI dream they sell.

replies(2): >>44573103 #>>44573707 #
173. sheiyei ◴[] No.44572818{8}[source]
> they're trusted to accurately represent the sources.

Which is still too much trust

174. fkyoureadthedoc ◴[] No.44572886{8}[source]
Not writing and committing code with GitHub Copilot, I'll tell you that. These things need to come a _long_ way before that's a reality.
175. Wowfunhappy ◴[] No.44572887[source]
...I do wonder what percent of ChatGPT usage is just students cheating on their homework, though.
replies(1): >>44573066 #
176. xnx ◴[] No.44572898[source]
> (the supersonic jetliner) ... (the microwave oven)

But have we ever had a general purpose technology (steam engine, electricity) that failed to change society?

replies(1): >>44573438 #
177. owlninja ◴[] No.44572951{5}[source]
I was chatting with Gemini about vacation ideas and could absolutely picture a world where if it lists some hotels I might like, the businesses that bought some LLM ad space could easily show up more often than others.
178. ebiester ◴[] No.44572986{5}[source]
Consider the general research - in all, it doesn't eliminate people, but let's say it shakes out to speeding up developers 10% over all tasks. (That includes creating tickets, writing documentation, unblocking bugs, writing scripts, building proof of concepts, and more rote refactoring, but does not solve the harder problems or stop us from doing the hard work of software engineering that doesn't involve lines of code.)

That means that it's worth up to 10% of a developer's salary as a tool. And more importantly, smaller teams go faster, so it might be worth that full 10%.

Now, assume other domains end up similar - some less, some more. So, that's a large TAM.

179. LordDragonfang ◴[] No.44573012{5}[source]
It very much does not assume that, only that some fraction will have become accustomed to using it to the point of not giving it up. In fact, they could probably remain profitable without a single new customer, given the number of subscribers they already have.
180. sheiyei ◴[] No.44573056{6}[source]
Your argument requires "Claude can replace a programme" to be true. Thus, your argument is false for the foreseeable future.
181. ◴[] No.44573061{5}[source]
182. genghisjahn ◴[] No.44573066{3}[source]
Neal Stephenson has a recent post that covers some of this. Also links to teachers talking about many students just putting all their work into chatgpt and turning it in.

https://nealstephenson.substack.com/p/emerson-ai-and-the-for...

replies(1): >>44573728 #
183. thewebguyd ◴[] No.44573078{6}[source]
I just hope that if it comes to that (and I have no doubt that it will), regulation will catch up and mandate any ad/product placement is labeled as such and not just slipped in with no disclosure whatsoever. But, given that we've never regulated influencer marketing which does the same thing, nor are TV placements explicitly called out as "sponsored" I have my doubts but one can hope.
184. xedrac ◴[] No.44573094{5}[source]
> especially if they can't pay trump to ban open source?

Huh? Do you mean for official government use?

185. 6510 ◴[] No.44573103{5}[source]
I see a real window this time to sell your soul.
186. roughly ◴[] No.44573104{4}[source]
Social networks finding profitability via advertising is what created the entire problem space of social media - the algorithmic timelines, the gaming, the dopamine circus, the depression, everything negative that’s come from social media has come from the revenue model, so yes, I think it’s worth being concerned about how LLMs make money, not because I’m worried they won’t, because I’m worried they Will.
replies(3): >>44573381 #>>44575502 #>>44577204 #
187. overfeed ◴[] No.44573125{5}[source]
> LLMs can’t do ads without compromising the product.

Spoiler: they are still going to do ads, their hand will be forced.

Sooner or later, investors are going to demand returns on the massive investments, and turn off the money faucet. There'll be consolidation, wind-downs and ads everywhere.

188. fkyoureadthedoc ◴[] No.44573132{8}[source]
> And since the market price of a programmer is going to be set at $200/m

You keep saying this, but I don't see it. The current tools just can't replace developers. They can't even be used in the same way you'd use a junior developer or intern. It's more akin to going from hand tools to power tools than it is getting an apprentice. The job has not been automated and hasn't been outsourced to LLMs.

Will it be? Who knows, but in my personal opinion, it's not looking like it will any time soon. There would need to be more improvement than we've seen from day 1 of ChatGPT until now before we could even be seriously considering this.

> Those high level languages still needed actual programmers.

So does the LLM from day one until now, and for the foreseeable future.

> This is a bit of a non-sequitor; what does that have to do with breaking the pipeline for actual juniors?

Who says the pipeline is even broken by LLMs? The job market went to shit with rising interest rates before LLMs hit the scene. Nobody was hiring them anyway.

replies(1): >>44578167 #
189. UncleOxidant ◴[] No.44573137[source]
Let's not ignore the technical aspects as well: LLMs are probably a local minima that we've gotten stuck in because of their rapid rise. Other areas in AI are being starved of investment because all of the capital is pouring into LLMs. We might have been better off in the long run if LLMs hadn't been so successful so fast.
190. xnx ◴[] No.44573175[source]
> AI apps like Perplexity are now beginning to challenge Google’s search dominance

Now that is a wild claim. ChatGPT might be challenging Google's dominance, but Perplexity is nothing.

replies(1): >>44577427 #
191. LtWorf ◴[] No.44573190{7}[source]
I'm not even in USA and I've never been in USA in my entire life.
replies(1): >>44580145 #
192. underdeserver ◴[] No.44573204[source]
> That is a such a wild claim. People like the output of LLMs so much that ChatGPT is the fastest growing app ever.

The people using ChatGPT like its output enough when they're the ones reading it.

The people reading ChatGPT output that other people asked for generally don't like it. Especially if it's not disclosed up front.

replies(2): >>44573861 #>>44575445 #
193. ants_everywhere ◴[] No.44573208[source]
> That is a such a wild claim.

Some people who hate LLMs are absolutely convinced everyone else hates them. I've talked with a few of them.

I think it's a form of filter bubble.

replies(1): >>44573910 #
194. pera ◴[] No.44573215[source]
> People like the output of LLMs so much that ChatGPT is the fastest growing app ever

While people seem to love the output of their own queries they seem to hate the output of other people's queries, so maybe what people actually love is to interact with chatbots.

If people loved LLM outputs in general then Google, OpenAI and Anthropic would be in the business of producing and selling content.

replies(5): >>44573668 #>>44574425 #>>44574602 #>>44575192 #>>44576161 #
195. JackFr ◴[] No.44573270{5}[source]
You sell them Copilot. You Sell them CursorAI. You sell them Windsurf. You sell them Devin. You sell the Claude Code.

Software guys are doing much, much more than treating LLM's like an improved Stack Overflow. And a lot of them are willing to pay.

196. cgh ◴[] No.44573279{5}[source]
There's another specific class of person that seems helped by them: the paralysis by analysis programmer. I work with someone really smart who simply cannot get started when given ordinary coding tasks. She researches, reads and understands the problem inside and out but cannot start actually writing code. LLMs have pushed her past this paralysis problem and given her the inertia to continue.

On the other end, I know a guy who writes deeply proprietary embedded code that lives in EV battery controllers and he's found LLMs useless.

197. gomox ◴[] No.44573292{6}[source]
I predict this comment to enter the Dropbox/iPod hall of shame of discussion forum skeptics.
198. TeMPOraL ◴[] No.44573316{6}[source]
I don't even get where this argument comes from. Pretraining is expensive, yes, but both LoRAs in diffusion models and finetunes of transformers show us that this is not the be-all, end-all; there's plenty of work being done on extensively tuning base models for cheap.

But inference? Inference is dirt cheap and keeps getting cheaper. You can run models lagging 6-12 years on consumer hardware, and by this I don't mean absolutely top-shelf specs, but more of "oh cool, turns out the {upper-range gaming GPU/Apple Silicon machine} I bought a year ago is actually great at running local {image generation/LLM inference}!" level. This is not to say you'll be able to run o3 or Opus 4 on a laptop next year - larger and more powerful models obviously require more hardware resources. But this should anchor expectations a bit.

We're measuring inference costs in multiples of gaming GPUs, so it's not an impending ecological disaster as some would like the world to believe - especially after accounting for data centers being significantly more efficient at this, with specialized hardware, near-100% utilization, countless of optimization hacks (including some underhanded ones).

199. pegasus ◴[] No.44573318{7}[source]
For one, it's not like we're at some CS conference, so we're engaging in ordinary speech here, as far as I can tell. For two, "trusted" doesn't have just one meaning, even in the narrower context of CS.
200. philomath_mn ◴[] No.44573370[source]
> most people agree that the output is trite and unpleasant to consume

This is likely a selection bias: you only notice the obviously bad outputs. I have created plenty of outputs myself that are good/passable -- you are likely surrounded by these types of outputs without noticing.

Not a panacea, but can be useful.

201. milesvp ◴[] No.44573381{5}[source]
I think this can't be understated. It also destroyed search. I listened to a podcast a few years ago with an early googler who talked about this very precipice in early google days. They did a lot of testing, and a lot of modeling of people's valuation of search. They figured that the average person got something like $50/yr of value out of search (I can't remember the exact number, I hope I'm not off by an order of magnitude). And that was the most they could ever realistically charge. Meanwhile, advertising for just Q4 was like 10 times the value. It meant that they knew that advertising on the platform was inevitable. They also acknowledged that it would lead to the very problem that Brin and Page wrote about in their seminal paper on search.

I see LLMs inevitably leading to the same place. There will undoubtedly be advertising baked into the models. It is too strong a financial incentive. I can only hope that an open source alternative will at least allow for a hobbled version to consume.

edit: I think this was the podcast https://freakonomics.com/podcast/is-google-getting-worse/

replies(1): >>44575943 #
202. immibis ◴[] No.44573390{4}[source]
Targeted banner ads based on chat history is last-two-decades thinking. The money with LLMs will be targeted answers. Have Coca-Cola pay you a few billion dollars to reinforce the model to say "Coke" instead of "soda". Train it the best source of information about political subjects is to watch Fox News. This even works with open-source models, too!
replies(1): >>44574196 #
203. ysavir ◴[] No.44573394{4}[source]
The thing about facebook/twitter/etc was that everyone knew how they achieve lock-in and build a moat (network effect), but the question was around where to source revenue.

With LLMs, we know what the revenue source is (subscription prices and ads), but the question is about the lock-in. Once each of the AI companies stops building new iterations and just offers a consistent product, how long until someone else builds the same product but charges less for it?

What people often miss is that building the LLM is actually the easy part. The hard part is getting sufficient data on which to train the LLM, which is why most companies just put ethics aside and steal and pirate as much as they can before any regulations cuts them off (if any regulations ever even do). But that same approach means that anyone else can build an LLM and train on that data, and pricing becomes a race to the bottom, if open source models don't cut them out completely.

replies(1): >>44574537 #
204. magic_hamster ◴[] No.44573406[source]
There are pretty hidden assumption in this comment. First of all, not every business in the AI space is _training_ models, and the difference between training and inference is massive - i.e. most businesses can easily afford inference, perhaps depending on model, but they definitely can.

Another several unfounded claims were made here, but I just wanted to say LLMs with MCP are definitely good enough for almost every use case you can come up with as long as you can provide them with high quality context. LLMs are absolutely the future and they will take over massive parts of our workflow in many industries. Try MCP for yourself and see. There's just no going back.

replies(2): >>44573427 #>>44573440 #
205. dontlikeyoueith ◴[] No.44573427[source]
> I just wanted to say LLMs with MCP are definitely good enough for almost every use case you can come up with as long as you can provide them with high quality context.

This just shows you lack imagination.

I have a lot of use cases that they are not good enough for.

206. blueflow ◴[] No.44573438[source]
It wouldn't be general purpose if it fails to bring change. I'd take every previous iteration of "AI" as example, IBM Watson, that stuff
207. ramoz ◴[] No.44573440[source]
LLMs with tools*

MCP isn’t inherently special. A Claude Code with Bash() tool can do nearly anything a MCP server will give you - much more efficiently.

Computer Use agents are here and are only going to get better.

The conversation shouldn’t be about LLMs any longer. Providers will be providing agents.

replies(1): >>44575692 #
208. throwawayoldie ◴[] No.44573459{5}[source]
What does it tell you then, that the interpretation of "A" as "annualized" is the interpretation Anthropic, to name one, has chosen?
209. DSingularity ◴[] No.44573467[source]
You aren’t extrapolating enough. Nearly the entire history of computing has been one that isolates between shared computing and personal computing. Give it time. These massive cloud bills are building the case for accelerators in phones. It’s going to happen just needs time.
replies(1): >>44574143 #
210. DSingularity ◴[] No.44573478{4}[source]
Try again.
replies(1): >>44575573 #
211. jahewson ◴[] No.44573506{3}[source]
Then cut off the free riders. Problem solved overnight.
212. Wowfunhappy ◴[] No.44573558{4}[source]
> This echoes a lot of the rhetoric around "but how will facebook/twitter/etc make money?" back in the mid 2000s.

The difference is that Facebook costs virtually nothing to run, at least on a per-user basis. (Sure, if you have a billion users, all of those individual rounding errors still add up somewhat.)

By contrast, if you're spending lots of money per user... well look at what happened to MoviePass!

The counterexample here might be Youtube; when it launched, streaming video was really expensive! It still is expensive too, but clearly Google has figured out the economics.

replies(1): >>44574284 #
213. lxgr ◴[] No.44573568{7}[source]
I meant it in the ordinary speech sense (which I don't even thing contradicts the "CS sense" fwiw).

Many people have a lot of trust in anything ChatGPT tells them.

214. jack_pp ◴[] No.44573615{3}[source]
"Here's what chatGPT said about..." Is the new lmgtfy
replies(1): >>44574001 #
215. cruffle_duffle ◴[] No.44573668{3}[source]
> While people seem to love the output of their own queries they seem to hate the output of other people's queries

Listening or trying to read other peoples chats with these things is like listening to somebody describe a dream. It’s just not that interesting most of the time. It’s remarkable for the person experiencing it but it is deeply personal.

216. ghc ◴[] No.44573670{7}[source]
> You need to train new models to advance the knowledge cutoff

That's a great point, but I think it's less important now with MCP and RAG. If VC money dried up and the bubble burst, we'd still have broadly useful models that wouldn't be obsolete for years. Releasing a new model every year might be a lot cheaper if a company converts GPU opex to capex and accepts a long training time.

> Also the semianalysis estimate is from Feb 2023,

Oh! I missed the date. You're right, that's a lot more expensive. On the other hand, inference has likely gotten a lot cheaper (in terms of GPU TOPS) too. Still, I think there's a profitable business model there if VC funding dries up and most of the model companies collapse.

217. tikhonj ◴[] No.44573682[source]
At some point, Groupon was the fastest growing company ever.
218. lxgr ◴[] No.44573705{6}[source]
Yup, and I wouldn't be willing to bet that any firewall between content and advertising would hold, long-term.

For example, the more product placement opportunities there are, the more products can be placed, so sooner or later that'll become an OKR to the "content side" of the business as well.

219. signatoremo ◴[] No.44573707{5}[source]
It won’t be ads. Social media target consumers, so advertising is dominant. We all love free services and don’t mind some attraction.

AI on the other hand target businesses and consumers alike. A bank using LLM won’t get ads. Using LLM will be cost of doing business. Do you know what they means to consumers? Price for ChatGPT will go down.

replies(2): >>44574025 #>>44579330 #
220. frozenseven ◴[] No.44573728{4}[source]
He links to Reddit, a site where most people are aggressively against AI. So, not necessarily a representative slice of reality.
replies(3): >>44573929 #>>44574014 #>>44574589 #
221. saratogacx ◴[] No.44573735[source]
We actually aren't too far off from that reality. There are several models you can run fully offline on your phone (phi-3, Gemma-3n-E2b-it, Qwen2.5-1.5b-instruct all run quite well on my Samsung S24 ultra). There are a few offline apps that also have tool calling (mostly for web search but I suspect this is extendable).

If you want to play around a bit and are on android there is PocketPal,ChatterUI, MyDeviceAI, SmolChat are good multi-model apps and Google's Edge gallery won't keep your chats but is a fun tech demo.

All are on github and can be installed using Obtainium if you don't want to

222. Gormo ◴[] No.44573813{4}[source]
> I am much more impressed by the opportunities I see in consulting rather than applying LLMs to dev tasks.

I expect there'll be a lot of consulting work in the near future in cleanup and recovery from LLM-generated disasters.

223. ohyes ◴[] No.44573861{3}[source]
Had someone put up a project plan for something that was not disclosed as LLM assisted output.

While technically correct it came to the wrong conclusions about the best path forward and inevitably hamstrung the project.

I only discovered this later when attempting to fix the mess and having my own chat with an LLM and getting mysteriously similar responses.

The problem was that the assumptions made when asking the LLM were incorrect.

LLMs do not think independently and do not have the ability to challenge your assumptions or think laterally. (yet, possibly ever, one that does may be a different thing).

Unfortunately, this still makes them as good as or better than a very large portion of the population.

I get pissed off not because of the new technology or the use of the LLM, but the lack of understanding of the technology and the laziness with which many choose to deliver the results of these services.

I am more often mad at the person for not doing their job than I am at the use of a model, the model merely makes it easier to hide the lack of competence.

replies(3): >>44574420 #>>44574450 #>>44574904 #
224. johnnyanmac ◴[] No.44573868[source]
People "like" or people "suffice" with the output? This "rise of whatever" as one blog put it gives me feelings that people are instead lowering their standards and cutting corners. Letting them cut through to stuff they actually want to do.
225. johnnyanmac ◴[] No.44573910{3}[source]
This isn't some niche outcry: https://www.forbes.com/sites/bernardmarr/2024/03/19/is-the-p...

And that was 18 months ago.

Yes, believe it or not, people eventually wake up and realize slop is slop. But like everything else with LLM development, tech is trying to brute force it on people anyway.

replies(2): >>44574228 #>>44577334 #
226. johnnyanmac ◴[] No.44573929{5}[source]
Given recent studies, that does seem to reflect reality. Trust in AI has been waning for 2 years now.
replies(1): >>44574383 #
227. johnnyanmac ◴[] No.44573961{4}[source]
Well, given the answers to the former: maybe we should stop now before we end up selling even more of our data off to technocrats. Or worse, your chatbot shilling to you between prompts.

And yes these are still businesses. If they can't find profitability they will drop it like it's hot. i.e. we hit another bubble burst that tech is known to do every decade or 2. There's no free money anymore to carry them anymore, so perfect time to burst.

228. zdragnar ◴[] No.44574001{4}[source]
lmgtfy was (from what I saw) always used as a snarky way to tell someone to do a little work on their own before asking someone else to do it for them.

I have seen people use "here's what chatGPT" said almost exclusively unironically, as if anyone else wants humans behaving like agents for chatbots in the middle of other people's discussion threads. That is to say, they offer no opinion or critical thought of their own, they just jump into a conversation with a wall of text.

replies(1): >>44574421 #
229. genghisjahn ◴[] No.44574014{5}[source]
He links to a post about a teacher’s expertise with students using AI. The fact that it’s on Reddit is irrelevant.
replies(1): >>44574238 #
230. thewebguyd ◴[] No.44574022{3}[source]
> I also imagine it’s hard to go to investors with “while all of our competitors are improving their models and either closing the gap or surpassing us, we’re just going to stabilize and see if people will pay for our current product.”

Yeah, no one wants to be the first to stop improving models. As long as investor money keeps flowing in there's no reason to - just keep burning it and try to outlast your competitors, figure out the business model later. We'll only start to see heavy monetization once the money dries up, if it ever does.

replies(1): >>44574405 #
231. johnnyanmac ◴[] No.44574025{6}[source]
>AI on the other hand target businesses and consumers alike.

Okay. So AI will be using ads for consumers and make deals with the billionaires. If window 11/12 still puts ads in what is a paid premium product, I see no optimism in thinking that a "free" chatbot will not also resort to it. Not as long as the people up top only see dollar signs and not long term longevity.

>Price for ChatGPT will go down.

Price for ChatGPT in reality, is going up in the meanwhile. This is like hoping grocery prices come down as inflation lessens. This never happens, you can only hope to be compensated more to make up for inflation.

replies(1): >>44576560 #
232. johnnyanmac ◴[] No.44574055{6}[source]
>what has been developed up until now has a great deal of utility and transformative power

I think we'd be more screwed than VR if development ceased today. They are little more than toys right now who's most successsful outings are grifts, and the the most useful tools are simply aiding existing tooling (auto-correct). It is not really "intelligence" as of now.

>I believe that "LLMs transforming society is inevitable" is a much more defensible assertion

Sure. But into what? We can't just talk about change for change's sake. Look at the US in 2025 with that mentality.

233. johnnyanmac ◴[] No.44574101{6}[source]
You still have faith in society after decades of ads being spit at them.
234. johnnyanmac ◴[] No.44574128{4}[source]
I mean, it adjusted down by having some hundreds of thousands of engineers laid off in he last 2+ years. they know slashing salaries is legal suicide, so they just make the existing workers work 3x as hard.
235. giancarlostoro ◴[] No.44574143{3}[source]
That's fine, that's what I want ;) I just grow tired of people hating on me for thinking that we really need to localize the models for them to take off.
replies(1): >>44577939 #
236. mvieira38 ◴[] No.44574169{3}[source]
LLMs are amazing at anything requiring text analysis (go figure). Everyone I know doing equity or economic research in finance is using it extensively for that, and from what I hear from doctors the LLMs are as good as that in their space if not better
237. __loam ◴[] No.44574184{6}[source]
Yeah callousness does seem to be the leaking area of improvement.
238. ericfr11 ◴[] No.44574196{5}[source]
It sounds quite scary that an LLM could be trained on a single source of news (specially FN).
239. __loam ◴[] No.44574220{6}[source]
Great, we've got mediocre writing from unprofitable companies that are subsidizing the cost of this technology.
replies(1): >>44575138 #
240. elictronic ◴[] No.44574228{4}[source]
You posted an article about investors trust in AI companies to deliver and societies strong distrust of large corporations.

You article isn’t making the point you seem to think it is.

replies(1): >>44577680 #
241. frozenseven ◴[] No.44574238{6}[source]
If you're going to champion something that comes from a place of extreme political bias, you could at least acknowledge it.
replies(2): >>44574993 #>>44576027 #
242. PebblesRox ◴[] No.44574260{8}[source]
If you believe this, people believe everything they read by default and have to apply a critical thinking filter on top of it to not believe the thing.

I know I don't have as much of a filter as I ought to!

https://www.lesswrong.com/s/pmHZDpak4NeRLLLCw/p/TiDGXt3WrQwt...

replies(1): >>44575741 #
243. nyarlathotep_ ◴[] No.44574269{7}[source]
> I constantly get junior developers handed to me from "strategic partners", they are just disguised as senior developers. I'm telling you brother, the LLMs aren't helping these guys do the job. I've let go 3 of them in July alone.

I find this surprising. I figured the opposite: that the quality of body shop type places would improve and the productivity increases would decrease as you went "up" the skill ladder.

I've worked on/inherited a few projects from the Big Name body shops and, frankly, I'd take some "vibe coded" LLM mess any day of the week. I really figured there was nowhere to go but "up" for those kinds of projects.

244. jsnell ◴[] No.44574284{5}[source]
You're either overestimating the cost of inference or underestimating the cost of running a service like Facebook at that scale. Meta's cost of revenue (i.e. just running the service, not R&D, not marketing, not admin, none of that) was about $30B/year in 2024. In the leaked OpenAI financials from last year, their 2024 inference costs were 1/10th of that.
replies(2): >>44575558 #>>44581093 #
245. doctorpangloss ◴[] No.44574360{3}[source]
No… dude… it’s a new household name. We haven’t had those in software for a long time, maybe since TikTok and Fortnite.
replies(1): >>44575533 #
246. frozenseven ◴[] No.44574383{6}[source]
By what relevant metric?

The userbase has grown by an order of magnitude over the past few years. Models have gotten noticeably smarter and see more use across a variety of fields and contexts.

replies(1): >>44574686 #
247. BolexNOLA ◴[] No.44574405{4}[source]
Maybe I’m naïve/ignorant of how things are done in the VC world, but given the absolutely enormous amount of money flowing into so many AI startups right now, I can’t imagine that the gravy train is going to continue for more than a few years. Especially not if we enter any sort of economic downturn/craziness from the very inconsistent and unpredictable decisions being made by the current administration
replies(1): >>44574587 #
248. thewebguyd ◴[] No.44574420{4}[source]
> do not have the ability to challenge your assumptions or think laterally.

Particularly on the challenging your assumptions part is where I think LLMs fail currently, though I won't pretend to know enough about how to even resolve that; but right now, I can put whatever nonsense I want into ChatGPT and it will happily go along telling me what a great idea that is. Even on the remote chance it does hint that I'm wrong, you can just prompt it into submission.

None of the for-profit AI companies are going to start letting their models tell users they're wrong out of fear of losing users (people generally don't like to be held accountable) but ironically I think it's critically important that LLMs start doing exactly that. But like you said, the LLM can't think so how can it determine what's incorrect or not, let alone if something is a bad idea or not.

Interesting problem space, for sure, but unleashing these tools to the masses with their current capabilities I think has done, and is going to continue to do more harm than good.

replies(2): >>44575562 #>>44575564 #
249. SoftTalker ◴[] No.44574421{5}[source]
Yeah I don't even read those. If someone can't be bothered to communicate their own thoughts in their own words, I have little belief that they are adding anything worth reading to the conversation.
replies(1): >>44574559 #
250. henryfjordan ◴[] No.44574425{3}[source]
Google does put AI output at the top of every search now, and sometimes it's helpful and sometimes it's crap. They have been trying since long before LLMs to not just provide the links for a search but also the content.

Google used to be interested in making sure you clicked either the paid link or the top link in the results, but for a few years now they'd prefer that a user doesn't even click a link after a search (at least to a non-Google site)

replies(1): >>44576141 #
251. justfix17 ◴[] No.44574450{4}[source]
> LLMs do not think

Yep.

More seriously, you described a great example of one of the challenges we haven't addressed. LLM output masquerades as thoughtful work products and wastes people's time (or worse tanks a project, hurts people, etc).

Now my job reviewing work is even harder because bad work has fewer warning signs to pick up on. Ugh.

I hope that your workplace developed a policy around LLM use that addressed the incident described. Unfortunately I think most places probably just ignore stuff like this in the faux scramble to "not be left behind".

replies(1): >>44574771 #
252. danielbln ◴[] No.44574494{4}[source]
You may want to pay for Claude Max outside of the Google or iOS ecosystem and save $40/month.
253. umpalumpaaa ◴[] No.44574537{5}[source]
ChatGPT also makes money via affiliate links. If you ask ChatGPT something like "what is the best airline approved cabin luggage you can buy?" you get affiliate links to Amazon and other sites. I use ChatGPT most of the time before I buy anything these days… From personal experience (I operated an app financed by affiliate links). I can tell you that this for sure generates a lot of money. My app was relatively tiny and I only got about 1% of the money I generated but that app pulled in about $50k per month.

Buying better things is one of my main use cases for GPT.

replies(1): >>44575812 #
254. thewebguyd ◴[] No.44574544{7}[source]
> a much larger number are using them as an excuse for layoffs they would have wanted to do anyway

It's a simple formula. Layoffs because of market conditions or company health = stock price go down. Layoffs because "AI took the jobs" = stock price go up.

255. Sharlin ◴[] No.44574559{6}[source]
Why communicate your own thoughts when ChatGPT can give you the Correct Answer? Saves everybody time and effort, right? I guess that’s the mental model of many people. That, or they’re just excited to be able to participate (in their eyes) productively in a conversation.
replies(1): >>44575420 #
256. thewebguyd ◴[] No.44574587{5}[source]
You would think so. Investors are eventually going to want a return on their money put in. But there seems to be a ton of hype and irrationality around AI, even worse than blockchain back in the day.

I think there's an element of FOMO - should someone actually get to AGI, or at least something good enough to actually impact the labor market and replace a lot of jobs, the investors of that company/product stand to make obscene amounts of money. So everyone pumps in, in hope of that far off future promise.

But like you said, how long can this keep going before it starts looking like that future promise will not be fulfilled in this lifetime and investors start wanting a return.

257. Sharlin ◴[] No.44574589{5}[source]
I’d like to see a statistically sound source for that claim. Given how many non-nerds there are on Reddit these days, it’s unlikely that there’s any particular strong bias in any direction compared to any similar demographic.
258. brokencode ◴[] No.44574602{3}[source]
I think the thing people hate about that is the lack of effort and attention to detail. It’s an incredible enabler for laziness if misused.

If somebody writes a design or a report, you expect that they’ve put in the time and effort to make sure it is correct and well thought out.

If you then find the person actually just had ChatGPT generate it and didn’t put any effort into editing it and checking for correctness, then that is very infuriating.

They are essentially farming out the process of creating the document to AI and farming out the process of reviewing it to their colleagues. So what is their job then, exactly?

These are tools, not a replacement for human thought and work. Maybe someday we can just have ChatGPT serve as an engineer or a lawyer, but certainly not today.

replies(1): >>44578396 #
259. satvikpendem ◴[] No.44574638[source]
> People like the output of LLMs so much that ChatGPT is the fastest growing app ever

And how much of that is free usage, like the parent said? Even when users are paying, ChatGPT's costs are larger than their revenue.

260. JTbane ◴[] No.44574686{7}[source]
> Models have gotten noticeably smarter and see more use across a variety of fields and contexts.

Is that really true? The papers I've read seem to indicate the hallucination rate is getting higher.

replies(1): >>44574978 #
261. ludicrousdispla ◴[] No.44574771{5}[source]
It's even worse than you suggest, for the following reason. The rare employee that cares enough to read through an entire report is more likely to encounter false information which they will take as fact (not knowing that LLM produced the report, or unaware that LLMs produce garbage). The lazy employees will be unaffected.
262. strange_quark ◴[] No.44574774[source]
> There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return (the supersonic jetliner)

I think this is a great analogy, not just to the current state of AI, but maybe even computers and the internet in general.

Supersonic transports must've seemed amazing, inevitable, and maybe even obvious to anyone alive at the time of their debut. But hiding under that amazing tech was a whole host of problems that were just not solvable with the technology of the era, let alone a profitable business model. I wonder if computers and the internet are following a similar trajectory to aerospace. Maybe we've basically peaked, and all that's left are optimizations around cost, efficiency, distribution, or convenience.

If you time traveled back to the 1970s and talked to most adults, they would have witnessed aerospace go from loud, smelly, and dangerous prop planes to the 707, 747 and Concorde. They would've witnessed the moon landings and were seeing the development of the Space Shuttle. I bet they would call you crazy if you told this person that 50 years later, in 2025, there would be no more supersonic commercial airliners, commercial aviation would basically look the same except more annoying, and also that we haven't been back to the moon. In the previous 50 years we went from the Wright Brothers to the 707! So maybe in 2075 we'll all be watching documentaries about LLMs (maybe even on our phones or laptops that look basically the same), and reminiscing about the mid-2020s and wondering why what seemed to be such a promising technology disappeared almost entirely.

replies(7): >>44575171 #>>44575370 #>>44575839 #>>44576734 #>>44578123 #>>44579514 #>>44585127 #
263. shpongled ◴[] No.44574872[source]
I would pay $5000 to never have to read another LLM-authored piece of text ever again.
264. 131012 ◴[] No.44574904{4}[source]
> LLMs do not think independently and do not have the ability to challenge your assumptions

It IS possible for a LLM to challenge your assumptions, as its training material may include critical thinking on many subjects.

The helpful assistant, being almost by definition a sycophant, cannot.

replies(1): >>44576617 #
265. frozenseven ◴[] No.44574978{8}[source]
Models from a few years ago are comparatively dumb. Basically useless when it comes to performing tasks you'd give to o3 or Gemini 2.5 Pro. Even smaller reasoning models can do things that would've been impossible in 2023.
266. Capricorn2481 ◴[] No.44574993{7}[source]
This is a baffling response. The politics are completely irrelevant to this topic. Pretty much every American is distrustful of big tech and is completely unaware of what the current administration has conceded to AI companies, with larger scandals taking the spotlight, so there hasn't been a chance for one party or the other to rally around a talking point with AI.

People don't like AI because its impact on the internet is filling it with garbage, not because of tribalism.

replies(1): >>44575328 #
267. reasonableklout ◴[] No.44575036{3}[source]
Last year, ChatGPT was 75% of OpenAI's revenue[1], not the API.

[1]: https://www.businessofapps.com/data/chatgpt-statistics/

268. dvfjsdhgfv ◴[] No.44575101{6}[source]
> Obviously you don't need to train new models to operate existing ones.

For a few months, maybe. Then they become obsolete and, in some cases like coding, useless.

269. lotsoweiners ◴[] No.44575104{5}[source]
To be fair, ads always compromise the product.
270. SpicyLemonZest ◴[] No.44575138{7}[source]
What specifically do you find to be mediocre? I feel like LLMs write better than most people I know, myself included.

There could be a mismatch on what the state of the art really is these days. In my experience, since the release of GPT-4 and especially 4o, ChatGPT has been able to do the vast majority of concrete things people tell me it can't do.

replies(1): >>44582880 #
271. kenjackson ◴[] No.44575171[source]
I think this is both right and wrong. There was a good book that came out probably 15 years ago about how technology never stops in aggregate, but individual technologies tend to grow quickly and then stall. Airplane jets were one example in the book. The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

A better example, also in the book, are skyscrapers. Each year they grew and new ones were taller than the ones last year. The ability to build them and traverse them increased each year with new technologies to support it. There wasn't a general consensus around issues that would stop growth (except at more extremes like air pressure). But the growth did stop. No one even has expectations of taller skyscrapers any more.

LLMs may fail to advance, but not because of any consensus reason that exists today. And it maybe that they serve their purpose to build something on top of them which ends up being far more revolutionary than LLMs. This is more like the path of electricity -- electricity in itself isn't that exciting nowadays, but almost every piece of technology built uses it.

I fundamentally find it odd that people seem so against AI. I get the potential dystopian future, which I also don't want. But the more mundane annoyance seems odd to me.

replies(7): >>44575296 #>>44575503 #>>44575767 #>>44575922 #>>44576225 #>>44576624 #>>44578021 #
272. rpdillon ◴[] No.44575180{4}[source]
Yep. Remember when Amazon could never make money and we kept trying to explain they were reinvesting their earnings into R&D and nobody believed it? All the rhetoric went from "Amazon can't be profitable" to "Amazon is a monopoly" practically overnight. It's like people don't understand the explore/exploit strategy trade-off.
replies(1): >>44575844 #
273. kenjackson ◴[] No.44575192{3}[source]
If I cared about the output from other people's queries then wouldn't they be my queries? I don't care about ChatGPTs response to your queries is because I don't care about your queries. I don't care if they came from ChatGPT or the world's foremost expert in whatever your query was about.
274. Earw0rm ◴[] No.44575296{3}[source]
There are sound math reasons for skyscrapers topping out, mostly due to elevator capacity and the inability to effectively get people in and out of the floorspace as you go past a few hundred ft. There's no construction engineering reason you can't go taller - the Burj Khalifa, for example, is three times taller than a typical Western major city skyscraper - it just doesn't make economic sense unless you're a newly rich nation looking to prove a point.
replies(1): >>44578334 #
275. frozenseven ◴[] No.44575328{8}[source]
>This is a baffling response.

Likewise.

95+% of the time I see a response like this, it's from one particular side of the political aisle. You know the one. Politics has everything to do with this.

>what the current administration has conceded to AI companies

lol, I unironically think that they're not lax enough when it comes to AI.

replies(2): >>44575665 #>>44577985 #
276. Earw0rm ◴[] No.44575370[source]
From a system optimisation perspective, SSTs solved the wrong problem.

Want to save people time flying? Solve the grotesque inefficiency pit that is airport transit and check-in.

Like, I'm sorry, STILL no high speed, direct to terminal rail at JFK, LAX and a dozen other major international airports? And that's before we get to the absolute joke of "border security" and luggage check-in.

Sure, supersonic afterburning engines are dope. But it's like some 10GHz single-core CPU that pulls 1.2kW out of the wall. Like it or not, an iPhone 16 delivers far more compute utility in far more scenarios.

replies(3): >>44575423 #>>44575790 #>>44576754 #
277. hyperbovine ◴[] No.44575415{3}[source]
It still doesn't pass the Turing test, and is not close. Five years ago me would be impressed but still adamant that this is not AI, nor is it on the path to AI.
278. SoftTalker ◴[] No.44575420{7}[source]
If I want the "correct answer" I'll research it, maybe even ask ChatGPT. If I'm having a conversation I'm interesed in what the other participants think.

If I don't know something, I'll say I don't know, and maybe learn something by trying to understand it. If I just pretend I know by pasting in what ChatGPT says, I'm not only a fraud but also lazy.

279. dingnuts ◴[] No.44575423{3}[source]
SST came and went in an era when none of that security theater existed to begin with
replies(1): >>44575795 #
280. JohnMakin ◴[] No.44575425[source]
> That is a such a wild claim. People like the output of LLMs so much that ChatGPT is the fastest growing app ever.

And this kind of meaningless factoid was immediately usurped by the Threads app release, which IMO is kind of a pointless app. Maybe let's find a more meaningful metric before saying someone else's claim is wild.

replies(1): >>44576935 #
281. LeifCarrotson ◴[] No.44575445{3}[source]
Especially if it's not disclosed up front, and especially when it supplants higher-value content. I've been shocked how little time it's taken for AI slop SEO optimized blogs to overtake the articles written by genuine human experts, especially in niche product reviews and technical discussions.

However, whether or not people like it is almost irrelevant. The thing that matters is not whether economics likes it.

At least so far, it looks like economics absolutely loves LLMs: Why hire expensive human customer support when you can just offload 90% of the work to a computer? Why pay expensive journalists when you can just have the AI summarize it? Why hire expensive technical writers to document your code when you can just give it to the AI and check the regulatory box with docs that are good enough?

replies(1): >>44575950 #
282. swat535 ◴[] No.44575452{5}[source]
> LLMs can’t do ads without compromising the product.

It depends on what you mean by "compromise" here but they sure can inject ads.. like make the user wait 5 seconds, show an ad, then reply..

They can delay the response times and promote "premium" plans, etc

Lots of ways to monetize, I suppose the question is: will users tolerate it?

Based on what I've seen, the answer is yes, people will tolerate anything as long as it's "free".

283. socalgal2 ◴[] No.44575502{5}[source]
Social networks will have all of those effects without any effort by the platform itself because the person with more followers has more influence so the people on the platform will do all they can to get more.

I'm not excusing the platforms for bad algorithms. Rather, I believe it's naive to think that, but for the behavior of the platform itself that things would be great and rosy.

No, they won't. The fact that nearly every person in the world can mass communicate to nearly every other person in the world is the core issue. It is not platform design.

284. bluefirebrand ◴[] No.44575503{3}[source]
> even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today

I think they pretty strongly do

The solution seems to be "just lower your standards for acceptable margin of error to whatever the LLM is capable of producing" which should be concerning and absolutely unacceptable to anyone calling themselves an Engineer

replies(2): >>44576994 #>>44579420 #
285. matthewdgreen ◴[] No.44575533{4}[source]
Lots of things had household recognition. Do you fondly remember the Snuggie? The question is whether it'll be durable. The lack of network effects is one reason to be skeptical.
replies(1): >>44578682 #
286. matthewdgreen ◴[] No.44575558{6}[source]
But their research costs are extremely high, and without a network effect that revenue is only safe until a better competitor emerges.
replies(1): >>44577903 #
287. DrewADesign ◴[] No.44575562{5}[source]
I'm no expert, but the most frequent recommendations I hear to address this are:

a) tell it that it's wrong and to give you the correct information.

b) use some magical incantation system prompt that will produce a more critical interlocutor.

The first requires knowing enough about the topic to know the chatbot is full of shit, which dramatically limits the utility of an information retrieval tool. The second assumes that the magical incantation correctly and completely does what you think it does, which is not even close to guaranteed. Both assume it even has the correct information and is capable of communicating it to you. While attempting to use various models to help modify code written in a less-popular language with a poorly-documented API, I learned how much time that can waste the hard way.

If your use case is trivial, or you're using it as a sounding board with a topic you're familiar with as you might with, say, a dunning-kruger-prone intern, then great. I haven't found a situation in which I find either of those use cases compelling.

288. myrryr ◴[] No.44575564{5}[source]
This is why once you are using to using them, you start asking them for there the plan goes wrong. They won't tell you off the bat, whuch can be frustrating, but they are really good at challenging your assumptions, if you ask them to do so.

They are good at telling you what else you should be asking, if you ask them to do so.

People don't use the tools effectively and then think that the tool can't be used effectively...

Which isn't true, you just have to know how the tool acts.

289. mrbungie ◴[] No.44575573{5}[source]
Any retries before nailing the prompt are still going to be billed, so this supports GP position about LLMs being expensive for trivial things.
290. intended ◴[] No.44575665{9}[source]
Based on your response and logic - no dem should read stuff written by repub voters, or if they do read it, dismiss their account because it cannot be … what?

Not sure how we get to dismissing the teacher subreddit, to be honest.

replies(3): >>44575830 #>>44582453 #>>44584237 #
291. brulard ◴[] No.44575667{6}[source]
Obviously YMMV, but it is extremely useful for me and for many people out there.
292. anthonypasq ◴[] No.44575692{3}[source]
correct and companies will be exposing their data via mcp instead of standard rest apis.
replies(1): >>44575756 #
293. killerstorm ◴[] No.44575728{4}[source]
Yes. I verified it myself. Best results from Opus 4 so far, Gemini might be OK too.
294. andrewflnr ◴[] No.44575741{9}[source]
That checks out with my experience. I don't think it's just reading either. Even deeper than stranger danger, we're inclined to assume other humans communicating with us are part of our tribe, on our side, and not trying to deceive us. Deception, and our defenses against deception, are a secondary phenomenon. It's the same reason that jokes like "the word 'gullible' is written in the ceiling", gesturing to wipe your face at someone with a clean face, etc, all work by default.
295. ramoz ◴[] No.44575756{4}[source]
That makes no sense. MCP at best is a protocol transpilation at runtime. It is not redefining things like DB drivers or connections. And I did not say rest apis enable agents. Computer use tooling does. APIs and everything else that already exists.

MCP is more like graphql. Not a new network paradigm.

The design of MCP right now is not very optimal esp when you can equip an agent with one tool vs 5-20 that bloat it's reasoning every prompt.

replies(1): >>44575872 #
296. da_chicken ◴[] No.44575767{3}[source]
> The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

The fundamental problem has already been mentioned: Nobody can figure out how to SELL it. Because few people are buying it.

It's useful for aggregation and summarization of large amounts of text, but it's not trustworthy. A good summary decreases noise and amplifies signal. LLMs don't do that. Without the capability to validate the output, it's not really generating output of lasting value. It's just a slightly better search engine.

It feels like, fundamentally, the primary invention here is teaching computers that it's okay to be wrong as long as you're convincing. That's very useful for propaganda or less savory aspects of business, but it's less useful for actual communication.

replies(2): >>44576213 #>>44579435 #
297. SJC_Hacker ◴[] No.44575780{6}[source]
That's pretty much what search engines are nowadays
298. mxschumacher ◴[] No.44575790{3}[source]
that's not a technology problem, many airports are super efficient, e.g. Singapore. Public transport in the US is held back by other forces.
replies(1): >>44575827 #
299. Earw0rm ◴[] No.44575795{4}[source]
It did for international, maybe not at the dawn of SSTs but after a string of hijackings in the 70s/80s they brought it in. Not for US internal flights, it's true.
300. ysavir ◴[] No.44575812{6}[source]
Makes you wonder whether the affiliate links are actual, valid affiliate links or just hallucinations from affiliate links it's come across in the wild
replies(1): >>44577465 #
301. MonkeyIsNull ◴[] No.44575820[source]
> 2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them,

I always think back to how Bezos and Amazon were railed against for losing money for years. People thought that would never work. And then when he started selling stuff other than books? People I know were like: please, he's desperate.

Someone, somewhere will figure out how to make money off it - just not most people.

302. mxschumacher ◴[] No.44575826{4}[source]
what I struggle with is that the top 10 providers of LLMs all have identical* products. The services have amazing capabilities, but no real moats.

The social media applications have strong network effects, this drives a lot of their profitability.

* sure, there are differences, see the benchmarks, but from a consumer perspective, there's no meaningful differentiation

replies(1): >>44578080 #
303. Earw0rm ◴[] No.44575827{4}[source]
It makes it all the dumber that we have the tech and still can't manage to implement the solution.

Like an org with crappy management and team structure shipping bloated, buggy code even though they've the budget to hire great engineers and the problems they're solving are largely known and well-trodden.

304. SJC_Hacker ◴[] No.44575839[source]
The problem with supersonic commercial jets was mainly one of marketing/politics. The so called "sonic boom" problem was vastly overhyped, as anyone who lives near an air force base can tell you.

The conspiracy theorist tells me the American aerospace manufacturers at the time (Boening, McDonnell-Douglas, etc.), did everything they could to kill the Concorde. With limited flyable routes (NYC and DC to Paris and London I think were the only ones), the financials didn't make sense. If overland routes were available, especially opening up LA, San Francisco and Chicago, it might have been a different story.

replies(3): >>44578304 #>>44581166 #>>44581314 #
305. mxschumacher ◴[] No.44575844{5}[source]
AWS is certainly super profitable, if the ecommerce business was standalone, would it really be such a cash-gusher?
replies(1): >>44576546 #
306. anthonypasq ◴[] No.44575872{5}[source]
why would you make an agent click around a web browser like a human when it could self discover the api and call it directly?
replies(1): >>44576047 #
307. citizenpaul ◴[] No.44575922{3}[source]
Was the problem that supersonic flight was expensive and the amount of customers willing to pay the price was even lower than the number of customers that could even if they wanted to?
replies(2): >>44576116 #>>44588993 #
308. SJC_Hacker ◴[] No.44575943{6}[source]
This is an interesting take - is my "attention" really worth several thousand a year? In that my purchasing decisions being influenced by advertising to that degree that someone is literally paying someone else for my attention ...

I wonder if instead, could I sell my "attention" instead of others profitting of it?

replies(1): >>44577375 #
309. davidcbc ◴[] No.44575950{4}[source]
Eventually the economics will correct themselves once people yet again learn the old "you get what you pay for" lesson (or the more modern FAFO lesson)
310. fireflash38 ◴[] No.44576027{7}[source]
Why? So you could discard it faster?

Read things from people that you disagree with.

replies(1): >>44576635 #
311. ramoz ◴[] No.44576047{6}[source]
self discovery via primitives is what works well today. I never discouraged that, only discouraged MCP sensationalism.

However, an agent that can see the screen and immediately click through whatever desired UI modality is immensely more efficient than swimming through protocols.

There is at least one frontier lab who has prepared enough foresight that agents running on VDI infrastructure is a major coming wave.

312. reddit_clone ◴[] No.44576116{4}[source]
From what I had read in passing and remember.

  - They were loud (sonic booms were nasty).

  - They were expensive to maintain and operate. Guzzlers. (Britain and France clung to them as a matter of pride/ego)

  - They were narrow and uncomfortable. I have seen videos where there is space only for one stewardess to walk. I had been inside of one in Seattle museum. Very cramped.

  - As you mentioned, ticket cost was high.

  - I suspect people traveled in these mostly for bragging rights.
replies(1): >>44577206 #
313. LtWorf ◴[] No.44576141{4}[source]
It made me switch away from google. The push I needed
314. reddit_clone ◴[] No.44576161{3}[source]
Low effort Youtube shorts with AI voice annoy the crap out of me.

After all this hype, they still can't do text to speech properly. Pause at the wrong part of the sentence all the time.

315. kenjackson ◴[] No.44576213{4}[source]
> Nobody can figure out how to SELL it. Because few people are buying it.

Just picking one company who basically just does AI, OpenAI. They reported it has 20 million PAID subscribers to ChatGPT. With revenue projected above $12b dollars (https://www.theverge.com/openai/640894/chatgpt-has-hit-20-mi...).

I think what you meant to say is that costs are high so they can't generate large profits. but saying that they can't figure out how to sell it seems absurd. Is it Netflix level of subscribers, no. But there can't be more than a couple of hundred products that have that type of subscription reach.

replies(1): >>44576435 #
316. z2 ◴[] No.44576225{3}[source]
Yeah, and with LLMs the thing I can't shake, however, is that this time it's pretty strongly (maybe parasitically) latched onto the aggregate progress of Moore's law. Few other technologies have enjoyed such relatively unfettered exponential improvement. It's like if skyscraper materials double in strength every n years, and their elevators approach teleportation speed, the water pumps get twice as powerful, etc., which would change the economics vs the reality that most of the physical world doesn't improve that fast.
317. strange_quark ◴[] No.44576435{5}[source]
Ok but isn’t 20 million subscribers out of what, 800 million or 1 billion monthly users or whatever they’re claiming, an absolutely abysmal conversion rate? Especially given that the industry and media have been proclaiming this as somewhere between the internet and the industrial revolution in terms of impact and advancement? Why can they not get more than 3% of users to convert to paying subscribers for such a supposedly world changing technology, even with a massive subsidy?
replies(3): >>44576611 #>>44576733 #>>44578184 #
318. rpdillon ◴[] No.44576546{6}[source]
Amazon is successful because of the insanely broad set of investments they've made - many of them compound well in a way that supports their primary business. Amazon Music isn't successful, but it makes Kindle tablets more successful. This is in contrast to Google, which makes money on ads, and everything else is a side quest. Amazon has side quests, but also has many more initiatives that create a cohesive whole from the business side.

So while I understand how it looks from a financial perspective, I think that perspective is distorted in terms of what causes those outcomes. Many of the unprofitable aspects directly support the profitable ones. Not always, though.

319. Geezus_42 ◴[] No.44576560{7}[source]
Has any SAAS product ever reduced their subscription cost?
replies(1): >>44577885 #
320. oarsinsync ◴[] No.44576611{6}[source]
Because they give too much of it away for free? Most casual use fits into the very generous free tier.
replies(1): >>44576677 #
321. newAccount2025 ◴[] No.44576617{5}[source]
Strong agree. If you simply ask an LLM to challenge your thinking, spot weaknesses in your argument, or what else you might consider, it can do a great job.

This is literally my favorite way to use it. Here’s an idea, tell me why it’s wrong.

322. overgard ◴[] No.44576624{3}[source]
>I think this is both right and wrong. There was a good book that came out probably 15 years ago about how technology never stops in aggregate, but individual technologies tend to grow quickly and then stall. Airplane jets were one example in the book. The reason why I partially note this as wrong is that even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today.

I don't see any solution to hallucinations, nor do I see any solution in sight. I think that could count as a concrete issue that would stop them.

replies(1): >>44578050 #
323. frozenseven ◴[] No.44576635{8}[source]
Because I'm not going to play a game where the other side gets to ignore the rules.
324. strange_quark ◴[] No.44576677{7}[source]
Ok so the argument is that all the model builders either suck at business or they are purposefully choosing to lose billions of dollars?
replies(1): >>44577087 #
325. kenjackson ◴[] No.44576733{6}[source]
As another commenter notes, because you get access to a lot of functionality for free. And other providers are also providing free alternatives. The ratio for their free/paid tier is about the same as YouTube's. And like YouTube, it's not that YouTube isn't providing great value, but rather that most people get what they need out of the free tier.

The better question is what if all LLM services stopped providing for free at all -- how many paid users would there then be?

326. Lu2025 ◴[] No.44576734[source]
Oh no, LLMs won't disappear but they will be a lot less loud.

Progress is often an S shaped curve and we are nearing saturation.

327. Lu2025 ◴[] No.44576754{3}[source]
They don't optimize for our convenience, they optimize for their profit.
replies(1): >>44577448 #
328. Geezus_42 ◴[] No.44576838{5}[source]
Social and search both compromised the product for ad revenue.
329. og_kalu ◴[] No.44576935{3}[source]
Asking your Instagram Users to hop on to your ready made TikTok Clone is hardly in the same sphere as spinning up that much users from nothing.

And while Threads growth and usage stalled, ChatGPT is very much still growing and has *far* more monthly visits than threads.

There's really nothing meaningless about ChatGPT being the 5th most visited site on the planet, not even 3 years after release. Threads doesn't make the top 50.

replies(1): >>44577923 #
330. Aeolun ◴[] No.44576994{4}[source]
> absolutely unacceptable to anyone calling themselves an Engineer

Isn’t that exactly what engineers do? Even very strong bridges aren’t designed to survive every possible eventuality.

replies(2): >>44577175 #>>44577251 #
331. alonsonic ◴[] No.44577087{8}[source]
They are purposely losing billions, this is a growth phase where all of the big AI companies are racing to grow their userbase, later down the line they will monetize that captured userbase.

This is very similar to Uber which lost money for 14 years before becoming profitable, but with significantly more upside.

Investors see the growth, user stickiness and potential for the tech; and are throwing money to burn to be part of the winning team, which will turn on the money switch on that userbase down the line.

The biggest companies and investors in the planet aren't all bad at business.

replies(1): >>44580581 #
332. eddythompson80 ◴[] No.44577153{7}[source]
(sorry I kept writing and didn't realize how long it got and don't have the time to summarize it better)

Here is how it sort of happens sometimes:

- You are an analyst at some hedge fund.

- You study the agriculture industry overall and understand the general macro view of the market segment and its parameters etc.

- You pick few random agriculture company (e.g: WeGrowPotatos Corp.) that did really really solid returns between 2001 and 2007 and analyze their performance.

- You try to see how you could have predicted the company's performance in 2001 based on all the random bits of data you have. You are not looking for something that makes sense per se. Investing based on metrics that make intuitive sense is extremely hard if not impossible because everyone is doing that which makes the results very unpredictable.

- You figure out that for whatever reason, if you sum the total sales for a company, subtract reserved cash, and divide that by the global inflation rate minus the current interest rate in the US; this company has a value that's an anomaly among all the other agriculture companies.

- You call that bullshit The SAGI™ ratio (Sales Adjusted for Global Inflation ratio)

- You calculate the SAGI™ ratio for other agriculture companies in different points in time and determine its actual historical performance and parameters compared to WeGrowPotatoes in 2001.

- You then calculate that SAGI™ ratio for all companies today and study the ones that match your desired number then invest in them. You might even start applying SAGI™ analysis to non-agriculture companies.

- (If you're successful) In few years you will have built a reputation. Everyone wants to learn from you how you value a company. You share your method with the world. You still investigate the business to see how much it diverges from your "WeGrowPotatoes" model you developed the SAGI ratio based on.

- People look at your returns, look at your (1) step of calculating SAGI, and proclaim that the SAGI ratio paramount. Everyone is talking about nothing but SAGI ratio. Someone creates a SAGIHeads.com and /r/SAGInation and now Google lists it under every stock for some reason.

It's all about that (sales - cash / inflation - interest). A formula that makes no sense; but people are gonna start working it backwards by trying to understand what does "sales - cash" actually mean for a company?

Like that SAGI is bullshit I just made up, but EV is an actual metric and it's generally calculated as (equity + debt - cash). What do you think that tells you about a company? and why do people look at it? How does it make any sense for a company to sum its assets and debt? what is that? According to financial folks it tells you the actual market operation size of the company. The cash a company holds is not in the market so it doesn't count. the assets are obviously important to count, but debt for a company can be positive if it's on path to convert into asset on a reasonable timeline.

I don't know why investors in the tech space focus too much on ARR. It's possible that it was a useful metric with traditional internet startups model like Google, Facebook, Twitter, Instagram, Reddit, etc where the general wisdom was it's impossible to expect people to pay a lot for online services. So generating any sort of revenue almost always correlated with how many contracts do you get to signup with advertisers or enterprises and those are usually pretty stable and lucrative.

I highly recommend listening to Warren Buffets investing Q&As or lectures. He got me to view companies and the entire economy differently.

replies(1): >>44577403 #
333. wrs ◴[] No.44577175{5}[source]
That's what a "margin of error" is. The margin of error of a bridge is predictable thanks to well-established techniques of physical analysis.

An LLM system, on the other hand, can fail because you moved some punctuation around.

replies(1): >>44585221 #
334. Centigonal ◴[] No.44577204{5}[source]
oh, I 100% agree with this. The way the social web was monetized is the root of a lot of evil. With AI, we have an opportunity to learn from the past. I think a lesson here is "don't wait to think critically about the societal consequences of the next Big Tech Thing's business model because you have doubts about its profitability or unit economics."
335. wrs ◴[] No.44577206{5}[source]
You made this point in passing, but it's so relevant to LLMs I wanted to highlight it: The development and operational cost was heavily subsidized by the British and French governments, because having an SST was a point of national prestige.
336. bluefirebrand ◴[] No.44577251{5}[source]
No

I'm talking about engineering a bridge for 50 cars that collapses at 51, not engineering a bridge for 500 cars that is only expected to get 50

Engineering does require tradeoffs of course. But that's not what the minimum possible quality is

337. brokencode ◴[] No.44577334{4}[source]
Yup, any day now people will suddenly realize that LLMs suck and you were right all along. Any day now..
replies(2): >>44577667 #>>44578752 #
338. lymbo ◴[] No.44577375{7}[source]
Yes, but your attention rapidly loses value the more that your subsequent behavior misaligns with the buyer’s desires. In other words, the ability to target unsuspecting, idle minds far exceeds the value of a willing and conscious attention seller.
339. jdiff ◴[] No.44577403{8}[source]
No worries about the length, I appreciate you taking the time and appreciate the insight! That does help start to work ARR into a mental model that, while still not sane, is at least as understandably insane as everything else in the financial space.
340. brokencode ◴[] No.44577427{3}[source]
It’s not a wild claim, though maybe your interpretation is wild.

I never said Perplexity individually is challenging Google, but rather as part of a group of apps including ChatGPT, which you conveniently left out of your quote.

341. scarface_74 ◴[] No.44577448{4}[source]
Such airports removing moving walkways because they reduce profits in stores…

https://nypost.com/2025/05/27/lifestyle/airports-are-removin...

342. umpalumpaaa ◴[] No.44577465{7}[source]
It clearly is a 100% custom UI logic implemented by OpenAI… They render the products in carrousels… They probably get a list of product and brand names from the LLM (for certain requests/responses) and render that in a separate UI after getting those affiliate links for those products… its not hard to do. Just slap on your affiliate ID to the links you found and you are done.
replies(1): >>44581004 #
343. scarface_74 ◴[] No.44577467{4}[source]
No one ever doubted that Facebook would make money. It was profitable early on, never lost that much money and was definitely profitable by the time it went public.

Twitter has never been consistently profitable

344. scarface_74 ◴[] No.44577474{4}[source]
No one ever doubted that Facebook would make money. It was profitable early on, never lost that much money and was definitely profitable by the time it went public.

Twitter has never been consistently profitable.

ChatGPT also has higher marginal costs than any of the software only tech companies did previously.

345. gonzobonzo ◴[] No.44577486[source]
My guess is that LLM's are bridge technology, the equivalent of cassette tapes. A big step forward, allowing things that we couldn't before. But before long they'll be surpassed by much better technology, and future generations will look back on them as primitive.

You have top scientists like LeCun arguing this position. I'd imagine all of these companies are desperately searching for the next big paradigm shift, but no one knows when that will be, and until then they need to squeeze everything they can out of LLMs.

346. johnnyanmac ◴[] No.44577667{5}[source]
Yup, I can wait a while. Took some 7-8 years for people to turn on Facebook.
347. johnnyanmac ◴[] No.44577680{5}[source]
What point do you think it means? Seems pretty clear to me.

1. Investors are pushing a lot of hype

2. People are not trusting the hype.

Hence why people's trust in LLM's are waning.

348. jittery41 ◴[] No.44577751[source]
Oh wow I forgot that the microwave oven was once marketed as the total replacement of cooking chores and in futuristic life people can just press a button and have a delicious good meal ( well you can now but microwave meals are often seen as worse than fastfood ).
349. koolba ◴[] No.44577885{8}[source]
Does S3 count as a SaaS? Or is that too low level?

How about tarsnap? https://www.daemonology.net/blog/2014-04-02-tarsnap-price-cu...

350. jsnell ◴[] No.44577903{7}[source]
You're moving the goalposts, given the original complaint was not about research costs but about the marginal cost of serving additional users...

I guess you'd be surprised to find out that Meta's R&D costs are an order of magnitude higher than OpenAI's training + research costs? ($45B in 2024, vs. about $5B for OpenAI according to the leaked financials.)

replies(1): >>44583653 #
351. alexpotato ◴[] No.44577911[source]
To use the Internet as a comparison:

Phase 1 - mid to late 1990s:

- "The Internet is going to change EVERYTHING!!!"

Phase 2 - late 1990s to early 2000s:

- "It's amazing and we are all making SO much money!"

- "Oh no! The bubble burst"

- "Of course everyone could see this coming: who is going to buy 40 lb bags of dogfood or their groceries over the Internet?!?!?"

Phase 3 - mid 2000s to 2020:

- "It is astounding the amount of money being by tech companies"

- "Who could have predicted that social media would change the ENTIRE landscape??"

352. JohnMakin ◴[] No.44577923{4}[source]
I think you just precisely explained why MAU / DAU growth is a meaningless metric in such discussions.
replies(1): >>44583580 #
353. DSingularity ◴[] No.44577939{4}[source]
I’m not sure why people are hating on you. If you love being free then you should love the idea of being independent when it comes to common computing. If LLM is to become common we should all be rooting for open weights and efficient local execution.

It’s gonna take some time but it’s inevitable I think.

354. Capricorn2481 ◴[] No.44577985{9}[source]
> 95+% of the time I see a response like this, it's from one particular side of the political aisle. You know the one. Politics has everything to do with this

I really don't, honestly you're being so vague and it's such a bipartisan issue I can't piece together who you're mad at. Godspeed.

355. ogogmad ◴[] No.44578020[source]
The next step is clearly improved vision and everyday-physics models. These can also solve hallucinations.
356. stickfigure ◴[] No.44578021{3}[source]
> even in the 70s people recognized that supersonic travel had real concrete issues with no solution in sight. I don't think LLMs share that characteristic today

I hate to dogpile on this statement but I can think of two major issues right now:

* Small context windows, and serious degradation when pushing the limits of existing context windows. A human can add large amounts of state to their "context window" every day.

* Realtime learning. My humans get smarter every day, especially in the context of working with a specific codebase.

Maybe the AI companies will figure this out, but they are not "same technique more processor power" kinds of problems.

357. ogogmad ◴[] No.44578050{4}[source]
Vision and everyday-physics models are the answer: hallucinations will stop when the models stop thinking in words and start thinking in physical reality.
replies(2): >>44580840 #>>44582257 #
358. ogogmad ◴[] No.44578080{5}[source]
This is perfect news for consumers and terrible news for investors. Which are you?
replies(1): >>44587315 #
359. hilux ◴[] No.44578123[source]
> So maybe in 2075 we'll all be watching documentaries about LLMs (maybe even on our phones or laptops that look basically the same), and reminiscing about the mid-2020s and wondering why what seemed to be such a promising technology disappeared almost entirely.

It's hard for me to believe that anyone who works with technology in general, and LLMs in particular, could think this.

360. ogogmad ◴[] No.44578138{3}[source]
Hallucinations might get solved by faster, cheaper and more accurate, vision and commonsense-physics models. Hypothesis: Hallucinations are a problem only because physical reality isn't text. Once people switch to models that predict physical states instead of missing text, then we'll have domestic robots and lower hallucination rates.
replies(1): >>44582912 #
361. bscphil ◴[] No.44578167{9}[source]
> The current tools just can't replace developers. They can't even be used in the same way you'd use a junior developer or intern. It's more akin to going from hand tools to power tools than it is getting an apprentice.

In that case it seems to depend on what you mean by "replacing", doesn't it? It doesn't mean a non-developer can do a developers job, but it does mean that one developer can do two developer's jobs. That leads to a lot more competition for the remaining jobs and presumably many competent developers will accept lower salaries in exchange for having a job at all.

362. tjwebbnorfolk ◴[] No.44578184{6}[source]
You could say the same of Dropbox. Or Gmail.
replies(2): >>44578326 #>>44578709 #
363. tptacek ◴[] No.44578304{3}[source]
Seems... wrong. Booms broke windows and drove zillions of complaints. Supersonic flight near airbases is controlled and happens on specific traffic corridors, right?
364. bucklybuck ◴[] No.44578326{7}[source]
True, although I don't think Dropbox or Gmail's operating costs to support those free users are anywhere near those of OpenAI.
365. seanmcdirmid ◴[] No.44578334{4}[source]
Economic Concrete construction (what China specializes in) typically tops out at 30-40 floors, so the vast majority of buildings in Asia are that height, a sweet spot so to speak especially for residential (even in limited space HK).
366. snackernews ◴[] No.44578396{4}[source]
This is the biggest impact I have noticed in my job.

The inundation of verbose, low SNR text and documents. Maybe someone put thought into all of those words. Maybe they vibed it into existence with a single prompt and it’s filled with irrelevant dot points and vague, generic observations.

There is no way to know which you’re dealing with until you read it, or can make assumptions based on who wrote it.

367. doctorpangloss ◴[] No.44578682{5}[source]
Lack of network effects... It's the biggest thing ever! Everyone is talking about it, all the time, nonstop! How is that not a network? Network effects do not exclusively mean multiplayer software, communications or social media. And anyway, it is almost certainly all three of these things, because content is being made (and often consumed) by ChatGPT in every digital network there is.

Anyway, I don't think it's possible in this forum to have a conversation about it, if "ChatGPT is humongous" is a controversial, downvotable POV.

replies(1): >>44578725 #
368. const_cast ◴[] No.44578709{7}[source]
A service like Gmail or Dropbox with low storage is close to free to operate. Same thing with iCloud - 50 gigs a month is what, 1 dollar? How is that possible?

Because 50 gigs is next to nothing, and you only need a rinky dink amount of compute to write files.

YouTube, on the other hand, is actually pretty expensive to operate. Takes a lot of storage to store videos, never mind handling uploads. But then streaming video? Man, the amount of bandwidth required for that makes file syncing look like nothing. I mean, how often does a single customer watch a YouTube video? And then, how often do people download files from Dropbox? It's orders of magnitude in difference.

But LLMs outshine both. They require stupid amounts of compute to run.

369. hattmall ◴[] No.44578725{6}[source]
I don't know if you are joking or not, but people were talking about ChatGPT non-stop in like March of 2023 in my social group. Now it's far less frequently mentioned, basically never. In fact mostly if it is, it's in some form of a sarcastic joke or reply.
370. const_cast ◴[] No.44578731[source]
> > “most people agree that the output is trite and unpleasant to consume”

> That is a such a wild claim.

I think when he said "consume" he meant in terms of content consumption. You know, media - the thing that makes Western society go round. Movies, TV, music, books.

Would I watch an AI generated movie? No. What about a TV show? Uh... no. What about AI music? I mean, Spotify is trying to be tricky with that one, but no. I'd rather listen to Remi Wolf's 2024 Album "Big Ideas", which I thought was, ironically, less inspired than "Juno" but easily one of the best albums of the year.

ChatGPT is a useful interface, sure, but it's not entertaining. It's not high-quality. It doesn't provoke thought or offer us some solace in times of sadness. It doesn't spark joy or make me want to get up and dance.

371. const_cast ◴[] No.44578752{5}[source]
It's not that LLMs are bad, they're very useful. It's that the media they produce is, in fact, slop.

I want to watch Breaking Bad, not AI generated YouTube shorts. I want to listen to "On the Radio" by Donna Summer, not some Spotify generated piano solo. I want to read a high quality blog post about tech with a unique perspective, not an LLM summary of said blog post that removes all the charm.

The gap in quality, when it comes to entertainment, is truly astronomical. I mean, it's not even kind of close. I would expect literal children to produce content - after all, Mozart was a prodigy.

372. overfeed ◴[] No.44579330{6}[source]
> Price for ChatGPT will go down.

As will the response quality, while maintaining the same product branding. Users will accept whatever response OpenAI gives them under the "4o", "6p","9x" or whatever brand of the day, even as they ship-of-Theseus the service for higher margins. I'm yet to see an AI service with QoS guarantees, or even that the model weights & infrastructure won't be "optimized" over time to the customer's disadvantage.

373. closewith ◴[] No.44579420{4}[source]
99% or more of software developers behave in ways that would be inconceivable in actual engineering. That's not to say there aren't software engineers, but most developers aren't engineers and aren't held to that standard.
replies(1): >>44580556 #
374. closewith ◴[] No.44579435{4}[source]
In my companies, AI subscriptions and API access are now the biggest costs after salaries and taxes. Don't know what makes you think these services aren't attracting paid customers?
375. codebolt ◴[] No.44579514[source]
The crucial point is that we simply do not know yet if there is an inherent limitation in the reasoning capabilities of LLMs, and if so whether we are currently near to pushing up against them. It seems clear that American firms are still going to increase the amount of compute by a lot more (with projects like the Stargate factory), so time will tell if that is the only bottleneck to further progress. There might also still be methodological innovations that can push capabilities further.
376. disgruntledphd2 ◴[] No.44580145{8}[source]
Wow, that's either a bug or an incredibly incompetent advertiser.
replies(1): >>44581329 #
377. skydhash ◴[] No.44580556{5}[source]
Code is not physical. While computation errors can have real effects, a lot of orgs and people are resilient about them.
378. vrighter ◴[] No.44580581{9}[source]
I'd say the userbase has grown. You can't claim half a billion users and simultaneously say you're still trying to grow. This isn't a month-old technology now. And they still can't turn a profit. (edit: and by "you" i meant "they")
replies(1): >>44583444 #
379. prime_ursid ◴[] No.44580840{5}[source]
The way you phrased it reveals how these model providers have framed the conversation in their favor. Models don’t “think.”
380. ysavir ◴[] No.44581004{8}[source]
ahh, okay. I don't use the service, I didn't realize they had a dedicated UI for it. I assumed it was all just embedded in the text.
381. satyrun ◴[] No.44581011{3}[source]
lol you must not be looking for a white collar job right now then outside of IT.

The only thing that is over hyped is there is no white collar bloodbath but a white collar slow bleed out.

Not mass firing events but transition by attrition over time. A bleed out in jobs that don't get back filled and absolutely nothing in terms of hiring reserve capacity for the future.

My current company is a sinking ship, I suspect it will go under in the next two years so I have been trying to get off but there is absolutely no place to go.

In 2-3 years I expect to be unemployed and unemployable, needing to retrain to do something I have never done before.

What is on display in this thread is that human's are largely denial machines. We have to be otherwise we would be paralyzed by our own inevitable demise.

It is more comforting to believe everything is fine and the language models are just some kind of doge coin tech hype bullshit.

382. Wowfunhappy ◴[] No.44581093{6}[source]
You're right, I was underestimating the cost of running Facebook! $30B spent / ~3B users = ~$10 per user per year. I'd thought it would be closer to 10¢.

Do you know why it's so expensive? I'd thought serving html would be cheaper, particularly at Facebook's scale. Does the $30B include the cost of human content moderators? I also guess Facebook does a lot of video now, do you think that's it?

Also, even still, $10 per user has got to be an order of magnitude less than what OpenAI is spending on its free users, no?

383. rightbyte ◴[] No.44581166{3}[source]
> The so called "sonic boom" problem was vastly overhyped, as anyone who lives near an air force base can tell you.

The pilots don't shit where they eat. Ask some farmer a bit further away how many sheep die a year from panic instead.

384. hollerith ◴[] No.44581314{3}[source]
>as anyone who lives near an air force base can tell you.

In the US, the Air Force is simply not allowed to fly supersonic anywhere near a city or a suburb with only a few exceptions.

One exception is Edwards Air Force Base in the California desert: there are houses nearby, but the base (and supersonic warplanes) preceded the construction of the homes, so the reasoning is that the home builders and home buyers knew what they were buying into.

Another exception (quoting Google Gemini):

>From 1964 to 1966, the FAA and U.S. Air Force conducted supersonic flights over St. Louis and other cities like Oklahoma City to gauge public reaction to daily sonic booms. The goal was to understand public tolerance for commercial supersonic transport (SST) operations. Reactions in St. Louis, as elsewhere, were largely negative, contributing to the eventual ban on commercial supersonic flight over land in the U.S.

Have you have experienced sonic booms? I have (when my family visited West Germany in 1970) and I certainly would not want to be subjected to them regularly.

385. LtWorf ◴[] No.44581329{9}[source]
Are we surprised someone working for trump is not competent?
386. peder ◴[] No.44581530{4}[source]
Utter nonsense. The scale of disruption with LLMs is almost unfathomable. Every small business in the country has basically abandoned the big platforms and expensive enterprises for IT support, marketing and digital content creation, HR, legal...

Patients are having detailed conversations about their health with LLMs. Office visits for routine questions are plummeting.

Software is written almost entirely by LLMs, producing a greater volume of code in a fraction of the time.

Rapidly, we are approaching a point where there is no need for junior employees in most organizations. It's not industry-specific, it's universal. This will reshape corporate Big Four accounting, software engineering, and medicine because revenue will shift so dramatically.

This is not just some marginally more effective use of computing resources.

replies(1): >>44582612 #
387. amanaplanacanal ◴[] No.44582257{5}[source]
They had easy access to a large corpus of writing to train on, way larger than any human being trained their own language model on. I can't see where they are going to find a large corpus of physical interaction with reality to train that kind of model.
replies(1): >>44583105 #
388. amanaplanacanal ◴[] No.44582453{10}[source]
I think there implication is that because the teacher posted on Reddit, they are some kind of socialist, and therefore shouldn't be listened to. I guess their story would be worth listening to if it was posted on truth social instead?
replies(2): >>44583563 #>>44584362 #
389. amanaplanacanal ◴[] No.44582494{4}[source]
I guess all they have to do is start injecting ads into chatgpt responses. I'm sure that will be fine.
390. Gormo ◴[] No.44582612{5}[source]
Do you have even a shred of evidence to suggest that anything you're describing has actually taking place at scale?

"Software is written almost entirely by LLMs" is obviously false. "Every small business in the country has basically abandoned the big platforms and expensive enterprises for IT support" is obviously false. And how would you even know what medical conversations people are having with either their doctors or LLMs?

Everything you're saying sounds like unsubstantiated wishful thinking from someone who's taken a big gulp of the LLM Kool-Aid.

391. amanaplanacanal ◴[] No.44582880{8}[source]
I suspect most people using LLMs to write things are writing a bunch of boilerplate that nobody is going to read anyway. I could be wrong.
392. amanaplanacanal ◴[] No.44582912{4}[source]
Where is the training data for that? LLMs work because we already had tons of text that could be obtained cheaply. Where is the training data for physical reality?
393. awuji ◴[] No.44583007{4}[source]
You can already run a large LLM (like sonnet 3.5) locally on CPU with 128GB of ram which is <300 USD, but can be offset by swap space. Obviously, response speed is going to be slower, but I can't imagine people will pay much more than 20 USD for waiting 30-60 seconds longer for a response.

And obviously consumer hardware is already being more optimized for running models locally.

394. yunwal ◴[] No.44583105{6}[source]
cameras are fairly cheap these days.
395. og_kalu ◴[] No.44583444{10}[source]
>You can't claim half a billion users and simultaneously say you're still trying to grow.

You can if you're still growing. ChatGPT is the 5th most visited site on the planet yes, but it is still growing hundreds of millions of visits with every passing month.

They aren't turning a profit because they aren't monetizing the vast majority of subscribers in any way (not even ads). LLM inference is cheap enough for ads to be viable.

396. intended ◴[] No.44583563{11}[source]
Ah! Nice point on truth social.
replies(1): >>44584513 #
397. og_kalu ◴[] No.44583580{5}[source]
Seems like it's only meaningless if you ignore basic context.
replies(1): >>44585340 #
398. matthewdgreen ◴[] No.44583653{8}[source]
Meta has a massively profitable social media business with an impenetrable network effect, so they're using that to subsidize the research. Whether that's a good decision or not is above my paygrade, but it's sustainable until something changes with the social media market.

I don't know what "moving the goalposts" means. Why were the goalposts there in the first place? The interesting questions here are whether OpenAI can sustain their current cost model long-term, and whether the revenue stream is sustainable without the costs. We'll see, I guess! It's fascinating.

399. frozenseven ◴[] No.44584237{10}[source]
Look, another one! Twist it however you want, I'm not going to accept the idea that far-lefty Reddit is some impartial representation of what teaching is or what the average person thinks of AI.
400. frozenseven ◴[] No.44584362{11}[source]
>they are some kind of socialist

Yes, that is accurate.

>I guess their story would be worth listening to if it was posted on truth social instead?

No, I don't take anti-AI nonsense seriously in the first place. That aside, the main point here was that Reddit has a very strong political leaning. If anyone tried to insist that the politics of Truth Social is irrelevant, you'd immediately call it out.

replies(2): >>44584713 #>>44589065 #
401. frozenseven ◴[] No.44584513{12}[source]
Nah, misses the entire point of what I was saying.

But thanks for recognizing that Truth Social has a noticeable political leaning. So close, yet so far.

402. amanaplanacanal ◴[] No.44584713{12}[source]
So instead of addressing the actual substance, you dismiss it because of your assumption of their political leanings.

Good luck navigating the world, I guess.

replies(1): >>44585313 #
403. yieldcrv ◴[] No.44585127[source]
slower, no fast option, no smoking in the cabins, less leg room, but with TVs plastered on the back of every chair, sometimes

its actually kind of scary to think of a world where generative AI in the cloud goes away due to costs, in favor of some other lesser chimera version that can't currently be predicted

but good news is that locally run generative AI is still getting better and better with fewer and fewer resources consumed to use

404. bluefirebrand ◴[] No.44585221{6}[source]
> An LLM system, on the other hand, can fail because you moved some punctuation around

An LLM system can fail without changing anything, it could just fail more or less randomly without any way to diagnose why it happened

405. frozenseven ◴[] No.44585313{13}[source]
They have the same political leanings as you. I notice these things.

>Good luck navigating the world, I guess.

Thanks. And a hardy "F you" to you too.

406. JohnMakin ◴[] No.44585340{6}[source]
What basic context is being ignored? Here's how the thread has gone:

"chatGPT has the fastest growing userbase in history which shows users really like the output!"

This unsourced (and wrong) claim was offered in rebuttal to another post saying people don't like the output of LLM's. This rebuttal offers DAU/MAU as a metric of how much people like the app, I presume, and thus the output of the app. Besides that being a wild jump on its own, it's incorrect. As I pointed out - threads almost immediately beat that DAU/MAU record, and I'd offer a claim it hasn't exactly been a tremendous success either in popularity or monetarily. Pointing out that they got that DAU/MAU by registering their own users to it is precisely the point that is being made - this metric is a meaningless gauge of how popular an app is, and especially when viewed from the context of this argument, which is whether the popularity of the app (as it relates to DAU/MAU growth) also suggests people love consuming the output of it.

No offense, but are you sure you're following this conversation?

407. mxschumacher ◴[] No.44587315{6}[source]
well, I'm surprised by the sky-high valuations I see in the context of the problem I have outlined above. This is great for consumers, sure.

The AI bubble is so big that if it pops, it will have dramatic effects on the economy.

408. habinero ◴[] No.44588993{4}[source]
Yeah, basically. Nobody wanted to pay $12,000 to be in a plane for three hours when they could pay ~$1200 to be in one for six hours. Plus, they used up a lot of fuel. That made them real vulnerable to oil price spikes.

Contrast that with modern widebody jets, which fly ~300 people plus paid cargo on much more fuel-efficient engines.

409. habinero ◴[] No.44589065{12}[source]
It really doesn't lol.

I don't get the reactionary right's hysteria about Reddit. It's so clearly not true it's just silly.

It's like when my brother let my little cousin watch a scary movie and she had hysterics about scary things for days. Y'all tell each other ghost stories and convince yourselves it's real.