Most active commenters
  • eric-burel(7)
  • __loam(4)

←back to thread

LLM Inevitabilism

(tomrenner.com)
1611 points SwoopsFromAbove | 36 comments | | HN request time: 0.005s | source | bottom
Show context
lsy ◴[] No.44568114[source]
I think two things can be true simultaneously:

1. LLMs are a new technology and it's hard to put the genie back in the bottle with that. It's difficult to imagine a future where they don't continue to exist in some form, with all the timesaving benefits and social issues that come with them.

2. Almost three years in, companies investing in LLMs have not yet discovered a business model that justifies the massive expenditure of training and hosting them, the majority of consumer usage is at the free tier, the industry is seeing the first signs of pulling back investments, and model capabilities are plateauing at a level where most people agree that the output is trite and unpleasant to consume.

There are many technologies that have seemed inevitable and seen retreats under the lack of commensurate business return (the supersonic jetliner), and several that seemed poised to displace both old tech and labor but have settled into specific use cases (the microwave oven). Given the lack of a sufficiently profitable business model, it feels as likely as not that LLMs settle somewhere a little less remarkable, and hopefully less annoying, than today's almost universally disliked attempts to cram it everywhere.

replies(26): >>44568145 #>>44568416 #>>44568799 #>>44569151 #>>44569734 #>>44570520 #>>44570663 #>>44570711 #>>44570870 #>>44571050 #>>44571189 #>>44571513 #>>44571570 #>>44572142 #>>44572326 #>>44572360 #>>44572627 #>>44572898 #>>44573137 #>>44573370 #>>44573406 #>>44574774 #>>44575820 #>>44577486 #>>44577751 #>>44577911 #
1. eric-burel ◴[] No.44568416[source]
Developers haven't even started extracting the value of LLMs with agent architectures yet. Using an LLM UI like open ai is like we just figured fire and you use it to warm you hands (still impressive when you think about it, but not worth the burns), while LLM development is about building car engines (here is you return on investment).
replies(8): >>44568647 #>>44568953 #>>44568969 #>>44569090 #>>44569602 #>>44569667 #>>44570029 #>>44570985 #
2. clarinificator ◴[] No.44568647[source]
Every booster argument is like this one. $trite_analogy triumphant smile
replies(1): >>44568826 #
3. Jensson ◴[] No.44568953[source]
> Developers haven't even started extracting the value of LLMs with agent architectures yet

There are thousands of startups doing exactly that right now, why do you think this will work when all evidence points towards it not working? Or why else would it not already have revolutionized everything a year or two ago when everyone started doing this?

replies(2): >>44569274 #>>44571260 #
4. pydry ◴[] No.44568969[source]
Theyre doing it so much it's practically a cliche.

There are underserved areas of the economy but agentic startups is not one.

5. mns ◴[] No.44569090[source]
>evelopers haven't even started extracting the value of LLMs with agent architectures yet.

Which is basically what? The infinite monkey theorem? Brute forcing solutions for problems at huge costs? Somehow people have been tricked to actually embrace and accept that now they have to pay subscriptions from 20$ to 300$ to freaking code? How insane is that, something that was a very low entry point and something that anyone could do, is now being turned into some sort of classist system where the future of code is subscriptions you pay for companies ran by sociopaths who don't care that the world burns around them, as long as their pockets are full.

replies(3): >>44569263 #>>44569305 #>>44570800 #
6. rrr_oh_man ◴[] No.44569198{3}[source]
Throwing genetic fallacies around instead of engaging with the comment at hand... :)
7. frizlab ◴[] No.44569263[source]
I cannot emphasize how much I agree with this comment. Thank you for writing it, I would never have had written it as well.
8. eric-burel ◴[] No.44569274[source]
Most of them are a bunch of prompts and don't even have actual developers. For the good reason that there is no training system yet and the wording of how you call the people that build these system isn't even there or clearly defined. Local companies haven't even setup a proper internal LLM or at least a contract with a provider. I am in France so probably lagging behind USA a bit especially NY/SF but the word "LLM developer" is just arriving now and mostly under the pressure of isolated developers and companies like me. This feel really really early stage.
replies(3): >>44569675 #>>44569759 #>>44570203 #
9. eric-burel ◴[] No.44569305[source]
I don't have a subscription not even an Open AI account (mostly cause they messed up their google account system). You can't extract value of an LLM by just using the official UI, you just scratch the surface of how they work. And yet there aren't much developers able to actually build an actual agent architecture that does deliver some value. I don't include the "thousands" of startups that are clearly suffer from a signaling bias: they don't exist in the economy and I don't care about them like at all in my reasonning. I am talking about actual LLM developers that you can recruit locally the same way you recruit a web developer today, and that can make sense out of "frontier" LLM garbage talk by using proper architectures. These devs are not there yet.
10. camillomiller ◴[] No.44569602[source]
>> Developers haven't even started extracting the value of LLMs with agent architectures yet.

What does this EVEN mean? Do words have any value still, or are we all just starting to treat them as the byproduct of probabilistic tokens?

"Agent architectures". Last time I checked an architecture needs predictability and constraints. Even in software engineering, a field for which the word "engineering" is already quite a stretch in comparison to construction, electronics, mechanics.

Yet we just spew the non-speak "Agentic architectures" as if the innate inability of LLMs in managing predictable quantitative operations is not an unsolved issue. As if putting more and more of these things together automagically will solves their fundamental and existential issue (hallucinations) and suddenly makes them viable for unchecked and automated integration.

replies(2): >>44571385 #>>44578138 #
11. __loam ◴[] No.44569667[source]
3 years into automating all white collar labor in 6 months.
replies(2): >>44571334 #>>44581011 #
12. __loam ◴[] No.44569675{3}[source]
The smartest and most well funded people on the planet have been trying and failing to get value out of this technology for years and the best we've come up with so far is some statistically unreliable coding assistants. Hardly the revolution its proponents keep eagerly insisting we're seeing.
replies(3): >>44570282 #>>44571073 #>>44571482 #
13. aquariusDue ◴[] No.44569759{3}[source]
Between the ridiculously optimistic and the cynically nihilistic I personally believe there is some value that extremely talented people at huge companies can't really provide because they're not in the right environment (too big a scale) but neither can grifters packaging a prompt in a vibecoded app.

In the last few months the building blocks for something useful for small companies (think less than 100 employees) have appeared, now it's time for developers or catch-all IT at those companies and freelancers serving small local companies to "up-skill".

Why do I believe this? Well for a start OCR became much more accessible this year cutting down on manual data entry compared to tesseract of yesteryear.

14. ◴[] No.44570029[source]
15. liveoneggs ◴[] No.44570203{3}[source]
is there a non-prompt way to interact with LLMs?
replies(1): >>44570271 #
16. eric-burel ◴[] No.44570271{4}[source]
In an agentic setup the value is half the prompts half how you plug them together. I am opposing for instance a big prompt that is supposed to write a dissertation vs a smart web scraper that builds a knowledge graph out of sources and outputs a specialized search engine for your task. The former is a free funny intern, the latter is growth percentage visible in the economy.
17. eric-burel ◴[] No.44570282{4}[source]
They try to get value at their scale, which is tough. Your local SME definitely sees value in an embedding-based semantic search engine over their 20 years of weird unstructured data.
18. pj_mukh ◴[] No.44570800[source]
I pay $300 to fly from SF to LA when I could've just walked for free. Its true. How classist!
19. dvfjsdhgfv ◴[] No.44570985[source]
> Developers haven't even started extracting the value of LLMs with agent architectures yet.

For sure there is a portion of developers who don't care about the future, are not interested in current developements and just live as before hoping nothing will change. But the rest already gave it a try and realized tools like Claude Code can give excellent results for small codebases to fail miserably at more complex tasks with the net result being negative as you get a codebase you don't understand, with many subtle bugs and inconsistencies created over a few days you will need weeks to discover and fix.

replies(1): >>44571363 #
20. liveoneggs ◴[] No.44571073{4}[source]
my company has already fired a bunch of people in favor of LLMs so they are realizing all kinds of value
replies(5): >>44571223 #>>44571241 #>>44571416 #>>44572040 #>>44574184 #
21. dasil003 ◴[] No.44571223{5}[source]
I don’t know your company but this thinking doesn’t necessarily follow logically. In a large company the value of developers is not distributed evenly across people and time, and also has a strong dependency on market realities in front of them.

While it’s true that lots of companies are getting some value out of LLMs, a much larger number are using them as an excuse for layoffs they would have wanted to do anyway—LLMs are just a golden opportunity to tie in an unmitigated success narrative.

replies(1): >>44574544 #
22. Capricorn2481 ◴[] No.44571241{5}[source]
Only as much as replacing all your devs with a frog is "realizing value"
23. ReptileMan ◴[] No.44571260[source]
>Or why else would it not already have revolutionized everything a year or two ago when everyone started doing this?

The internet needed 20 years to take over the world. All of the companies of the first dot com bust are in the past. The tech is solid.

24. erfgh ◴[] No.44571334[source]
It's 3 years away.
25. eric-burel ◴[] No.44571363[source]
This is a bit developer centric, I am much more impressed by the opportunities I see in consulting rather than applying LLMs to dev tasks. And I am still impressed by the code it can output eventhough we are still in the funny intern stage in this area.
replies(1): >>44573813 #
26. eric-burel ◴[] No.44571385[source]
This means I believe we currently underuse LLM capabilities and their empirical nature makes it difficult to assess their limitations without trying. I've been studying LLMs from various angles during a few months before coming to this conclusion, as an experienced software engineer and consultant. I must admit it is however biased towards my experience as an SME and in my local ecosystem.
27. SketchySeaBeast ◴[] No.44571416{5}[source]
I imagine they HOPE they'll realize value. A lot of people are acting on what might be, rather than what is, which makes sense given that the AI "thought leaders" (CEOs with billions invested that need to start turning a profit) are all promising great things soon™.
28. SpicyLemonZest ◴[] No.44571482{4}[source]
The best they've come up with is the LLM chatbot, which both OpenAI and Anthropic have as their flagship product because many people find it extremely valuable. Many people I know routinely use ChatGPT to help them write things, even those who were already good at writing, and if you don't think that's true at your workplace I strongly suspect it's because people aren't telling you about it.
replies(1): >>44574220 #
29. sensanaty ◴[] No.44572040{5}[source]
So has mine, and quite predictably our product has gone into the shitter and breaks constantly, requiring reverts almost daily. They've armed a couple of Juniors with Cursor and given them the workload of all those people they fired / have quit since the firings, some of which have been at the company for years and held a lot of institutional knowledge that is now biting them in the ass.

Now sure, "Just don't fire the useful people and get rid of the juniors and supercharge the good devs with AI tooling" or whatever, except the whole reason the C-level is obsessed with this AI shit is because they're sold on the idea of replacing their most expensive asset, devs, because they've been told by people who sell AI as a job that it can replace those pesky expensive devs and be replaced by any random person in the company prompting up a storm and vibecoding it all.

Churn rates are up, we're burning unfathomable amounts of money on the shitty AI tooling and the project has somehow regressed after we've finally managed to get a good foothold on it and start making real progress for once. Oh and the real funny part is they're starting to backpedal a bit and have tried to get some people back in.

I expect to hear a LOT more of this type of thing happening in the near future. As the idiots in charge start slowly realizing all the marketing sold to them on LinkedIn or wherever the fuck it is they get these moronic ideas from are literal, actual literal lies.

30. Gormo ◴[] No.44573813{3}[source]
> I am much more impressed by the opportunities I see in consulting rather than applying LLMs to dev tasks.

I expect there'll be a lot of consulting work in the near future in cleanup and recovery from LLM-generated disasters.

31. __loam ◴[] No.44574184{5}[source]
Yeah callousness does seem to be the leaking area of improvement.
32. __loam ◴[] No.44574220{5}[source]
Great, we've got mediocre writing from unprofitable companies that are subsidizing the cost of this technology.
replies(1): >>44575138 #
33. thewebguyd ◴[] No.44574544{6}[source]
> a much larger number are using them as an excuse for layoffs they would have wanted to do anyway

It's a simple formula. Layoffs because of market conditions or company health = stock price go down. Layoffs because "AI took the jobs" = stock price go up.

34. SpicyLemonZest ◴[] No.44575138{6}[source]
What specifically do you find to be mediocre? I feel like LLMs write better than most people I know, myself included.

There could be a mismatch on what the state of the art really is these days. In my experience, since the release of GPT-4 and especially 4o, ChatGPT has been able to do the vast majority of concrete things people tell me it can't do.

35. ogogmad ◴[] No.44578138[source]
Hallucinations might get solved by faster, cheaper and more accurate, vision and commonsense-physics models. Hypothesis: Hallucinations are a problem only because physical reality isn't text. Once people switch to models that predict physical states instead of missing text, then we'll have domestic robots and lower hallucination rates.
36. satyrun ◴[] No.44581011[source]
lol you must not be looking for a white collar job right now then outside of IT.

The only thing that is over hyped is there is no white collar bloodbath but a white collar slow bleed out.

Not mass firing events but transition by attrition over time. A bleed out in jobs that don't get back filled and absolutely nothing in terms of hiring reserve capacity for the future.

My current company is a sinking ship, I suspect it will go under in the next two years so I have been trying to get off but there is absolutely no place to go.

In 2-3 years I expect to be unemployed and unemployable, needing to retrain to do something I have never done before.

What is on display in this thread is that human's are largely denial machines. We have to be otherwise we would be paralyzed by our own inevitable demise.

It is more comforting to believe everything is fine and the language models are just some kind of doge coin tech hype bullshit.