←back to thread

LLM Inevitabilism

(tomrenner.com)
1613 points SwoopsFromAbove | 1 comments | | HN request time: 0.206s | source
Show context
Animats ◴[] No.44568076[source]
There may be an "LLM Winter" as people discover that LLMs can't be trusted to do anything. Look for frantic efforts by companies to offload responsibility for LLM mistakes onto consumers. We've got to have something that has solid "I don't know" and "I don't know how to do this" outputs. We're starting to see reports of LLM usage having negative value for programmers, even though they think it's helping. Too much effort goes into cleaning up LLM messes.
replies(5): >>44568232 #>>44568321 #>>44568785 #>>44570451 #>>44578122 #
1. 01100011 ◴[] No.44578122[source]
I find these LLM doomer takes as silly as LLM maximalist takes.

LLMs are literally performing useful functions today and they're not going away. Are they AGI? No, but so what?

There is waaay too much projecting and philosophizing going on in these comments and not enough engineering-minded comments from objective observers.

Is AI hyped? Sure. Are LLMs overshadowing other approaches? Sure. Are LLMs inefficient? Somewhat. Do they have problems like hallucinations? Yes. Do they produce useful output? Yes.