←back to thread

337 points throw0101c | 8 comments | | HN request time: 1.951s | source | bottom
Show context
oytis ◴[] No.44609364[source]
I just hope when (if) the hype is over, we can repurpose the capacities for something useful (e.g. drug discovery etc.)
replies(16): >>44609452 #>>44609461 #>>44609463 #>>44609471 #>>44609489 #>>44609580 #>>44609632 #>>44609635 #>>44609712 #>>44609785 #>>44609958 #>>44609979 #>>44610227 #>>44610522 #>>44610554 #>>44610755 #
1. alphazard ◴[] No.44609712[source]
The rest of the world has not caught up to current LLM capabilities. If it all stopped tomorrow and we couldn't build anything more intelligent than what we have now: there would be years of work automating away toil across various industries.
replies(3): >>44609814 #>>44609887 #>>44614001 #
2. Terr_ ◴[] No.44609814[source]
Creating oodles of new jobs in internally QAing LLM results, or finding and suing companies for reckless outcomes. :p
replies(1): >>44609973 #
3. sterlind ◴[] No.44609887[source]
my experience using LLM-powered tools (e.g. copilot in agent mode) has been underwhelming. like, shockingly so. like not cd-ing to the wrong dir where a script is located, and getting lost, disregarding my instructions to run ./tests.ps1 and running `dotnet test`, writing syntactically incorrect scripts and failing to correct them, particularly being overwhelmed by verbose logs. sometimes it even fails to understand the semantic meaning of my prompts.

whereas my experience describing my problem and actually asking the AI is much, much smoother.

I'm not convinced the "LLM+scaffolding" paradigm will work all that well. sanity degrades with context length, and even the models with huge context windows don't seem to use it all that effectively. RAG searches often give lackluster results. the models fundamentally seem to do poorly with using commands to accomplish tasks.

I think fundamental model advances are needed to make most things more than superficially automatable: better planning/goal-directed behavior, a more organic connection to RAG context, automatic gym synthesis, and RL-based fine tuning (that holds up to distribution shift.)

I think that will come, but I think if LLMs plateau here they won't have much more impact than Google Search did in the '90s.

replies(2): >>44610156 #>>44612609 #
4. alphazard ◴[] No.44609973[source]
As long as liability is clearly assigned, it doesn't have an economic impact. The ambiguity of liability is what creates negative economic impact. Once it's assigned initially through law, then it can be reassigned via contract in exchange for cash to ensure the most productive outcome.

e.g. if OpenAI is responsible for any damages caused by ChatGPT then the service shuts down until you waive liability and then it's back up. Similarly if companies are responsible for the chat bots they deploy then they can buy insurance or put up guard rails around the chat bot, or not use it.

replies(1): >>44610453 #
5. break_the_bank ◴[] No.44610156[source]
I’m curious which was the model you used when you ran into the cd-ing bug?

I’d give building with sonnet 4 a fair shot. It’s really good, not accurate all the time but pretty good.

6. Terr_ ◴[] No.44610453{3}[source]
> As long as liability is clearly assigned, it doesn't have an economic impact

In a reality with perfect knowledge, complete laws always applied, and populated by un-bankrupt-able immortals with infinite lines of credit, yes. :P

7. fragmede ◴[] No.44612609[source]
> won't have much more impact than Google Search did in the '90s.

Given that Google IPOd in 99, and is one of the biggest tech companies in the world, I'm not sure what you mean by that.

8. Winsaucerer ◴[] No.44614001[source]
I'm one of those people who thinks simultaneously that (a) current AI cannot replace developers, it just isn't good enough (and I don't think it's good for it to write much code), and (b) AI is simply an incredible invention and will go down as one of the top 5 or 10 in history.

I've said the same thing as you, that there is a LOT left to be done with current AI capabilities, and we've barely scratched the surface.