Most active commenters
  • ramesh31(3)

←back to thread

419 points serjester | 12 comments | | HN request time: 0.398s | source | bottom
1. ramesh31 ◴[] No.43536049[source]
More capability, less reliability please. I want something that can achieve superhuman results 1 out of 10 times, not something that gives mediocre human results 9 out of 10 times.

All of reality is probabilistic. Expecting that to map deterministically to solving open ended complex problems is absurd. It's vectors all the way down.

replies(5): >>43536239 #>>43536358 #>>43536391 #>>43537262 #>>43537360 #
2. soulofmischief ◴[] No.43536239[source]
Stability is the bedrock of the evolution of stable systems. LLMs will not democratize software until an average person can get consistently decent and useful results without needing to be a senior engineer capable of a thorough audit.
replies(1): >>43536325 #
3. ramesh31 ◴[] No.43536325[source]
>Stability is the bedrock of the evolution of stable systems.

So we also thought with AI in general, and spent decades toiling on rules based systems. Until interpretability was thrown out the window and we just started letting deep learning algorithms run wild with endless compute, and looked at the actual results. This will be very similar.

replies(3): >>43537298 #>>43537320 #>>43537703 #
4. deprave ◴[] No.43536358[source]
What would be a superhuman result for booking a flight?
replies(1): >>43536585 #
5. Jianghong94 ◴[] No.43536391[source]
Superhuman results 1/10 are, in fact, a very strong reliability guarantee (maybe not up to today's nth 9 decimal standard that we are accustomed to, but probably much higher than any agent in real-world workflow).
6. mjmsmith ◴[] No.43536585[source]
10% of the time the seat on either side of you is empty, 90% of the time you land in the wrong country.
7. klabb3 ◴[] No.43537262[source]
Reality is probabilistic yes but it’s not black box. We can improve our systems by understanding and addressing the flaws in our engineering. Do you want probabilistic black-box banking? Flight controls? Insurance?

”It works when it works” is fine when stakes are low and human is in the loop, like artwork for a blog post. And so in a way, I agree with you. AI doesn’t belong in intermediate computer-to-computer interactions, unless the stakes are low. What scares me is that the AI optimists are desperately looking to apply LLMs to domains and tasks where the cost of mistakes are high.

8. skydhash ◴[] No.43537298{3}[source]
Rules based systems are quite useful, not for interacting with an untrained human, but for getting things done. Deep learning can be good at exploring the edges of a problem space, but when a solution is found, we can actually get to the doing part.
9. klabb3 ◴[] No.43537320{3}[source]
This can be explained easily – there are simply some domains that were hard to model, and those are the ones where AI is outperforming humans. Natural language is the canonical example of this. Just because we focus on those domains now due to the recent advancements, doesn’t mean that AI will be better at every domain, especially the ones we understand exceptionally well. In fact, all evidence suggests that AI excels at some tasks and struggles with others. The null hypothesis should be that it continues to be the case, even as capability improves. Not all computation is the same.
10. recursive ◴[] No.43537360[source]
> Expecting that to map deterministically to solving open ended complex problems is absurd.

TCP creates an abstraction layer with more reliability than what it's built on. If you can detect failure, you can create a retry loop, assuming you can understand the rules of the environment you're operating in.

replies(1): >>43541202 #
11. soulofmischief ◴[] No.43537703{3}[source]
Stability and probability are orthogonal concepts. You can have stable probabilistic systems. Look no further than our own universe, where everything is ultimately probabilistic and not "rules-based".
12. ramesh31 ◴[] No.43541202[source]
>If you can detect failure, you can create a retry loop, assuming you can understand the rules of the environment you're operating in

Indeed, this is what makes autonomous agentic tool using systems robust as well. Those retry loops become ad-hoc where needed, and the agent can self correct based on error responses, compared to a defined workflow that would get stuck in said loop if it couldn't figure things out, or just error out the whole process.