Most active commenters
  • furyofantares(5)
  • namaria(4)
  • TeMPOraL(4)
  • shakna(3)

←back to thread

279 points nnx | 22 comments | | HN request time: 0.001s | source | bottom
Show context
ChuckMcM ◴[] No.43543501[source]
This clearly elucidated a number of things I've tried to explain to people who are so excited about "conversations" with computers. The example I've used (with varying levels of effectiveness) was to get someone to think about driving their car by only talking to it. Not a self driving car that does the driving for you, but telling it things like: turn, accelerate, stop, slow down, speed up, put on the blinker, turn off the blinker, etc. It would be annoying and painful and you couldn't talk to your passenger while you were "driving" because that might make the car do something weird. My point, and I think it was the author's as well, is that you aren't "conversing" with your computer, you are making it do what you want. There are simpler, faster, and more effective ways to do that then to talk at it with natural language.
replies(11): >>43543657 #>>43543721 #>>43543740 #>>43543791 #>>43543890 #>>43544393 #>>43544444 #>>43545239 #>>43546342 #>>43547161 #>>43551139 #
phyzix5761 ◴[] No.43543740[source]
You're onto something. We've learned to make computers and electronic devices feel like extensions of ourselves. We move our bodies and they do what we expect. Having to switch now to using our voice breaks that connection. Its no longer an extension of ourselves but a thing we interact with.
replies(1): >>43543986 #
1. namaria ◴[] No.43543986[source]
Two key things that make computers useful, specificity and exactitude, are thrown out of the window by interposing NLP between the person and the computer.

I don't get it at all.

replies(3): >>43544143 #>>43546069 #>>43546495 #
2. TeMPOraL ◴[] No.43544143[source]

   [imprecise thinking]
         v <--- LLMs do this for you
   [specific and exact commands]
         v
   [computers]
         v
   [specific and exact output]
         v <--- LLMs do this for you
   [contextualized output]
In many cases, you don't want or need that. In some, you do. Use right tool for the job, etc.
replies(2): >>43544577 #>>43551590 #
3. shakna ◴[] No.43544577[source]
I don't think they give a specific and exact output, considering how nondeterminism plays a role in most models.
replies(2): >>43545096 #>>43545818 #
4. furyofantares ◴[] No.43545096{3}[source]
The diagram you're replying to agrees with this.
replies(1): >>43545198 #
5. walthamstow ◴[] No.43545198{4}[source]
Does it? The way I'm reading it, the first step is LLM turning human imprecise thinking into specific and exact commands
replies(1): >>43545359 #
6. furyofantares ◴[] No.43545359{5}[source]
That's true, but that is the input section of the diagram, not the output section where [specific and exact output] is labeled, so I believe there was legitimate confusion I was responding to.

To your point, which I think is separate but related, that IS a case where LLMs are good at producing specific and exact commands. The models + the right prompt are pretty reliable at tool calling by themselves, because you give them a list of specific and exact things they can do. And they can be fully specific and exact at inference time with constrained output (although you may still wish it called a different tool.)

replies(1): >>43546984 #
7. TeMPOraL ◴[] No.43545818{3}[source]
I'll need to work on the diagram to make it clearer next time.

What it's trying to communicate is, in general, a human operating a computer has to turn their imprecise thinking into "specific and exact commands", and subsequently, understand the "specific and exact output" in whatever terms they're thinking off, prioritizing and filtering out data based on situational context. LLMs enter the picture in two places:

1) In many situations, they can do the "imprecise thinking" -> "specific and exact commands" step for the user;

2) In many situations, they can do the "specific and exact output" -> contextualized output step for the user;

In such scenarios, LLMs are not replacing software, they're being slotted as intermediary between user and classical software, so the user can operate closer to what's natural for them, vs. translating between it and rigid computer language.

This is not applicable everywhere, but then, this is also not the only way LLMs are useful - it's just one broad class of scenarios in which they are.

8. brookst ◴[] No.43546069[source]
I also don’t like command like interfaces for all things, but there are cases where they excel, or where they are necessary due to technical constraints. But when the man page for a simple command runs to 10 screens of options I sometimes wonder.
9. grbsh ◴[] No.43546495[source]
Why would you ever hire a human to perform some task for you in a company? They're known for having problems with ambiguity and precision in communication.

Humans require a lot of back and forth effort for "alignment" with regular "syncs" and "iterations" and "I'll get that to you by EOD". If you approach the potential of natural interfaces with expectations that frame them the same way as 2000s era software, you'll fail to be creative about new ways humans interact with these systems in the future.

10. shakna ◴[] No.43546984{6}[source]
The tool may not even exist. LLMs are really terrible at admitting where the limits of the training are. They will imagine a tool into being. They will also claim the knowledge is within their realm, when it isn't.
replies(1): >>43548118 #
11. furyofantares ◴[] No.43548118{7}[source]
At inference time you can constrain output to a strict json schema that only includes valid tools.
replies(2): >>43548907 #>>43550784 #
12. shakna ◴[] No.43548907{8}[source]
That would only be possible, if you could prevent hallucinations from ever occurring. Which you can't. Even if you supply a strict schema, the model will sometimes act outside of it - and infer the existence of "something similar".
replies(1): >>43549754 #
13. furyofantares ◴[] No.43549754{9}[source]
That's not true. You say the model will sometimes act outside of the schema, but models don't act at all, they don't hallucinate by themselves, they don't produce text at all, they do all of this in conjunction with your inference engine.

The model's output is a probability for every token. Constrained output is a feature of the inference engine. With a strict schema the inference engine can ignore every token that doesn't adhere to the schema and select the top token that does adhere to the schema.

14. xigoi ◴[] No.43550784{8}[source]
Just because the answer adheres to the schema does not mean that it’s correct.
replies(1): >>43551241 #
15. furyofantares ◴[] No.43551241{9}[source]
Yes, we've been discussing "specific and exact" output. As I said, you might wish it called at different tool; nothing in this discussion is addressing that.
16. namaria ◴[] No.43551590[source]
Despite feeling like a "let me draw it for you" answer is a tad condescending, I want to address something here.

This would be great if LLMs did not tend to output nonsense. Truly it would be grand. But they do. So it isn't. It's wasting resources hoping for a good outcome and risking frustration, misapprehensions, prompt injection attacks... It's non-deterministic algorithms hoping P=NP, except instead of branching at every decision you're doing search by tweaking vectors whose values you don't even know and whose influence on the outcome is impossible to foresee.

Sure, a VC subsidized LLM is a great way to make CVs in LaTeX (I do it all the time), translating text, maybe even generating some code if you know what you need and can describe it well. I will give you that. I even created a few - very mediocre - songs. Am I contradicting myself? I don't think I am, because I would love to live in a hotel if I only had to pay a tiny fraction of the cost. But I would still think that building hotels would be a horrible way to address the housing crisis in modern metropolises.

replies(2): >>43552390 #>>43552872 #
17. TeMPOraL ◴[] No.43552390{3}[source]
> Despite feeling like a "let me draw it for you" answer is a tad condescending, I want to address something here.

I didn't mean it to be condescending - though I can see how it can come across as such. FWIW, I opted for a diagram after I typed half a page worth of "normal" text and realized I'm still not able to elucidate my point - so I deleted it and drew something matching my message more closely.

> This would be great if LLMs did not tend to output nonsense. Truly it would be grand. But they do. So it isn't.

I find this critique to be tiring at this point - it's just as wrong as assuming LLMs work perfectly and all is fine. Both views are too definite, too binary. In reality, LLMs are just non-deterministic - that is, they have an error rate. How big it is, and how small can it get in practice for a given tasks - those are the important questions.

Pretty much every aspect of computing is only probabilistically correct - either because the algorithm is explicitly so (UUIDs and primality testing, for starters), or just because it runs on real hardware, and physics happen. Most people get away with pretending that our systems are either correct or not, but that's only possible because the error rate is low enough. But it's never that low by accident - it got pushed there by careful design at every level, hardware and software. LLMs are just another probabilistically correct system that, over time, we'll learn how to use in ways that gets the error rate low enough to stop worrying about it.

How can we get there - now, that is an interesting challenge.

replies(1): >>43554183 #
18. fragmede ◴[] No.43552872{3}[source]
Using the word hotel has a lot of baggage, but having a large quantity of rooms for rent, for cheap, with a bathroom but no dedicated kitchen would be amazing for the housing crisis. If they were high quality and sound isolated, with high speed elevators, and communal spaces for residents, it could work. I'm not an architect though.
replies(2): >>43553074 #>>43554142 #
19. saratogacx ◴[] No.43553074{4}[source]
In the early 2000's there was a push for building apodments which were a room, bathroom, and shared kitchen area. Some people liked them but it isn't for everyone.
20. namaria ◴[] No.43554142{4}[source]
You're describing student housing and if you ever lived in one you'd know how bad of an idea you're musing with.
replies(1): >>43554270 #
21. namaria ◴[] No.43554183{4}[source]
Natural language has a high entropy floor. It's a very noisy channel. This isn't anything like bit flipping or component failure. This is a whole different league. And we've been pouring outrageous amounts of resources into diminishing returns. OpenAI keeps touting AGI and burning cash. It's being pushed everywhere as a silver bullet, helping spin lay offs as a good thing.

LLMs are cool technology sure. There's a lot of cool things in the ML space. I love it.

But don't pretend like the context of this conversation isn't the current hype and that it isn't reaching absurd levels.

So yeah we're all tired. Tired of the hype, of pushing LLMs, agents, whatever, as some sort of silver bullet. Tired of the corporate smoke screen around it. NLP is still a hard problem, we're nowhere near solving it, and bolting it on everything is not a better idea now than it was before transformers and scaling laws.

On the other hand my security research business is booming and hey the rational thing for me to say is: by all means keep putting NLP everywhere.

22. TeMPOraL ◴[] No.43554270{5}[source]
He's also describing hotels, and aparthostels, and officers' quarters on a ship and bunch of other stuff. The devil is in the details - specifically, how much it costs to rent per sqm, and what stops the price from going up to the point it forces multiple people to share the room? What stops the landlords from subdividing the rooms further and renting them out apiece? What stops already shoddy construction from getting even worse?

Those are the big challenges of housing. Not just how many units there are, but what they are, and how much the "how many" is plain cheating.