←back to thread

46 points petethomas | 2 comments | | HN request time: 0.408s | source
Show context
Azkron ◴[] No.44397822[source]
| "Not even AI’s creators understand why these systems produce the output they do."

I am so tired of this "NoBody kNows hoW LLMs WoRk". It fucking software. Sophisticated probability tables with self correction. Not magic. Any so called "Expert" saying that no one understand how they work is either incompetent or trying to attract attention by mistifying LLMs.

replies(5): >>44397894 #>>44397909 #>>44397952 #>>44398104 #>>44398578 #
lappa ◴[] No.44397894[source]
This isn't suggesting no one understands how these models are architected, nor is anyone saying that SDPA / matrix multiplication isn't understood by those who create these systems.

What's being said is that the result of training and the way in which information is processed in latent space is opaque.

There are strategies to dissect a models inner workings, but this is an active field of research and incomplete.

replies(1): >>44400063 #
1. Azkron ◴[] No.44400063[source]
Whatever comes out of any LLM will directly depend upon the data you fed it and which answers your reinforced as correct. There is nothing unknown or mystical about it.
replies(1): >>44403166 #
2. richardatlarge ◴[] No.44403166[source]
The same could be said of people, revealing the emptiness of this idea. Knowing the process at a mechanism level says nothing about the outcome. Some people output German, some English. It’s sub-mechanisms are plastic and emergent