←back to thread

724 points simonw | 1 comments | | HN request time: 0.207s | source
Show context
xnx ◴[] No.44527256[source]
> It’s worth noting that LLMs are non-deterministic,

This is probably better phrased as "LLMs may not provide consistent answers due to changing data and built-in randomness."

Barring rare(?) GPU race conditions, LLMs produce the same output given the same inputs.

replies(7): >>44527264 #>>44527395 #>>44527458 #>>44528870 #>>44530104 #>>44533038 #>>44536027 #
kcb ◴[] No.44527458[source]
FP multiplication is non-commutative.
replies(2): >>44527482 #>>44528992 #
1. DemocracyFTW2 ◴[] No.44528992[source]
That's like you can't deduce the input t from a cryptographic hash h but the same input always gives you the same hash, so t->h is deterministic. h->t is, in practice, not a way that you can or want to walk (because it's so expensive to do) and because there may be / must be collisions (given that a typical hash is much smaller than the typical inputs), so the inverse is not h->t with a single input but h->{t1,t2,...}, a practically open set of possible inputs that is still deterministic.