←back to thread

688 points crescit_eundo | 1 comments | | HN request time: 0s | source
Show context
snickerbockers ◴[] No.42144943[source]
Does it ever try an illegal move? OP didn't mention this and I think it's inevitable that it should happen at least once, since the rules of chess are fairly arbitrary and LLMs are notorious for bullshitting their way through difficult problems when we'd rather they just admit that they don't have the answer.
replies(2): >>42145004 #>>42145793 #
sethherr ◴[] No.42145004[source]
Yes, he discusses using a grammar to restrict to only legal moves
replies(4): >>42147380 #>>42148708 #>>42150800 #>>42152205 #
topaz0 ◴[] No.42147380[source]
Still an interesting direction of questioning. Maybe could be rephrased as "how much work is the grammar doing"? Are the results with the grammar very different than without? If/when a grammar is not used (like in the openai case), how many illegal moves does it try on average before finding a legal one?
replies(3): >>42147422 #>>42150017 #>>42151815 #
1. int_19h ◴[] No.42150017[source]
A grammar is really just a special case of the more general issue of how to pick a single token given the probabilities that the model spits out for every possible one. In that sense, filters like temperature / top_p / top_k are already hacks that "do the work" (since always taking the most likely predicted token does not give good results in practice), and grammars are just a more complicated way to make such decisions.