I think the explanation is pretty simple, as I said in my earlier comment: https://news.ycombinator.com/item?id=44904107
I also believe the problem is we don't know what we want: https://news.ycombinator.com/item?id=45509015
If we could make LLMs to apply a modest set of logic rules consistently, it would be a win.
replies(1):