←back to thread

66 points appwiz | 1 comments | | HN request time: 0.209s | source
Show context
simonw ◴[] No.44383691[source]
I still don't think hallucinations in generated code matter very much. They show up the moment you try to run the code, and with the current batch of "coding agent" systems it's the LLM itself that spots the error when it attempts to run the code.

I was surprised that this paper talked more about RAG solutions than tool-use based solutions. Those seem to me like a proven solution at this point.

replies(4): >>44384474 #>>44384576 #>>44387027 #>>44388124 #
mucha ◴[] No.44387027[source]
Interesting. How do existing systems catch Task Requirement hallucinations?
replies(1): >>44387759 #
1. simonw ◴[] No.44387759[source]
They don't. My comment was about "hallucinations in generated code".