←back to thread

A PM's Guide to AI Agent Architecture

(www.productcurious.com)
205 points umangsehgal93 | 1 comments | | HN request time: 0.201s | source
Show context
barbazoo ◴[] No.45130243[source]
> Confidence calibration: When your agent says it's 60% confident, it should be right about 60% of the time. Not 90%, not 30%. Actual 60%.

With current technology (LLM), how can an agent ever be sure about its confidence?

replies(2): >>45130587 #>>45131981 #
1. fumeux_fume ◴[] No.45131981[source]
The author's inner PM comes out here and makes some wild claims. Calibration is something we can do with traditional, classification models, but not with most off-the-shelf LLMs. Even if you devised a way to determine if the LLM's confidence claim matched it's actual performance, you wouldn't be able to calibrate or tune it like you would a more traditional model.