> Confidence calibration: When your agent says it's 60% confident, it should be right about 60% of the time. Not 90%, not 30%. Actual 60%.
With current technology (LLM), how can an agent ever be sure about its confidence?
replies(2):
With current technology (LLM), how can an agent ever be sure about its confidence?
Calibrated Language Models Must Hallucinate