Back to glossaryExternal reference
AI GLOSSARY
Calibration
Evaluation & Performance
A measure of how well a model's confidence scores reflect the true likelihood of its predictions being correct. A well-calibrated model that says it is 80% confident should be right about 80% of the time. Poor calibration, being systematically overconfident or underconfident, is a significant problem in high-stakes applications like medical diagnosis or risk assessment.
See also: Evaluation, Uncertainty Quantification.