Back to glossary

AI GLOSSARY

Calibration

Evaluation & Performance

A measure of how well a model's confidence scores reflect the true likelihood of its predictions being correct. A well-calibrated model that says it is 80% confident should be right about 80% of the time. Poor calibration, being systematically overconfident or underconfident, is a significant problem in high-stakes applications like medical diagnosis or risk assessment.
See also: Evaluation, Uncertainty Quantification.

External reference