Learn Then Test: Calibrating Predictive Algorithms to Achieve Risk Control
Anastasios N. Angelopoulos,
Stephen Bates,
Emmanuel J. Candes,
Michael I. Jordan and
Lihua Lei
Additional contact information
Anastasios N. Angelopoulos: ?
Stephen Bates: ?
Emmanuel J. Candes: ?
Michael I. Jordan: ?
Lihua Lei: Stanford U
Research Papers from Stanford University, Graduate School of Business
Abstract:
We introduce a framework for calibrating machine learning models so that their predictions satisfy explicit, finite-sample statistical guarantees. Our calibration algorithm works with any underlying model and (unknown) data-generating distribution and does not require model refitting. The framework addresses, among other examples, false discovery rate control in multi-label classification, intersection-over-union control in instance segmentation, and the simultaneous control of the type-1 error of outlier detection and confidence set coverage in classification or regression. Our main insight is to reframe the risk-control problem as multiple hypothesis testing, enabling techniques and mathematical arguments different from those in the previous literature. We use our framework to provide new calibration methods for several core machine learning tasks with detailed worked examples in computer vision and tabular medical data.
Date: 2022-04
New Economics Papers: this item is included in nep-big, nep-cmp, nep-ecm and nep-rmg
References: Add references at CitEc
Citations:
Downloads: (external link)
https://doi.org/10.48550/arXiv.2110.01052
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:ecl:stabus:4030
Access Statistics for this paper
More papers in Research Papers from Stanford University, Graduate School of Business Contact information at EDIRC.
Bibliographic data for series maintained by ().