Smooth Calibration, Leaky Forecasts, Finite Recall, and Nash Dynamics
Dean P. Foster and
Sergiu Hart
Papers from arXiv.org
Abstract:
We propose to smooth out the calibration score, which measures how good a forecaster is, by combining nearby forecasts. While regular calibration can be guaranteed only by randomized forecasting procedures, we show that smooth calibration can be guaranteed by deterministic procedures. As a consequence, it does not matter if the forecasts are leaked, i.e., made known in advance: smooth calibration can nevertheless be guaranteed (while regular calibration cannot). Moreover, our procedure has finite recall, is stationary, and all forecasts lie on a finite grid. To construct the procedure, we deal also with the related setups of online linear regression and weak calibration. Finally, we show that smooth calibration yields uncoupled finite-memory dynamics in n-person games "smooth calibrated learning" in which the players play approximate Nash equilibria in almost all periods (by contrast, calibrated learning, which uses regular calibration, yields only that the time-averages of play are approximate correlated equilibria).
Date: 2022-10
New Economics Papers: this item is included in nep-gth and nep-mic
References: View references in EconPapers View complete reference list from CitEc
Citations:
Published in Games and Economic Behavior 109 (May 2018), 271-293
Downloads: (external link)
http://arxiv.org/pdf/2210.07152 Latest version (application/pdf)
Related works:
Journal Article: Smooth calibration, leaky forecasts, finite recall, and Nash dynamics (2018) 
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:arx:papers:2210.07152
Access Statistics for this paper
More papers in Papers from arXiv.org
Bibliographic data for series maintained by arXiv administrators ().