Finite sample weighting of recursive forecast errors
Simon P. Burke and
International Journal of Forecasting, 2016, vol. 32, issue 2, 458-474
This paper proposes and tests a new framework for weighting recursive out-of-sample prediction errors according to their corresponding levels of in-sample estimation uncertainty. In essence, we show how to use the maximum possible amount of information from the sample in the evaluation of the prediction accuracy, by commencing the forecasts at the earliest opportunity and weighting the prediction errors. Via a Monte Carlo study, we demonstrate that the proposed framework selects the correct model from a set of candidate models considerably more often than the existing standard approach when only a small sample is available. We also show that the proposed weighting approaches result in tests of equal predictive accuracy that have much better sizes than the standard approach. An application to an exchange rate dataset highlights relevant differences in the results of tests of predictive accuracy based on the standard approach versus the framework proposed in this paper.
Keywords: Forecast evaluation; Forecast comparison; Recursive model estimation; Mean squared error; Forecast weighting scheme (search for similar items in EconPapers)
References: View references in EconPapers View complete reference list from CitEc
Citations: Track citations by RSS feed
Downloads: (external link)
Full text for ScienceDirect subscribers only
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
Persistent link: https://EconPapers.repec.org/RePEc:eee:intfor:v:32:y:2016:i:2:p:458-474
Access Statistics for this article
International Journal of Forecasting is currently edited by R. J. Hyndman
More articles in International Journal of Forecasting from Elsevier
Bibliographic data for series maintained by Dana Niculescu ().