Title: Optimal probabilistic forecasts: When they work
Authors: Worapree Ole Maneesoonthorn - University of Melbourne (Australia) [presenting]
Gael Martin - Monash University (Australia)
David Frazier - Monash University (Australia)
Ruben Laoiza Maya - Monash University (Australia)
Andres Ramirez Hassan - Universidad EAFIT (Colombia)
Abstract: Proper scoring rules are used to assess the out-of-sample accuracy of probabilistic forecasts, with different scoring rules rewarding distinct aspects of forecast performance. We re-investigate the practice of using proper scoring rules to produce probabilistic forecasts that are `optimal' according to a given score, and assess when their out-of-sample accuracy is superior to alternative forecasts, according to that score. Particular attention is paid to relative predictive performance under misspecification of the predictive model. Using numerical illustrations, we document several novel findings within this paradigm that highlights the important interplay between the true data generating process. The assumed predictive model and the scoring rule. Notably, we show that only when a predictive model is sufficiently compatible with the true process to allow a particular score criterion to reward what it is designed to reward, will this approach to forecasting reap benefits. Subject to this compatibility; however, the superiority of the optimal forecast will be greater, the greater is the degree of misspecification. We explore these issues under a range of different scenarios and using both artificially simulated and empirical data.