text stringlengths 1 298 |
|---|
formation considers the orthogonal directions formed by the abscissa and ordinate axes and evaluates how |
the variogram changes between these directions. The transformation leads to negative or zero quantities |
with values close to zero characterizing isotropy and negative values corresponding to the anisotropy of the |
variograms in the directions and at the scale involved. |
4.3 Other transformations |
Transformations other than projections or summary statistics can be used to target forecast characteristics. |
For example, a transformation in the form of a change of coordinates or a change of scale (e.g., a logarithmic |
scale) can be used to obtain proper scoring rules. We highlight two families of scoring rules that can be seen |
as transformation-based scoring rules: wavelet-based scoring rules and threshold-weighted scoring rules. |
Generally speaking, wavelet-based scoring rules are built thanks to a projection of forecast and observation |
f |
ields onto a wavelet basis. Based on the wavelet coefficients, dimension reduction might be performed to |
target specific characteristics such as the dependence structure or the location. The resulting coefficients of |
the forecast fields are compared to the coefficients of the observations fields using scoring rules (e.g., squared |
error (SE) or energy score (ES)). Wavelet transformations are (complex) transformations, and thus, the |
scoring rules associated fall within the scope of Proposition 1. In particular, Buschow et al. (2019) used a |
dimension reduction procedure resulting in the obtention of a mean and a scale spectra and used scoring |
rules to compare forecasts and observation spectra. For example, the ES of the mean spectrum is used and |
shows good discrimination ability when the scale structure is misspecified. |
Note that Buschow et al. (2019) proposed two other wavelet-based scoring rules: one based on the earth |
mover’s distance (EMD) of the scale histograms and one based on the distance in the scale histograms’ center |
of mass. The EMD-based scoring rules are not proper since the EMD is not a proper scoring rule (Thorarins |
dottir et al., 2013) and the so-called distance between centers of mass is not a distance but rather a difference |
of position leading to an improper scoring rule. However, the ES-based scoring rules are proper and could be |
derived from scale histograms. Despite their apparent complexity, wavelet transformations allow to target |
interpretable characteristics such as the location (Buschow, 2022), the scale structure (Buschow et al., 2019; |
Buschow and Friederichs, 2020) or the anisotropy (Buschow and Friederichs, 2021). The transformations |
proposed for the deterministic forecasts setting in most of these articles could be used as foundations for |
future work willing to propose wavelet-based proper scoring rules targeting the location, the scale structure |
17 |
or the anisotropy. |
As showcased in Heinrich-Mertsching et al. (2021) for a specific example and hinted in Allen et al. (2024), |
transformations can also be used to emphasize certain outputs. Threshold weighting is one of the three main |
types of weighting conserving the propriety of scoring rules. Its name come from the fact that it corresponds |
to a weighting over different thresholds in the case of CRPS (7) (Gneiting, 2011). Recall that given a |
conditionally negative definite kernel ρ, the kernel scoring associated Sρ (15) is proper relative to Pρ. Many |
popular scoring rules are kernel scores such as the BS (5), the CRPS (6), the ES (13) and the VS (14). By |
definition (Allen et al., 2023b, Definition 4), threshold-weighted kernel scores are constructed as |
twSρ(F,y;v) = EF[ρ(v(X),v(y))] − 1 |
2EF[ρ(v(X),v(X′))] − 1 |
2ρ(v(y),v(y)); |
=Sρ(v(F),v(y)), |
where v is the chaining function capturing how the emphasis is put on certain outputs. With this explicit |
definition, it is obvious that threshold-weighted kernel scores are covered by the framework of Proposition 1. |
It can be noted that Proposition 4 in Allen et al. (2023b) states that strict propriety of the kernel scoring |
rule is preserved by the chaining function v if and only if v is injective. Weighted scoring rules allow to |
emphasize particular outcomes: when studying extreme events, it is often of particular interest to focus |
on values larger than a given threshold t and this can be achieved using the chaining rule v(x) = 1x≥t. |
Threshold-weighted scoring rules have been used in verification procedures in the literature; we illustrate its |
use through three different studies. Lerch and Thorarinsdottir (2013) aggregated across stations twCRPS |
to compare the upper tail performance of different daily maximum wind speed forecasts. Chapman et al. |
(2022) aggregated the threshold-weighted CRPS across locations to study the improvement of statistical |
postprocessing techniques, the importance of predictors and the influence of the size of the training set |
on the performance. Allen et al. (2023a) used threshold-weighted versions of the CRPS, the ES, and the |
VS to compare the predictive performance of forecasts regarding heatwave severity; the scoring rules were |
aggregated across stations. Readers may refer to Allen et al. (2023a) and Allen et al. (2023b) for insightful |
reviews of weighted scoring rules in both univariate and multivariate settings. |
5 Simulation study |
This section provides simulated examples to showcase the different uses of the framework introduced in |
Section 3 to construct interpretable proper scoring rules for multivariate forecasts. Four examples are |
developed. Firstly, a setup where the emphasis is put on 1-marginal verification is proposed. This setup |
serves as a means of recalling and showing the limitations of strictly proper scoring rules and the benefits |
of interpretable scoring rules in a concrete setting. Secondly, a standard multivariate setup is studied |
where popular multivariate scoring rules (i.e., VS and ES) are compared to a multivariate scoring rule |
aggregated over patches and an aggregation-and-transformation-based scoring rule in their discrimination |
ability regarding the dependence structure. Thirdly, a setup introducing anisotropy in both observations |
and forecasts is introduced. The anisotropic score is constructed based on the transformation principle |
with the goal of discriminating differences of anisotropy in the dependence structure between forecast and |
observations. Fourthly, we propose a setup to test the sensitivity of scoring rules to the double-penalty effect |
and we introduce scoring rules that can be built to be resilient to some manifestation of the double-penalty |
effect. |
In these four numerical experiments, the spatial field is observed and predicted on a regular 20×20 grid |
D={1,...,20}×{1,...,20}. Observations are realizations of a Gaussian random field (G(s))s∈D with zero |
mean and power-exponential covariance defined as |
cov(G(s),G(s′)) = σ02 exp − ∥s−s′∥ |
λ0 |
The parameters are taken equal to σ0 = 1, λ0 = 3 and β0 = 1. |
β0 |
, |
s, s′ ∈ D. |
(20) |
18 |
In each numerical experiment, we compare a few predictive distributions, including the distribution |
generating observations and other ones deviating from the generative distributions in a specific way. These |
different predictive distributions are evaluated with different scoring rules and the aim is to illustrate the |
discriminatory ability of the different scoring rules. |
The simulation study uses 500 observations of the random field (G(s))s∈D. The scoring rules are com |
puted using exact formulas when possible (see Appendix E), and, when exact formulas are not available, |
they are computed based on a sample of size 100 (i.e., ensemble forecasts) of the probabilistic forecast. |
Estimated expectations over the 500 observations are computed and the experiment is repeated 10 times. |
The corresponding results are represented by boxplots. The units of the scoring rules are rescaled by the |
average expected score of the true distribution (i.e., the ideal forecast). The statistical significativity of |
the ranking between forecasts is tested using a Diebold-Mariano test (Diebold and Mariano, 1995). When |
deemed necessary, statistical significativity is mentioned for a confidence level of 95%. |
The code used for the different numerical experiments is publicly available1. |
5.1 Marginals |
This first numerical experiment focuses on the prediction of the 1-dimensional marginal distributions and |
the aggregation of univariate scoring rules. For simplicity, we consider only stationary random fields so that |
the 1-marginal distribution is the same at all grid points. Although similar conclusions could be drawn |
from an univariate framework (i.e., with independent 1-dimensional rather than spatial observations), this |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.