text
stringlengths 1
298
|
|---|
for y such that f(y) > 0. The Hyvärinen score is proper relative to the subclass of P(R) such that the density
|
f exists, is twice continuously differentiable and satisfies f′(x)/f(x) → 0 as |x| → ∞. It is worth noticing that
|
the Hyvärinen score can be computed even if f is only known up to a scale factor (e.g., up to a normalizing
|
constant). This property allows circumventing the use of Monte Carlo methods or approximations of the
|
normalizing constant when it is unavailable or hard to compute. This is a property of local proper scoring
|
rules except for the logarithmic score (Parry et al., 2012). Readers eager to learn more about local proper
|
scoring rules may refer to Parry et al. (2012) and Ehm and Gneiting (2012).
|
The logarithmic score and the Hyvärinen score do not allow f to be zero. To overcome this limitation,
|
scoring rules expressed in terms of the Lα-norm have been proposed. The quadratic score is defined as
|
QuadS(F,y) = ∥f∥2
|
2 −2f(y),
|
6
|
where ∥f∥2
|
2 = R
|
f(y)2dy. The quadratic score is strictly proper relative to the class L2(R).
|
The pseudospherical score is defined as
|
PseudoS(F,y) = −f(y)α−1/∥f∥α−1
|
α ,
|
with α > 1. For α = 2, it reduces to the spherical score (see, e.g., Jose 2007). The pseudospherical score
|
is strictly proper relative to the class Lα(R). The four scoring rules presented above have been criticized as
|
they do not encourage a high probability in the vicinity of the observation y (Gneiting and Raftery, 2007).
|
In particular, as the logarithmic score is more sensitive to outliers, probabilistic forecasts inferred by its
|
minimization may be overdispersive (Gneiting et al., 2005). Moreover, the pdf is not always available, for
|
example in the case of ensemble forecasts.
|
Readers may refer to the various reviews of scoring rules available (see, e.g., Bröcker and Smith 2007;
|
Gneiting and Raftery 2007; Gneiting and Katzfuss 2014; Thorarinsdottir and Schuhen 2018; Alexander et al.
|
2022). Formulas of the expected scoring rules presented are available in Appendix A.
|
Strictly proper scoring rules can be seen as more powerful than proper scoring rules. This is theoretically
|
true when the interest is in identifying the ideal forecast (i.e., the true distribution). Regardless, in practice,
|
scoring rules are also used to rank probabilistic forecasts and with that in mind, a given ranking of forecasts
|
in terms of the expectation of a strictly proper scoring rule (such as the CRPS) is harder to interpret than
|
a ranking in terms of the expectation of a proper but more interpretable scoring rule (such as the SE). The
|
SE is known to discriminate the mean, and thus, a better rank in terms of expected SE implies a better
|
prediction of the mean of the true distribution. Conversely, a better ranking in terms of CRPS implies a
|
better prediction of the whole prediction, but it might not be useful as is, and other verification tools will be
|
needed to know what caused this ranking. When forecasts are not calibrated, there seems to be a trade-off
|
between interpretability and discriminatory power and this becomes more prominent in a multivariate set
|
ting. However, simpler interpretable tools and discriminatory-powerful tools can be used complementarily.
|
The framework proposed in Section 3 aims at helping the construction of interpretable proper scoring rules.
|
2.3 Multivariate scoring rules
|
In a multivariate setting, forecasters cannot solely use univariate scoring rules as they are not able to
|
discriminate forecasts beyond their 1-dimensional marginals. Univariate scoring rules cannot discriminate
|
the dependence structure between the univariate margins. Multivariate forecasts can be applied in different
|
setups: spatial forecasts, temporal forecasts, multivariable forecasts or any combination of these categories
|
(e.g., spatio-temporal forecasts of multiple variables). Considering weather forecasting, a spatial forecast
|
could aim at predicting temperatures across multiple locations. A temporal forecast could be focused on
|
predicting rainfall at multiple lead times at a given location. A multivariable forecast could predict both
|
eastward and northward components of the wind. In the following, we consider F ∈ F ⊆ P(Rd) a multivariate
|
probabilistic forecast and y ∈ Rd an observation.
|
Even if there is no natural ordering in the multivariate case, the notions of median and quantile can
|
be adapted using level sets, and then scoring rules using these quantities can be constructed (see, e.g.,
|
Meng et al. 2023). Nonetheless, as the mean is well-defined, the squared error (SE) can be defined in the
|
multivariate setting :
|
SE(F,y) = ∥µF −y∥2
|
2,
|
(12)
|
where µF is the mean vector of the distribution F. Similar to the univariate case, the SE is proper relative
|
to P2(Rd). Moments are well-defined in the multivariate case allowing the multivariate version of the Dawid
|
Sebastiani score to be defined. The Dawid-Sebastiani score (DSS) was proposed in Dawid and Sebastiani
|
(1999) as
|
DSS(F,y) = log(detΣF)+(µF −y)TΣ−1
|
F (µF −y),
|
where µF and ΣF are the mean vector and the covariance matrix of the distribution F. The DSS is proper
|
relative to P2(Rd) and it becomes strictly proper relative to any convex class of probability measures charac
|
terized by their first two moments (Gneiting and Raftery, 2007). The second term in the DSS is the squared
|
7
|
Mahalanobis distance between y and µF.
|
To define a strictly proper scoring rule for multivariate forecast, Gneiting and Raftery (2007) proposed
|
the energy score (ES) as a generalization of the CRPS to the multivariate case. The ES is defined by
|
ESα(F,y) = EF∥X −y∥α
|
2−1
|
2EF∥X −X′∥α
|
2,
|
(13)
|
where α ∈ (0,2) and F ∈ Pα(Rd), the class of Borel probability measures on Rd such that the moment of
|
order α is finite. The definition of the ES is related to the kernel form of the CRPS (6), to which the ES
|
reduces for d = 1 and α = 1. As pointed out in Gneiting and Raftery (2007), in the limiting case α = 2, the
|
ES becomes the SE (12). The ES is strictly proper relative to Pα(Rd) (Székely, 2003; Gneiting and Raftery,
|
2007) and is suited for ensemble forecasts (Gneiting et al., 2008). Moreover, the parameter α gives some
|
f
|
lexibility: a small value of α can be chosen and still lead to a strictly proper scoring rule, for example, when
|
higher-order moments are ill-defined. The discrimination ability of the ES has been studied in numerous
|
studies (see, e.g., Pinson and Girard 2012; Pinson and Tastu 2013; Scheuerer and Hamill 2015). Pinson and
|
Girard (2012) studied the ability of the ES to discriminate among rival sets of scenarios (i.e., forecasts) of
|
wind power generation. In the case of bivariate Gaussian processes, Pinson and Tastu (2013) illustrated that
|
the ES appears to be more sensitive to misspecifications of the mean rather than misspecifications of the
|
variance or dependence structure. The lack of sensitivity to misspecifications of the dependence structure
|
has been confirmed in Scheuerer and Hamill (2015) using multivariate Gaussian random vectors of higher
|
dimension. Moreover, the discriminatory power of the ES deteriorates in higher dimensions (Pinson and
|
Tastu, 2013).
|
To overcome the discriminatory limitation of the ES, Scheuerer and Hamill (2015) proposed the variogram
|
score (VS), a score targeting the verification of the dependence structure. The VS of order p is defined as
|
d
|
VSp(F,y) =
|
i,j=1
|
wij (EF [|Xi −Xj|p] −|yi −yj|p)2
|
(14)
|
where Xi is the i-th component of the random vector X following F, wij are nonnegative weights and p > 0.
|
The variogram score capitalizes on the variogram, used in spatial statistics to access the dependence struc
|
ture. The VS cannot detect an equal bias across all components. The VS of order p is proper relative to
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.