1. Introduction
Different skill measures present differing dependence on the intrinsic predictability level as well as differing dependence on sample size. Kumar (2009) computed the expected skill of idealized forecasts with varying levels of predictability using the anomaly correlation (AC), the Heidke skill score (HSS), and the ranked probability skill score (RPSS) as skill measures. An important consideration for seasonal climate forecasts, as demonstrated by Kumar (2009), is that these skill measures may vary significantly from their expected values when they are computed from relatively short forecast histories. In carrying out his skill measure characterizations, Kumar (2009) assumed that the forecast variance was equal to the climatological variance. Additionally, when computing average skill scores, forecast distributions were assumed to have identical means (signals) for a given predictability level. Here we examine some implications of those assumptions.
This comment is organized as follows. In section 2 we present a perfect model framework for examining predictability. A natural requirement for predictability is that the forecast and climatological distributions be different. In the case of joint normal distributions, the existence of predictability (positive signal variance) implies that the forecast variance is less than the climatological variance. In this case, inflating the forecast variance to be equal to the climatological variance results in underconfident forecasts and lower probabilistic skill scores. In section 3 we compute the three skill scores for a given forecast signal (single initial condition) and for a set of forecasts with a specified signal to noise variance ratio (multiple initial conditions). The expected skill of a single forecast depends on both signal level and forecast variance. The AC depends on the ratio of squared signal to forecast variance, while the dependence of HSS and RPSS on signal level and forecast variance is more complex. A consequence of this functional dependence is that the AC of a set of forecasts is equal to that of a single forecast with the signal equal to the signal standard deviation. However, there is no similar relation for HSS and RPSS, and assuming a fixed signal as in Kumar (2009) generally overestimates HSS and always overestimates RPSS. We also present a useful approximation relating the RPSS and AC of a set of forecasts.
2. Predictability framework
A basic question of predictability studies is whether the future (verification) state υ of some quantity is predictable given the current (initial) state i, and if so, with what level of skill. Potential predictability and climate change studies examine the extent to which the climate system is determined by the specification of a boundary condition (e.g., sea surface temperature or land surface properties) or some other forcing (e.g., solar irradiance, aerosols or greenhouse gases). The predictability framework presented here can be applied to questions of potential predictability by interpreting i as a forcing or boundary condition and υ as the associated climate response.
The relation between the initial condition i and the verification υ is completely described by the joint probability distribution p(υ, i); we use the convention that the argument of p determines the probability distribution function in question. The most complete description of υ given i is the conditional distribution p(υ|i). We consider the conditional distribution p(υ|i) to be the “perfect” forecast distribution and base our discussion on it. If υ and i are independent, then υ is not predictable from i, and the conditional distribution p(υ|i) is equal to the unconditional or climatological distribution p(υ). This characterization is consistent with that of Lorenz who described the absence of predictability as when a forecast is no better than a random draw from the climatological distribution (Lorenz 1969; DelSole and Tippett 2007). Therefore, a fundamental property of predictability is that predictability exists only when the conditional distribution p(υ|i) differs from the climatological distribution p(υ). In practice, the climatological distribution is often estimated from recent observations, for instance, a recent 30-yr period in the case of seasonal climate prediction.



In the context of predictability studies based on ensemble integrations, μυ|i is the ensemble mean, and



3. Predictability and skill scores
The above framework allows us to examine the dependence of various skill scores on the level of predictability, and in the case when i and υ have a joint normal distribution, to obtain fairly explicit formulas. Following Kumar (2009) we examine three skill scores: AC, HSS, and RPSS. First, similar to Kumar et al. (2001) we examine the expected skill of a forecast with a specified signal level. Here, however, we do not assume that the noise variance
Without loss of generality, we take unit climatological variance συ = 1 so that the forecast variance is given by
a. Expected skill for a given signal







b. Expected skill for a given signal-to-noise ratio






We note that computing
c. Variance of the skill for a given signal-to-noise ratio
4. Summary and conclusions
The difference between the climatological and forecast probability distribution functions is an indication of predictability. In the case of variables with a joint normal distribution, a necessary and sufficient condition for predictability is that the climatological and forecast variances are different. Making the forecast variance equal to the climatological variance results in underconfident forecasts and lower probabilistic skill scores. The expected skill as measured by the AC, HSS, and RPSS of a forecast with a specified signal level depends on both the signal level and the forecast variance. However, for forecast variances consistent with modest skill levels (ρ ≤ 0.5), the dependence on forecast variance is weak.
We compute the AC, HSS, and RPSS for a set of forecasts with specified signal-to-noise ratio; the forecast variance is constant and the signal is allowed to vary from one forecast to another, consistent with the signal-to-noise ratio. The HSS and RPSS values obtained in this manner are lower than those found in Kumar (2009), which for a given level of predictability used a set of forecasts in which all the forecasts had identical means equal to the signal standard deviation. Assuming a fixed signal results in overestimates of HSS and RPSS. We also provide a useful approximation that expresses expected RPSS values in terms of correlation values.
The variability of the three skill scores is computed when the sample size is finite. The variance of the AC for a finite set of forecasts with constant variance and varying signal can be expressed as a function of the signal-to-noise ratio, or correlation. The variances of the HSS and RPSS were found by Monte Carlo simulation.
Acknowledgments
The author is supported by a grant/cooperative agreement from the National Oceanic and Atmospheric Administration (NA05OAR4311004). The views expressed herein are those of the authors and do not necessarily reflect the views of NOAA or any of its subagencies. The authors gratefully acknowledge Maria R. D’Orsogna for her generous help with the Taylor series approximations.
REFERENCES
DelSole, T. , and M. K. Tippett , 2007: Predictability: Recent insights from information theory. Rev. Geophys., 45 , RG4002. doi:10.1029/2006RG000202.
Fisher, R. A. , 1915: Frequency distribution of the values of the correlation coefficient in samples of an indefinitely large population. Biometrika, 10 , 507–521.
Hotelling, H. , 1953: New light on the correlation coefficient and its transforms. J. Roy. Stat. Soc. B Meteor., 15 , 193–225.
Kumar, A. , 2009: Finite samples and uncertainty estimates for skill measures for seasonal prediction. Mon. Wea. Rev., 137 , 2622–2631.
Kumar, A. , A. G. Barnston , and M. P. Hoerling , 2001: Seasonal predictions, probabilistic verifications, and ensemble size. J. Climate, 14 , 1671–1676.
Lorenz, E. N. , 1969: The predictability of a flow which possesses many scales of motion. Tellus, 21 , 289–307.
Rowell, D. P. , 1998: Assessing potential seasonal predictability with an ensemble of multidecadal GCM simulations. J. Climate, 11 , 109–120.
Sardeshmukh, P. D. , G. P. Compo , and C. Penland , 2000: Changes of probability associated with El Niño. J. Climate, 13 , 4268–4286.
Shukla, J. , 1981: Dynamical predictability of monthly means. J. Atmos. Sci., 38 , 2547–2572.
Tippett, M. K. , A. G. Barnston , and A. W. Robertson , 2007: Estimation of seasonal precipitation tercile-based categorical probabilities from ensembles. J. Climate, 20 , 2210–2228.
Van den Dool, H. M. , and Z. Toth , 1991: Why do forecasts for near normal often fail? Wea. Forecasting, 6 , 76–85.
APPENDIX
Approximations of the RPSS

(a) AC, (b) HSS, and (c) RPSS as a function of the conditional mean μυ|i for ρ = 0.0, 0.5, 0.7, and 0.9; the skill is an increasing function of correlation. (d) AC, (e) HSS, and (f) RPSS for a set of forecasts with signal-to-noise ratio S; the thick lines are based on analytical expression and thin lines are based on 5000 Monte Carlo simulations. In (f), the dotted line is based on the approximation in (21). (g)–(j) The standard deviation of the quantities in (d)–(f) as a function of their expected values when a sample size of n = 30 is used; the thick and thin lines are as in (d)–(f).
Citation: Monthly Weather Review 138, 4; 10.1175/2009MWR3214.1