Probabilistic Verification of Monthly Temperature Forecasts

Andreas P. Weigel Federal Office of Meteorology and Climatology, MeteoSwiss, Zurich, Switzerland

Search for other papers by Andreas P. Weigel in
Current site
Google Scholar
PubMed
Close
,
Daniel Baggenstos Federal Office of Meteorology and Climatology, MeteoSwiss, Zurich, Switzerland

Search for other papers by Daniel Baggenstos in
Current site
Google Scholar
PubMed
Close
,
Mark A. Liniger Federal Office of Meteorology and Climatology, MeteoSwiss, Zurich, Switzerland

Search for other papers by Mark A. Liniger in
Current site
Google Scholar
PubMed
Close
,
Frédéric Vitart European Centre for Medium-Range Weather Forecasts, Reading, United Kingdom

Search for other papers by Frédéric Vitart in
Current site
Google Scholar
PubMed
Close
, and
Christof Appenzeller Federal Office of Meteorology and Climatology, MeteoSwiss, Zurich, Switzerland

Search for other papers by Christof Appenzeller in
Current site
Google Scholar
PubMed
Close
Full access

Abstract

Monthly forecasting bridges the gap between medium-range weather forecasting and seasonal predictions. While such forecasts in the prediction range of 1–4 weeks are vital to many applications in the context of weather and climate risk management, surprisingly little has been published on the actual monthly prediction skill of existing global circulation models. Since 2004, the European Centre for Medium-Range Weather Forecasts has operationally run a dynamical monthly forecasting system (MOFC). It is the aim of this study to provide a systematic and fully probabilistic evaluation of MOFC prediction skill for weekly averaged forecasts of surface temperature in dependence of lead time, region, and season. This requires the careful setup of an appropriate verification context, given that the verification period is short and ensemble sizes small. This study considers the annual cycle of operational temperature forecasts issued in 2006, as well as the corresponding 12 yr of reforecasts (hindcasts). The debiased ranked probability skill score (RPSSD) is applied for verification. This probabilistic skill metric has the advantage of being insensitive to the intrinsic unreliability due to small ensemble sizes—an issue that is relevant in the present context since MOFC hindcasts only have five ensemble members. The formulation of the RPSSD is generalized here such that the small hindcast ensembles and the large operational forecast ensembles can be jointly considered in the verification. A bootstrap method is applied to estimate confidence intervals. The results show that (i) MOFC forecasts are generally not worse than climatology and do outperform persistence, (ii) MOFC forecasts are skillful beyond a lead time of 18 days over some ocean regions and to a small degree also over tropical South America and Africa, (iii) extratropical continental predictability essentially vanishes after 18 days of integration, and (iv) even when the average predictability is low there can nevertheless be climatic conditions under which the forecasts contain useful information. With the present model, a significant skill improvement beyond 18 days of integration can only be achieved by increasing the averaging interval. Recalibration methods are expected to be without effect since the forecasts are essentially reliable.

Corresponding author address: Andreas Weigel, MeteoSwiss, Krähbühlstrasse 58, P.O. Box 514, CH-8044 Zürich, Switzerland. Email: andreas.weigel@meteoswiss.ch

Abstract

Monthly forecasting bridges the gap between medium-range weather forecasting and seasonal predictions. While such forecasts in the prediction range of 1–4 weeks are vital to many applications in the context of weather and climate risk management, surprisingly little has been published on the actual monthly prediction skill of existing global circulation models. Since 2004, the European Centre for Medium-Range Weather Forecasts has operationally run a dynamical monthly forecasting system (MOFC). It is the aim of this study to provide a systematic and fully probabilistic evaluation of MOFC prediction skill for weekly averaged forecasts of surface temperature in dependence of lead time, region, and season. This requires the careful setup of an appropriate verification context, given that the verification period is short and ensemble sizes small. This study considers the annual cycle of operational temperature forecasts issued in 2006, as well as the corresponding 12 yr of reforecasts (hindcasts). The debiased ranked probability skill score (RPSSD) is applied for verification. This probabilistic skill metric has the advantage of being insensitive to the intrinsic unreliability due to small ensemble sizes—an issue that is relevant in the present context since MOFC hindcasts only have five ensemble members. The formulation of the RPSSD is generalized here such that the small hindcast ensembles and the large operational forecast ensembles can be jointly considered in the verification. A bootstrap method is applied to estimate confidence intervals. The results show that (i) MOFC forecasts are generally not worse than climatology and do outperform persistence, (ii) MOFC forecasts are skillful beyond a lead time of 18 days over some ocean regions and to a small degree also over tropical South America and Africa, (iii) extratropical continental predictability essentially vanishes after 18 days of integration, and (iv) even when the average predictability is low there can nevertheless be climatic conditions under which the forecasts contain useful information. With the present model, a significant skill improvement beyond 18 days of integration can only be achieved by increasing the averaging interval. Recalibration methods are expected to be without effect since the forecasts are essentially reliable.

Corresponding author address: Andreas Weigel, MeteoSwiss, Krähbühlstrasse 58, P.O. Box 514, CH-8044 Zürich, Switzerland. Email: andreas.weigel@meteoswiss.ch

1. Introduction

Probabilistic ensemble forecasts have become a standard technique in numerical weather and climate forecasting. Particularly in the context of weather and climate risk management, the probability information provided by ensembles is important in that it allows us to foresee forecast uncertainties and thus the limits of predictability. Depending on the time scales considered, and ignoring uncertainties due to model errors, ensembles sample two kinds of forecast uncertainty (Schneider and Griffies 1999; Schwierz et al. 2006): In short-term and medium-range weather forecasting, they address the forecast uncertainties arising from uncertain initial conditions. In seasonal forecasting, on the other hand, ensembles sample the distribution of states that are consistent with an evolving boundary condition, such as anomalies in sea surface temperature or even greenhouse gas forcing (Liniger et al. 2007). Both medium-range weather forecasts (e.g., Molteni et al. 1996; Buizza et al. 2005; Toth et al. 1997) and seasonal climate predictions (e.g., Anderson et al. 2003; Graham et al. 2005; Barnston et al. 2003) are meanwhile well-established techniques and operationally issued by many weather and climate services. However, on the intraseasonal or monthly time scale (i.e., in between the two “extremes” of the initial value problem of medium-range weather forecasting and the boundary value problem of seasonal forecasting) there is still a pronounced gap in prediction capability (Waliser 2006).

The central problems of monthly predictions are (i) that the monthly time scale is too long for conventional weather forecasting, since the atmospheric system has typically lost its memory of the initial conditions after less than 14 days; and (ii) that it is too short for the influence of the ocean state to differ significantly from its initial state and thus to beat persistence forecasts (Vitart 2004). While there actually are processes with pronounced variability on the intraseasonal time scale, particularly the Madden–Julian oscillation (MJO; see e.g., Madden and Julian 1971; Waliser 2006; Vitart et al. 2007) and stratosphere–troposphere interactions (e.g., Baldwin et al. 2003), and while these processes do have the potential to act as sources of predictability, it is fair to say that they are not yet well reproduced by operational global circulation models (e.g., Vitart et al. 2007; (World Climate Research Programme) WCRP 2008). It is probably for these reasons that, to the authors’ knowledge, there are currently only two weather centers that operationally issue monthly forecasts on the basis of a global circulation model. These are the European Centre for Medium-Range Weather Forecasts (ECMWF; see Vitart 2004), and the Japan Meteorological Agency (JMA; see Kurihara 2002). This is in striking contrast to the wide range of application fields in weather and climate risk management that would profit from decision support on the monthly time scale, such as agriculture (e.g., Meinke and Stone 2005), insurance companies (e.g., Zeng 2000), the energy sector (e.g., Cherry et al. 2005), and the health services (e.g., McGregor et al. 2006).

So far, only very little has been published on the actual monthly prediction skill of existing global circulation models. Most notably, Vitart (2004) has provided an overview of the design and capability of the ECMWF monthly ensemble forecasting system (MOFC). Based on the evaluation of 45 forecast samples, covering a span of 2 yr, he concludes that the MOFC performs better than climatology or persistence until day 18, and that even after 20 days of integration some regions are skillful. However, it is not clear how representative these results are, given that no confidence intervals are provided, and given that extended-range predictability may vary from year to year and, for example, strongly depends on the phase of El Niño (Shukla et al. 2000). Indeed, a longer period covering more verification years would be desirable. While the MOFC actually does provide 12 yr of reforecasts (or “hindcasts”), these data have so far not been considered for verification, apart from a brief estimation of hemispherically averaged skill based on the relative operating characteristics (ROC) skill metric, which is more a measure of potential predictability rather than of “true” prediction skill (Doblas-Reyes et al. 2005).

The central problem of evaluating a MOFC-type prediction system in a fully probabilistic way, considering all hindcast data available, is the small size of the hindcast ensembles (only 5 members) in comparison to the actual forecasts (51 members). This difference has severe consequences on any probabilistic skill metric that is sensitive to reliability. Consider for example a 5-member ensemble prediction system, with probability forecasts being issued by taking the fraction of ensemble members that predict a given event (e.g., “warmer than normal”): only a finite number of discrete probability forecasts can be issued that way, namely, 0%, 20%, 40%, 60%, 80%, and 100%. On the other hand, the observed frequency of this event, conditioned on the forecast probabilities, is in principle not restricted to these probabilities. Therefore, such small-sized ensemble hindcasts are inherently miscalibrated, or intrinsically unreliable (Richardson 2001; Weigel et al. 2007a, b; Ferro et al. 2008), leading to a strong underestimation of the true prediction skill.

It is therefore one of the objectives of this study to define and set up a fully probabilistic verification context that corrects for the problem of intrinsic unreliability without ignoring true (i.e., model induced) reliability deficits, and that allows to jointly consider small hindcast ensembles and large forecast ensembles in the verification. Building upon the work of Vitart (2004), we then apply this verification approach to systematically assess the skill of probabilistic MOFC near-surface temperature predictions in dependence of region, lead time , and season.

The paper is structured as follows: in section 2, the operational ECMWF monthly prediction system is described, together with the observational dataset used for verification. Section 3 provides a detailed description of the verification methodology applied. Section 4 gives a systematic overview of the prediction skill, and skill uncertainty, of MOFC temperature forecasts. A discussion on the potential for further skill improvement is given in section 5, and concluding remarks are provided in section 6.

2. Data

a. The ECMWF monthly prediction system

A full and detailed description of the MOFC can be found in Vitart (2004) and at ECMWF (2007). Here a short summary is provided. The MOFC is a coupled atmosphere–ocean global circulation model. The atmospheric component is the ECMWF atmospheric model Integrated Forecast System (IFS), which is run at TL159L62 resolution. This corresponds to a horizontal resolution of about 125 km and 62 levels in the vertical. The ocean component is the Hamburg Primitive Equation Model (HOPE; Wolff et al. 1997) with 29 levels in the vertical. The atmosphere and the ocean component are coupled through the Ocean–Atmosphere–Sea Ice–Soil coupler (OASIS; Terray et al. 1995).

Since 7 October 2004, the MOFC is initialized and integrated forward in time every Thursday to provide forecasts of the upcoming 32 days. It is set up as an ensemble with 51 members. Along with each real-time forecast, 5-member ensemble hindcasts are initialized at the same day of the previous 12 yr and equally integrated for 32 days. For example, the real-time 51-member forecast that was initialized on 19 January 2006 was accompanied by the integration of 5-member hindcasts that were initialized from data of 19 January 2005, 19 January 2004, . . . , 19 January 1994 (illustrated in Fig. 1).

The atmospheric component of the real-time forecasts is initialized from the ECMWF operational analysis; the hindcasts are initialized from the 40-yr ECMWF reanalysis (ERA-40; Uppala et al. 2005) if available (i.e., before August 2002) and from the ECMWF operational analysis afterward. Note that the operational analysis and the ERA-40 data do not share the same land surface representation (e.g., different soil roughness, different soil model versions, change in snow analysis, and different resolution). Most of those changes do not seem to have a significant impact on the monthly forecasts except those related to the changes in snow analysis that can generate spurious anomalies. However, the number of grid points affected by this inconsistency is small and should not affect the results of the paper. The ocean initialization stems from the ECMWF ocean data assimilation system (ODASYS), which is also used for the seasonal forecasts. However, since some of the data necessary for the ocean assimilation arrive with considerable delay, the ocean initial conditions lag the real time by about 12 days. To obtain a real-time estimate of the ocean conditions to be used for initialization, the ocean model is integrated from the respectively last analysis, thus “predicting” the actual ocean analysis.

The atmospheric perturbations for ensemble generation are produced in the same way as for the medium-range ensemble prediction system [i.e., by applying the singular vector method; Buizza and Palmer (1995); and including perturbations by targeting tropical cyclones; Puri et al. (2001)]. Additionally, the concept of stochastic physics (Palmer 2001) is applied to capture at least part of the uncertainties due to model formulation. The oceanic perturbations are generated in the same way as in the operational ECMWF seasonal forecasts (for more details see Vialard et al. 2005).

The ECMWF issues its monthly forecasts as weekly means; the predicted variables are averaged over four 7-day intervals (Monday–Sunday), which correspond to days 5–11 [week 1 (W1)], days 12–18 [week 2 (W2)], 19–25 [week 3 (W3)], and days 26–32 [week 4 (W4)]. To keep our evaluations consistent with the operational forecasts and with the verification of Vitart (2004), we apply the same averaging convention (see also Fig. 2) unless stated differently.

The forecasts considered in the present study cover the entire year 2006 (providing 52 samples forecast realizations), and the corresponding hindcasts cover the years 1994–2005, providing another 12 sets of 52 predictions each.

b. Observations

The forecasts are verified against ERA-40 data (Uppala et al. 2005) if available (i.e., for all hindcasts prior to August 2002) and against the ECMWF operational analysis otherwise. At least for ERA-40, several comparison studies have shown that near-surface temperatures are well represented by this dataset, justifying its use for the present study (Simmons et al. 2004; Kunz et al. 2007). For the verification, both forecasts and “observations” are interpolated on a grid with 1° × 1° resolution. One of the problems in this context is that the resolution and thus the gridded orography of the operational analysis has repeatedly slightly changed during the verification period considered. To avoid inhomogeneities in the temperature time series, we therefore extrapolate all surface temperatures of the operational analysis onto the ERA-40 orography, applying a lapse rate of 0.006 K m−1.

3. Verification context

a. The RPSSD skill metric

Our verification is based on a modified version of the widely used ranked probability skill score (RPSS; Epstein 1969; Murphy 1969,1971). The classical RPSS is a squared measure comparing the cumulative probabilities of categorical forecast and observation vectors relative to a climatological forecast strategy. It is defined by
i1520-0493-136-12-5162-e1
The brackets 〈. . .〉 denote the average of the RPS and RPSCl values over a given number of forecast–observation pairs. Here Yk is the kth component of a cumulative categorical forecast vector Y, and Ok is the kth component of the corresponding cumulative observation vector O the forecast is verified against. That is, Yk = Σki=1yi, with yi being the probabilistic forecast for the event to happen in category i, and Ok = Σki=1oi with oi = 1 if the observation is in category i and oi = 0 if the observation falls into a category ji. Analogously, Pk is the cumulative climatological probability of the kth category. A more detailed description of the RPSS is provided in Wilks (2006). The RPSS is a favorable probabilistic skill score in that it is sensitive to distance (i.e., a forecast is increasingly penalized the more its cumulative probabilities differ from the actual outcome). Moreover, the RPSS is strictly proper, meaning that it cannot be optimized by hedging the probabilistic forecasts toward other values against the forecaster’s true belief. A big caveat of the RPSS is its strong negative bias for small ensemble sizes (e.g., Buizza and Palmer 1998; Richardson 2001; Kumar et al. 2001; Mason 2004). The reason for this bias is the intrinsic unreliability (Weigel et al. 2007a) of small ensembles, leading to inconsistencies in the formulation of the RPSS. However, particularly when the skill of large-sized ensemble forecasts is to be estimated on the basis of small-sized ensemble hindcasts, a “biasless” skill score is required (i.e., a skill score that is insensitive to ensemble size).
Müller et al. (2005) and Weigel et al. (2007a, b) have derived a debiased version of the RPSS, the so-called discrete ranked probability skill score (RPSSD). The RPSSD lacks the RPSS’s strong dependence on ensemble size while retaining its favorable properties, in particular the strict propriety, making it the skill score of choice for the present study. If K forecast categories are considered, and if these K categories are equiprobable (as is chosen to be the case in this study), the RPSSD assumes a relatively simple analytical form and is given by
i1520-0493-136-12-5162-e2
i1520-0493-136-12-5162-e3
with M being the ensemble size of the prediction system. If predictions of different ensemble sizes are to be jointly evaluated in one verification sample, this can be considered adequately by replacing the ensemble size M in Eq. (2) by an appropriate effective ensemble size Meff. If, for example, a set of predictions feeding into one verification consists of Nf forecasts with ensemble size Mf , and Nh hindcasts with ensemble size Mh (in the MOFC we have Nf = 1, Mf = 51, Nh = 12, and Mh = 5), then Meff is given by (see the appendix for the derivation):
i1520-0493-136-12-5162-e4

b. The quantiles

In this study, MOFC forecasts of weekly averaged temperature are evaluated for three climatologically equiprobable categories: “colder than normal,” “normal,” and “warmer than normal”. The terciles separating these three categories are determined separately for each week of the year to adequately account for the seasonal cycle of temperatures. Otherwise, artifical “false skill” could be introduced as described by Hamill and Juras (2006). To account for systematic model errors and model drift (see section 4a), the terciles defining model climatology are not determined from the observations but from the pool of forecasts and hindcasts available. Since the model climatology directly after initialization is different from the state of the model after, say, three weeks of integration, the model terciles are calculated individually for each lead time considered.

To guarantee that no information from a given prediction is used in the verification process of that very prediction, we apply the quantile calculation in a leave-one-out cross-validation mode (e.g., Wilks 2006). This means, for each year considered the quantiles are calculated separately, using information from all years except the one under consideration.

The comparatively short hindcast period of the MOFC imposes a major problem for estimating terciles: to obtain terciles in weekly resolution, they need to be estimated on the basis of only 13 yr of data, which corresponds to 12 samples due to the cross validation applied. To enhance the statistical robustness, we therefore additionally include the week before and after the week of interest to estimate the terciles. That way, the effective sample size is increased from 12 to 36. This procedure is applied both for the observations and for the MOFC predictions for each lead time considered. This procedure results in a smoother seasonal cycle of the terciles (not shown).

c. Confidence intervals

To make inferences about the true value of a skill score, it is essential to estimate confidence intervals around the measured skill value (Joliffe 2007). This is particularly important in the context of the MOFC, where both the sample size (only 13 yr of hindcast data) and the ensemble size (only 5 members in the hindcasts) are low. As a straightforward nonparametric approach to obtain such confidence intervals, we apply a bootstrap method as described in Efron and Gong (1983) and Wilks (2006).

Assume, a skill score S (in our case the RPSSD), is calculated at a given grid point for a given week of the year. The skill score S is then a function of Ns forecast–observation pairs d1, d2, . . . , dNs, with Ns being the number of sample years feeding into the verification. In the present case, and applying the terminology of Eq. (4), Ns = Nh + Nf = 13. The sampling uncertainty of S can be estimated by recomputing the value of S Nb times on the basis of Nb random samples (with replacement) of sample size Ns from {d1, d2, . . . , dNs}. From the resulting distribution of Nb S values, the 90% percent confidence interval is then estimated by calculating the 5%–95% interquantile range. In the present study Nb = 1000.

Now assume that a skill score average, 〈S〉, is calculated (as will be done frequently in the remainder of this study). This could be a spatial average over the skill scores S obtained at Ng grid points (e.g., the RPSSD values at all European grid points for a given week of the year), or a temporal average over the skill scores S obtained for Nt forecast weeks (e.g., the RPSSD values of all summer weeks at a given grid point), or both. Thus, in the general case, the test statistic 〈S〉 is based on Ns × Nt × Ng data pairs. However, following Candille et al. (2007), also in this case the resampling is performed only over the Ns verification years rather than all Ns × Nt × Ng data pairs, since that way the spatial and temporal correlation of forecast errors can be crudely accounted for. Otherwise, the number of independent samples would be overestimated and the confidence interval size underestimated.

4. Results

a. Model drift

We start our evaluations by examining the systematic absolute deviations of MOFC output from the verifying observations. Figure 3 shows the annual mean bias of 2-m temperature for forecasting weeks W1–W4. The model develops a substantial bias in many regions already in W1 (Fig. 3a). For parts of Greenland, South America, Southeast Asia, and Antarctica the model is more than 1 K colder than the observations. Also southern Europe is affected by a cold bias. Large areas in continental North America and the polar oceans, on the other hand, show a warm anomaly of at least 1 K. Other areas of large differences between the model and the observations include many coastal regions, such as the entire western coast of the two Americas and southwest Africa. Off the coastlines, the oceans are generally too cold. During forecast weeks W2 until W4 the anomalies further amplify, while revealing the same spatial pattern as in W1 (Figs. 3b–d). In particular, the cold anomalies over Greenland are nearly doubled from W1 to W2, and over Antarctica a maximum bias of nearly 2 K is developed by W4. For the nonpolar oceans, an increase of the negative bias from 0.3 to 0.5 K on average can be detected after the full integration time. Generally the error growth saturates after about 20 forecast days. This result, as well as the stationarity of the spatial bias distribution, is consistent with the results of an analysis of the MOFC bias in 500-hPa geopotential height as described by Jung (2005). A more in-depth analysis (Baggenstos 2007) shows that the spatial distribution and magnitude of biases also reveals some visible seasonal variability, but their large-scale spatial structure remains similar throughout the year.

The observation of a pronounced model drift per se is not surprising, given that no “artificial” measures are applied to suppress model drift, and given that imbalances in the initial state of the model are not corrected (Vitart 2004). What is more surprising is that the bias is very pronounced already during W1. This contrasts with findings of Vitart (2004), who evaluated the MOFC bias of 500-hPa geopotential height and found that only after 10 days of model integration the model output begins to systematically deviate from the analysis. This indicates that processes not directly linked to circulation are probably contributing to the observed systematic temperature errors (e.g., a wrong representation of surface processes or problems in the ocean–atmosphere coupling). While this topic is not further evaluated here, the existence of such strong biases emphasizes the need for reforecasts and an appropriate verification, so that systematic model errors can be diagnosed and corrected.

b. Prediction skill

We now proceed with the probabilistic verification of the MOFC temperature forecasts in dependence of lead time, region, and season, applying the verification context described in section 3. This is done for each grid point, each week of the year, and each of the four lead times W1–W4. We start by evaluating annually averaged skill.

1) Annual mean skill

Figure 4 shows maps of the RPSSD for forecast weeks W1–W4, averaged over all 52 weeks of the year.

For W1 forecasts, prediction skill exceeds a value of 0.3 almost everywhere. Particularly high values on the order of 0.7 are assumed over the central and eastern equatorial Pacific [i.e., the region of the El Niño–Southern Oscillation (ENSO)]. On the other hand, comparatively low skill of only 0.1 or even below is found over the tropical Atlantic, Indian, and western Pacific Oceans. The low skill values over the tropics are likely because of two reasons. First, outside the ENSO region, interannual variability is very low. Indeed, the average tercile boundaries for a week in January over the equatorial Atlantic are at 299.50 and 229.80 K. This means, the difference between a warm and a cold week is only 0.3 K, making categorical temperature forecasts very error prone. For comparison, over the ENSO region, the terciles are 1.6 K apart. The second reason is that, in relative terms to the interannual variability, the error growth in the tropics is larger than in the extratropics, since the errors are mainly associated with convection rather than baroclinic instability as in the extratropics (Shukla 1981). This, together with low daily fluctuations, leads to a shorter upper limit of predictability.

In W2, prediction skill is substantially reduced almost everywhere. This particularly applies for continental skill, which now only more assumes values on the order of 0.1. Southern America and Africa thereby reveal slightly better prediction skill than North America, Eurasia, and Australia. Considerably more skill is found over the oceans, but also here the decrease w.r.t. W1 is pronounced.

Both W3 and W4 are characterized by the loss of basically all skill over the continents except for some areas in the tropics, in particular the northern part of South America, and Indonesia. Over the oceans, the ENSO region retains its high predictability well into W4. Generally, the patterns of skill distribution observed in W2 remain, but their magnitudes decrease.

It is important to note that there are only very few regions (e.g., the tropical Atlantic) where skill is negative, meaning that the forecasts are only rarely worse than climatology. This is in clear contrast to seasonal forecasts, where because of ensemble overconfidence, wide areas of negative RPSSD skill are found (Weigel et al. 2008). A further discussion of this aspect is provided in section 5.

So far, the focus has been on annual averages of MOFC prediction skill. We continue with an evaluation of the seasonal variability in prediction skill.

2) Seasonal and regional dependence of skill

Figure 5 displays the annual cycle of MOFC temperature prediction skill, averaged over all grid points of the northern and southern extratropics as well as the tropics. The time series are smoothed by applying a centered five-point moving-average filter with symmetric weights (1/9, 2/9, 3/9, 2/9, 1/9) (e.g., Chatfield 2004) to identify the major intra-annual characteristics. In all panels, the absolute skill drop between forecast weeks W1 and W2 is seen to be independent of the time of the year, but not independent of the region. Indeed, the loss of skill is less pronounced in the tropics where, as has been discussed with Fig. 4, skill is either already low from the beginning (tropical Indian and Atlantic Ocean), or remains high for all lead times (ENSO region in the equatorial Pacific).

Over the northern extratropics, the minimum in W1 skill is found in boreal summer, which is consistent with common experience in conventional weather forecasting. The skill of W2, W3, and W4, on the other hand, does not reveal a significant annual cycle. If anything, W2 skill is seen to be slightly enhanced in late winter (February and March), a result that is consistent with the findings of Vitart (2004). A similar, but more pronounced, pattern of seasonal skill variability is found in the southern extratropics, where even W4 predictability reveals a significant annual cycle with a minimum in late boreal summer and a maximum in boreal winter. Over the tropics, on the other hand, this cycle is less pronounced, but still visible. The reasons for this enhanced winter predictability are not evaluated here, but they may well be linked with the magnitude of ENSO-induced SST anomalies, which are strongest in boreal winter.

To obtain more regional information of the prediction skill for different times of the year, global maps of seasonally averaged W2-skill of temperature have been calculated and are displayed in Figs. 6a–d for boreal winter, spring, summer, and autumn, respectively. The corresponding maps of W3 and W4 prediction skill are not shown here but can be found in Baggenstos (2007). They look qualitatively similar to Fig. 6, but with strongly reduced magnitude. Figures 7 and 8 display W2 prediction skill in greater detail for two economically important extratropical regions, namely, North America and Europe. Finally, in Table 1, the average prediction skill as well as its uncertainty have been calculated in dependence of season and lead time, averaged over the entire globe, the Niño-3.4 region (5°S–5°N, 170°–120°W), and over the continental climate regions as defined by Giorgi and Francisco (2000).

The figures and table are not discussed individually, but a summary and synthesis of the key results on MOFC temperature prediction skill is presented in the following.

  1. The figures for global mean temperature prediction skill in the first row of Table 1 confirm the conclusions drawn above from Fig. 5, namely, that global predictability is stronger during boreal winter and weakest in boreal summer for all lead times considered.

  2. The skill reveals a very pronounced seasonal dependence in the eastern and tropical Pacific, as can be seen from Fig. 6. The skill is seen to be stronger in autumn and winter, while being weaker in spring and summer. This is consistent with the skill values for the Niño-3.4 region in Table 1, which, for example, show a W2 skill of 0.67 ± 0.10 in winter, but only 0.34 ± 0.14 in spring.

  3. Over most continental regions, the skill of W3 and W4 forecasts is at best marginal, but usually not even significant. Exceptions are some tropical regions such as the Amazon basin during all seasons, southern South America in spring, eastern and western Africa in winter, and southern Africa in spring, revealing W3 and W4 skill on the order of 0.1.

  4. The lowest value of continental W1 prediction skill is found over Southeast Asia during all seasons and over tropical Africa and South Asia in summer (on the order of 0.1–0.2). The reason is probably the enhanced error growth rate and the reduced interannual variability as discussed in section 4b(1).

  5. The highest value of continental W1 and W2 skill is found in winter over eastern North America, with a skill value of 0.50 ± 0.05 and 0.25 ± 0.04, respectively. The enhanced W2 winter predictability over the eastern United States is also clearly visible in Fig. 7a. Note that a different picture is obtained with the National Centers for Environmental Prediction (NCEP) medium-range model where maximum W2 winter skill is found in the upper Midwest (Hamill et al. 2004). During the rest of the year, North American W2 skill is much lower in spring and weakest in summer and autumn. No pronounced W3 prediction skill is found over any of the North American Giorgi regions (see Table 1), contrasting with the results of Vitart (2004) who found predictability in North America even beyond day 18.

  6. Also in the European domain, W1 and W2 prediction skill tends to be higher in winter as compared with the other seasons. However, the seasonal variability in prediction skill is less pronounced than in North America and not always significant. Higher skill is observed if also ocean grid points are considered. For example, Fig. 8 shows that W2 prediction skill of the Mediterranian Sea is more skillful in autumn and winter, while over the North Sea exactly the opposite is found with a pronounced maximum in summer. As over most other extratropical continental regions, the W3 and W4 prediction skill of mean near-surface temperature is very low.

c. Persistence forecasts

The evaluations presented above are based on climatology as a reference with which to be compared. Another common reference strategy, which is sometimes even considered to be a harder test, is persistence. To evaluate the prediction skill of persistence forecasts, we follow the approach of Vitart (2004) and persist W1 probability forecasts to “predict” weeks W2, W3, and W4. In other words, the W2 (W3, W4) persistence forecast is a “normal” W1 probability forecast that is verified against the observations of W2 (W3, W4). The verification is carried out in the same way as described above (i.e., by applying the debiased RPSSD skill metric with climatology as a reference).

Maps of the annually averaged persistence prediction skill are shown in Fig. 9 for forecast weeks W2–W4. Already in W2, persistence performs worse than climatology on all continental landmasses except for equatorial South America. Over the oceans, significant positive skill is only observed over the ENSO region and parts of the Atlantic. A comparison with Fig. 4 reveals that it is only in these oceanic regions that persistence performs almost equally well as the MOFC predictions. The observed contrasts become even more pronounced during forecast weeks W3 and W4: only the ENSO region over the Pacific and the trade wind zones over the Atlantic retain their positive skill, indicating that the high MOFC skill observed in these regions (see Figs. 4c,d) is to a large degree a consequence of SST persistence effects. On the other hand, basically all continental areas as well as the extratropical oceans are characterized by significantly negative persistence skill. This is mainly because ensemble spread usually widens as the lead time increases (i.e., the W1 forecasts are typically less dispersed than forecasts for W2 and later; Vitart 2004). This implies that W1 ensemble distributions tend to be overconfident given the weak predictability found in forecast weeks W2–W4. Overconfidence, in turn, is a property that is heavily penalized by the RPSSD skill metric as has been shown by Weigel et al. (2008).

All in all, the results show that, in a probabilistic prediction context, and for these time scales, persistence is not a competitive alternative for MOFC temperature forecasts or even climatology (over the continents).

5. Discussion

It is not the aim of this study to evaluate the physical reasons behind the observed spatial and seasonal variability in prediction skill. Rather, in our discussion we want to focus on a key result, which is most relevant from a user perspective; namely, the finding that extratropical continental prediction skill essentially vanishes after forecast week W2 (i.e., that the MOFC does not seem to provide additional value with respect to climatology for lead times that are beyond the prediction range of medium-range weather forecasts). In this context, we want to discuss the following two questions: Could the limit of predictability be extended beyond W2 by applying appropriate postprocessing methods? Could the limit of predictability be extended by increasing the averaging period? Apart from that, we want to discuss an example that shows that, even if the average prediction skill is low, there may nevertheless be some forecasts that contain useful information.

a. Can recalibration improve the monthly forecasts beyond forecast week W2?

For most probabilistic skill scores, including the RPSSD, zero skill as observed in forecast weeks W3 and W4 can be because of two reasons: (i) the absence of a physically predictive signal per se, or (ii) wrong ensemble spread (i.e., poor reliability). In the latter case, some improvement in prediction skill could at least in principle be achieved a posteriori by recalibration [Toth et al. (2003), e.g., by inflating the ensemble spread as for example described in Doblas-Reyes et al. (2005)]. To estimate the potentially achievable skill, we apply a method published in Hamill et al. (2004). By random sampling from bivariate normal distributions having specified correlation, Hamill et al. (2004) generate correlated time series of synthetic observations and ensemble means. From these, they establish an empirical relationship between the expected RPSS of reliable forecasts on the one hand and the correlation on the other hand (see their Fig. 5d). To investigate whether the MOFC forecasts follow this curve, the correlation coefficients between the MOFC ensemble mean forecasts and the corresponding obsverations have been calculated for each week of the year and each grid point for forecast weeks W1–W4. Maps of the resulting annually averaged correlation coefficients are shown in Fig. 10. In Fig. 11, all annually averaged continental RPSSD values of Fig. 4 are plotted against the corresponding correlation values of Fig. 10 for the four lead times. The method of Hamill et al. (2004) has been adjusted for finite-sized ensembles, and the resulting RPSSD correlation curve is shown in Fig. 11 as a gray line. Assuming Gaussian behavior of the observed and simulated weekly temperature averages, this relationship essentially displays the RPSSD values that would be obtained were the MOFC forecasts reliable. Admittedly, the Gaussian assumption is a very simplifying one, but it can be justified as a first rough estimate for the variable considered (Wilks 2002, 2006). Figure 11 shows that particularly during W1 many points are below the gray line, implying that the prediction skill can be potentially improved by recalibration. However, during W3 and particularly W4 the MOFC prediction skill follows the gray line (i.e., the continental W3 and W4 forecasts are essentially reliable and can hardly be further improved by recalibration).

While this result may appear disappointing, it also has a positive facet in that the RPSSD is generally not negative. In other words, in the worst case MOFC forecasts are, due to their reliability, essentially identical to the climatological forecast, but not worse. This implies that the use of MOFC forecasts, while not always being valuable, is generally not harmful either. This contrasts with longer-term seasonal forecasts, which often do reveal negative skill due to strong overconfidence and can therefore be worse than climatological guessing (Weigel et al. 2008).

b. Can larger averaging periods improve prediction skill?

In all our analyses so far, we have only considered 7-day temperature averages. The question rises whether predictability can be increased and the limit of predictability extended by choosing another averaging period. We evaluate this question for two example regions, namely, Europe and the Niño-3.4 region. For these regions, annually averaged prediction skill is now not only calculated for 7-day temperature averages, but also for 3- and 14-day averages. The results are displayed in Fig. 12 as a function of the number of days between initialization and the end of the respective prediction interval. Note that the lead time is now incremented by day rather than by week as before. The picture shows that both for Europe and the Niño-3.4 region skill increases as the length of the averaging interval becomes larger. This is consistent with the study of Rodwell and Doblas-Reyes (2006) who evaluated the effect of averaging period on temperature forecasts during the European heat wave of 2003. There are two reasons for this behavior. First, increasing the averaging period implies that the higher skill of earlier lead times contributes to the averages. For example, the 3-day prediction interval that ends 20 days after initialization covers the forecasting days 18–20, while the corresponding 7-day (14 day) interval includes days 14–20 (7–20; i.e., contains more of the high predictability found for shorter lead times). Particularly in Europe this can have a big effect, where skill is observed to drop rapidly after only one week from initialization. Second, nonpredictable high-frequency noise is filtered out as the prediction intervals grow. Particularly for the Niño-3.4 region, the latter effect seems to be dominant, since skill drops too slowly with lead time as to explain the observed gain in prediction skill solely by the inclusion of days with shorter lead and higher skill. The observation that skill can be enhanced by increasing the averaging intervals is certainly a result that is not very surprising; on the contrary, it may appear rather trivial. However, this property is only rarely explicitly mentioned, even though it may have quite some relevance from a user perspective: it implies that the limit of predictability can be significantly extended into the future for users who do not require forecast information in temporal resolutions as high as weekly. The same applies for users who feed raw forecasts into application models with some degree of memory or inertia (i.e., with integrating characteristics that carry the high prediction skill of short-lead forecasts further into the future). For instance, preliminary results from a case study where MOFC forecasts over Switzerland have been fed into a soil moisture prediction model, have revealed that the limit of predictability is extended by almost a week if the output of the soil moisture predictions is evaluated rather than the raw MOFC forecasts (Bolius et al. 2007).

c. A further comment

From seasonal forecasting, it is known that prediction skill can be closely linked to the presence or absence of certain climatic boundary conditions. For example, seasonal winter prediction skill in the North American–Pacific region is higher during ENSO years than during the other years (Shukla et al. 2000). Similarly, it may be possible to find situations of enhanced predictability on the monthly time scale beyond W2, even if the average predictability is low. As an example, we consider MOFC winter forecasts (December–March) for temperature in continental Europe. We stratify the verification samples on two subsets, depending on whether the prediction week falls into a phase of positive or negative North Atlantic Oscillation (NAO) index [the NAO index is defined as the sea level pressure difference between the Azores and Iceland and is closely connected to the storminess characteristics in Europe (Hurrell 1995; Jung et al. 2003); data of the Climate Research Unit (CRU) Web site page have been used (see online at http://www.cru.uea.ac.uk/cru/data/nao.htm)]. These two subsets are verified separately and displayed in Fig. 13 as a function of prediction time (i.e., the time between initialization and the end of the prediction interval), again with lead time being incremented daywise as in Fig. 12a rather than weekwise as before. The results show that for prediction times of 12–28-day skill during negative NAO (NAO) phases is higher than during positive NAO (NAO+) phases. The reasons for this observation may be linked to the augmented occurrence of persistent blockings observed during NAO phases (e.g., Pavan et al. 2000; Scherrer et al. 2006), but are not further investigated here. To quantify the significance of this finding, the 10% and 90% percentiles of the skill difference between NAO and NAO+ phases have been estimated by bootstrapping (cf. section 3c). The skill difference and the corresponding confidence range is plotted in Fig. 13b, showing that predictability during NAO phases outperforms NAO+ skill on the 90% confidence level at least for prediction times from 15 to 19 days. Of course, in practical terms this result is not very useful, since there is no value in predicting the predictability when the forecasts are reliable anyway. This example should rather be understood as a demonstration that, even when the average predictability is close to zero, there may be climatological conditions under which the skill is enhanced and the forecasts actually do contain more useful information than suggested by the averaged figures in Table 1. This conclusion is probably not only restricted to the presence or absence of certain flow regimes, but may also apply to other aspects such as the surface conditions. For example, it is imaginable that the MOFC predictability of the European summer may depend on the amplitude of the initial soil moisture anomalies.

All in all, the aforementioned decline of MOFC prediction skill for lead times beyond forecasting week W2 would be less stringent than suggested by Table 1, if larger averaging intervals were considered. However, the value of increased averaging periods is highly user specific and to some degree only hypothetical. Because of the surprisingly good reliability, which by itself is a very favorable property, it is not expected that any form of recalibration would significantly improve the MOFC prediction skill.

6. Conclusions

In this paper we presented a probabilistic verification of surface temperature forecasts obtained from the operational ECMWF monthly forecasting system, a coupled atmosphere–ocean global circulation model. The main objectives of this study have been to (i) define and set up an appropriate probabilistic verification framework, and (ii) to apply this verification framework for a systematic and fully probabilistic skill assessment of MOFC temperature forecasts in dependence of lead time, region, and season. The verification is based on the debiased ranked probability skill score (RPSSD), a probabilistic and strictly proper skill score that has the advantage of not penalizing the intrinsic unreliability induced by small ensemble sizes—a property that is favorable in the present context, since the ensemble size of MOFC hindcasts is very small. The RPSSD has been generalized such that the small-sized hindcast ensembles and large-sized forecast ensembles can be jointly considered in the verification. Moreover, in contrast to earlier work of this type, the sampling uncertainty of the skill values obtained has been estimated by a bootstrapping approach. The following results have been found.

  1. Already from the first forecasting week, the MOFC reveals significant systematic biases on the order 1 K and more, also over land. While this result is not surprising, given that no measures are applied to suppress model drift, it manifests the need for hindcasts to estimate the model climatology and to correct for such systematic errors.

  2. The seasonal cycle of global predictability reveals higher skill in boreal autumn and winter and lower skill in boreal spring and summer for all lead times.

  3. Continental prediction skill during forecast week W1 (days 5–11 from initialization) is on the order of 0.4–0.6 in the extratropics and on the order of 0.2–0.3 in the tropics. In forecast week W2 (days 12–18), continental skill drops to much lower values of about 0.1–0.2. The highest continental W1 and W2 prediction skill is found over the eastern United States in boreal winter.

  4. Beyond forecast week W2, pronounced prediction skill is predominantly found over the oceans, particularly over the ENSO region in the central and eastern Pacific. Continental skill, however, essentially vanishes after W2, apart from some areas in tropical Africa and South America.

  5. An alternative forecasting strategy that is based on persisting W1 forecasts turns out to be worse than climatological guessing, at least over land grid points and in a probabilistic verification context.

  6. Only little can be done to improve continental prediction skill of near-surface temperature beyond forecast week W2 for the investigated model. The potential of a posteriori recalibration is limited because the forecasts are surprisingly reliable. Some skill improvement can be achieved by increasing the averaging period considered. While this averaging effect is to some degree trivial in that simply the high-skill forecasts of shorter lead times are carried further into the future, this may still be of relevance for specific user groups.

  7. When the average prediction skill is low, there may nevertheless be climatologic conditions under which the forecasts contain some useful information. For example, there is evidence that during NAO conditions in the winter months, the limit of temperature predictability for continental Europe is extended by about 4–5 days.

All in all, the results presented may appear disappointing in that they imply that the MOFC does not have much additional information on time scales that are not already covered by (slightly extended) medium-range weather forecasts. While only temperature forecasts have been considered here, we do not expect fundamentally better results for other variables such as precipitation, which is often considered to be even harder to predict. However, the observed low prediction skill of tercile-based weekly averages does not necessarily imply an equally low prediction skill of extreme events, and it could be worthwhile to further investigate this aspect.

Generally, we believe that a major improvement in monthly predictability might be feasible once the Madden–Julian oscillation (MJO), the most important source of intraseasonal variability, is better represented by the model, at least in the tropics. Having said that, there are indications that the MJO might even affect the intraseasonal variability in the extratropics (Ferranti et al. 1990; Cassou 2008). Currently, however, the MOFC is not able to predict the MJO beyond 14 days and has problems maintaining the amplitude of an MJO event. Vitart et al. (2007) stress that better parameterizations and a better representation of ocean mixing is needed. This, together with a better representation of other slowly evolving boundaries the troposphere interacts with (e.g., the sea surface, soil moisture, snow cover and the stratosphere) might eventually effect some improvement in MOFC prediction skill.

Acknowledgments

This study was supported by the Swiss National Science Foundation through the National Center for Competence in Research (NCCR-Climate) and by the ENSEMBLES project (EU FP6 Contract GOCE-CT-2003-505539).

REFERENCES

  • Anderson, D. L. T., and Coauthors, 2003: Comparison of the ECMWF seasonal forecast systems 1 and 2, including the relative performance for the 1997/8 El Niño. ECMWF Tech. Memo 404, 93 pp.

  • Baggenstos, D., 2007: Probabilistic verification of operational monthly temperature forecasts. Tech. Rep. 76, Federal Bureau of Meteorology and Climatology MeteoSwiss, Zurich, Switzerland, 52 pp.

  • Baldwin, M. P., D. B. Stephenson, D. W. J. Thompson, T. J. Dunkerton, A. J. Charlton, and A. O’Neill, 2003: Stratospheric memory and skill of extended-range weather forecasts. Science, 301 , 636640.

    • Search Google Scholar
    • Export Citation
  • Barnston, A. G., S. J. Mason, L. Goddard, D. G. DeWitt, and S. E. Zebiak, 2003: Multimodel ensembling in seasonal climate forecasting at IRI. Bull. Amer. Meteor. Soc., 84 , 17831796.

    • Search Google Scholar
    • Export Citation
  • Bolius, D., P. Calanca, A. Weigel, and M. A. Liniger, 2007: Prediction of moisture availability in agricultural soils using probabilistic monthly forecasts. Geophysical Research Abstracts, Vol. 9, EGU2007-A-02175, European Geosciences Union, 1 p. [Available online at http://www.cosis.net/abstracts/EGU2007/02175/EGU2007-J-02175.pdf.].

  • Buizza, R., and T. N. Palmer, 1995: The singular-vector structure of the atmospheric global circulation. J. Atmos. Sci., 52 , 14341456.

    • Search Google Scholar
    • Export Citation
  • Buizza, R., and T. N. Palmer, 1998: Impact of ensemble size on ensemble prediction. Mon. Wea. Rev., 126 , 25082518.

  • Buizza, R., P. L. Houtekamer, Z. Toth, G. Pellerin, M. Wei, and Y. Zhu, 2005: A comparison of the ECMWF, MSC, and NCEP global ensemble prediction systems. Mon. Wea. Rev., 133 , 10761097.

    • Search Google Scholar
    • Export Citation
  • Candille, G., C. Côté, P. L. Houtekamer, and G. Pellerin, 2007: Verification of ensemble prediction systems against observations. Mon. Wea. Rev., 135 , 26882699.

    • Search Google Scholar
    • Export Citation
  • Cassou, C., 2008: Madden-Julian Oscillation influence on North Atlantic weather regimes at medium-range timescales. Geophysical Research Abstracts, Vol. 10, EGU2008-A-11008. [Available online at http://www.cosis.net/abstracts/EGU2008/11008/EGU2008-A-11008.pdf.].

    • Search Google Scholar
    • Export Citation
  • Chatfield, C., 2004: The Analysis of Time Series. 6th ed. Chapman and Hall/CRC, 333 pp.

  • Cherry, J., H. Cullen, M. Visbeck, A. Small, and C. Uvo, 2005: Impacts of the North Atlantic Oscillation on Scandinavian hydropower production and energy markets. Water Res. Manage., 19 , 15731650.

    • Search Google Scholar
    • Export Citation
  • Doblas-Reyes, F. J., R. Hagedorn, and T. N. Palmer, 2005: The rational behind the success of multi-model ensembles in seasonal forecasting. Part II: Calibration and combination. Tellus, 57A , 234252.

    • Search Google Scholar
    • Export Citation
  • ECMWF, cited. 2007: Monthly forecasting. [Available online at http://www.ecmwf.int/research/monthly_forecasting/index.html.].

  • Efron, B., and G. Gong, 1983: A leisurely look at the bootstrap, the jackknife, and cross-validation. Amer. Stat., 37 , 3648.

  • Epstein, E. S., 1969: A scoring system for probability forecasts of ranked categories. J. Appl. Meteor., 8 , 985987.

  • Ferranti, L., T. N. Palmer, F. Molteni, and E. Klinker, 1990: Tropical–extratropical interaction associated with the 30–60 day oscillation and its impact on medium and extended range prediction. J. Atmos. Sci., 47 , 21772199.

    • Search Google Scholar
    • Export Citation
  • Ferro, C. A. T., D. S. Richardson, and A. P. Weigel, 2008: On the effect of ensemble size on the discrete and continuous ranked probability scores. Meteor. Appl., 15 , 1924.

    • Search Google Scholar
    • Export Citation
  • Giorgi, F., and R. Francisco, 2000: Uncertainties in regional climate change prediction: A regional analysis of ensemble simulations with the HADCM2 coupled AOGCM. Climate Dyn., 16 , 169182.

    • Search Google Scholar
    • Export Citation
  • Graham, R. J., M. Gordon, P. J. McLean, S. Ineson, M. R. Huddleston, M. K. Davey, A. Brookshaw, and R. T. H. Barnes, 2005: A performance comparison of coupled and uncoupled versions of the Met Office seasonal prediction general circulation model. Tellus, 57A , 320339.

    • Search Google Scholar
    • Export Citation
  • Hamill, T. M., and J. Juras, 2006: Measuring forecast skill: Is it real skill or is it the varying climatology? Quart. J. Roy. Meteor. Soc., 132 , 29052923.

    • Search Google Scholar
    • Export Citation
  • Hamill, T. M., J. S. Whitaker, and X. Wei, 2004: Ensemble reforecasting: Improving medium-range forecast skill using retrospective forecasts. Mon. Wea. Rev., 132 , 14341447.

    • Search Google Scholar
    • Export Citation
  • Hurrell, J. W., 1995: Decadal trends in the North Atlantic Oscillation: Regional temperatures and precipitation. Science, 269 , 676679.

    • Search Google Scholar
    • Export Citation
  • Joliffe, I. T., 2007: Uncertainty and inference for verification measures. Wea. Forecasting, 22 , 637650.

  • Jung, T., 2005: Systematic errors of the atmospheric circulation in the ECMWF forecasting system. Quart. J. Roy. Meteor. Soc., 131 , 10451073.

    • Search Google Scholar
    • Export Citation
  • Jung, T., M. Hilmer, E. Ruprecht, S. Kleppek, S. K. Gulev, and O. Zolina, 2003: Characteristics of the recent eastward shift of interannual NAO variability. J. Climate, 16 , 33713382.

    • Search Google Scholar
    • Export Citation
  • Kumar, A., A. G. Barnston, and M. P. Hoerling, 2001: Seasonal predictions, probabilistic verifications, and ensemble size. J. Climate, 14 , 16711676.

    • Search Google Scholar
    • Export Citation
  • Kunz, H., S. C. Scherrer, M. A. Liniger, and C. Appenzeller, 2007: The evolution of ERA-40 surface temperatures and total ozone compared to Swiss time series. Meteor. Z., 16 , 171181.

    • Search Google Scholar
    • Export Citation
  • Kurihara, K., 2002: Climate-related service at the Japan Meteorology Agency. Extended Abstracts, Asia-Pacific Economic Cooperation Climate Network (APCN) Second Working Group Meeting, Seoul, South Korea, APCN. [Available online at http://www.apcc21.net/common/download.php?filename=sem/(12)Koich%20Kurihara.pdf.].

  • Liniger, M. A., H. Mathis, C. Appenzeller, and F. J. Doblas-Reyes, 2007: Realistic greenhouse gas forcing and seasonal forecasts. Geophys. Res. Lett., 34 .L04705, doi:10.1029/2006GL028335.

    • Search Google Scholar
    • Export Citation
  • Madden, R. A., and P. R. Julian, 1971: Detection of a 40–50 day oscillation in the zonal wind in the tropical Pacific. J. Atmos. Sci., 28 , 702708.

    • Search Google Scholar
    • Export Citation
  • Mason, S. J., 2004: On using “climatology” as a reference strategy in the Brier and ranked probability skill scores. Mon. Wea. Rev., 132 , 18911895.

    • Search Google Scholar
    • Export Citation
  • McGregor, G. R., M. Cox, Y. Cui, Z. Cui, M. K. Davey, R. F. Graham, and A. Brookshaw, 2006: Winter-season climate prediction for the U.K. health sector. J. Appl. Meteor. Climatol., 45 , 17821792.

    • Search Google Scholar
    • Export Citation
  • Meinke, H., and R. C. Stone, 2005: Seasonal and inter-annual climate forecasting: The new tool for increasing preparedness to climate variability and change in agricultural planning and operations. Climatic Change, 70 , 221253.

    • Search Google Scholar
    • Export Citation
  • Molteni, F., R. Buizza, T. N. Palmer, and T. Petroliagis, 1996: The new ECMWF Ensemble Prediction System: Methodology and validation. Quart. J. Roy. Meteor. Soc., 122 , 73119.

    • Search Google Scholar
    • Export Citation
  • Müller, W. A., C. Appenzeller, F. J. Doblas-Reyes, and M. A. Liniger, 2005: A debiased ranked probability skill score to evaluate probabilistic ensemble forecasts with small ensemble sizes. J. Climate, 18 , 15131523.

    • Search Google Scholar
    • Export Citation
  • Murphy, A. H., 1969: On the ranked probability skill score. J. Appl. Meteor., 8 , 988989.

  • Murphy, A. H., 1971: A note on the ranked probability skill score. J. Appl. Meteor., 10 , 155156.

  • Palmer, T. N., 2001: A nonlinear dynamical perspective on model error: A proposal for non-local stochastic-dynamic parameterization in weather and climate prediction models. Quart. J. Roy. Meteor. Soc., 127 , 279304.

    • Search Google Scholar
    • Export Citation
  • Pavan, V., S. Tibaldi, and C. Brankovic, 2000: Seasonal prediction of blocking frequency: Results from winter ensemble experiments. Quart. J. Roy. Meteor. Soc., 126 , 21252142.

    • Search Google Scholar
    • Export Citation
  • Puri, K., J. Barkmeijer, and T. N. Palmer, 2001: Tropical singular vectors computed with linearized diabatic physics. Quart. J. Roy. Meteor. Soc., 127 , 709731.

    • Search Google Scholar
    • Export Citation
  • Richardson, D. S., 2001: Measures of skill and value of ensemble prediction systems, their interrelationship and the effect of ensemble size. Quart. J. Roy. Meteor. Soc., 127 , 24732489.

    • Search Google Scholar
    • Export Citation
  • Rodwell, M. J., and F. J. Doblas-Reyes, 2006: Medium-range, monthly, and seasonal prediction for Europe and the use of forecast information. J. Climate, 19 , 60256046.

    • Search Google Scholar
    • Export Citation
  • Scherrer, S. C., M. Croci-Maspoli, C. Schwierz, and C. Appenzeller, 2006: Two-dimensional indices of atmospheric blocking and their statistical relationship with winter climate patterns in the Euro-Atlantic region. Int. J. Climatol., 26 , 233249.

    • Search Google Scholar
    • Export Citation
  • Schneider, T., and S. M. Griffies, 1999: A conceptual framework for predictability studies. J. Climate, 12 , 31333155.

  • Schwierz, C., C. Appenzeller, H. C. Davies, M. A. Liniger, W. Muller, T. F. Stocker, and M. Yoshimore, 2006: Challenges posed by and approaches to the study of seasonal-to-decadal climate variability. Climatic Change, 79 , 3163.

    • Search Google Scholar
    • Export Citation
  • Shukla, J., 1981: Dynamical predictability of monthly means. J. Atmos. Sci., 38 , 25472572.

  • Shukla, J., and Coauthors, 2000: Dynamical seasonal prediction. Bull. Amer. Meteor. Soc., 81 , 25932606.

  • Simmons, A. J., and Coauthors, 2004: Comparison of trends and low-frequency variability in CRU, ERA-40, and NCEP/NCAR analyses of surface air temperature. J. Geophys. Res., 109 .D24115, doi:10.1029/2004JD005306.

    • Search Google Scholar
    • Export Citation
  • Terray, L., E. Sevault, E. Guilyardi, and O. Thual, 1995: The OASIS coupler user guide version 2.0. CERFACS Tech. Rep. TR/CMGC/95-46, 123 pp.

  • Toth, Z., E. Kalnay, S. Tracton, R. Wobus, and J. Irwin, 1997: A synoptic evaluation of the NCEP ensemble. Wea. Forecasting, 12 , 140153.

    • Search Google Scholar
    • Export Citation
  • Toth, Z., O. Talagrand, G. Candille, and Y. Zhu, 2003: Probability and ensemble forecasts. Forecast VerificationA Practitioner’s Guide in Atmospheric Science, I. T. Joliffe and D. B. Stephenson, Eds., John Wiley and Sons, 137–163.

    • Search Google Scholar
    • Export Citation
  • Uppala, S. M., and Coauthors, 2005: The ERA-40 re-analysis. Quart. J. Roy. Meteor. Soc., 131 , 29613012.

  • Vialard, J., F. Vitart, M. Balmaseda, T. Stockdale, and D. Anderson, 2005: An ensemble generation method for seasonal forecasting with an ocean–atmosphere coupled model. Mon. Wea. Rev., 133 , 441453.

    • Search Google Scholar
    • Export Citation
  • Vitart, F., 2004: Monthly forecasting at ECMWF. Mon. Wea. Rev., 132 , 27612779.

  • Vitart, F., S. Woolnough, M. A. Balmaseda, and A. M. Tompkins, 2007: Monthly forecast of the Madden–Julian oscillation using a coupled GCM. Mon. Wea. Rev., 135 , 27002715.

    • Search Google Scholar
    • Export Citation
  • Waliser, D. E., 2006: Predictability of tropical intraseasonal variability. Predictability of Weather and Climate, T. Palmer and R. Hagedorn, Eds., Cambridge University Press, 275–305.

    • Search Google Scholar
    • Export Citation
  • WCRP, 2008: WCRP position paper on seasonal prediction. Report from the First WCRP Seasonal Prediction Workshop (Barcelona, Spain, 4–7 June 2007). ICPO Publ. 127, 24 pp.

  • Weigel, A. P., M. A. Liniger, and C. Appenzeller, 2007a: The discrete Brier and ranked probability skill scores. Mon. Wea. Rev., 135 , 118124.

    • Search Google Scholar
    • Export Citation
  • Weigel, A. P., M. A. Liniger, and C. Appenzeller, 2007b: Generalization of the discrete brier and ranked probability skill scores for weighted multimodel ensemble forecasts. Mon. Wea. Rev., 135 , 27782785.

    • Search Google Scholar
    • Export Citation
  • Weigel, A. P., M. A. Liniger, and C. Appenzeller, 2008: Can multi-model combination really enhance the prediction skill of ensemble forecasts? Quart. J. Roy. Meteor. Soc., 134 , 241260.

    • Search Google Scholar
    • Export Citation
  • Wilks, D. S., 2002: Smoothing forecast ensembles with fitted probability distributions. Quart. J. Roy. Meteor. Soc., 128 , 28212836.

  • Wilks, D. S., 2006: Statistical Methods in the Atmospheric Sciences. International Geophysics Series, Vol. 91, Academic Press, 627 pp.

  • Wolff, J. O., E. Maier-Raimer, and S. Legutke, 1997: The Hamburg ocean primitve equation model. Deutsches Klimarechenzentrum, Tech. Rep. 13, Hamburg, Germany, 98 pp.

  • Zeng, L., 2000: Weather derivatives and weather insurance: Concept, application, and analysis. Bull. Amer. Meteor. Soc., 81 , 20752082.

    • Search Google Scholar
    • Export Citation

APPENDIX

Effective Ensemble Size

Weigel et al. (2007a) have shown that the RPSSD for ensemble forecasts with ensemble size M is derived by replacing the 〈RPSCl〉 in Eq. (1) with the expectation ε of the scores RPSran,M, which an M-member ensemble prediction system would produce in the case of merely random climatological resamples:
i1520-0493-136-12-5162-ea1
Assume that a skill score is now calculated on the basis of Nh hindcasts with ensemble size Mh, and Nf forecasts with ensemble size Mf . This composition must then be equally represented in the reference:
i1520-0493-136-12-5162-eqa1
Applying Eqs. (10)–(15) of Weigel et al. (2007a), together with Eq. (3) of the present paper, one obtains the following:
i1520-0493-136-12-5162-eqa2
with
i1520-0493-136-12-5162-eqa3

Fig. 1.
Fig. 1.

Schematic of the ECMWF monthly prediction system. Each Thursday, a real-time forecast ensemble with 51 members is initialized and integrated forward in time. At the same time, 12 hindcast ensembles with 5 members each are initialized and integrated over the corresponding period of the previous 12 yr.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 2.
Fig. 2.

Definition of the forecast intervals. In the present study, mainly weekly averages are considered and evaluated as illustrated above.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 3.
Fig. 3.

Annual mean bias (ensemble mean − observation) of 2-m temperature (K) for forecast weeks W1–W4. Data from all 13 yr of forecasts and hindcasts available (1994–2006) are used.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 4.
Fig. 4.

Annually averaged RPSSD of all available MOFC forecasts of 2-m temperature (K) for forecast weeks (a) W1, (b)W2, (c) W3, and (d) W4.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 5.
Fig. 5.

Annual cylce of average skill in (a) the northern extratropics (30°–85°N), (b) the southern extratropics (30°–85°S), and (c) the tropics (30°S–30°N). A five-point symmetric moving-average filter has been applied as described in the text. Both land and sea points are considered. A few typical confidence intervals are plotted to illustrate the range of uncertainty of the skill values obtained.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 6.
Fig. 6.

Seasonally averaged RPSSD skill of all W2 forecasts available for (a) boreal winter, (b) spring, (c) summer, and (d) autumn.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 7.
Fig. 7.

As in Fig. 6, but for the North American domain.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 8.
Fig. 8.

As in Fig. 6, but for the European domain.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 9.
Fig. 9.

Annually averaged RPSSD of persisted W1 forecasts, verified against observations from (a) W2, (b) W3, and (c) W4.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 10.
Fig. 10.

As in Fig. 4, but showing the annually averaged ensemble mean correlation.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 11.
Fig. 11.

Annually averaged RPSSD plotted against annually averaged ensemble mean correlation for all continental grid points for forecast weeks (a) W1, (b) W2, (c) W3, and (d) W4. The gray lines show the RPSSD values that would be obtained were the forecasts reliable.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 12.
Fig. 12.

Dependence of temperature prediction skill (RPSSD) on the length of the forecast averaging interval. Annually averaged skill is plotted against the prediction time (i.e., the time between initialization and the end of the prediction interval) for the Niño-3.4 region and continental Europe. Forecast averaging intervals of 3, 7, and 14 days are considered. The sampling uncertainty of the skill values obtained is on the order of ±0.10 in the Niño-3.4 region and on the order of ±0.02 over the European domain.

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Fig. 13.
Fig. 13.

Dependence of average continental European temperature prediction skill (RPSSD) in winter on the sign of the NAO index. Forecast–observation pairs have been stratified on periods of positive and negative NAO index. (a) Skill has been calculated separately for these two subsets and is plotted against the time between initialization and the end of the 7-day prediction intervals. (b) The skill difference between these two subsets is displayed (thick line), together with the 10%–90% confidence range (thin lines).

Citation: Monthly Weather Review 136, 12; 10.1175/2008MWR2551.1

Table 1.

Local skill in dependence of season and forecast week averaged over the entire world (first row, considering both land and sea grid points), the Niño-3.4 region (second row), as well as the 21 climate regions (only land points) defined by Giorgi and Francisco (2000). The numbers in brackets represent the half-width of the 5%–95% confidence intervals of the skill estimates. Months are abbreviated by first letter of each month [e.g., December–February (DJF)].

Table 1.
Save
  • Anderson, D. L. T., and Coauthors, 2003: Comparison of the ECMWF seasonal forecast systems 1 and 2, including the relative performance for the 1997/8 El Niño. ECMWF Tech. Memo 404, 93 pp.

  • Baggenstos, D., 2007: Probabilistic verification of operational monthly temperature forecasts. Tech. Rep. 76, Federal Bureau of Meteorology and Climatology MeteoSwiss, Zurich, Switzerland, 52 pp.

  • Baldwin, M. P., D. B. Stephenson, D. W. J. Thompson, T. J. Dunkerton, A. J. Charlton, and A. O’Neill, 2003: Stratospheric memory and skill of extended-range weather forecasts. Science, 301 , 636640.

    • Search Google Scholar
    • Export Citation
  • Barnston, A. G., S. J. Mason, L. Goddard, D. G. DeWitt, and S. E. Zebiak, 2003: Multimodel ensembling in seasonal climate forecasting at IRI. Bull. Amer. Meteor. Soc., 84 , 17831796.

    • Search Google Scholar
    • Export Citation
  • Bolius, D., P. Calanca, A. Weigel, and M. A. Liniger, 2007: Prediction of moisture availability in agricultural soils using probabilistic monthly forecasts. Geophysical Research Abstracts, Vol. 9, EGU2007-A-02175, European Geosciences Union, 1 p. [Available online at http://www.cosis.net/abstracts/EGU2007/02175/EGU2007-J-02175.pdf.].

  • Buizza, R., and T. N. Palmer, 1995: The singular-vector structure of the atmospheric global circulation. J. Atmos. Sci., 52 , 14341456.

    • Search Google Scholar
    • Export Citation
  • Buizza, R., and T. N. Palmer, 1998: Impact of ensemble size on ensemble prediction. Mon. Wea. Rev., 126 , 25082518.

  • Buizza, R., P. L. Houtekamer, Z. Toth, G. Pellerin, M. Wei, and Y. Zhu, 2005: A comparison of the ECMWF, MSC, and NCEP global ensemble prediction systems. Mon. Wea. Rev., 133 , 10761097.

    • Search Google Scholar
    • Export Citation
  • Candille, G., C. Côté, P. L. Houtekamer, and G. Pellerin, 2007: Verification of ensemble prediction systems against observations. Mon. Wea. Rev., 135 , 26882699.

    • Search Google Scholar
    • Export Citation
  • Cassou, C., 2008: Madden-Julian Oscillation influence on North Atlantic weather regimes at medium-range timescales. Geophysical Research Abstracts, Vol. 10, EGU2008-A-11008. [Available online at http://www.cosis.net/abstracts/EGU2008/11008/EGU2008-A-11008.pdf.].

    • Search Google Scholar
    • Export Citation
  • Chatfield, C., 2004: The Analysis of Time Series. 6th ed. Chapman and Hall/CRC, 333 pp.

  • Cherry, J., H. Cullen, M. Visbeck, A. Small, and C. Uvo, 2005: Impacts of the North Atlantic Oscillation on Scandinavian hydropower production and energy markets. Water Res. Manage., 19 , 15731650.

    • Search Google Scholar
    • Export Citation
  • Doblas-Reyes, F. J., R. Hagedorn, and T. N. Palmer, 2005: The rational behind the success of multi-model ensembles in seasonal forecasting. Part II: Calibration and combination. Tellus, 57A , 234252.

    • Search Google Scholar
    • Export Citation
  • ECMWF, cited. 2007: Monthly forecasting. [Available online at http://www.ecmwf.int/research/monthly_forecasting/index.html.].

  • Efron, B., and G. Gong, 1983: A leisurely look at the bootstrap, the jackknife, and cross-validation. Amer. Stat., 37 , 3648.

  • Epstein, E. S., 1969: A scoring system for probability forecasts of ranked categories. J. Appl. Meteor., 8 , 985987.

  • Ferranti, L., T. N. Palmer, F. Molteni, and E. Klinker, 1990: Tropical–extratropical interaction associated with the 30–60 day oscillation and its impact on medium and extended range prediction. J. Atmos. Sci., 47 , 21772199.

    • Search Google Scholar
    • Export Citation
  • Ferro, C. A. T., D. S. Richardson, and A. P. Weigel, 2008: On the effect of ensemble size on the discrete and continuous ranked probability scores. Meteor. Appl., 15 , 1924.

    • Search Google Scholar
    • Export Citation
  • Giorgi, F., and R. Francisco, 2000: Uncertainties in regional climate change prediction: A regional analysis of ensemble simulations with the HADCM2 coupled AOGCM. Climate Dyn., 16 , 169182.

    • Search Google Scholar
    • Export Citation
  • Graham, R. J., M. Gordon, P. J. McLean, S. Ineson, M. R. Huddleston, M. K. Davey, A. Brookshaw, and R. T. H. Barnes, 2005: A performance comparison of coupled and uncoupled versions of the Met Office seasonal prediction general circulation model. Tellus, 57A , 320339.

    • Search Google Scholar
    • Export Citation
  • Hamill, T. M., and J. Juras, 2006: Measuring forecast skill: Is it real skill or is it the varying climatology? Quart. J. Roy. Meteor. Soc., 132 , 29052923.

    • Search Google Scholar
    • Export Citation
  • Hamill, T. M., J. S. Whitaker, and X. Wei, 2004: Ensemble reforecasting: Improving medium-range forecast skill using retrospective forecasts. Mon. Wea. Rev., 132 , 14341447.

    • Search Google Scholar
    • Export Citation
  • Hurrell, J. W., 1995: Decadal trends in the North Atlantic Oscillation: Regional temperatures and precipitation. Science, 269 , 676679.

    • Search Google Scholar
    • Export Citation
  • Joliffe, I. T., 2007: Uncertainty and inference for verification measures. Wea. Forecasting, 22 , 637650.

  • Jung, T., 2005: Systematic errors of the atmospheric circulation in the ECMWF forecasting system. Quart. J. Roy. Meteor. Soc., 131 , 10451073.

    • Search Google Scholar
    • Export Citation
  • Jung, T., M. Hilmer, E. Ruprecht, S. Kleppek, S. K. Gulev, and O. Zolina, 2003: Characteristics of the recent eastward shift of interannual NAO variability. J. Climate, 16 , 33713382.

    • Search Google Scholar
    • Export Citation
  • Kumar, A., A. G. Barnston, and M. P. Hoerling, 2001: Seasonal predictions, probabilistic verifications, and ensemble size. J. Climate, 14 , 16711676.

    • Search Google Scholar
    • Export Citation
  • Kunz, H., S. C. Scherrer, M. A. Liniger, and C. Appenzeller, 2007: The evolution of ERA-40 surface temperatures and total ozone compared to Swiss time series. Meteor. Z., 16 , 171181.

    • Search Google Scholar
    • Export Citation
  • Kurihara, K., 2002: Climate-related service at the Japan Meteorology Agency. Extended Abstracts, Asia-Pacific Economic Cooperation Climate Network (APCN) Second Working Group Meeting, Seoul, South Korea, APCN. [Available online at http://www.apcc21.net/common/download.php?filename=sem/(12)Koich%20Kurihara.pdf.].

  • Liniger, M. A., H. Mathis, C. Appenzeller, and F. J. Doblas-Reyes, 2007: Realistic greenhouse gas forcing and seasonal forecasts. Geophys. Res. Lett., 34 .L04705, doi:10.1029/2006GL028335.

    • Search Google Scholar
    • Export Citation
  • Madden, R. A., and P. R. Julian, 1971: Detection of a 40–50 day oscillation in the zonal wind in the tropical Pacific. J. Atmos. Sci., 28 , 702708.

    • Search Google Scholar
    • Export Citation
  • Mason, S. J., 2004: On using “climatology” as a reference strategy in the Brier and ranked probability skill scores. Mon. Wea. Rev., 132 , 18911895.

    • Search Google Scholar
    • Export Citation
  • McGregor, G. R., M. Cox, Y. Cui, Z. Cui, M. K. Davey, R. F. Graham, and A. Brookshaw, 2006: Winter-season climate prediction for the U.K. health sector. J. Appl. Meteor. Climatol., 45 , 17821792.

    • Search Google Scholar
    • Export Citation
  • Meinke, H., and R. C. Stone, 2005: Seasonal and inter-annual climate forecasting: The new tool for increasing preparedness to climate variability and change in agricultural planning and operations. Climatic Change, 70 , 221253.

    • Search Google Scholar
    • Export Citation
  • Molteni, F., R. Buizza, T. N. Palmer, and T. Petroliagis, 1996: The new ECMWF Ensemble Prediction System: Methodology and validation. Quart. J. Roy. Meteor. Soc., 122 , 73119.

    • Search Google Scholar
    • Export Citation
  • Müller, W. A., C. Appenzeller, F. J. Doblas-Reyes, and M. A. Liniger, 2005: A debiased ranked probability skill score to evaluate probabilistic ensemble forecasts with small ensemble sizes. J. Climate, 18 , 15131523.

    • Search Google Scholar
    • Export Citation
  • Murphy, A. H., 1969: On the ranked probability skill score. J. Appl. Meteor., 8 , 988989.

  • Murphy, A. H., 1971: A note on the ranked probability skill score. J. Appl. Meteor., 10 , 155156.

  • Palmer, T. N., 2001: A nonlinear dynamical perspective on model error: A proposal for non-local stochastic-dynamic parameterization in weather and climate prediction models. Quart. J. Roy. Meteor. Soc., 127 , 279304.

    • Search Google Scholar
    • Export Citation
  • Pavan, V., S. Tibaldi, and C. Brankovic, 2000: Seasonal prediction of blocking frequency: Results from winter ensemble experiments. Quart. J. Roy. Meteor. Soc., 126 , 21252142.

    • Search Google Scholar
    • Export Citation
  • Puri, K., J. Barkmeijer, and T. N. Palmer, 2001: Tropical singular vectors computed with linearized diabatic physics. Quart. J. Roy. Meteor. Soc., 127 , 709731.

    • Search Google Scholar
    • Export Citation
  • Richardson, D. S., 2001: Measures of skill and value of ensemble prediction systems, their interrelationship and the effect of ensemble size. Quart. J. Roy. Meteor. Soc., 127 , 24732489.

    • Search Google Scholar
    • Export Citation
  • Rodwell, M. J., and F. J. Doblas-Reyes, 2006: Medium-range, monthly, and seasonal prediction for Europe and the use of forecast information. J. Climate, 19 , 60256046.

    • Search Google Scholar
    • Export Citation
  • Scherrer, S. C., M. Croci-Maspoli, C. Schwierz, and C. Appenzeller, 2006: Two-dimensional indices of atmospheric blocking and their statistical relationship with winter climate patterns in the Euro-Atlantic region. Int. J. Climatol., 26 , 233249.

    • Search Google Scholar
    • Export Citation
  • Schneider, T., and S. M. Griffies, 1999: A conceptual framework for predictability studies. J. Climate, 12 , 31333155.

  • Schwierz, C., C. Appenzeller, H. C. Davies, M. A. Liniger, W. Muller, T. F. Stocker, and M. Yoshimore, 2006: Challenges posed by and approaches to the study of seasonal-to-decadal climate variability. Climatic Change, 79 , 3163.

    • Search Google Scholar
    • Export Citation
  • Shukla, J., 1981: Dynamical predictability of monthly means. J. Atmos. Sci., 38 , 25472572.

  • Shukla, J., and Coauthors, 2000: Dynamical seasonal prediction. Bull. Amer. Meteor. Soc., 81 , 25932606.

  • Simmons, A. J., and Coauthors, 2004: Comparison of trends and low-frequency variability in CRU, ERA-40, and NCEP/NCAR analyses of surface air temperature. J. Geophys. Res., 109 .D24115, doi:10.1029/2004JD005306.

    • Search Google Scholar
    • Export Citation
  • Terray, L., E. Sevault, E. Guilyardi, and O. Thual, 1995: The OASIS coupler user guide version 2.0. CERFACS Tech. Rep. TR/CMGC/95-46, 123 pp.

  • Toth, Z., E. Kalnay, S. Tracton, R. Wobus, and J. Irwin, 1997: A synoptic evaluation of the NCEP ensemble. Wea. Forecasting, 12 , 140153.

    • Search Google Scholar
    • Export Citation
  • Toth, Z., O. Talagrand, G. Candille, and Y. Zhu, 2003: Probability and ensemble forecasts. Forecast VerificationA Practitioner’s Guide in Atmospheric Science, I. T. Joliffe and D. B. Stephenson, Eds., John Wiley and Sons, 137–163.

    • Search Google Scholar
    • Export Citation
  • Uppala, S. M., and Coauthors, 2005: The ERA-40 re-analysis. Quart. J. Roy. Meteor. Soc., 131 , 29613012.

  • Vialard, J., F. Vitart, M. Balmaseda, T. Stockdale, and D. Anderson, 2005: An ensemble generation method for seasonal forecasting with an ocean–atmosphere coupled model. Mon. Wea. Rev., 133 , 441453.

    • Search Google Scholar
    • Export Citation
  • Vitart, F., 2004: Monthly forecasting at ECMWF. Mon. Wea. Rev., 132 , 27612779.

  • Vitart, F., S. Woolnough, M. A. Balmaseda, and A. M. Tompkins, 2007: Monthly forecast of the Madden–Julian oscillation using a coupled GCM. Mon. Wea. Rev., 135 , 27002715.

    • Search Google Scholar
    • Export Citation
  • Waliser, D. E., 2006: Predictability of tropical intraseasonal variability. Predictability of Weather and Climate, T. Palmer and R. Hagedorn, Eds., Cambridge University Press, 275–305.

    • Search Google Scholar
    • Export Citation
  • WCRP, 2008: WCRP position paper on seasonal prediction. Report from the First WCRP Seasonal Prediction Workshop (Barcelona, Spain, 4–7 June 2007). ICPO Publ. 127, 24 pp.

  • Weigel, A. P., M. A. Liniger, and C. Appenzeller, 2007a: The discrete Brier and ranked probability skill scores. Mon. Wea. Rev., 135 , 118124.

    • Search Google Scholar
    • Export Citation
  • Weigel, A. P., M. A. Liniger, and C. Appenzeller, 2007b: Generalization of the discrete brier and ranked probability skill scores for weighted multimodel ensemble forecasts. Mon. Wea. Rev., 135 , 27782785.

    • Search Google Scholar
    • Export Citation
  • Weigel, A. P., M. A. Liniger, and C. Appenzeller, 2008: Can multi-model combination really enhance the prediction skill of ensemble forecasts? Quart. J. Roy. Meteor. Soc., 134 , 241260.

    • Search Google Scholar
    • Export Citation
  • Wilks, D. S., 2002: Smoothing forecast ensembles with fitted probability distributions. Quart. J. Roy. Meteor. Soc., 128 , 28212836.

  • Wilks, D. S., 2006: Statistical Methods in the Atmospheric Sciences. International Geophysics Series, Vol. 91, Academic Press, 627 pp.

  • Wolff, J. O., E. Maier-Raimer, and S. Legutke, 1997: The Hamburg ocean primitve equation model. Deutsches Klimarechenzentrum, Tech. Rep. 13, Hamburg, Germany, 98 pp.

  • Zeng, L., 2000: Weather derivatives and weather insurance: Concept, application, and analysis. Bull. Amer. Meteor. Soc., 81 , 20752082.

    • Search Google Scholar
    • Export Citation
  • Fig. 1.

    Schematic of the ECMWF monthly prediction system. Each Thursday, a real-time forecast ensemble with 51 members is initialized and integrated forward in time. At the same time, 12 hindcast ensembles with 5 members each are initialized and integrated over the corresponding period of the previous 12 yr.

  • Fig. 2.

    Definition of the forecast intervals. In the present study, mainly weekly averages are considered and evaluated as illustrated above.

  • Fig. 3.

    Annual mean bias (ensemble mean − observation) of 2-m temperature (K) for forecast weeks W1–W4. Data from all 13 yr of forecasts and hindcasts available (1994–2006) are used.

  • Fig. 4.

    Annually averaged RPSSD of all available MOFC forecasts of 2-m temperature (K) for forecast weeks (a) W1, (b)W2, (c) W3, and (d) W4.

  • Fig. 5.

    Annual cylce of average skill in (a) the northern extratropics (30°–85°N), (b) the southern extratropics (30°–85°S), and (c) the tropics (30°S–30°N). A five-point symmetric moving-average filter has been applied as described in the text. Both land and sea points are considered. A few typical confidence intervals are plotted to illustrate the range of uncertainty of the skill values obtained.

  • Fig. 6.

    Seasonally averaged RPSSD skill of all W2 forecasts available for (a) boreal winter, (b) spring, (c) summer, and (d) autumn.

  • Fig. 7.

    As in Fig. 6, but for the North American domain.

  • Fig. 8.

    As in Fig. 6, but for the European domain.

  • Fig. 9.

    Annually averaged RPSSD of persisted W1 forecasts, verified against observations from (a) W2, (b) W3, and (c) W4.

  • Fig. 10.

    As in Fig. 4, but showing the annually averaged ensemble mean correlation.

  • Fig. 11.

    Annually averaged RPSSD plotted against annually averaged ensemble mean correlation for all continental grid points for forecast weeks (a) W1, (b) W2, (c) W3, and (d) W4. The gray lines show the RPSSD values that would be obtained were the forecasts reliable.

  • Fig. 12.

    Dependence of temperature prediction skill (RPSSD) on the length of the forecast averaging interval. Annually averaged skill is plotted against the prediction time (i.e., the time between initialization and the end of the prediction interval) for the Niño-3.4 region and continental Europe. Forecast averaging intervals of 3, 7, and 14 days are considered. The sampling uncertainty of the skill values obtained is on the order of ±0.10 in the Niño-3.4 region and on the order of ±0.02 over the European domain.

  • Fig. 13.

    Dependence of average continental European temperature prediction skill (RPSSD) in winter on the sign of the NAO index. Forecast–observation pairs have been stratified on periods of positive and negative NAO index. (a) Skill has been calculated separately for these two subsets and is plotted against the time between initialization and the end of the 7-day prediction intervals. (b) The skill difference between these two subsets is displayed (thick line), together with the 10%–90% confidence range (thin lines).

All Time Past Year Past 30 Days
Abstract Views 0 0 0
Full Text Views 5124 4386 1657
PDF Downloads 674 239 22