1. Introduction
Since the pioneering work of Barnett and Preisendorfer (1987) and Barnston (1994), statistical seasonal forecasting primarily has been based on the relationships between sea surface temperatures (SSTs) and subsequent seasonal time averages of predictands such as near-surface temperatures. The physical basis for the predictive relationships captured by these statistical forecasting approaches is that the effects of slowly varying boundary conditions such as SSTs are subsequently felt through their influences on the large-scale atmospheric circulation (e.g., Palmer and Anderson 1994; Goddard et al. 2001).
One of the issues that must be confronted when computing seasonal temperature forecasts is the nonstationarity of the climate and, in particular, the clear warming trend that has been evident, especially over the past 40 or so years. Neglecting the effect of this nonstationarity on seasonal forecasts has led to pronounced cold biases, for both dynamical (Barnston and Mason 2011; Peng et al. 2012; Wilks and Godfrey 2002) and statistical (Wilks 2008; 2014a,b) methods. Reduction of these biases can be extremely important for some users of seasonal temperature forecasts (Wilks and Livezey 2013).
Except during strong ENSO episodes, incorporation of the warming trend into the forecast formulation appears to provide the dominant source of current seasonal predictive skill for midlatitude temperatures, improving skill scores by reducing the forecast cold bias (Livezey and Timofeyeva 2008; Peng et al. 2012). The importance of representing trend in seasonal forecasts is manifested by the weak decreases in forecast skill with increasing lead time (Livezey and Timofeyeva 2008; Wilks 2008, 2014a), because the trend estimators can be computed nearly 1 year ahead. A straightforward method for incorporating the trend information into seasonal forecasts is to add an estimate of the current, evolving climate mean to forecasts of climate anomalies that are expressed relative to a more traditional 30-yr average (Higgins et al. 2004; Livezey et al. 2007; Wilks 2014a). Wilks (2013) and Wilks and Livezey (2013) compared a variety of estimators for computing current and near-future climate means for this purpose and concluded that the most recent 15-yr averages are presently the most successful for representing the nonstationarity of U.S. seasonal temperatures.
Although SST is the boundary condition that most often provides the predictor data for seasonal forecasts, other persistent surface conditions may also contribute skill. In particular, autumn Eurasian snow cover has been found to provide skill in statistical forecasts of North American winter temperatures (Foster et al. 1983; Cohen and Fletcher 2007; Lin and Wu 2011; Brands 2013). The physical basis for this predictive skill appears to be the connection between Eurasian snow cover and the Arctic Oscillation (Cohen and Entekhabi 1999; Cohen et al. 2007; Gong et al. 2007; Cohen and Jones 2011; Smith et al. 2011), and possibly also the Pacific–North America pattern (Clark and Serreze 2000; Wu et al. 2011). The spatial pattern of this predictability may complement that of SSTs for North American seasonal temperatures (Cohen and Fletcher 2007). The utility of antecedent snow cover as a predictor for seasonal forecasts may be limited by its short data record, which begins in November of 1966. Experience with seasonal forecasting on the basis of SSTs has been that using training data (i.e., the data used to fit statistical forecasting equations, which must be distinct from the data being forecast so as to avoid spurious artificial skill) from 1895 onward substantially improves forecast skill over use of the conventional 1950-onward training period, even though the older reconstructed SST data are of lesser quality (Wilks 2008, 2014a). Statistical forecasting equations using snow-cover predictors can be trained using data from 1966 onward only, and it is not clear whether these will be sufficiently skillful to compensate for the inability to use simultaneously SST training data prior to 1966 also.
The purpose of this paper is to investigate the use of SSTs, a representation of nonstationarity in the mean, and antecedent Northern Hemisphere snow cover, alone and in combination, for statistical forecasting of North American seasonal temperatures. The data sources are described in section 2. Section 3 documents the forecasting procedures. Section 4 presents the results, and section 5 concludes the paper.
2. Data
The SST predictor data have been derived from the extended reconstructed ERSST.v3b dataset (Smith et al. 2008; http://www.ncdc.noaa.gov/ersst/#grid). As in Wilks (2014a), the original 2° × 2° data have been averaged onto a 4° × 4° grid to reduce the computational burden, yielding 708 SST grid boxes over the Indo-Pacific basin (13°S–55°N, 51°E–81°W). These data have been expressed as anomalies relative to the 1951–80 period. Leading principal components of the seasonal averages (running 3-month means) from January–March (JFM) 1894 through December–February 2013/14 were computed as described in section 3. The reconstructed SST data are serially complete.
Monthly Eurasian and North American areal snow-extent data (Dewey and Heim 1982; Robinson and Frei 2000) were downloaded from the Rutgers University Global Snow Laboratory (http://climate.rutgers.edu/snowcover). Again, seasonal (running 3 month) averages have been computed for November–January (NDJ) 1966/67 through NDJ 2013/14. These predictor data have been expressed as anomalies relative to the 1966–80 reference period. The nine missing months early in the record were set to zero anomaly before computing the seasonal averages.




3. Forecast procedures
a. Maximum covariance analysis
The forecasts here have been computed using maximum covariance analysis (MCA), which is a linear multivariate statistical procedure that maximizes covariances between linear combinations of predictor and predictand vectors (e.g., Bretherton et al. 1992; Wilks 2011). This method is similar to the more familiar canonical correlation analysis (CCA) and the related method of redundancy analysis. Although the three methods often yield similar results, MCA has been chosen here because it has yielded more accurate forecasts for shorter training-data series (Wilks 2014a), since use of the snow-extent predictors strongly constrains the training-data sample sizes. The appendix contains the mathematical details underlying construction of probabilistic MCA forecasts.
b. Data processing and predictor selection
Out-of-sample seasonal temperature hindcasts were computed for, and verification results were aggregated over, the period JFM 1991–FMA 2014, at lead times of 0 (the end of the initial-time predictor season occurs immediately before the beginning of the target season) through 9 months (the end of the predictor season occurs 9 months before the beginning of the target season). The construction of the hindcasts avoids any use of the target-year data or subsequent data, and therefore the protocol emulates results that could have been achieved operationally. The 1991–2014 forecast period is shorter than the 1981–2013 period used in Wilks (2014a) because the snow-extent predictor records begin only in November of 1966.
The SST data enter the predictor vector x through the leading principal components for the appropriate initial-time season, and these have been recomputed for each forecast year using the additional data that would have become available most recently; thus, the training period is extended by 1 year for each successively later target year. These analyses were based on covariance matrices for the entire training-data records beginning in 1894, even when used in conjunction with the shorter snow-extent series, with the underlying SST anomalies being weighted using the square roots of latitude cosines so that the variances being analyzed are scaled according to the cosines of latitudes (North et al. 1982; Wilks 2011).
The snow-extent predictors enter the predictor vector x as the three raw monthly anomalies within the predictor season, separately for Eurasian and North American snow cover, so that linear combinations of these anomalies in appendix Eq. (A2a) can represent both seasonal averages and also rates of change of snow cover within the season that would correlate positively with within-month trends that are based on daily data (Cohen and Jones 2011). Unlike the SST predictors, the snow-extent data are not represented as principal components relating to spatial patterns, and therefore no attempt has been made to identify possible effects of subcontinental snow-cover variations, as was done by Mote and Kutney (2012).
As noted in section 2, the predictand vector y consists of the leading PCs of the seasonal gridbox temperature averages, computed using anomalies centered with the most recent 30-yr averages not including the target year,
Separately for each forecast year, season, lead time, and predictor variable combination, the numbers and identities of predictor variables have been chosen using a leave-one-out cross-validation procedure over each available training period, optimizing the continuous ranked probability score (CRPS; Matheson and Winkler 1976; Wilks 2011), again as in Wilks (2014a). For forecasts that are based on SST only, up to I = 10 leading SST PCs and at least J = 3 predictand PCs were made available for the MCA calculations, enforcing the constraint 3 ≤ J ≤ I ≤ 10. For forecasts that include snow-extent predictors, the most recent, the two most recent, or all three months of the initial-time season, individually for Eurasian and North American snow extents, were made available in the cross validation, again allowing at least three predictand PCs in y but no more than the number I of predictor variables in x. For forecasts that involve both snow and SST predictors, the two leading SST PCs were allowed as predictors in addition to the (up to) six monthly snow-extent anomalies. That combination of numbers and identities of predictor variables and number of predictand PCs, minimizing the CRPS averaged over training years and grid boxes, was selected to be used in a given forecast. Gaussian predictive distributions were assumed, on the basis of appendix Eq. (A4) for the gridbox means and the diagonal elements of Eq. (A7) for the corresponding variances, forecasting the observed anomalies centered using
c. Formulation and verification of probability forecasts























The forecasts are evaluated using the ranked probability skill score (RPSS), which compares the ranked probability score (RPS; Epstein 1969; Wilks 2011) in the conventional way with the RPS for the (preceding 30-yr) climatological forecast, for which
4. Results
The black bars in Fig. 1 show RPSS for the three-category probability forecasts constructed without adjusting for the recent climate nonstationarity [Eq. (5)], with the whiskers indicating 90% bootstrap confidence intervals. Within each of the four seasonal stratifications, each group of four bars aggregates results over the three lead-time stratifications: 0–1, 2–5, and 6–9 months. The four bars within each lead-time stratification show results for (from left to right): snow-extent predictors only, SST predictors only (with the MCA trained using data from 1966 onward so as to compare fairly with the snow-extent-based forecasts), both snow extent and SST allowed as predictors (training data from 1966 onward), and SST predictors only (training data from 1895 onward).

RPSS (black bars) and forecast bias (gray bars) for forecasts computed using Eq. (5) (no nonstationarity adjustment): (a) from NDJ through JFM, (b) from FMA through April–June (AMJ), (c) from May–July (MJJ) through July–September (JAS), and (d) from August–October (ASO) through October–December (OND). The left, center, and right groups of four bars stratify the lead times into 0–1, 2–5, and 6–9 months. Individual bars within each group of four represent results for (from left to right) snow predictors only, SST only (training data from 1966 onward), snow + SST, and SST only (training data from 1895 onward). Whiskers in each case show 90% bootstrap confidence intervals.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1

RPSS (black bars) and forecast bias (gray bars) for forecasts computed using Eq. (5) (no nonstationarity adjustment): (a) from NDJ through JFM, (b) from FMA through April–June (AMJ), (c) from May–July (MJJ) through July–September (JAS), and (d) from August–October (ASO) through October–December (OND). The left, center, and right groups of four bars stratify the lead times into 0–1, 2–5, and 6–9 months. Individual bars within each group of four represent results for (from left to right) snow predictors only, SST only (training data from 1966 onward), snow + SST, and SST only (training data from 1895 onward). Whiskers in each case show 90% bootstrap confidence intervals.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
RPSS (black bars) and forecast bias (gray bars) for forecasts computed using Eq. (5) (no nonstationarity adjustment): (a) from NDJ through JFM, (b) from FMA through April–June (AMJ), (c) from May–July (MJJ) through July–September (JAS), and (d) from August–October (ASO) through October–December (OND). The left, center, and right groups of four bars stratify the lead times into 0–1, 2–5, and 6–9 months. Individual bars within each group of four represent results for (from left to right) snow predictors only, SST only (training data from 1966 onward), snow + SST, and SST only (training data from 1895 onward). Whiskers in each case show 90% bootstrap confidence intervals.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
The RPSS results in Fig. 1 are modest overall, particularly for the winter and spring seasons, but the snow + SST forecasts at 0–1 months in spring and 2–5 months in autumn and the (1895 onward) SST-based forecasts in summer and autumn exhibit skill levels that are reasonably large in relation to the sampling variability indicated by the bootstrap confidence limits. The gray bars and 90% bootstrap confidence intervals in Fig. 1 show the corresponding forecast biases (forecast minus observed). These biases are large and uniformly negative (cold), reflecting the fact that defining the predictive distributions using the conventional 30-yr climatological means fails to capture much of the recent climate nonstationarity.







As in Fig. 1, but for forecasts computed with the CPC-15 nonstationarity adjustment [Eq. (6)]. Single bars at the extreme right of each panel show results for CPC-15 alone [Eq. (7)], which forecasts are the same for all lead times. Stars indicate forecasts that improve on the CPC-15-only forecasts with respect to RPSS.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1

As in Fig. 1, but for forecasts computed with the CPC-15 nonstationarity adjustment [Eq. (6)]. Single bars at the extreme right of each panel show results for CPC-15 alone [Eq. (7)], which forecasts are the same for all lead times. Stars indicate forecasts that improve on the CPC-15-only forecasts with respect to RPSS.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
As in Fig. 1, but for forecasts computed with the CPC-15 nonstationarity adjustment [Eq. (6)]. Single bars at the extreme right of each panel show results for CPC-15 alone [Eq. (7)], which forecasts are the same for all lead times. Stars indicate forecasts that improve on the CPC-15-only forecasts with respect to RPSS.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
Figure 2 shows that the somewhat naive forecast formulation in Eq. (7) is a strong competitor, exhibiting RPSS that is exceeded by the MCA-based forecasts only in the 12 cases indicated by the star symbols. Note that these include the snow-only forecasts for winter and spring and the snow + SST forecasts for spring, at the 0–1-month lead times. These snow-extent-based forecasts also exhibit larger skill than do the 1895-onward SST forecasts, indicating that, in these cases at least, there is sufficient predictive information in the snow-extent data to justify their use even though the available long SST training series cannot be used with them. Other instances outperforming the CPC-15-only [Eq. (7)] RPSS include the SST-based MCA forecasts trained from 1895 onward for all lead times in summer and autumn and the snow + SST forecasts at the two longer lead times for autumn.
Spatial distributions of nonnegative RPSS for the 0–1-month lead times during the winter and spring seasons, for which the snow-extent predictors contribute skill to the forecasts, are shown in Figs. 3 and 4, respectively. Figure 3a shows generally modest but widespread positive RPSS for winter MCA forecasts that are based on the snow-extent predictors only, in a spatial pattern that is reminiscent of the correlations with snow extent and the AO index in Cohen and Fletcher (2007) and Cohen and Jones (2011), and of areas of positive skill for heating degree-days in Brands (2013). Figure 3b shows that the skill of winter MCA forecasts that are based only on the Indo-Pacific SST predictors (trained from 1966 onward) is weak and concentrated along the northwestern portion of the domain but is generally geographically complementary to the nonnegative skill distribution for the snow-only forecasts in Fig. 3a. The winter CPC-15-only forecasts [from Eq. (7)] in Fig. 3c show relatively strong skill throughout the domain. Figure 3d shows the result when the snow-extent forecasts are combined with the CPC-15 nonstationarity adjustment using Eq. (6), which yielded the best winter forecasts at these lead times (Fig. 2a) of all methods considered.

Nonnegative RPSS across the 58 predictand grid boxes for 0–1-month lead time in winter (NDJ–JFM): (a) snow predictors only using Eq. (5); (b) SST predictors trained from 1966 onward, using Eq. (5); (c) CPC-15-only predictions using Eq. (7); and (d) snow predictors using Eq. (6).
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1

Nonnegative RPSS across the 58 predictand grid boxes for 0–1-month lead time in winter (NDJ–JFM): (a) snow predictors only using Eq. (5); (b) SST predictors trained from 1966 onward, using Eq. (5); (c) CPC-15-only predictions using Eq. (7); and (d) snow predictors using Eq. (6).
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
Nonnegative RPSS across the 58 predictand grid boxes for 0–1-month lead time in winter (NDJ–JFM): (a) snow predictors only using Eq. (5); (b) SST predictors trained from 1966 onward, using Eq. (5); (c) CPC-15-only predictions using Eq. (7); and (d) snow predictors using Eq. (6).
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1

As in Fig. 3, but for spring (FMA–AMJ), except that (d) shows results for snow and SST predictors using the Eq. (6) nonstationarity adjustment.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1

As in Fig. 3, but for spring (FMA–AMJ), except that (d) shows results for snow and SST predictors using the Eq. (6) nonstationarity adjustment.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
As in Fig. 3, but for spring (FMA–AMJ), except that (d) shows results for snow and SST predictors using the Eq. (6) nonstationarity adjustment.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
Figure 4a shows grid boxes having positive RPSS when the spring MCA forecasts are based on the snow-extent data only, which are displaced westward and exhibit generally weaker skill relative to their winter counterparts in Fig. 3a. Figure 4b shows that positive RPSS for spring forecasts that are based only on the (1966 onward) Indo-Pacific SST predictors are concentrated along the northern Pacific and Gulf coasts, again complementing the spatial pattern in Fig. 4a. In Fig. 4c the spring CPC-15-only forecasts show relatively strong skill except in the north-central and northeastern portions of the domain. Figure 4d shows the result when all of the predictors are combined, which appears to draw strength from each of the three underlying predictor sources and which yielded the best forecasts overall for spring at these lead times (Fig. 2b).
Both Eurasian and North American snow-extent data are chosen by the cross-validation predictor-selection procedure for nearly all forecasts involving the snow-extent predictors. For the overwhelming majority of cases in which appendix Eq. (A4) does not yield a zero anomaly vector, snow extents in all three of the antecedent months for both Eurasian and North American snow cover are selected as predictors, regardless of whether the SST predictors are being used also (results not shown). Figure 5 shows spatial distributions of nonnegative RPSS for forecasts using only one or the other of Eurasian or North American monthly snow extents as MCA predictors, again for the 0–1-month lead times in winter and spring. Skill for the winter forecasts using only Eurasian snow extent (Fig. 5a) is stronger than for North American (Fig. 5c) snow extent alone, with the pattern of positive skill in Fig. 5a again resembling the spatial extent of positive correlations in Cohen and Fletcher (2007) and Cohen and Jones (2011). Both are individually weak, however, with neither being as skillful as the winter forecasts using snow extents from both continents jointly (Fig. 3a). In a similar way, neither the Eurasian (Fig. 5b) nor North American (Fig. 5d) snow-extent predictors alone perform as well for spring as both do together (Fig. 4a). The Eurasian snow-extent results for spring in Fig. 5c are weaker overall than those for winter in Fig. 5a, whereas there are relatively strong results in the mountainous southwestern United States in Fig. 5d using the North American snow predictors, the physical mechanism behind which might derive from a positive correlation between late-winter snow extent and subsequent spring soil moisture (Walsh et al. 1982; Bao et al. 2011).

Nonnegative RPSS across the 58 predictand grid boxes for forecasts computed using Eq. (5) at 0–1-month lead times for (a) winter using Eurasian snow cover only, (b) spring using Eurasian snow cover only, (c) winter using North American snow cover only, and (d) spring using North American snow cover only.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1

Nonnegative RPSS across the 58 predictand grid boxes for forecasts computed using Eq. (5) at 0–1-month lead times for (a) winter using Eurasian snow cover only, (b) spring using Eurasian snow cover only, (c) winter using North American snow cover only, and (d) spring using North American snow cover only.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
Nonnegative RPSS across the 58 predictand grid boxes for forecasts computed using Eq. (5) at 0–1-month lead times for (a) winter using Eurasian snow cover only, (b) spring using Eurasian snow cover only, (c) winter using North American snow cover only, and (d) spring using North American snow cover only.
Citation: Journal of Applied Meteorology and Climatology 54, 1; 10.1175/JAMC-D-14-0215.1
5. Conclusions
Eurasian and North American snow extents, a simple representation of recent climate nonstationarity, and the more conventional antecedent SST conditions, individually and in combination, have been investigated as statistical predictors for seasonal North American temperatures. Because of the clear warming trend during recent decades, the strongest contribution to forecast skill was derived from adding the CPC-15 statistic [Eq. (2)], which is simply the average of the previous 15 years’ predictand values, to the forecast anomalies computed using MCA. It is evident that the increased RPS skill contributed by this more responsive representation of the current climate mean derives from its strongly reducing the cold biases that are typically exhibited by seasonal forecasts and that were exhibited here by the MCA forecasts computed using the more conventional 30-yr means. Indeed, in many instances MCA forecasts involving the SST and/or snow-extent predictors did not outperform the simple mean extrapolations computed using CPC-15 together with climatological standard deviations.
That recent trends often provide the dominant source of forecast skill is consistent with the results of Livezey and Timofeyeva (2008), who noted seasonal forecast skill diminishing only slightly or not at all with increasing lead time. Using CPC-15 to represent the recent climate warming is more effective than extrapolating a simple linear trend over the period of record because the warming trajectory has accelerated beginning in the mid-1970s (Livezey et al. 2007; Wilks and Livezey 2013). More-sophisticated nonstationarity adjustments may become preferable as the current warming trend continues to unfold, and therefore this issue should be revisited periodically in the coming years (Wilks 2013; Wilks and Livezey 2013).
The data history for snow extent is relatively short but nevertheless was seen to contribute positive predictive skill for North American temperatures in winter and spring at lead times of 0–1 months. Both Eurasian and North American snow cover, and for all three months of the initial-time seasons, appeared to contribute positively to the overall skill. Best results for winter and spring at 0–1-month lead time were achieved using the snow-extent predictors, even though their inclusion precluded simultaneous use of the long (1894 onward) training history of SSTs that prior work has found to improve seasonal forecast skill relative to using only the higher-quality recent (1950 onward) SST training data (Wilks 2008, 2014a). For the summer and autumn seasons, the best forecast skill was achieved using only the longer SST training data together with the CPC-15 nonstationarity adjustment. The snow-extent predictors appeared to provide similar overall skill levels at the 2–5- and 6–9- month lead times for autumn when used together with (1966 onward) SSTs and CPC-15, however. This result should be interpreted cautiously since it is not clear what physical mechanism might provide the basis for this skill, but it is consistent with the correlation results reported in Fig. 1a of Saito and Cohen (2003).
A clear limitation of using existing snow predictors for statistical seasonal forecasts is that they extend only from November of 1966, because they are based on satellite observations. Experience with extending the training-data sample size for SST predictors (Wilks 2008, 2014a) has shown that this straightforward change to the forecast protocol can greatly increase the skill, even though the older training data are known to be of lesser quality. Since snow-extent predictors add to seasonal forecast skill in some circumstances and in ways that appear to be complementary to the more usual SST predictors, extending the hemispheric snow-cover data record farther back in time using land-based measurements (Brown and Robinson 2011) might similarly improve seasonal-forecast skill. This result would be expected both because of the additional training data for the snow extents and also because it would allow fuller simultaneous use of longer SST training series. Poor availability of instrumental Eurasian snow-cover data (e.g., Brown 2000) appears to limit the prospects for such a data extension, however.
Acknowledgments
I thank both anonymous reviewers for comments that led to improved clarity of the presentation. This research was supported by the National Science Foundation under Grant AGS-1112200.
APPENDIX
Probabilistic MCA Forecasts































REFERENCES
Bao, Z., R. Kelly, and R. Wu, 2011: Variability of regional snow cover in spring over western Canada and its relationship to temperature and circulation anomalies. Int. J. Climatol., 31, 1280–1294, doi:10.1002/joc.2155.
Barnett, T. P., and R. W. Preisendorfer, 1987: Origins and levels of monthly and seasonal forecast skill for United States surface air temperatures determined by canonical correlation analysis. Mon. Wea. Rev., 115, 1825–1850, doi:10.1175/1520-0493(1987)115<1825:OALOMA>2.0.CO;2.
Barnston, A. G., 1994: Linear statistical short-term climate predictive skill in the Northern Hemisphere. J. Climate, 7, 1513–1564, doi:10.1175/1520-0442(1994)007<1513:LSSTCP>2.0.CO;2.
Barnston, A. G., and S. J. Mason, 2011: Evaluation of IRI’s seasonal climate forecasts for the extreme 15% tails. Wea. Forecasting, 26, 545–554, doi:10.1175/WAF-D-10-05009.1.
Brands, S., 2013: Skillful seasonal predictions of boreal winter accumulated heating degree-days and relevance for the weather derivative market. J. Appl. Meteor. Climatol., 52, 1297–1302, doi:10.1175/JAMC-D-12-0303.1.
Bretherton, C. S., C. Smith, and J. M. Wallace, 1992: An intercomparison of methods for finding coupled patterns in climate data. J. Climate, 5, 541–560, doi:10.1175/1520-0442(1992)005<0541:AIOMFF>2.0.CO;2.
Brown, R. D., 2000: Northern Hemisphere snow cover variability and change, 1915–97. J. Climate, 13, 2339–2355, doi:10.1175/1520-0442(2000)013<2339:NHSCVA>2.0.CO;2.
Brown, R. D., and D. A. Robinson, 2011: Northern Hemisphere spring snow cover variability and change over 1922–2010 including an assessment of uncertainty. Cryosphere, 5, 219–229, doi:10.5194/tc-5-219-2011.
Clark, M. P., and M. C. Serreze, 2000: Effects of variations in East Asian snow cover on modulating atmospheric circulation over the North Pacific Ocean. J. Climate, 13, 3700–3710, doi:10.1175/1520-0442(2000)013<3700:EOVIEA>2.0.CO;2.
Cohen, J., and D. Entekhabi, 1999: Eurasian snow cover variability and Northern Hemisphere climate predictability. Geophys. Res. Lett., 26, 345–348, doi:10.1029/1998GL900321.
Cohen, J., and C. Fletcher, 2007: Improved skill of Northern Hemisphere winter surface temperature predictions based on land–atmosphere fall anomalies. J. Climate, 20, 4118–4132, doi:10.1175/JCLI4241.1.
Cohen, J., and J. Jones, 2011: A new index for more accurate winter predictions. Geophys. Res. Lett., 38, L21701, doi:10.1029/2011GL049626.
Cohen, J., M. Barlow, P. J. Kushner, and K. Saito, 2007: Stratosphere–troposphere coupling and links with Eurasian land surface variability. J. Climate, 20, 5335–5343, doi:10.1175/2007JCLI1725.1.
Dewey, K. F., and R. Heim Jr., 1982: A digital archive of Northern Hemisphere snow cover, November 1966 through December 1980. Bull. Amer. Meteor. Soc., 63, 1132–1141, doi:10.1175/1520-0477(1982)063<1132:ADAONH>2.0.CO;2.
Epstein, E. S., 1969: A scoring system for probability forecasts of ranked categories. J. Appl. Meteor., 8, 985–987, doi:10.1175/1520-0450(1969)008<0985:ASSFPF>2.0.CO;2.
Foster, J., M. Owe, and A. Rango, 1983: Snow cover and temperature relationships in North America and Eurasia. J. Climate Appl. Meteor., 22, 460–469, doi:10.1175/1520-0450(1983)022<0460:SCATRI>2.0.CO;2.
Goddard, L., S. J. Mason, S. E. Zebiak, C. F. Ropelewski, R. Basher, and M. A. Cane, 2001: Current approaches to seasonal-to-interannual climate predictions. Int. J. Climatol., 21, 1111–1152, doi:10.1002/joc.636.
Gong, G., J. Cohen, D. Entekhabi, and Y. Ge, 2007: Hemispheric-scale climate response to northern Eurasia land surface characteristics and snow anomalies. Global Planet. Change, 56, 359–370, doi:10.1016/j.gloplacha.2006.07.025.
Higgins, R. W., H.-K. Kim, and D. Unger, 2004: Long-lead seasonal temperature and precipitation prediction using tropical Pacific SST consolidation forecasts. J. Climate, 17, 3398–3414, doi:10.1175/1520-0442(2004)017<3398:LSTAPP>2.0.CO;2.
Johnson, R. A., and D. W. Wichern, 2002: Applied Multivariate Statistical Analysis. 5th ed. Prentice Hall, 767 pp.
Lin, H., and Z. Wu, 2011: Contribution of the autumn Tibetan Plateau snow cover to seasonal prediction of North American winter temperature. J. Climate, 24, 2801–2813, doi:10.1175/2010JCLI3889.1.
Livezey, R. E., and M. M. Timofeyeva, 2008: The first decade of long-lead U.S. seasonal forecasts: Insights from a skill analysis. Bull. Amer. Meteor. Soc., 89, 843–854, doi:10.1175/2008BAMS2488.1.
Livezey, R. E., K. Y. Vinnkov, M. M. Timofeyeva, R. Tinker, and H. M. van den Dool, 2007: Estimation and extrapolation of climate normals and climatic trends. J. Appl. Meteor. Climatol., 46, 1759–1776, doi:10.1175/2007JAMC1666.1.
Matheson, J. E., and R. L. Winkler, 1976: Scoring rules for continuous probability distributions. Manage. Sci., 22, 1087–1096, doi:10.1287/mnsc.22.10.1087.
Mote, T. L., and E. R. Kutney, 2012: Regions of autumn Eurasian snow cover and associations with North American winter temperatures. Int. J. Climatol., 32, 1164–1177, doi:10.1002/joc.2341.
North, G. R., T. L. Bell, R. F. Cahalan, and F. J. Moeng, 1982: Sampling errors in the estimation of empirical orthogonal functions. Mon. Wea. Rev., 110, 699–706, doi:10.1175/1520-0493(1982)110<0699:SEITEO>2.0.CO;2.
Palmer, T. N., and D. L. T. Anderson, 1994: The prospects for seasonal forecasting—A review paper. Quart. J. Roy. Meteor. Soc., 120, 755–793, doi:10.1002/qj.49712051802.
Peng, P., A. Kumar, and M. S. Halpert, 2012: An analysis of CPC’s operational 0.5-month lead seasonal outlooks. Wea. Forecasting, 27, 898–917, doi:10.1175/WAF-D-11-00143.1.
Peterson, T. C., and R. S. Vose, 1997: An overview of the Global Historical Climatology Network temperature database. Bull. Amer. Meteor. Soc., 78, 2837–2849, doi:10.1175/1520-0477(1997)078<2837:AOOTGH>2.0.CO;2.
Robinson, D. A., and A. Frei, 2000: Seasonal variability of Northern Hemisphere snow extent using visible satellite data. Prof. Geogr., 52, 307–315, doi:10.1111/0033-0124.00226.
Saito, K., and J. Cohen, 2003: The potential role of snow cover in forcing interannual variability of the major Northern Hemisphere mode. Geophys. Res. Lett., 30, 1302, doi:10.1029/2002GL016341.
Smith, K. L., P. J. Kushner, and J. Cohen, 2011: The role of linear interference in northern annular mode variability associated with Eurasian snow cover extent. J. Climate, 24, 6185–6202, doi:10.1175/JCLI-D-11-00055.1.
Smith, T. M., R. W. Reynolds, T. C. Peterson, and J. Lawrimore, 2008: Improvements to NOAA’s historical merged land–ocean surface temperature analysis (1880–2006). J. Climate, 21, 2283–2296, doi:10.1175/2007JCLI2100.1.
Walsh, J. E., D. R. Tucek, and M. R. Peterson, 1982: Seasonal snow cover and short-term climatic fluctuations over the United States. Mon. Wea. Rev., 110, 1474–1485, doi:10.1175/1520-0493(1982)110<1474:SSCAST>2.0.CO;2.
Wilks, D. S., 1996: Statistical significance of long-range “optimal climate normal” temperature and precipitation forecasts. J. Climate, 9, 827–839, doi:10.1175/1520-0442(1996)009<0827:SSOLRC>2.0.CO;2.
Wilks, D. S., 2008: Improved statistical seasonal forecasts using extended training data. Int. J. Climatol., 28, 1589–1598, doi:10.1002/joc.1661.
Wilks, D. S., 2011: Statistical Methods in the Atmospheric Sciences. 3rd ed. Academic Press, 676 pp.
Wilks, D. S., 2013: Projecting “normals” in a nonstationary climate. J. Appl. Meteor. Climatol., 52, 289–302, doi:10.1175/JAMC-D-11-0267.1.
Wilks, D. S., 2014a: Comparison of probabilistic statistical forecast and trend adjustment methods for North American seasonal temperatures. J. Appl. Meteor. Climatol., 53, 935–949, doi:10.1175/JAMC-D-13-0294.1.
Wilks, D. S., 2014b: Probabilistic canonical correlation analysis forecasts, with application to tropical Pacific sea-surface temperatures. Int. J. Climatol., 34, 1405–1413, doi:10.1002/joc.3771.
Wilks, D. S., and C. M. Godfrey, 2002: Diagnostic verification of the IRI net assessment forecasts, 1997–2000. J. Climate, 15, 1369–1377, doi:10.1175/1520-0442(2002)015<1369:DVOTIN>2.0.CO;2.
Wilks, D. S., and R. E. Livezey, 2013: Performance of alternative “normals” for tracking climate changes, using homogenized and nonhomogenized seasonal U.S. surface temperatures. J. Appl. Meteor. Climatol., 52, 1677–1687, doi:10.1175/JAMC-D-13-026.1.
Wu, Q., H. Hu, and L. Zhang, 2011: Observed influences of autumn–early winter Eurasian snow cover anomalies on the hemispheric PNA-like variability in winter. J. Climate, 24, 2017–2023, doi:10.1175/2011JCLI4236.1.