1. Introduction
The southeastern United States (SEUS) has the fastest growing population in the country (Seager et al. 2009) and has a multibillion-dollar agricultural industry, which is largely rain fed (Hansen et al. 1998). Therefore, the SEUS requires a reliable prediction of streamflow in its watersheds. Several studies have already indicated the benefits of skillful streamflow prediction for water supply, agriculture, and hydropower generation (e.g., Broad et al. 2007; Yao and Georgakakos 2001). The SEUS is the wettest region in the continental United States during the boreal summer season (Chan and Misra 2010). Chan and Misra (Chan and Misra 2010) further indicate that the SEUS also exhibits the largest seasonal variance in the summer. However, most global climate models exhibit a poor seasonal prediction skill, especially for rainfall during the summer and fall seasons over the SEUS (Stefanova et al. 2012b). In the winter and spring seasons, large-scale variations of global SST (e.g., El Niño and the Southern Oscillation) have significant influence on the rainfall variability in the SEUS (Ropelewski and Halpert 1986; Ropelewski and Halpert 1987). There is an absence of such teleconnection in the summer and fall seasons (especially at seasonal to interannual time scales), which leads to corresponding reduction of the seasonal climate prediction skill from climate models (Stefanova et al. 2012b).
Seasonal climate predictability is, however, rapidly evolving, with state-of-the-art climate models that have increased in complexity and resolution over time (Saha et al. 2006; Saha et al. 2010; Gent et al. 2010; Gent et al. 2011), with updated data-assimilated products available for initializing the various components of the climate system (Derber and Rosati 1989; Rosati et al. 1997; Balmaseda et al. 2008; Cazes-Boezio et al. 2008; Saha et al. 2010; Xue et al. 2011; Paolino et al. 2012), with newer ways of initializing the climate models (Zhang et al. 2007; Yang et al. 2009; Balmaseda and Anderson 2009), and with increased participation from several modeling groups (Palmer et al. 2004; Kirtman and Min 2009) to generate multimodel estimates of seasonal climate. More recently, two multi-institutional real-time projects in seasonal prediction have been instituted: the U.S. National Multimodel Ensemble (NMME; http://www.cpc.ncep.noaa.gov/products/ctb/MMEWhitePaperCPO_revised.pdf) and the European Seasonal to Interannual Prediction (EUROSIP; http://cosmos.enes.org/uploads/media/TStockdale.pdf) project, which is anticipated to further improve seasonal climate prediction skill and more importantly produce robust estimates of seasonal forecast uncertainty.
Much of hydrological forecasting is based on empirical methods, like linear regression, which use initial conditions and information on future climate conditions as predictors (Rosenberg et al. 2011; Pagano et al. 2009). However, since the introduction of extended streamflow prediction (Day 1985), the field of surface hydrological forecasting has developed relatively fast with ensemble streamflow prediction (ESP; Franz et al. 2003; McEnery et al. 2005; Wood and Schaake 2008; Bohn et al. 2010). ESP makes probabilistic forecasts by running the calibrated hydrological models with a number of realizations of meteorological forcing resampled from historical observations based on matching analogs of the forecasted climate state. Calibrated hydrological models are first run with historical observed climate data until the time of forecast to obtain initial conditions for the hydrological model. Then the weather sequence resampled from observed historical record corresponding to the closest analog of the forecasted climate state is used to force the hydrological models. Generations of daily precipitation by resampling from historical record are widely used in hydrological, ecological, and agricultural applications (Lobell et al. 2006; Wood and Lettenmaier 2006). The physical and conceptual basis of hydrological models allows ESP to overcome some of the limitations of the regression-based method. Furthermore, ESP is more flexible in the sense that it can use output from different sources (e.g., different climate models). Unlike the deterministic approach, the ESP framework provides information on the uncertainty surrounding the point estimate, which may benefit water resources management (Krzysztofowicz 2001). However, ESP is based on the premise that historical records are representative of the possible future.
Some popular approaches to applying climate forecast in hydrology include using observed resampling from historical data [e.g., the Schaake shuffle (Clark et al. 2004), weather generators (Wilks 2002; Voisin et al. 2011), and climate downscaling (Wood et al. 2004; Bastola and Misra 2013)]. In a resampling-based approach (e.g., the Schaake shuffle), the number of years of historical data potentially restricts the number of ensembles that can be generated. Furthermore, the Schaake shuffle reorders the samples to preserve the observed spatial and temporal structure (Clark et al. 2004). The size of the ensemble further reduces when resampling is made conditional upon the seasonal forecast information. This makes the weather generator attractive. Weather generators can generate daily weather sequences conditioned on climate forecasts. Weather generators, however, may have limitations in reproducing certain observed statistics (e.g., underestimation of interannual variance of monthly mean, the distribution of hot and cold spells; Semenov and Barrow 1997). Furthermore, point weather generators are inappropriate for multisite extension of the meteorological forcing that would become essential if they were to be used for distributed or semidistributed hydrological model applications (Wilks 2002). In the climate model-based approach, the outputs from global climate models are downscaled to finer resolutions and bias corrected to produce the forcing for the hydrological model (e.g., Wood et al. 2005; Wood et al. 2002). Examining the prospect of the National Centers for Environmental Prediction (NCEP) medium-range forecast (MRF) for streamflow prediction over the United States, Clark and Hay (Clark and Hay 2004) observed significantly large biases in precipitation and temperature over many areas. However, they observed improvement in streamflow prediction over some watersheds with statistical postprocessing of the NCEP MRF output.
Apart from the skill of the seasonal climate forecasts, the fidelity of a hydrological forecast also relies on the quality of the hydrological model in simulating the streamflow. However, the predictions from hydrological models are also plagued with uncertainties. These uncertainties stem from a variety of sources such as uncertainty in the parameters and structure of hydrological models. In response to such challenges, some studies in hydrological literature have adopted the multimodel approach in which streamflows from different hydrological models are combined using ranges of statistical postprocessing tools (e.g., Ajami et al. 2006; Duan et al. 2006; Bastola et al. 2011; Bastola and Misra 2013).
In this study we examine the results of a comprehensive set of seasonal streamflow predictions over 28 watersheds distributed throughout the SEUS with the aim of exploring the utility of the Florida Climate Institute–Florida State University Seasonal Hindcasts at 50-km grid resolution (FISH50; Misra et al. 2013) as a tool for seasonal hydrologic forecasting. FISH50, by way of its spatial resolution and its overall seasonal prediction skill compared to the NMME models that also include the operational seasonal climate forecasts from the National Centers for Environmental Prediction (NCEP), represents state-of-the-art seasonal prediction skill (Misra et al. 2013). Misra et al. (Misra et al. 2013) showed that seasonal prediction skill for the boreal summer and fall seasons is superior to most of the NMME models in the subtropical latitudes of the SEUS. The streamflow forecasting is based on calibrated hydrological models. We examine both the deterministic and the probabilistic forecast skill of rainfall and streamflow and discuss the shortcomings and virtues of our forecast system.
2. Study region and data
In the SEUS region, which is one of the wettest parts of the United States in the boreal summer and fall seasons (Chan and Misra 2010), surface streamflow is a major source of the river system. We have a priori selected 28 watersheds that are minimally affected by water management and are part of the Model Parameter Estimation Experiment (MOPEX; Schaake et al. 2006). Among the selected watersheds, six (viz., 2165000, 2217500, 210200, 3451500, 3454000, and 2156500) are affected to some extent by power plants and small ponds on the tributaries, suggesting that they may be more suitable for analysis at monthly time intervals (for details, see http://pubs.usgs.gov/wri/wri934076/stations/). Figure 1 shows the map of the watersheds (and their subbasins) included in this study. Further information on the characteristics of these watersheds can be found in Bastola and Misra (Bastola and Misra 2013).
The hydrological models used in this study are forced from the seasonal climate hindcast data of FISH50 (Misra et al. 2013). A brief outline of FISH50 data is provided in Table 1 (for additonal details on FISH50, see Misra et al. 2013). The retrospective seasonal forecasts of FISH50 are conducted from 1 June to 1 December of each year from 1982 to 2008. For each season there are six ensemble members of the seasonal climate hindcast. The observed rainfall dataset follows from the unified daily U.S. precipitation analysis of the Climate Prediction Center (CPC) available at 0.25° resolution (Higgins et al. 2000). The CPC data, available from 1948 onward, are used as our observed rainfall. The number of meteorological pixels, for both FISH50 and CPC, contained within each basin is shown in Figure 2. It is clear that the resolutions of FISH50 and CPC are relatively coarse and the size of the watersheds is comparatively small. This is a reminder that FISH50, which offers the highest spatial resolution among the existing seasonal climate hindcast data sources (e.g., NMME, EUROSIP), is far from sufficient for hydrological applications.
A brief outline of FISH50.
3. Generation of the forcing for the hydrological models
Hydrological models used for hydrological forecasting usually require meteorological forcing at relatively high (typically daily) temporal resolution. However, given the uncertainty, coarse resolution, and comparatively low fidelity of the climate models, the direct use of the climate model forecast variables to force the hydrological model does not yield reliable results (e.g., Clark and Hay 2004; Wood et al. 2002). Therefore, downscaling (either statistical or dynamical) the output from climate models at locally relevant spatial resolution is often considered for hydrological applications. In this study, three different variants of FISH50 forcing are used for seasonal hydrologic hindcast experiment: FISH50 daily data (FISH50), bias-corrected FISH50 (FISH50_BC), and resampling from observed historical record on the basis of the closest analog of the forecasted climate state (FISH50_Resamp).
In lieu of statistical bias correction, resampling historical observations on the basis of matching observed analogs of a certain forecasted climate state is another commonly adopted approach to force hydrological models to circumvent the bias in climate models. FISH50 is shown to produce reasonable conditional probability skill of seasonal anomalies of precipitation (Misra et al. 2013). For FISH50_Resamp, we leverage this feature of FISH50 to isolate observed analogs on the basis of the forecasted seasonal (6 months) anomalies of precipitation averaged over the watershed in consideration, which are then used to obtain the meteorological forcing at high (daily) temporal and spatial (subbasin) resolution. Such resampling methods preserve various moments of a time series (e.g., Efron 1979).
Following the methodology of Prudhomme and Davies (Prudhomme and Davies 2009), we use block resampling without replacement. Here a block is defined as a continuous 6-month duration of daily rainfall. The block could have been made smaller (i.e., 1 or 3 months) for the generation of the time series. However, resampling with such a small block is likely to affect the seasonal structure and, consequently, introduce large biases (e.g., Prudhomme and Davies 2009). Furthermore, FISH50 are 6-month-long integrations and therefore we have the ability to extend this block to 6 months.
The procedure to generate a conditioned daily sequence of weather data for the semidistributed hydrological models [Equation (3)] is as follows:
Obtain the quartile category of the 6-month averaged (June–November) rainfall forecast [x in Equation (2)] for a given year and watershed region.
Select q (=10 for this study) times a block (6 months: June–November) of weather sequence from observed historical records that has the same forecast category as defined in step 1. Repeat this procedure independently for each ensemble member of FISH50 for a given year. At each instance q of selecting the observed analog, store the corresponding date (i.e., year, month, and day).
In correspondence with the selected date from step 2, select for each subbasin the precipitation, temperature, and PET from the observed historical record so that the disaggregation from watershed to subbasin scale is limited by the resolution of the observed rainfall. (For very small watersheds, the values of daily rainfall may be alike for the subbasins.)
4. Description of the hydrological models and experimental setup
In hydrological modeling, uncertainty stems from a variety of sources. Readers are directed to Beven et al. (Beven et al. 2012) and Clark et al. (Clark et al. 2012) for an ongoing debate on how to assess hydrological model uncertainty. Implementation of a robust statistical framework for assessing uncertainty in a hydrological model is vital for water resource planning and management (e.g., Steinschneider et al. 2012). Despite their known limitations, conceptual rainfall–runoff (RR) models continue to be widely used for assessing the impacts of climate change on water resources and for projecting potential ranges of future climate change impacts (e.g., Bastola et al. 2011). In this study the uncertainty in simulation is accounted for by combining simulations obtained from three RR models [the Hydrologic Model (HyMOD; Wagener et al. 2001), the Nedbør–Afstrømnings Model (NAM; Madsen 2000), and the tank model (Sugawara 1995)]. These models have been widely used to simulate hydrological responses in watersheds located in various climatic regions (e.g., Bastola et al. 2011; Wagener et al. 2001; Tingsanchali and Gautam 2000). All three models quantify different components of the hydrological cycle by accounting for moisture content in different but mutually interrelated storage.
The RR models discussed above require the calibration of key parameters to yield reliable predictions (Gupta et al. 2005). In the present application, HyMOD, NAM, and tank require 6, 10, and 16 parameters, respectively, to be estimated through model calibration. Several studies discuss problems associated with model calibration and parameter uncertainty (e.g., Kuczera 1997; Beven and Binley 1992; Duan et al. 1992). There is growing agreement in the hydrological modeling community that a large combination of parameters results in reasonable model simulation (Beven 2005). We implement a multimodel and multiparameter simulation to conduct the hydrological forecast experiments. The generalized likelihood uncertainty estimation (GLUE) method (Beven and Binley 1992) is implemented to account for parametrically and structurally different hydrological models (i.e., multiparameter and multimodel). The individual model’s likelihood measure is used to weight the model prediction for the ensemble simulation in GLUE (Beven and Binley 1992).
The calibrated parameters for the 28 watersheds (Figure 1) and the three selected models are taken from Bastola and Misra (Bastola and Misra 2013). Bastola and Misra (Bastola and Misra 2013) calibrated the three conceptual models for the period 1948–68 using CPC rainfall data. The hydrological models were then validated for the period of 1969–79. The GLUE method was used to calibrate a suite of models based on a selected threshold that differentiates good and poor models according to selected likelihood measures. They used a prediction interval (width of the prediction interval), median model performance, and count efficiency (percentage of observation points lying within the prediction interval) to demonstrate the performance of the selected hydrological model. Bastola and Misra (Bastola and Misra 2013) also reported that, on average (across watersheds), the performance of the model measured in terms of Nash–Sutcliffe efficiency (NSE) is 0.72 and the average volume error is around 10% for both calibration (20-yr block) and validation (10-yr block). Furthermore, the large fractions of observation points were well encapsulated within the simulated prediction interval. Therefore, the same set of model parameters are used as calibrated model parameters in this study. The output from these calibrated hydrological models forced with CPC rainfall data is used as control or truth for verification of the hydrological forecasts.
The seasonal hydrological forecast experiments for a 6-month period (June–November) are initialized in summer (first week of June) for a period of 20 seasons (1982–2001) that coincides with the hydrometeorological data available for verification from MOPEX. The seasonal hydrological forecast experiment in this study is carried out following the ESP framework (Resamp; Wood and Lettenmaier 2006). ESP entails initializing the hydrological models by forcing them with observed meteorological forcing up to the start of the forecast. Shukla and Lettenmaier (Shukla and Lettenmaier 2011) and Wood and Lettenmaier (Wood and Lettenmaier 2006) suggested that initialization of hydrological models is important, especially for short-lead hydrologic forecasts such as those attempted here.
5. Analysis of forecast skill
A Brier score (Wilks 1995), a ranked probability score (Wilks 1995), and a relative operating characteristic or receiver operating characteristic curves (ROC) (Wilks 2001) are a few widely used probabilistic skill scores. These measures show the skill of forecast in discriminating occurrence and nonoccurrence of events (Zhang and Casey 2000; Wilks 2001). ROC, which is defined as the plot of sensitivity to specificity, is constructed from a plot of probability of hit rate against probability of false alarm rate for a given probability threshold. The area under the ROC (AROC) is a probabilistic forecast skill metric (Marzban 2004). The AROC values range between 0 and 1, with 1 being a perfect forecast and AROC values ≤ 0.5 regarded as no better than climatology.
6. Results and discussion
6.1. Skill analysis: Deterministic
In this section, we focus on the multiensemble member mean of FISH50 and the likelihood weighted average of the hydrological multimodel mean based on GLUE (multimodel mean hereafter). The probabilistic skill is examined in the subsequent subsection.
Over the 28 watersheds, the estimates of spatially averaged climatological rainfall (averaged from June through November) from FISH50 are significantly higher than the corresponding observations (Figure 3). On average, over most of these SEUS watersheds, FISH50 overestimates the summer and fall rainfall total by nearly 65%. For hydrological applications, this error could prove to be grave. For example, in Figure 4 it can be seen that using FISH50 rainfall produces extremely large biases in the forecasted streamflow. It should be mentioned that the volume error is computed with respect to the multimodel mean of the streamflow in the control simulation (i.e., flow simulated with the CPC precipitation dataset). In comparison, the bias-corrected FISH50 rainfall (FISH50_BC) and FISH50_Resamp appreciably reduce the bias. The results for FISH50 and FISH50_BC are based on average output from six ensemble members of FISH50 and the multimodel mean (based on GLUE) is from three hydrological models. However, the result for FISH50_Resamp is based on six ensemble members of FISH50, the three hydrological models, and 10 realizations of observed resampling from historical records for each ensemble member. This means that there are 180 realizations of streamflow prediction for every season using the FISH50_Resamp as opposed to 18 realizations in FISH50 and FISH50_BC forcing.
The elasticity of rainfall on streamflow (calculated as proportional change in the mean annual streamflow to proportional change in mean annual rainfall) varies from 1.5 to 3.5 with an average value of 2.7. This is consistent with the estimates of the elasticity of rainfall on the streamflow for the SEUS region from Sankarasubramanian et al. (Sankarasubramanian et al. 2001). Although we examine only 6 months of the season, most of the annual mean rainfall in the SEUS occurs in the boreal summer and fall seasons. Therefore, given this large elasticity, the use of the raw FISH50, with its large bias in rainfall, may amplify the errors further when it is propagated through a hydrological model of a watershed. For the sake of brevity, in Figure 5 we show the monthly mean climatological streamflow for six watersheds distributed across our study region (two in Florida, two in Georgia, one in North Carolina, and one in South Carolina). These watersheds have been chosen across latitudinal bands in the SEUS domain (Figure 1) that exhibit the largest contrasts of seasonality in rainfall (Misra and DiNapoli 2013) and therefore are most representative of the SEUS region. The bias in predicted flow measured with respect to control simulation (i.e., flow simulated with the reference precipitation dataset) is significantly large over the selected watersheds, irrespective of their latitudinal location. This bias is greatly reduced both in FISH50_Resamp and FISH50_BC. Interestingly, the volume errors of FISH50_Resamp are comparable to those of FISH50_BC. In other words, this result suggests that, despite the large wet bias in FISH50, the seasonal climate state as defined by the precipitation anomaly (in quartile categories) is skillful enough to isolate observed analogs that can yield relatively more realistic rainfall time series for these watersheds than the raw FISH50 can.
Figure 6 shows the normalized root-mean-square errors of the ensemble average streamflow based on two reference forecasts: the climatological (NSE) and the 1-yr lag forecast [or persistent forecast, wherein the forecast from the previous year is persisted through the following year, which is defined as the persistence efficiency measure (PEM)]. With respect to the climatological forecast, the flow simulated with FISH50_BC has little or no skill [i.e., NRMSE has a value less than or equal to zero in Equation (4), which means that the forecast is inferior to the reference forecast]. FISH50 and FISH50_Resamp forcing are equally poor (not shown). However, FISH50_BC and FISH50_Resamp exhibit relatively higher skill with respect to persistence (Figure 6). In fact, in many of the watersheds, PEM_FISH50_Resamp shows higher skill than PEM_FISH50_BC (Figure 6). However, there is no systematic decrease in NRMSE with increase in lead time of the forecast1 to suggest a significant impact of the initial conditions on the forecast errors. The results in Figure 6 suggest that the observed (or control simulation) variability of streamflow is very large, which makes the climatological forecast of streamflow superior to the persistence forecast. However, the deterministic skill of streamflow forecast using variants of FISH50 forcing (FISH50, FISH50_BC, and FISH50_Resample) is unskillful, which is not what we expected.
6.2. Skill analysis: Probabilistic
FISH50 shows some skill in discriminating different (quartile) categories of the seasonal (June–November) rainfall. The bubble plot in Figure 6 shows the distribution of AROC (with the size of the circles representing relative values above 0.5) for the seasonal mean (June–November) rainfall over each of the 28 watersheds. The latitudinal gradient of AROC in Figure 7 is likely artificial as there are more watersheds in the northern latitudes of the SEUS than in the southern latitudes. It is apparent in Figure 7 that the extreme quartiles (for both wet and dry seasonal anomalies) have higher skill than the medium quartile seasonal rainfall categories. In fact, this result is consistent with the seasonal prediction skill of the tercile rainfall anomalies analyzed in Misra et al. (Misra et al. 2013). However, there are exceptions, such as the watersheds in Florida (e.g., the St. Johns and the Peace River watersheds), which show higher skill for the medium quartile categories of seasonal rainfall anomalies than for the extreme quartile categories. However, more watersheds display higher skill for the extreme quartiles than for the middle quartiles (Figure 7). Most of the watersheds in Georgia, North Carolina, and South Carolina show higher skill than climatology in extreme quartiles. Of the 28 SEUS watersheds, FISH50 rainfall shows skill (better than climatology) in the very wet (18 watersheds), medium wet (16 watersheds), medium dry (13 watersheds), and very dry (18 watersheds) rainfall categories.
The 180 realizations of FISH50_Resamp for each watershed are used to calculate the AROC to evaluate the experimental hydrological streamflow forecasts probabilistically. The AROC values for the predicted seasonal (June–November) streamflow with FISH50_Resamp are shown in Figures 8 and 9. The results in Figures 8 and 9 are summarized in Figure 10. In Figures 8–10, it is clear that AROC is better for the extreme quartiles than for the middle quartiles. The AROC for upper medium (i.e., wet medium) is the least skillful. Considering that the SEUS watersheds have high elasticity of rainfall on streamflow, the high value of the AROC for the extreme quartile of streamflow can be attributed to the fact that the majority of the watersheds in the SEUS showed similar rainfall skill for extreme wet quartiles. Contrary to our expectation that skill would decrease with lead time, the relationship between skill and lead time is not apparent in Figures 8–10. The comparable AROC values for streamflow are consistently less than 0.5 for FISH50 and FISH50_BC (not shown). These results suggest that, despite the bias in FISH50 rainfall, it could be usefully applied for predicting the extreme quartiles, while for medium categories the forecasts remain more uncertain.
7. Conclusions
In this study we have examined the deterministic and probabilistic skill of seasonal hydrological predictions over 28 watersheds in the SEUS forced by an extensive set of retrospective seasonal climate forecasts known as FISH50. It is observed that FISH50 exhibits a relatively large wet bias in June–November over all 28 watersheds in the SEUS. Because of the large elasticity exhibited by most of the 28 watersheds, the errors in rainfall are translated to a disproportionate response in streamflow. Therefore, direct application of FISH50 is not suggested for hydrological application in the SEUS. The deterministic skill analysis as depicted by the NSE reveals that the streamflow predictions forced with FISH50 have lower skill than those forced with observed climatological rainfall over all 28 watersheds. This result holds true, irrespective of whether FISH50, FISH50_BC, or FISH50_Resamp is used to force the hydrological models.
FISH50_Resamp, however, offers a relatively large ensemble of streamflow estimation for every seasonal forecast from the climate model that potentially could be exploited for probabilistic forecasts. This large ensemble stems from the multiple hydrological models used in this study and the multiple (10 in our case) plausible matching observed analogs for the forecasted climate state (which in our case is the June–November rainfall quartile anomaly over the watershed). This gives rise to 180 independent estimations of streamflow per season, considering that there are six ensemble members of FISH50 for each season. Our study reveals that, for extreme quartiles, there is significantly higher probabilistic streamflow skill from FISH50_Resamp than from the climatological forecast for the majority of the 28 watersheds in the SEUS.
In conclusion, this study provides evidence of useful probabilistic hydrological prediction over the SEUS watersheds during the boreal summer and fall seasons. However, the experiment yielded poor corresponding deterministic forecast skills. We contend that using a multimodel framework with multiple ensemble members for the meteorological forcing to generate a total of 180 forecasts per watershed per forecast period allows for robust estimates of hydrological forecast uncertainty. This is portrayed in the probabilistic prediction skill, in contrast to its total neglect in the deterministic forecast prediction. Furthermore, probabilistic hydrological prediction augurs well when we realize that seasonal climate prediction is indeterminate, especially in the summer season when external factors like El Niño and the Southern Oscillation have little influence on the SEUS climate.
In other words, our study shows that, despite the grave wet bias in FISH50, it still could be exploited for harvesting useful streamflow predictions over a majority of these watersheds. In fact, for the medium quartile categories, the uncertainty of the rainfall and streamflow forecast is higher, which can also serve as useful information for managing water resources. We believe that this is a significant result given the challenge of seasonal climate prediction of the boreal summer and fall seasons over the SEUS.
Acknowledgments
This work was supported by grants from NOAA (NA12OAR4310078, NA10OAR4310215, and NA11OAR4310110), USGS (06HQGR0125), and USDA (027865). FISH50 integrations used in this paper were completed on the computational resources provided by the Extreme Science and Engineering Discovery Environment (XSEDE) under TG-ATM120017 and TG-ATM120010. We acknowledge Kathy Fearon of COAPS for the help with editing this manuscript.
References
Ajami, N. K., Q. Duan, X. Geo, and S. Sorooshian, 2006: Multimodel combination techniques for analysis of hydrological simulations: Application to distributed model intercomparison project results. J. Hydrometeor., 8, 755–768.
Balmaseda, M. A., and D. Anderson, 2009: Impact of initialization strategies and observations on seasonal forecast skill. Geophys. Res. Lett., 36, L01701, doi:10.1029/2008GL035561.
Balmaseda, M. A., A. Vidard, and D. L. T. Anderson, 2008: The ECMWF ocean analysis system: ORA-S3. Mon. Wea. Rev., 136, 3018–3034.
Bastola, S., and V. Misra, 2013: Evaluation of dynamically downscaled reanalysis precipitation data for hydrological application. Hydrol. Processes, doi:10.1002/hyp.9734.
Bastola, S., C. Murphy, and J. Sweeney, 2011: Role of hydrological model uncertainty in climate change impact studies. Adv. Water Resour., 34, 562–576.
Beven, K., 2005: A manifesto for the equifinality thesis. J. Hydrol.,20, 1–19.
Beven, K., and A. Binley, 1992: The future of distributed models: Model calibration and uncertainty prediction. Hydrol. Processes, 6, 279–298.
Beven, K., P. Smith, I. Westerberg, and J. Freer, 2012: Comment on “Pursuing the method of multiple working hypotheses for hydrological modeling” by P. Clark et al. Water Resour. Res., 48, W11801, doi:10.1029/2012WR012282.
Bohn, T. J., M. Y. Sonessa, and D. P. Lettenmaier, 2010: Seasonal hydrologic forecasting: Do multimodel ensemble averages always yield improvements in forecast skill? J. Hydrometeor., 11, 1358–1372.
Broad, K., A. Pfaff, R. Taddei, A. Sankarasubramanian, U. Lall, and F. de Assis de Souza Filho, 2007: Climate, stream flow prediction and water management in northeast Brazil: Societal trends and forecast value. Climatic Change, 84, 217–239.
Cazes-Boezio, G., D. Menemenlis, and C. R. Mechoso, 2008: Impact of ECCO ocean-state estimates on the initialization of seasonal climate forecasts. J. Climate, 21, 1929–1947.
Chan, S., and V. Misra, 2010: A diagnosis of the 1979–2005 extreme rainfall events in the southeastern United States with isentropic moisture tracing. Mon. Wea. Rev., 138, 1172–1185.
Clark, M. P., and L. Hay, 2004: Use of medium-range numerical weather prediction model output to produce forecasts of streamflow. J. Hydrometeor., 5, 15–32.
Clark, M. P., S. Gangopadhyay, D. Brandon, K. Werner, L. Hay, B. Rajagopalan, and D. Yates, 2004: A resampling procedure for generating conditioned daily weather sequences. Water Resour. Res., 40, W04304, doi:10.1029/2003WR002747.
Clark, M. P., D. Kavetski, and F. Fenicia, 2012: Reply to comment by K. Beven et al. on “Pursuing the method of multiple working hypotheses for hydrological modeling.” Water Resour. Res., 48, W11802, doi:10.1029/2012WR012547.
Day, G. N., 1985: Extended streamflow forecasting using NWSRFS. J. Water Resour. Plann. Manage.,111, 157–170.
Derber, J., and A. Rosati, 1989: A global oceanic data assimilation system. J. Phys. Oceanogr., 19, 1333–1347.
Duan, Q., S. Sorooshian, and V. K. Gupta, 1992: Effective and efficient global optimization for conceptual rainfall-runoff models. Water Resour. Res., 28, 1015–1031.
Duan, Q., N. K. Ajami, X. Gao, and S. Sorooshian, 2006: Multimodel ensemble hydrologic prediction using Bayesian model averaging. Adv. Water Resour., 30, 1371–1386.
Efron, B., 1979: Bootstrap methods: Another look at the jackknife. Ann. Stat., 7, 1–26.
Franz, K. J., H. C. Hartmann, S. Sorooshian, and R. Bales, 2003: Verification of National Weather Service ensemble streamflow predictions for water supply forecasting in the Colorado River basin. J. Hydrometeor., 4, 1105–1118.
Gent, P. R., S. G. Yeager, R. B. Neale, S. Levis, and D. A. Baily, 2010: Improvements in a half degree atmosphere/land version of the CCSM. Climate Dyn., 34, 819–833.
Gent, P. R., and Coauthors, 2011: The Community Climate System Model version 4. J. Climate, 24, 4973–4991.
Gupta, H. V., K. Beven, and T. Wagener, 2005: Model calibration and uncertainty estimation. Encyclopedia of Hydrological Sciences, M. G. Anderson, Ed., John Wiley and Sons, doi:10.1002/0470848944.hsa138.
Hansen, J. W., A. W. Hodges, and J. W. Jones, 1998: ENSO influences on agriculture in the southeastern United States. J. Climate, 11, 404–411.
Higgins, R. W., W. Shi, E. Yarosh, and R. Joyce, 2000: Improved United States precipitation quality control system and analysis. NCEP/CPC ATLAS 7. [Available online at http://www.cpc.ncep.noaa.gov/research_papers/ncep_cpc_atlas/7/index.html.]
Kirtman, B. P., and D. Min, 2009: Multimodel ensemble ENSO prediction with CCSM and CFS. Mon. Wea. Rev., 137, 2908–2930.
Krzysztofowicz, R., 2001: The case for probabilistic forecasting in hydrology. J. Hydrol., 249, 2–9.
Kuczera, G., 1997: Efficient subspace probabilistic parameter optimization for catchment models. Water Resour. Res., 33, 177–185.
Li, H., J. Sheffield, and E. F. Wood, 2010: Bias correction of monthly precipitation and temperature fields from Intergovernmental Panel on Climate Change AR4 models using equidistant quantile matching. J. Geophys. Res., 115, D10101, doi:10.1029/2009JD012882.
Lobell, D. B., C. B. Field, K. N. Cahill, and C. Bonfils, 2006: Impacts of future climate change on California perennial crop yields: Model projections with climate and crop uncertainties. Agric. Meteor., 141, 208–218.
Madsen, H., 2000: Automatic calibration of a conceptual rainfall–runoff model using multiple objectives. J. Hydrol., 235, 276–288.
Maraun, D., 2013: Bias correction, quantile mapping, and downscaling: Revisiting the inflation issue. J. Climate, 26, 2137–2143.
Marzban, C., 2004: The ROC curve and the area under it as performance measures. Wea. Forecasting,19, 1106–1114.
McEnery, J., J. Ingram, Q. Duan, T. Adams, and L. Anderson, 2005: NOAA’s Advanced Hydrologic Prediction Service: Building pathways for better science in water forecasting. Bull. Amer. Meteor. Soc., 86, 375–385.
Misra, V., and S. DiNapoli, 2013: Understanding the wet season variations over Florida. Climate Dyn., 40, 1361–1372.
Misra, V., H. Li, Z. Wu, S. DiNapoli, and Z. Wu, 2013: Global seasonal climate predictability in a two tiered forecast system: Part I: Boreal summer and fall seasons. Climate Dyn., doi:10.1007/s00382-013-1812-y, in press.
Murphy, A. H., 1988: Skill scores based on the mean square error and their relationships to the correlation coefficient. Mon. Wea. Rev., 116, 2417–2424.
Pagano, T. C., D. C. Garen, T. R. Perkins, and P. A. Pasteris, 2009: Daily updating of operational statistical seasonal water supply forecasts for the western US. J. Amer. Water Resour. Assoc., 45, 767–778.
Palmer, T. N., and Coauthors, 2004: Development of a European Multimodel Ensemble System for Seasonal-to-Interannual Prediction (DEMETER). Bull. Amer. Meteor. Soc., 85, 853–872.
Paolino, D. A., J. L. Kinter III, B. P. Kirtman, D. Min, and D. M. Straus, 2012: The impact of land surface initialization on seasonal forecasts with CCSM. J. Climate, 25, 1007–1021.
Prudhomme, C., and H. Davies, 2009: Assessing uncertainties in climate change impact analyses on the river flow regimes in the UK. Part 1: Baseline climate. Climatic Change, 93, 177–195.
Ropelewski, C. F., and M. S. Halpert, 1986: North American precipitation and temperature patterns associated with the El Niño/Southern Oscillation (ENSO). Mon. Wea. Rev., 114, 2352–2362.
Ropelewski, C. F., and M. S. Halpert, 1987: Global and regional scale precipitation patterns associated with the El Niño/Southern Oscillation. Mon. Wea. Rev., 115, 1606–1626.
Rosati, A., K. Miyakoda, and R. Gudgel, 1997: The impact of ocean initial conditions on ENSO forecasting with a coupled model. Mon. Wea. Rev., 125, 754–772.
Rosenberg, E. A., A. W. Wood, and A. C. Steinemann, 2011: Statistical applications of physically based hydrologic models to seasonal streamflow forecasts. Water Resour. Res., 47, W00H14, doi:10.1029/2010WR010101.
Saha, S., and Coauthors, 2006: The NCEP Climate Forecast System. J. Climate, 19, 3483–3517.
Saha, S., and Coauthors, 2010: The NCEP Climate Forecast Reanalysis. Bull. Amer. Meteor. Soc., 91, 1015–1057.
Sankarasubramanian, A., R. M. Vogel, and J. F. Limbrunner, 2001: The climate elasticity of streamflow in the United States. Water Resour. Res., 37, 1771–1781.
Schaake, J., S. Cong, and Q. Duan, 2006: U.S. MOPEX data set. IAHS Publication UCRL-JRNL-221228, 25 pp. [Available online at https://e-reports-ext.llnl.gov/pdf/333681.pdf.]
Seager, R., A. Tzanova, and J. Nakamura, 2009: Drought in the southeastern United States: Causes, variability over the last millennium, and the potential for future hydroclimatic change. J. Climate, 22, 5021–5045.
Semenov, M. A., and E. M. Barrow, 1997: Use of stochastic weather generator in the development of climate change scenarios. Climatic Change, 35, 397–414.
Shukla, S., and D. P. Lettenmaier, 2011: Seasonal hydrologic prediction in the United States: Understanding the role of initial hydrologic conditions and seasonal climate forecast skill. Hydrol. Earth Syst. Sci., 15, 3529–3538, doi:10.5194/hess-15-3529-2011.
Stefanova, L., V. Misra, S. Chan, M. Griffin, J. J. O’Brien, and T. J. Smith III, 2012a: A proxy for high-resolution regional reanalysis for the southeast United States: Assessment of precipitation variability in dynamically downscaled reanalyses. Climate Dyn.,38, 2449–2446, doi:10.1007/s00382-011-1230-y.
Stefanova, L., V. Misra, J. J. O’Brien, E. P. Chassignet, and S. Hameed, 2012b: Hindcast skill and predictability for precipitation and two-meter air temperature anomalies in global circulation models over the southeast United States. Climate Dyn., 38, 161–173.
Steinschneider, S., A. Polebitski, C. Brown, and B. H. Letcher, 2012: Toward a statistical framework to quantify the uncertainties of hydrologic response under climate change. Water Resour. Res., 48, W11525, doi:10.1029/2011WR011318.
Sugawara, M., 1995: Tank model. Computer Models of Watershed Hydrology, V. P. Singh, Ed., Water Resources, 165–214.
Tingsanchali, T., and M. R. Gautam, 2000: Application of tank, NAM, ARMA and neural network models to flood forecasting. Hydrol. Processes, 14, 2473–2487.
Voisin, N., F. Pappenberger, D. P. Lettenmaier, R. Buizza, and J. Schaake, 2011: Application of a medium-range global hydrologic probabilistic forecast scheme to the Ohio River basin. Wea. Forecasting,26, 425–446.
Wagener, T., D. P. Boyle, M. J. Lees, H. S. Wheater, H. V. Gupta, and S. Sorooshian, 2001: A framework for development and application of hydrological models. Hydrol. Earth Syst. Sci., 51, 13–26.
Wilks, D. S., 1995: Statistical Methods in the Atmospheric Sciences. Cambridge Press. 547 pp.
Wilks, D. S., 2001: A skill score based on economic value for probability forecasts. Meteor. Appl., 8, 209–219.
Wilks, D. S., 2002: Realizations of daily weather in forecast seasonal climate. J. Hydrometeor., 3, 195–207.
Wood, A. W., and D. P. Lettenmaier, 2006: A test bed for new seasonal hydrologic forecasting approaches in the western United States. Bull. Amer. Meteor. Soc., 87, 1699–1712.
Wood, A. W., and J. C. Schaake, 2008: Correcting errors in streamflow forecast ensemble mean and spread. J. Hydrometeor., 9, 132–148.
Wood, A. W., E. P. Maurer, A. Kumar, and D. P. Lettenmaier, 2002: Long-range experimental hydrologic forecasting for the eastern United States. J. Geophys. Res., 107, 4429, doi:10.1029/2001JD000659.
Wood, A. W., L. R. Leung, V. Sridhar, and D. P. Lettenmaier, 2004: Hydrologic implications of dynamical and statistical approaches to downscaling climate model outputs. Climatic Change, 62, 189–216.
Wood, A. W., A. Kumar, and D. P. Lettenmaier, 2005: A retrospective assessment of National Centers for Environmental Prediction climate model–based ensemble hydrologic forecasting in the western United States. J. Geophys. Res., 110, D04105, doi:10.1029/2004JD004508.
Xue, Y., and Coauthors, 2011: Comparative analysis of upper ocean heat content variability from ensemble operational ocean analysis. CLIVAR Variations, Vol. 9, No. 1, U.S. CLIVAR, Washington, D.C., 7–10.
Yang, S.-C., C. Keppenne, M. Rienecker, and E. Kalnay, 2009: Application of coupled bred vectors to seasonal-to-interannual forecasting and ocean data assimilation. J. Climate, 22, 2850–2870.
Yao, H., and A. Georgakakos, 2001: Assessment of Folsom Lake response to historical and potential future climate scenarios. J. Hydrol., 249, 176–196.
Zhang, H., and T. Casey, 2000: Verification of categorical probability forecasts. Wea. Forecasting, 15, 80–89.
Zhang, S., M. J. Harrison, A. Rosati, and A. Wittenberg, 2007: System design and evaluation of coupled ensemble data assimilation for global oceanic studies. Mon. Wea. Rev., 135, 3541–3564.