Browse

You are looking at 131 - 140 of 2,609 items for :

  • Weather and Forecasting x
  • Refine by Access: Content accessible to me x
Clear All
Hyeyum Hailey Shin
,
Wiebke Deierling
, and
Robert Sharman

Abstract

The skill of operational deterministic turbulence forecasts is impacted by the uncertainties in both weather forecasts from the underlying numerical weather prediction (NWP) models and diagnoses of turbulence from the NWP model output. This study compares various probabilistic turbulence forecasting approaches to quantify these uncertainties and provides recommendations on the most suitable approach for operational implementation. The approaches considered are all based on ensembles of NWP forecasts and/or turbulence diagnostics, and include a multi-diagnostic ensemble (MDE), a time-lagged NWP ensemble (TLE), a forecast-model NWP ensemble (FME), and combined time-lagged MDE (TMDE) and forecast-model MDE (FMDE). Both case studies and statistical analyses are provided. The case studies show that the MDE approach that represents the uncertainty in turbulence diagnostics provides a larger ensemble spread than the TLE and FME approaches that represent the uncertainty in NWP forecasts. The larger spreads of MDE, TMDE, and FMDE allow for higher probabilities of detection for low percentage thresholds at the cost of increased false alarms. The small spreads of TLE and FME result in either hits with higher confidence or missed events, highly dependent on the performance of the underlying NWP model. Statistical evaluations reveal that increasing the number of diagnostics in MDE is a cost-effective and powerful method for describing the uncertainty of turbulence forecasts, considering trade-offs between accuracy and computational cost associated with using NWP ensembles. Combining either time-lagged or forecast-model NWP ensembles with MDE can further improve prediction skill and could be considered if sufficient computational resources are available.

Free access
Weiguo Wang
,
Lin Zhu
,
Bin Liu
,
Zhan Zhang
,
Avichal Mehra
, and
Vijay Tallapragada

Abstract

An evaluation framework for tropical cyclone rapid intensification (RI) forecasts is introduced and applied to evaluate the performance of RI forecasts by the operational Hurricane Weather Research and Forecasting (HWRF) Model. The framework is based on the performance of each 5-day forecast cycle, while the conventional RI evaluation is based on the statistics of successful or false RI forecasts at individual lead times. The framework can be used to compare RI forecasts of different cycles, which helps model developers and forecasters to characterize RI forecasts under different scenarios. It also can provide the evaluation of statistical performance in the context of 5-day forecast cycles. The RI forecast of each cycle is assessed using a modified probability-based approach that takes the absolute errors in intensity changes into account. The overall performance of RI forecasts during a given period is assessed based on the fractions of the individual forecast cycles during which RI events are successfully or falsely predicted. The framework is applied to evaluate the performance of RI forecasts by the HWRF Model for the whole life cycle of a single hurricane, as well as for each of the hurricane seasons from 2009 to 2021. The metric based on the probabilities of detection and false alarm rate of RI is compared with that based on the absolute errors in the intensity and intensity change during RI events.

Significance Statement

An evaluation framework for tropical cyclone rapid intensification (RI) forecasts is introduced, focusing on the performance of RI forecasts in each 5-day forecast cycle. The cycle-based approach can help to characterize RI forecasts under different conditions such as certain synoptic scenarios, initial conditions, or vortex structures. It also can be used to assess the overall performance of RI forecasts in terms of the percentages of individual forecast cycles that successfully or falsely predict RI events.

Free access
Brett Roberts
,
Adam J. Clark
,
Israel L. Jirak
,
Burkely T. Gallo
,
Caroline Bain
,
David L. A. Flack
,
James Warner
,
Craig S. Schwartz
, and
Larissa J. Reames

Abstract

As part of NOAA’s Hazardous Weather Testbed Spring Forecasting Experiment (SFE) in 2020, an international collaboration yielded a set of real-time convection-allowing model (CAM) forecasts over the contiguous United States in which the model configurations and initial/boundary conditions were varied in a controlled manner. Three model configurations were employed, among which the Finite Volume Cubed-Sphere (FV3), Unified Model (UM), and Advanced Research version of the Weather Research and Forecasting (WRF-ARW) Model dynamical cores were represented. Two runs were produced for each configuration: one driven by NOAA’s Global Forecast System for initial and boundary conditions, and the other driven by the Met Office’s operational global UM. For 32 cases during SFE2020, these runs were initialized at 0000 UTC and integrated for 36 h. Objective verification of model fields relevant to convective forecasting illuminates differences in the influence of configuration versus driving model pertinent to the ongoing problem of optimizing spread and skill in CAM ensembles. The UM and WRF configurations tend to outperform FV3 for forecasts of precipitation, thermodynamics, and simulated radar reflectivity; using a driving model with the native CAM core also tends to produce better skill in aggregate. Reflectivity and thermodynamic forecasts were found to cluster more by configuration than by driving model at lead times greater than 18 h. The two UM configuration experiments had notably similar solutions that, despite competitive aggregate skill, had large errors in the diurnal convective cycle.

Free access
Xiping Zhang
,
Juan Fang
, and
Zifeng Yu

Abstract

Tropical cyclone (TC) genesis forecasts during 2018–20 from two operational global ensemble prediction systems (EPSs) are evaluated over three basins in this study. The two ensembles are from the European Centre for Medium-Range Weather Forecasts (ECMWF-EPS) and the MetOffice in the United Kingdom (UKMO-EPS). The three basins include the northwest Pacific, northeast Pacific, and the North Atlantic. It is found that the ensemble members in each EPS show a good level of agreement in forecast skill, but their forecasts are complementary. Probability of detection (POD) can be doubled by taking all the member forecasts in the EPS into account. Even if an ensemble member does not make a hit forecast, it may predict the presence of cyclonic vortices. Statistically, a hit forecast has more nearby disturbance forecasts in the ensemble than a false alarm. Based on the above analysis, we grouped the nearby forecasts at each model initialization time to define ensemble genesis forecasts, and verified these forecasts to represent the performance of the ensemble system. The PODs are found to be more than twice that of the individual ensemble members at most lead times, which is about 59% and 38% at the 5-day lead time in UKMO-EPS and ECMWF-EPS, respectively; while the success ratios are smaller compared with that of the ensemble members. In addition, predictability differs in different basins, and genesis events in the North Atlantic basin are the most difficult to forecast in EPS, and its POD at the 5-day lead time is only 46% and 23% in UKMO-EPS and ECMWF-EPS, respectively.

Significance Statement

Operational forecasting of tropical cyclone (TC) genesis relies greatly on numerical models. Compared with deterministic forecasts, ensemble prediction systems (EPSs) can provide uncertainty information for forecasters. This study examined the predictability of TC genesis in two operational EPSs. We found that the forecasts of ensemble members complement each other, and the detection ratio of observed genesis will be doubled by considering the forecasts of all members, as multiple simulations conducted by the EPS partially reflect the inherent uncertainties of the genesis process. Successful forecasts are surrounded by more cyclonic vortices in the ensemble than false alarms, so the vortex information is used to group the nearby forecasts at each model initialization to define ensemble genesis forecasts when evaluating the ensemble performance. The results demonstrate that the global ensemble models can serve as a valuable reference for TC genesis forecasting.

Free access
Patrik Benáček
,
Aleš Farda
, and
Petr Štěpánek

Abstract

Producing an accurate and calibrated probabilistic forecast has high social and economic value. Systematic errors or biases in the ensemble weather forecast can be corrected by postprocessing models whose development is an urgent challenge. Traditionally, the bias correction is done by employing linear regression models that estimate the conditional probability distribution of the forecast. Although this model framework works well, it is restricted to a prespecified model form that often relies on a limited set of predictors only. Most machine learning (ML) methods can tackle these problems with a point prediction, but only a few of them can be applied effectively in a probabilistic manner. The tree-based ML techniques, namely, natural gradient boosting (NGB), quantile random forests (QRF), and distributional regression forests (DRF), are used to adjust hourly 2-m temperature ensemble prediction at lead times of 1–10 days. The ensemble model output statistics (EMOS) and its boosting version are used as benchmark models. The model forecast is based on the European Centre for Medium-Range Weather Forecasts (ECMWF) for the Czech Republic domain. Two training periods 2015–18 and 2018 only were used to learn the models, and their prediction skill was evaluated in 2019. The results show that the QRF and NGB methods provide the best performance for 1–2-day forecasts, while the EMOS method outperforms other methods for 8–10-day forecasts. Key components to improving short-term forecasting are additional atmospheric/surface state predictors and the 4-yr training sample size.

Significance Statement

Machine learning methods have great potential and are beginning to be widely applied in meteorology in recent years. A new technique called natural gradient boosting (NGB) has been released and used in this paper to refine the probabilistic forecast of surface temperature. It was found that the NGB has better prediction skills than the traditional ensemble model output statistics in forecasting 1 and 2 days in advance. The NGB has similar prediction skills with lower computational demands compared to other advanced machine learning methods such as the quantile random forests. We showed a path to employ the NGB method in this task, which can be followed for refining other and more challenging meteorological variables such as wind speed or precipitation.

Free access
Matthew D. Brothers
and
Christopher L. Hammer

Abstract

High winds are one of the key forecast challenges across southeast Wyoming. The complex mountainous terrain across the region frequently results in strong gap winds in localized areas, as well as more widespread bora and chinook winds in the winter season (October–March). The predictors and general weather patterns that result in strong winds across the region are well understood by local forecasters. However, no single predictor provides notable skill by itself in separating warning-level events from others. Random forest (RF) classifier models were developed to improve upon high wind prediction using a training dataset constructed of archived observations and model parameters from the North American Regional Reanalysis (NARR). Three locations were selected for initial RF model development, including the city of Cheyenne, Wyoming, and two gap regions along Interstate 80 (Arlington) and Interstate 25 (Bordeaux). Verification scores over two winters suggested the RF models were beneficial relative to current operational tools when predicting warning-criteria high wind events. Three case studies of high wind events provide examples of the RF models’ effectiveness to forecast operations over current forecast tools. The first case explores a classic, widespread high wind scenario, which was well anticipated by local forecasters. A more marginal scenario is explored in the second case, which presented greater forecast challenges relating to timing and intensity of the strongest winds. The final case study carefully uses Global Forecast System (GFS) data as input into the RF models, further supporting real-time implementation into forecast operations.

Free access
Stephen J. Lord
,
Xingren Wu
,
Vijay Tallapragada
, and
F. M. Ralph

Abstract

The impact of assimilating dropsonde data from the 2020 Atmospheric River (AR) Reconnaissance (ARR) field campaign on operational numerical precipitation forecasts was assessed. Two experiments were executed for the period from 24 January to 18 March 2020 using the NCEP Global Forecast System, version 15 (GFSv15), with a four-dimensional hybrid ensemble–variational (4DEnVar) data assimilation system. The control run (CTRL) used all the routinely assimilated data and included ARR dropsonde data, whereas the denial run (DENY) excluded the dropsonde data. There were 17 intensive observing periods (IOPs) totaling 46 Air Force C-130 and 16 NOAA G-IV missions to deploy dropsondes over targeted regions with potential for downstream high-impact weather associated with the ARs. Data from a total of 628 dropsondes were assimilated in the CTRL. The dropsonde data impact on precipitation forecasts over U.S. West Coast domains is largely positive, especially for day-5 lead time, and appears driven by different model variables on a case-by-case basis. These results suggest that data gaps associated with ARs can be addressed with targeted ARR field campaigns providing vital observations needed for improving U.S. West Coast precipitation forecasts.

Free access
Daniel D. Tripp
,
Joseph E. Trujillo-Falcón
,
Kim E. Klockow-McClain
,
Heather D. Reeves
,
Kodi L. Berry
,
Jeff S. Waldstreicher
, and
James A. Nelson

Abstract

This study explores forecaster perceptions of emerging needs for probabilistic forecasting of winter weather hazards through a nationwide survey disseminated to National Weather Service (NWS) forecasters. Questions addressed four relevant thematic areas: 1) messaging timelines for specific hazards, 2) modeling needs, 3) current preparedness to interpret and communicate probabilistic winter information, and 4) winter forecasting tools. The results suggest that winter hazards are messaged on varying time scales that sometimes do not match the needs of stakeholders. Most participants responded favorably to the idea of incorporating new hazard-specific regional ensemble guidance to fill gaps in the winter forecasting process. Forecasters provided recommendations for ensemble run length and output frequencies that would be needed to capture individual winter hazards. Qualitatively, forecasters expressed more difficulties communicating, rather than interpreting, probabilistic winter hazard information. Differences in training and the need for social-science-driven practices were identified as a few of the drivers limiting forecasters’ ability to provide strategic winter messaging. In the future, forecasters are looking for new winter tools to address forecasting difficulties, enhance stakeholder partnerships, and also be useful to the local community. On the regional scale, an ensemble system could potentially accommodate these needs and provide specialized guidance on timing and sensitive/high-impact winter events.

Significance Statement

Probabilistic information gives forecasters the ability to see a range of potential outcomes so that they can know how much confidence to place in the forecast. In this study, we surveyed forecasters so that we can understand how the research community can support probabilistic forecasting in winter. We found that forecasters want new technologies that help them understand hard forecast situations, improve their communication skills, and that are useful to their local communities. Most forecasters feel comfortable interpreting probabilistic information, but sometimes are not sure how to communicate it to the public. We asked forecasters to share their recommendations for new weather models and tools and we provide an overview of how the research community can support probabilistic winter forecasting efforts.

Free access
William A. Gallus Jr.
and
Anna C. Duhachek

Abstract

Because bow echoes are often associated with damaging wind, accurate prediction of their severity is important. Recent work by Mauri and Gallus showed that despite increased challenges in forecasting nocturnal bows due to an incomplete understanding of how elevated convection interacts with the nocturnal stable boundary layer, several near-storm environmental parameters worked well to distinguish between bow echoes not producing severe winds (NS), those only producing low-intensity severe winds [LS; 50–55 kt (1 kt ≈ 0.51 m s−1)], and those associated with high-intensity (HS; >70 kt) severe winds. The present study performs a similar comparison for daytime warm-season bow echoes examining the same 43 SPC mesoanalysis parameters for 158 events occurring from 2010 to 2018. Although low-level shear and the meridional component of the wind discriminate well for nocturnal bow severity, they do not significantly differ in daytime bows. CAPE parameters discriminate well between daytime NS events and severe ones, but not between LS and HS, differing from nocturnal events where they discriminate between HS and the other types. The 500–850-hPa layer lapse rate works better to differentiate daytime bow severity, whereas the 500–700-hPa layer works better at night. Composite parameters work well to differentiate between all three severity types for daytime bow echoes, just as they do for nighttime ones, with the derecho composite parameter performing especially well. Heidke skill scores indicate that both individual and pairs of parameters generally are not as skillful at predicting daytime bow echo wind severity as they are at predicting nocturnal bow wind severity.

Free access
Hai Zhang
,
Zigang Wei
,
Barron H. Henderson
,
Susan C. Anenberg
,
Katelyn O’Dell
, and
Shobha Kondragunta

Abstract

The mass concentration of fine particulate matter (PM2.5; diameters less than 2.5 μm) estimated from geostationary satellite aerosol optical depth (AOD) data can supplement the network of ground monitors with high temporal (hourly) resolution. Estimates of PM2.5 over the United States were derived from NOAA’s operational geostationary satellites’ Advanced Baseline Imager (ABI) AOD data using a geographically weighted regression with hourly and daily temporal resolution. Validation versus ground observations shows a mean bias of −21.4% and −15.3% for hourly and daily PM2.5 estimates, respectively, for concentrations ranging from 0 to 1000 μg m−3. Because satellites only observe AOD in the daytime, the relation between observed daytime PM2.5 and daily mean PM2.5 was evaluated using ground measurements; PM2.5 estimated from ABI AODs were also examined to study this relationship. The ground measurements show that daytime mean PM2.5 has good correlation (r > 0.8) with daily mean PM2.5 in most areas of the United States, but with pronounced differences in the western United States due to temporal variations caused by wildfire smoke; the relation between the daytime and daily PM2.5 estimated from the ABI AODs has a similar pattern. While daily or daytime estimated PM2.5 provides exposure information in the context of the PM2.5 standard (>35 μg m−3), the hourly estimates of PM2.5 used in nowcasting show promise for alerts and warnings of harmful air quality. The geostationary satellite based PM2.5 estimates inform the public of harmful air quality 10 times more than standard ground observations (1.8 versus 0.17 million people per hour).

Significance Statement

Fine particulate matter (PM2.5; diameters less than 2.5 μm) are generated from smoke, dust, and emissions from industrial, transportation, and other sectors. They are harmful to human health and even lead to premature mortality. Data from geostationary satellites can help estimate surface PM2.5 exposure by filling in gaps that are not covered by ground monitors. With this information, people can plan their outdoor activities accordingly. This study shows that availability of hourly PM2.5 observations covering the entire continental United States is more informative to the public about harmful exposure to pollution. On average, 1.8 million people per hour can be informed using satellite data compared to 0.17 million people per hour based on ground observations alone.

Open access