Search Results

You are looking at 1 - 9 of 9 items for :

  • Author or Editor: Christopher D. Karstens x
  • Weather and Forecasting x
  • Refine by Access: All Content x
Clear All Modify Search
David R. Harrison
and
Christopher D. Karstens

Abstract

This study provides a quantitative climatological analysis of the fundamental geospatial components of storm-based warnings and offers insight into how the National Weather Service (NWS) uses the current storm-based warning system under the established directives and policies. From October 2007 through May 2016, the NWS issued over 500 000 storm-based warnings and severe weather statements (SVSs), primarily concentrated east of the Rocky Mountains. A geospatial analysis of these warning counts by county warning area (CWA) shows local maxima in the lower Mississippi valley, southern plains, central plains, and the southern Appalachians. Regional uniformity exists in the patterns of average speed and direction provided by the time/motion/location tags, while the mean duration and polygon area varies significantly by CWA and region. These observed consistencies and inconsistencies may be indicative of how local weather forecast office (WFO) policy and end-user needs factor into the warning issuance and update process. This research concludes with a comparison of storm-based warnings to NWS policy and an analysis of CWAs with the greatest number of warnings issued during a single convective day.

Full access
Eric D. Loken
,
Adam J. Clark
, and
Christopher D. Karstens

Abstract

Extracting explicit severe weather forecast guidance from convection-allowing ensembles (CAEs) is challenging since CAEs cannot directly simulate individual severe weather hazards. Currently, CAE-based severe weather probabilities must be inferred from one or more storm-related variables, which may require extensive calibration and/or contain limited information. Machine learning (ML) offers a way to obtain severe weather forecast probabilities from CAEs by relating CAE forecast variables to observed severe weather reports. This paper develops and verifies a random forest (RF)-based ML method for creating day 1 (1200–1200 UTC) severe weather hazard probabilities and categorical outlooks based on 0000 UTC Storm-Scale Ensemble of Opportunity (SSEO) forecast data and observed Storm Prediction Center (SPC) storm reports. RF forecast probabilities are compared against severe weather forecasts from calibrated SSEO 2–5-km updraft helicity (UH) forecasts and SPC convective outlooks issued at 0600 UTC. Continuous RF probabilities routinely have the highest Brier skill scores (BSSs), regardless of whether the forecasts are evaluated over the full domain or regional/seasonal subsets. Even when RF probabilities are truncated at the probability levels issued by the SPC, the RF forecasts often have BSSs better than or comparable to corresponding UH and SPC forecasts. Relative to the UH and SPC forecasts, the RF approach performs best for severe wind and hail prediction during the spring and summer (i.e., March–August). Overall, it is concluded that the RF method presented here provides skillful, reliable CAE-derived severe weather probabilities that may be useful to severe weather forecasters and decision-makers.

Free access
Amy McGovern
,
Christopher D. Karstens
,
Travis Smith
, and
Ryan Lagerquist

Abstract

Real-time prediction of storm longevity is a critical challenge for National Weather Service (NWS) forecasters. These predictions can guide forecasters when they issue warnings and implicitly inform them about the potential severity of a storm. This paper presents a machine-learning (ML) system that was used for real-time prediction of storm longevity in the Probabilistic Hazard Information (PHI) tool, making it a Research-to-Operations (R2O) project. Currently, PHI provides forecasters with real-time storm variables and severity predictions from the ProbSevere system, but these predictions do not include storm longevity. We specifically designed our system to be tested in PHI during the 2016 and 2017 Hazardous Weather Testbed (HWT) experiments, which are a quasi-operational naturalistic environment. We considered three ML methods that have proven in prior work to be strong predictors for many weather prediction tasks: elastic nets, random forests, and gradient-boosted regression trees. We present experiments comparing the three ML methods with different types of input data, discuss trade-offs between forecast quality and requirements for real-time deployment, and present both subjective (human-based) and objective evaluation of real-time deployment in the HWT. Results demonstrate that the ML system has lower error than human forecasters, which suggests that it could be used to guide future storm-based warnings, enabling forecasters to focus on other aspects of the warning system.

Open access
Bryan T. Smith
,
Richard L. Thompson
,
Douglas A. Speheger
,
Andrew R. Dean
,
Christopher D. Karstens
, and
Alexandra K. Anderson-Frey

Abstract

The Storm Prediction Center (SPC) has developed a database of damage-surveyed tornadoes in the contiguous United States (2009–17) that relates environmental and radar-derived storm attributes to damage ratings that change during a tornado life cycle. Damage indicators (DIs), and the associated wind speed estimates from tornado damage surveys compiled in the Damage Assessment Toolkit (DAT) dataset, were linked to the nearest manual calculations of 0.5° tilt angle maximum rotational velocity V rot from single-site WSR-88D data. For each radar scan, the maximum wind speed from the highest-rated DI, V rot, and the significant tornado parameter (STP) from the SPC hourly objective mesoscale analysis archive were recorded and analyzed. Results from examining V rot and STP data indicate an increasing conditional probability for higher-rated DIs (i.e., EF-scale wind speed estimate) as both STP and V rot increase. This work suggests that tornadic wind speed exceedance probabilities can be estimated in real time, on a scan-by-scan basis, via V rot and STP for ongoing tornadoes.

Full access
Bryan T. Smith
,
Richard L. Thompson
,
Douglas A. Speheger
,
Andrew R. Dean
,
Christopher D. Karstens
, and
Alexandra K. Anderson-Frey

Abstract

A sample of damage-surveyed tornadoes in the contiguous United States (2009–17), containing specific wind speed estimates from damage indicators (DIs) within the Damage Assessment Toolkit dataset, were linked to radar-observed circulations using the nearest WSR-88D data in Part I of this work. The maximum wind speed associated with the highest-rated DI for each radar scan, corresponding 0.5° tilt angle rotational velocity V rot, significant tornado parameter (STP), and National Weather Service (NWS) convective impact-based warning (IBW) type, are analyzed herein for the sample of cases in Part I and an independent case sample from parts of 2019–20. As V rot and STP both increase, peak DI-estimated wind speeds and IBW warning type also tend to increase. Different combinations of V rot, STP, and population density—related to ranges of peak DI wind speed—exhibited a strong ability to discriminate across the tornado damage intensity spectrum. Furthermore, longer duration of high V rot (i.e., ≥70 kt) in significant tornado environments (i.e., STP ≥ 6) corresponds to increasing chances that DIs will reveal the occurrence of an intense tornado (i.e., EF3+). These findings were corroborated via the independent sample from parts of 2019–20, and can be applied in a real-time operational setting to assist in determining a potential range of wind speeds. This work provides evidence-based support for creating an objective and consistent, real-time framework for assessing and differentiating tornadoes across the tornado intensity spectrum.

Full access
Joseph J. J. James
,
Chen Ling
,
Christopher D. Karstens
,
James Correia Jr.
,
Kristin Calhoun
,
Tiffany Meyer
, and
Daphne LaDue

Abstract

During spring 2016 the Probabilistic Hazard Information (PHI) prototype experiment was run in the National Oceanic and Atmospheric Administration (NOAA) Hazardous Weather Testbed (HWT) as part of the Forecasting a Continuum of Environmental Threats (FACETS) program. Nine National Weather Service forecasters were trained to use the web-based PHI prototype tool to produce dynamic PHI for severe weather threats. Archived and real-time weather scenarios were used to test this new paradigm of issuing probabilistic information, rather than deterministic information. The forecasters’ mental workload was evaluated after each scenario using the NASA-Task Load Index (TLX) questionnaire. This study summarizes the analysis results of mental workload experienced by forecasters while using the PHI prototype. Six subdimensions of mental workload: mental demand, physical demand, temporal demand, performance, effort, and frustration were analyzed to derive top contributing factors to workload. Average mental workload was 46.6 (out of 100, standard deviation: 19, range 70.8). Top contributing factors to workload included using automated guidance, PHI object quantity, multiple displays, and formulating probabilities in the new paradigm. Automated guidance provided support to forecasters in maintaining situational awareness and managing increased quantities of threats. The results of this study provided understanding of forecasters’ mental workload and task strategies and developed insights to improve usability of the PHI prototype tool.

Free access
Christopher D. Karstens
,
Greg Stumpf
,
Chen Ling
,
Lesheng Hua
,
Darrel Kingfield
,
Travis M. Smith
,
James Correia Jr.
,
Kristin Calhoun
,
Kiel Ortega
,
Chris Melick
, and
Lans P. Rothfusz

Abstract

A proposed new method for hazard identification and prediction was evaluated with forecasters in the National Oceanic and Atmospheric Administration Hazardous Weather Testbed during 2014. This method combines hazard-following objects with forecaster-issued trends of exceedance probabilities to produce probabilistic hazard information, as opposed to the static, deterministic polygon and attendant text product methodology presently employed by the National Weather Service to issue severe thunderstorm and tornado warnings. Three components of the test bed activities are discussed: usage of the new tools, verification of storm-based warnings and probabilistic forecasts from a control–test experiment, and subjective feedback on the proposed paradigm change. Forecasters were able to quickly adapt to the new tools and concepts and ultimately produced probabilistic hazard information in a timely manner. The probabilistic forecasts from two severe hail events tested in a control–test experiment were more skillful than storm-based warnings and were found to have reliability in the low-probability spectrum. False alarm area decreased while the traditional verification metrics degraded with increasing probability thresholds. The latter finding is attributable to a limitation in applying the current verification methodology to probabilistic forecasts. Relaxation of on-the-fence decisions exposed a need to provide information for hazard areas below the decision-point thresholds of current warnings. Automated guidance information was helpful in combating potential workload issues, and forecasters raised a need for improved guidance and training to inform consistent and reliable forecasts.

Full access
Christopher D. Karstens
,
James Correia Jr.
,
Daphne S. LaDue
,
Jonathan Wolfe
,
Tiffany C. Meyer
,
David R. Harrison
,
John L. Cintineo
,
Kristin M. Calhoun
,
Travis M. Smith
,
Alan E. Gerard
, and
Lans P. Rothfusz

Abstract

Providing advance warning for impending severe convective weather events (i.e., tornadoes, hail, wind) fundamentally requires an ability to predict and/or detect these hazards and subsequently communicate their potential threat in real time. The National Weather Service (NWS) provides advance warning for severe convective weather through the issuance of tornado and severe thunderstorm warnings, a system that has remained relatively unchanged for approximately the past 65 years. Forecasting a Continuum of Environmental Threats (FACETs) proposes a reinvention of this system, transitioning from a deterministic product-centric paradigm to one based on probabilistic hazard information (PHI) for hazardous weather events. Four years of iterative development and rapid prototyping in the National Oceanic and Atmospheric Administration (NOAA) Hazardous Weather Testbed (HWT) with NWS forecasters and partners has yielded insights into this new paradigm by discovering efficient ways to generate, inform, and utilize a continuous flow of information through the development of a human–machine mix. Forecasters conditionally used automated object-based guidance within four levels of automation to issue deterministic products containing PHI. Forecasters accomplished this task in a timely manner while focusing on communication and conveying forecast confidence, elements considered necessary by emergency managers. Observed annual increases in the usage of first-guess probabilistic guidance by forecasters were related to improvements made to the prototyped software, guidance, and techniques. However, increasing usage of automation requires improvements in guidance, data integration, and data visualization to garner trust more effectively. Additional opportunities exist to address limitations in procedures for motion derivation and geospatial mapping of subjective probability.

Full access
Burkely T. Gallo
,
Adam J. Clark
,
Israel Jirak
,
John S. Kain
,
Steven J. Weiss
,
Michael Coniglio
,
Kent Knopfmeier
,
James Correia Jr.
,
Christopher J. Melick
,
Christopher D. Karstens
,
Eswar Iyer
,
Andrew R. Dean
,
Ming Xue
,
Fanyou Kong
,
Youngsun Jung
,
Feifei Shen
,
Kevin W. Thomas
,
Keith Brewster
,
Derek Stratman
,
Gregory W. Carbin
,
William Line
,
Rebecca Adams-Selin
, and
Steve Willington

Abstract

Led by NOAA’s Storm Prediction Center and National Severe Storms Laboratory, annual spring forecasting experiments (SFEs) in the Hazardous Weather Testbed test and evaluate cutting-edge technologies and concepts for improving severe weather prediction through intensive real-time forecasting and evaluation activities. Experimental forecast guidance is provided through collaborations with several U.S. government and academic institutions, as well as the Met Office. The purpose of this article is to summarize activities, insights, and preliminary findings from recent SFEs, emphasizing SFE 2015. Several innovative aspects of recent experiments are discussed, including the 1) use of convection-allowing model (CAM) ensembles with advanced ensemble data assimilation, 2) generation of severe weather outlooks valid at time periods shorter than those issued operationally (e.g., 1–4 h), 3) use of CAMs to issue outlooks beyond the day 1 period, 4) increased interaction through software allowing participants to create individual severe weather outlooks, and 5) tests of newly developed storm-attribute-based diagnostics for predicting tornadoes and hail size. Additionally, plans for future experiments will be discussed, including the creation of a Community Leveraged Unified Ensemble (CLUE) system, which will test various strategies for CAM ensemble design using carefully designed sets of ensemble members contributed by different agencies to drive evidence-based decision-making for near-future operational systems.

Full access