1. Introduction
Meteorological radars are an integral piece of the weather enterprise’s toolset for observing and predicting weather with relatively high spatial and temporal resolution and across broad swaths of area as compared with in situ observation methods. Weather radar observations are essential for a wide range of meteorological applications including quantitative precipitation estimation (QPE; Hong and Gourley 2014), tornado detection and warning (e.g., Brown et al. 1978), identifying severe wind and hail hazards (e.g., Burgess and Lemon 1990; Ortega et al. 2016), characterizing tornado/mesoscale dynamics (e.g., Wurman et al. 1996), forecasting real-time mesoscale snowbands (e.g., Kristovich et al. 2003), and hurricane rainbands (e.g., Biggerstaff et al. 2021), and assimilation into numerical weather prediction systems (e.g., Stensrud et al. 2009).
To enable these applications, the U.S. National Weather Service (NWS) operates a network of Weather Surveillance Radar-1988 Doppler (WSR-88D) meteorological radar sets that provide real-time atmospheric observations over the United States and outlying territories. The ability of WSR-88D estimates to successfully support the NWS mission is contingent on a series of processing steps that convert raw in-phase and quadrature receiver base-band samples (I/Q; termed “Level-I” data) into a collection of standard radar data. The most basic of these data for a dual-polarization radar are the six “Level-II” variables that are calculated through time series processing of the Level-I signals: horizontal reflectivity factor, radial velocity, Doppler spectrum width, differential reflectivity, copolar correlation coefficient, and differential phase. These Level-II variables contain information relating to the size, shape, abundance, composition, and motion of objects in the atmosphere. A number of additional “Level-III” products are generated within the WSR-88D Open Radar Product Generator (ORPG) using more complex workflows and often incorporate two-, three-, or four-dimensional analyses across multiple Level-II variables, sometimes in combination with ancillary datasets beyond just radar (e.g., model data). For example, the ORPG QPE algorithm combines multiple Level-II variables and integrates estimated rainfall over time in a Level-III product. As a result of these sophisticated processing techniques, Level-III products often have a greater degree of quality control, are more directly relatable to common users of radar data, and provide more intuitive visual interpretation of the phenomena that they are meant to characterize, making them well suited for use by weather forecasters and the general public.
While the WSR-88D VDA demonstrates exceptional performance and ongoing improvements within each ORPG build, the algorithm is relatively complex and requires close to 1 s per tilt (constant-elevation-angle 360° azimuthal scan) for complex storm environments. While this is acceptable on appropriate hardware and for the current scan rates for the WSR-88D (which peak at 4.8 revolutions per minute and must keep up in “real time”), there are several reasons why a simpler/faster implementation would be highly desirable. For example, future radar systems, such as phased arrays, may scan much faster (Weber et al. 2021), resulting in the need for algorithms that are capable of running more efficiently. Several ORPG algorithms take much more than 1 s to run, making more efficient performance for future faster-scanning radars a must. In addition, the ORPG VDA utilizes environmental wind information, which is not always accessible in real time or when processing large batches of archived data. An accurate, fast, and highly portable algorithm that only requires a single velocity image would be of great value in situations like these.
For these reasons, we utilize machine learning (ML), specifically deep learning (DL), to emulate the two-dimensional WSR-88D VDA using purely data-driven approaches. DL approaches have been successfully demonstrated in diverse meteorological applications such as precipitation nowcasting (Ravuri et al. 2021; Xingjian et al. 2015), subseasonal forecasting (Weyn et al. 2021), satellite precipitation estimation, (Wang et al. 2021), synthetic weather radar (Veillette et al. 2018; Hilburn et al. 2021), mesocyclone classification (Krinitskiy et al. 2018), extreme weather event detection (Racah et al. 2017), and several others. Within radar meteorology specifically, a range of different ML algorithms has been used, and many algorithms were summarized by Chandrasekar (2020), with focuses on QPE, precipitation classification, and nowcasting. For example, Chen et al. (2020) use deep neural networks to estimate precipitation using both satellite-borne and ground-based radar networks. Other QPE-based DL approaches include those by Yo et al. (2021), Husnoo et al. (2021) (using the U.K. weather radar network), and Peng et al. (2022). Although not DL-centric, precipitation classification has used other forms of ML, including fuzzy logic (Park et al. 2009) and Bayesian approaches (Yang et al. 2019).
Currently, only one operational ORPG algorithm utilizes modern machine learning techniques (the chaff detection algorithm; Kurdzo et al. 2017a), but, because of the wide accessibility to years of archived data, the resources exist to enable other ML-based algorithms (e.g., Jatau et al. 2021). A DL-based VDA should enable a faster algorithm, provide a WSR-88D-quality dealiaser for users of other radar datasets, and possibly be implementable within open-source software packages. The VDA is an ideal algorithm to attempt to replicate within the ORPG, because it uses just one moment (the velocity field), is computationally complex (with a desire to decrease resources, especially for non-ORPG hardware), is a critical need for non-WSR-88D instruments, and performs well enough to use as an unsupervised truthed dataset. In this work, we specifically attempt to emulate the WSR-88D velocity dealiasing algorithm. This is an important distinction, because we are not expecting to necessarily improve upon the algorithm. The reason for this approach is that the existing algorithm can serve as an exceptionally large truthed dataset. The Level-III velocity product can be considered truth at all times if we are attempting to emulate/recreate the algorithm, eliminating the need for any manual dealiasing.
While improving upon the algorithm would be ideal, this approach sets the stage for future recreation of other ORPG algorithms using ML techniques for both computational efficiency and portability. We explore different techniques to optimize the algorithm, approaches for improving upon baseline results, and applications beyond the WSR-88D to other radars. The paper is organized as follows: section 2 discusses prior work in the VDA space; section 3 describes the data and the approach; section 4 discusses the chosen model architecture; section 5 presents results; and section 6 provides a discussion of the applicability of the results, along with some conclusions.
2. Prior work
Since some of the earliest work on velocity ambiguities by Ray and Ziegler (1977), several approaches to dealiasing constant-PRF radar radial velocity data have been presented in the literature, and a subset of these has been deployed operationally in the ORPG. The initial ORPG VDA was based on Eilts and Smith (1990); it takes a radial-by-radial approach to unfolding velocities. Each radial is compared with the previous radial using a local environment dealiasing (LED) technique. These radial continuity checks are based only in the radial dimension, making the algorithm computationally efficient. As discussed in Witt et al. (2009), as well as Zittel et al. (2011), the Eilts and Smith (1990) approach worked well in an initial ORPG implementation with limited computational resources but often failed in cases of lower Nyquists (i.e., volume coverage pattern 31) and strong shear (mesocyclones, gust fronts, etc.). As new computational capabilities were added, it became possible to implement a new ORPG VDA based on the work in Jing and Wiener (1993), which takes a two-dimensional approach to solving a linear system of equations in a connected two-dimensional region by minimizing discontinuities in a least squares problem. The ORPG version of the algorithm adds a weighting factor to Eq. (5) of Jing and Wiener (1993), which aids with issues in noisy data (Witt et al. 2009). A global dealiaser using environmental data is also applied as part of Jing and Wiener (1993) and is used in the ORPG VDA as the second step of a “two-phased dealiasing approach” (Witt et al. 2009). Anecdotally, the ORPG VDA performs admirably in most situations; however, the two-dimensional VDA is still relatively complex and computationally expensive (Witt et al. 2009; Zittel et al. 2011; NWS2021).
The current WSR-88D VDA is available via the ORPG Common Operations and Development Environment (CODE) software package (CODE 2021), an open-source version of the ORPG. As of this writing, the CODE ORPG Build 20 source code (written mainly in the C programming language) is available to the public (https://www.weather.gov/code88d/). To the authors’ knowledge, there is no widespread use of the WSR-88D VDA in other operational radars or research radars.
Beyond 2D VDAs (the main focus of this work), other approaches that utilize additional dimensions have been developed. For example, James and Houze (2001) present a real-time, four-dimensional dealiasing (4DD) scheme that takes a top-tilt-down approach using the elevation and time dimensions in addition to investigating neighboring gates and radials. This technique is implemented along with other dealiasing algorithms in the Python ARM Radar Toolkit (Py-ART) (Helmus and Collis 2016). These algorithms are also often similarly (or more) computationally complex in comparison with the WSR-88D algorithm, frequently taking a nontrivial amount of time to run, especially in complex velocity field situations. Additionally, at C-band and higher frequencies, with low Nyquist values, automated dealiasers perform particularly poorly, especially in tornadoes (e.g., Bluestein et al. 2007; Yu et al. 2007). For this reason, most research radars (e.g., mobile C-band, X-band, Ka-band, and W-band radars) that collect tornado data must have their velocities manually dealiased, step-by-step in time and elevation (e.g., Oye et al. 1995). This onerous task is extremely time consuming and often unavoidable. Even operational radars at C band, especially those in areas of sparse data and high environmental shear, can suffer from dealiaser failures, such as the Australian weather radar network (Louf et al. 2020).
After the 4DD approach was developed, additional dimensions were explored for dealiasing, especially with low-Nyquist radar systems. For example, Zhang and Wang (2006) discuss multipass filters that determine confidence levels that subsequently increase with additional passes of dealiasing. Similar to other multidimensional approaches, multiple passes add time and complexity to the processing steps despite performing exceptionally well. Xu et al. (2011) used VAD profiles to dealias radar data for numerical weather prediction assimilation. While their methods perform well in many case examples, they tend to break down in areas of localized high shear (mesocyclones, tornadoes, hurricane eyewalls, etc.). He et al. (2012) developed a C-band-specific approach to dealiasing for the Chinese CINRAD radar network, which operates at a single PRF with a high frequency, making them particularly susceptible to velocity aliasing issues.
More recently, the Unfold Radar Velocity (UNRAVEL) algorithm takes a modular approach to a VDA, allowing for an optimization of results that can balance between performance needs and computational complexity (Louf et al. 2020). UNRAVEL currently contains two different (but expandable) dealiasing strategies: a tilt-by-tilt (two dimensional) and a three-dimensional continuity check. Louf et al. (2020) use artificially aliased S-band data as a reference for analyzing performance at C band, a technique that is adopted in a similar fashion in this study. Finally, Feldmann et al. (2020) introduce the region-based recursive Doppler dealiasing method (R2D2) for dealiasing in highly sheared environments, a historically difficult area for VDAs. A four-dimensional procedure with a two-dimensional continuity check iterates until improvement is no longer achieved, and the highly sheared areas are given a “focus” with an image mask and a spatial buffer, resulting in exceptional performance.
3. Data description
This proposed VDA is entirely data driven and requires a large number of aligned samples containing input aliased velocity, associated Nyquist velocities, and dealiased target velocity: (Va,i, Vn,i, Vi). Below, we describe the sources used to build datasets for the development and evaluation of a DNN-based VDA. In addition, we describe the processing pipeline implemented to automatically construct labeled training samples required for DNN training and quantitative assessments.
a. Training data description
Training data for dealiasing were derived entirely from WSR-88D data products. As described in the introduction, WSR-88D data are available from two tiers of products, Level-II variables and Level-III products that are derived from Level-II variables. Of particular interest for this work is the Level-III radial velocity product, which is a dealiased and data quality edited version of the (potentially aliased) Level-II velocity.
While no perfect “truth” exists for this task (outside manually labeled data), Level-III velocity was chosen as truth for a DL implementation because it offers a high-quality, operationally hardened reference and is anecdotally seen as a gold standard among forecasters. Archives of both Level-II and Level-III products can be downloaded for free from various cloud providers—for example, Google Cloud Storage (https://cloud.google.com/storage/docs/public-datasets/nexrad).
1) Preprocessing and labeling
Training data requires that pairs of input velocity Va and target dealiased velocity V be coaligned gate by gate. For this reason, in an approach similar to Louf et al. (2020), training targets were created by artificially aliasing Level-III data using an input Nyquist velocity Vn (assumed in this work to be constant over each two-dimensional radar tilt).
The processing pipeline for creating a training pair is described as follows: given a WSR-88D site ID and a time stamp, base Level-III velocity data (code N0U within the ORPG used by the Radar Operations Center) are downloaded and extracted into a two-dimensional array with a height dimension of Naz = 360 azimuths and width dimension truncated to the first Nrng = 1152 radial gates (this choice of radial dimension made training more convenient but is not required during testing). The azimuthal dimension of each array is sorted such that the radial with minimum azimuth is located at the top of the array (index 0), ensuring a consistent azimuthal indexing across the training data. Since the Level-III velocity algorithm clips data to the range of [−63.5, +63.5 m s−1], regions in which the end points of this clipping range are observed need to be filled in using data from the nearest Level-II velocity field, which is downsampled by averaging radials in each 1° arc. Note that the dealiasing of Level-II data for filling data gaps is possible since it is known that the magnitude of velocity within clipped regions exceeds the limits of Level-III data, and hence, within clipped regions, the Nyquist velocity can be added/subtracted until values in the region exceed the clipped value.
After extracting, reindexing, and backfilling clipped regions, the resulting two-dimensional V arrays and associated Nyquist velocities are saved to disk using Tensorflow Record format (https://www.tensorflow.org/tutorials/load_data/tfrecord), which is a simple and efficient format for training DNNs using large datasets. At train time, a data loader reads a batch of two-dimensional dealiased velocity grids V and Nyquist velocities Vn and constructs a batch of artificially folded velocities Va along with label arrays F containing the target fold numbers n. To increase the difficultly of dealiasing, the Nyquist velocities used operationally were lowered by an amount of 10 m s−1 (a hyperparameter) so that the DNN would be exposed to more folded regions. This was also done to increase the likelihood that the DNN will generalize to radars that commonly use lower Nyquist velocities than does the S-band WSR-88D (such as radars at higher frequencies).
The label array F used for training is of shape Naz × Nrng × 6, where the last dimension is a “one-hot” encoding representing the categories shown in Table 1. Note that fold values with |n| > 2 are not included as a separate category as they are exceptionally rare in WSR-88D data. This makes use of the algorithm in some situations and with certain radar types less tenable, such as higher-frequency mobile radars in cases of tornadoes (e.g., Kurdzo et al. 2017b) and hurricanes (e.g., Alford et al. 2019). However, some examples of successful unfolding with these types of radars are shown in section 6. Using the fold array F, the artificially folded array Va was constructed by subtracting or adding the appropriate number of Nyquist velocities. The data were padded by 12° in the azimuthal dimension with a periodic boundary condition to create arrays with shape Naz = 384. An example of this processing can be seen in Fig. 2, where the original Level-III velocity is shown on the left, with regions exceeding the Nyquist velocity contoured in black. The artificially aliased velocity is shown in the center panel, and the resulting fold number n is shown on the right.
Translation of dealiased Level-III velocity V into one-hot encoded labels used for DNN training.
The result of this pipeline is four variables which are grouped as inputs: X = (Va, Vn) and outputs: Y = (F, V). These samples are grouped into minibatches and used for DNN training.
2) Event selection
Selecting a training dataset that includes sufficient representation of strong radial velocities and highly sheared environments is critical for the success of the algorithm. Storm events that contain strong and dynamic winds, such as microbursts, mesocyclones, tornadoes, and hurricanes, are not only more difficult to dealias due to sudden changes in wind magnitude and direction, but also represent the greatest impact operationally. Since these events are statistically rare, random sampling of events will likely underrepresent them. To ensure the training and testing datasets provide adequate coverage of these events, the NCEI Storm Event Database (https://www.ncdc.noaa.gov/stormevents/) was used to identify radar sites and times where these high-impact storms were reported. In addition to events with confirmed severe storm occurrences, a set of “null events” was also selected where either no storms were reported or warnings were issued and no confirmed cases were surveyed.
In all, the resulting dataset derived from WSR-88D was selected from 6350 separate events from the Storm Event Database that fall between January 2017 and June 2022 (66 months). For each of these events, several velocity images were sampled, totaling 427 491 processed velocity images of shape 360 × 1152. This dataset was split into three subsets: 80% for training sampled between January 2017 and October 2020, 8% for validation sampled between November 2020 and May 2021, and 12% for testing sampled between June 2021and June 2022. Partitioning the data by time was done to avoid potential overlaps in storm events and/or radars. When saved to disk and compressed, this combined dataset totaled 229 gigabytes in size.
b. Testing data description
In addition to the processed Level-III test data mentioned in the previous section, Level-II velocity as well as velocity data from other radars were also used for testing. The other radars included for verification purposes are the PX-1000 X-band radar (Cheong et al. 2013) and the C-band Terminal Doppler Weather Radar (TDWR; Michelson et al. 1990; Cho and Weber 2010). PX-1000 is a transportable, solid-state, polarimetric radar that has been used to study severe weather, tornadoes, and rainfall, as well as serve as a system for radar engineering, teaching, and research. The TDWR is an operational radar that is part of a network of 45 systems across the United States, and it serves as a primary tool for detecting wind shear and hazardous convection at large airports (Evans and Turnbull 1989).
The characteristics of the different test sets are described in Table 2. One challenge in using non-Level-III data is that there is no ideal dealiased reference to use for comparison, so verification for non-Level-III-derived velocity will be mostly qualitative and will utilize multiple views of the same event to increase confidence that the DNN VDA is correctly identifying the number of folds in the observed data. Quantitative scoring will be performed only in the case of artificially aliased Level-III data, where a truth is available.
Summary of different radar products used for testing.
4. Model description
This section describes the architecture of the DNN model f(·; θ). Dealiasing is framed here as an image-segmentation task, which amounts to classifying each pixel in a velocity image as one of 6 categories corresponding to fold numbers n = −2, −1, 0, 1, 2, or background. After the most likely fold number is identified for each pixel, the input aliased velocity is modified by the correct scaling of Nyquist velocity using Eq. (2).
The particular DNN model chosen for this task is a U-Net (Ronneberger et al. 2015), which is a popular neural network architecture that has been applied in a number of areas, including image segmentation (Iglovikov and Shvets 2018), biomedical image analysis (Haque and Neubert 2020), and image-to-image translation (Isola et al. 2017). The U-Net has also been widely adopted in meteorological applications, including nowcasting (Agrawal et al. 2019), precipitation estimation (Wang et al. 2021), satellite image gap filling (Xing et al. 2022), and several others. What makes the U-Net an effective model is its ability to learn and process features at multiple spatial scales across an image. This property makes the architecture ideal for dealiasing as it requires identifying sudden changes in velocities at fine scales and identifying wind patterns across the whole image without the need for environmental wind information, in order to determine the correct fold numbers.
Another beneficial property of the U-Net architecture is that it is fully convolutional, making it possible to perform inference on images that are of a different radial shape Nrng when compared with the training data. Different azimuthal shapes in testing data can also be used; however, the resolution should be in 1° steps, and it is recommended that inputs contain a full 360° + 12° of periodic boundary conditions (Naz = 384). The portability of the U-Net architecture will be demonstrated in a later section by applying the trained model to velocity data of different shapes. It is important to note that cases with sector scans, as can be common in mobile radars, are not yet technically supported with this technique. However, modifications can be made to the data and the algorithm in the future, including zero-padding of azimuths that are missing. While this has not been specifically tested, it would be a straightforward approach to using the U-Net on these types of datasets. A more difficult problem to solve would be the limitation on 1° azimuthal steps, which, as of the writing of this study, will limit the algorithm’s utility in some research settings. Future work, including input from the research community, will include these types of adjustments.
The details of the U-Net architecture used for dealiasing are shown in Fig. 3. The network takes two inputs. The first input is a batch of observed (aliased) velocity Va with shape [B, Naz, Nrng, 1], where B denotes the batch size, Naz is the (padded) azimuthal dimension, Nrng is the radial dimension, which may be of arbitrary size, and the final 1 denotes the singleton channel dimension of the input. The second input to the model is the array of Nyquist velocities for each element of the batch used to artificially fold the target images, which has shape [B, 1]. The aliased velocity is initially scaled by 1/Vn, which produces values in the interval [−1, 1]. After normalization, background values in the original velocity are set to a placeholder value of −3.
After normalization, the downward portion of the U-Net, made up of blocks labeled D1, …, D4, extracts features from the input and maps the data to lower and lower resolutions by cutting the sizes of each dimension by factors of 2 or 4 at each stage. Each D-block is made up of a series of standard deep learning operations summarized in the top panel of Fig. 4. The D-block is composed of two paths, the first being 2 two-dimensional convolutional layers with kernel size k with a leaky rectified linear unit (ReLU) activation function that uses α = 0.1. The output of the second convolution is downsampled with a d1, d2 average pooling layer. The number of channels (features) output by each D-block doubles after the initial block, and the channel dimension of the first output block D1 is a configurable parameter called the starting neurons or sn. The values of k, d1, and d2 for the four D-blocks are shown in the left portion of Table 3.
At the base of the dealiasing U-Net, low-resolution feature maps are processed by three parallel convolutional layers with kernel sizes of k =3, 5, and 7. These additional convolutions were added to the standard U-Net architecture to maximize the receptive field of the convolutions, which allows nonlocal features to be incorporated into the resulting segmentation. Larger receptive fields should help correctly identify predominate winds without the need for a reference velocity field. Experimentation with different network architectures showed that U-Nets with only a single 3 × 3 convolutional layer at the base produced validation losses that were approximately 2 times as large as architectures that use the parallel convolutions.
The right portion of the U-Net is made up of four upsampling blocks, Ui, i = 1, …, 4. These blocks take two inputs. The first input is the output of each parallel D-block, and the second is the output of the preceding U-block. The steps performed in the U-block are shown in the bottom panel of Fig. 4. The two inputs are concatenated along the channel dimension and followed by two applications of batch normalization (Ioffe and Szegedy 2015), ReLU, upsampling, and two-dimensional convolution. In contrast to the D-blocks, the output of each U-block halves the number of features. The values of k, d1, and d2 for the four U-blocks are shown in the right portion of Table 3.
The output of the final U-block is passed to a 3 × 3 convolutional layer with linear activation to produce the one-hot encoded array F with spatial dimensions equal to the input velocity. The output array F has six output channels that can be interpreted as log odds of the six potential categories of each pixel. The correct fold number is computed using the argmax along the channel dimension of this array. It is worth pointing out that this output can also be combined with a softmax activation to create probabilistic predictions across different fold numbers; however, this idea was not explored further.
5. Results
a. Model training
The model described in section 4 was implemented using the Tensorflow Keras DL framework (Bisong 2019) and trained on two NVIDIA Volta V100 graphical processing units (GPUs) located within the MIT Lincoln Laboratory TX-Green Super Computing Center (Kepner et al. 2017). Multiple hyper-parameter settings of the model were tested during development, and results are provided for a U-Net with starting neuron choices sn = 8, 16, and 24. A categorical cross-entropy loss function was used with an Adam (Kingma and Ba 2014) optimizer for weight updates. Computations utilized mixed-precision training (Samsi et al. 2020), where network operations use a 16-bit floating point, and weight updates were performed using a 32-bit floating point. An initial learning rate of 0.0025 was used and was decreased by a factor of 0.99 every epoch. Training iterations used a data-distributed model with a batch size of 32 input images spread across the two GPUs. The model was allowed to train for 200 epochs with 600 batches used per epoch. The total time for training averaged approximately 25 h per model.
Figure 5 shows the evolution of two metrics during training for a model trained with sn = 16. The left panel shows categorical cross-entropy loss, which describes how well the model classifies fold number across each image, and the right panel shows velocity RMSE computed between the true velocity and dealiased velocity. The training curve shows performance on the training set, which includes an artificially lowered Nyquist velocity. To ensure accuracy of the target dataset, the validation curve reflects Nyquist velocities that were used operationally for the WSR-88D, which explains why the validation results are slightly better than training curves. As can be seen by the validation loss curves, the model seems to fully converge by around epoch 125. The training epoch with the lowest-velocity RMSE on the validation set was selected for further testing.
b. Test set results
The trained U-Net with sn = 16 was tested on several instances of the 0.5° elevation Level-II velocity data with four sample cases visualized in Fig. 6. Note that these tilts of the Level-II velocity have a 0.5° azimuthal resolution, which is half that of the training data. Since the U-Net was trained on 1° data, a single scan of Level-II V was split into a batch of two separate 1° scans by splitting every other radial. After both of these splits were dealiased using the U-Net, the result was recombined into a 0.5° dataset.
These results demonstrate the ability of the U-Net to handle challenging dealiasing situations that involve global large-scale features, as well as resolve finer-scale aliasing in situations of high shear (e.g., severe storms). The four cases shown include zoomed-out (case a) and zoomed-in (case b) views of a tornado near KGWX (Greenwood Springs, Mississippi) at 0412 UTC 14 April 2019, Hurricane Ida as it moved past KLIX (Slidell, Louisiana) at 1639 UTC 29 August 2021 (case c), and a widespread area of convection with several isolated cells near KPAH (West Paducah, Kentucky) at 0352 UTC 11 December 2021 (case d). The highly sheared tornadic signature in case b, as well as the eyewall during Hurricane Ida in case c, shows examples of how the U-Net can correctly dealias areas with rapidly shifting, strong winds in complex environments. Cases a and d show examples of how the U-Net can identify prevailing winds across an entire PPI in order to dealias isolated cells where the correct direction is not necessarily obvious without external environmental data. In each case shown, the U-Net results are consistent with the Level-III reference image shown in the right-most column.
The comparison of performance of the U-Net and region-based VDAs on the Level-III test set is shown in Table 4. Across all metrics, the U-Net performance exceeds that of the baseline considered, with the best performance being observed for the U-Net with the most learnable parameters (sn = 24); however, the differences in performance metrics between the three U-Nets considered is small. When compared with the region-based baseline, the U-Net improves the velocity RMSE by over 300% and provides an approximately 30% increase in CSI. A further benefit of the U-Net is its computational efficiency, running 1200% faster than the region-based VDA on equivalent CPU, and an impressive 9800% faster on GPU (which is not an option for the region-based VDA). Also of note is the trade-off between skill and computation time for different sn. Larger models also require additional memory to run, which may limit the ability to use these models on certain hardware.
Performance of the U-Net VDA for different network sizes (controlled by the hyperparameter sn) on the Level-III test set. Results from the region-based VDA implemented in the Py-ART library are also shown as a baseline. The U-Net VDA scores high across all metrics and shows small increases as the number of parameters in the model increases. The run times of different VDAs are shown, illustrating the computational speed of the DL approach in comparison with the baseline. For each metric, the arrows indicate the direction of increasing performance and the boldface values indicate the best performance.
Note that comparisons of skill scores between the U-Net and region-based VDAs reported in Table 4 should be interpreted with some care, as the U-Net has the advantage of being trained to emulate the same algorithm used to create the test set. Since the ORPG VDA utilizes environmental wind data, it is reasonable to assume that its output will be more accurate, on average, than the region-based baseline that was not supplied this additional information. However, it is possible that the U-Net learned and reproduced the same mistakes as the ORPG, leading to overly optimistic skill scores in this case. To explore this possibility further, several case studies where the ORPG and region-based VDAs disagreed were examined by human experts to judge which are correct. In virtually all cases that were examined, results similar to the example shown in Fig. 7 were observed. Three consecutive frames of Level-II velocity (left column) are dealiased by the U-Net and region-based algorithms (second and third columns). In the second row, without providing environmental wind information, the region-based algorithm produces incorrect results for isolated regions in the north and northeast portions of the scan, which disagrees with both the U-Net and Level-III reference (right column), which provide consistent results across all frames.
Case studies
This section considers two case studies composed of tornadic storms that were simultaneously sampled by multiple radar systems. As such, each radar provides a unique perspective of the storm velocity structures, affected by its unique system’s aliasing characteristics. Application of U-Net VDA with sn = 16 in these cases provides intercomparison between the resulting dealiased wind fields and highlights the portability of this algorithm across S-, C-, and X-band weather radar systems. Note that no additional training of the U-Net was performed prior to applying the algorithm to these other radars.
The first case is an EF5 tornado that impacted Moore, Oklahoma, on 20 May 2013, as observed by the University of Oklahoma PX-1000 research radar and the Oklahoma City TDWR (Kurdzo et al. 2015). The selected PX-1000 time is 2001:44 UTC, and the TDWR time is 2002:00 UTC. At these times, the differential velocity within the tornado via PX-1000 was approximately 65 m s−1 (Kurdzo et al. 2015) and the tornado was producing EF4 damage in western Newcastle (Burgess et al. 2014; Atkins et al. 2014). When compared with the manual dealiasing in Figs. 2a and 2b in Kurdzo et al. (2015), the U-Net VDA accurately dealiases the broadscale mesocyclone (Fig. 8). The VDA on TDWR data also performs qualitatively well in the mesocyclone near the rear-flank gust front where there are strong inbound radial velocities. The VDA on both radars also accurately dealiases the inflow to the supercell (areas of outbounds/red colors northwest of each radar).
Within the tornadic core circulation, the VDA has difficulty dealiasing velocities properly. There are several reasons for these difficulties. First, it should be noted that dealiasing in high-shear environments such as tornadoes is exceptionally difficult, as recently described by Feldmann et al. (2020). This is particularly true at high frequencies due to the lower Nyquist velocity. Additionally, when the Nyquist velocity is low, the wide spectrum width in tornadoes makes it difficult for the spectral estimator to determine an accurate velocity (Yu et al. 2007). In Fig. 8, a qualitative assessment shows that the VDA applied to PX-1000 is not able to dealias velocities within the core circulation, specifically in the region of outbounds (red colors). This is likely due to the highest area of spectrum width being in that region but may also be due to the existence of three folds, which this iteration of the VDA is not capable of unfolding. The VDA applied to TDWR performs almost identically, with the area in the rear flank (inbounds/blue colors) dealiased properly, but the core circulation dealiased improperly. In the case of the TDWR, the core circulation is improperly dealiased as inbounds rather than outbounds.
It is important to note that, for the TDWR, the velocity estimates were properly dealiased in the raw data because of the dual–pulse repetitition time (PRT) VDA on the operational TDWR radar signal processor (Cho 2010). This dual-PRT approach had an aliasing velocity over 40 m s−1. To produce the images in Fig. 8, an artificial Nyquist velocity of 16.5 m s−1 was applied to the raw velocity data to simulate aliased velocities.
The second case is a long-lived supercell that moved across northern Illinois and Indiana on the evening of 13 June 2022, producing two brief EF0 tornadoes in the Chicago, Illinois, suburbs (https://www.weather.gov/lot/2022jun13). The posttornadic storm is shown in Fig. 9 as it passed through an area located along approximately the same radial direction from two nearby radars, a TDWR (TORD in Romeoville, Illinois; Fig. 9, top row) and a WSR-88D (KLOT in Romeoville; Fig. 9, bottom row), producing comparable radial velocity perspectives. Across both systems, the U-Net VDA performs qualitatively well, with two minor regions of errors when dealiasing the TDWR at ranges close to the radar (Fig. 9, top row). The WSR-88D is dealiased without any apparent errors (Fig. 9, bottom row).
6. Discussion, future directions, and conclusions
The results in section 5 demonstrate that DL has the potential to be an effective algorithmic tool for weather radar product development. A major challenge for supervised learning approaches like the one taken here is that they require large quantities of “truth” in order to train. In many cases (particularly with weather radar), large quantities of truth are not available, so if DL techniques are to be widely adopted in this space, alternative sources and approaches need to be leveraged.
In the case of dealiasing, this issue was overcome by utilizing an operational algorithm as truth, namely, the Level-III velocity product dealiased by the ORPG VDA. As a result, the DNN trained here can be viewed as a kind of ML “surrogate” to an already operational algorithm that is capable of generating very similar output (CSIs > 95% and RMSEs near 1.5 m s−1) in a fraction of the run time, and without access to environmental wind information. These faster implementations enabled by the DL approach could be leveraged in real-time settings where traditional dealiasing techniques fail to keep up with certain rapid scanning strategies. For example, if real-time dealiasing were desired for an extremely rapidly updating research radar [e.g., as fast as 2-s updates such as presented in Pazmany et al. (2013)], this implementation of a VDA would be capable of keeping up in real time with relatively modest hardware. The DL models may also be useful in research settings by allowing massive velocity datasets to be dealiased in minutes, rather than hours or days using traditional approaches. Last, more complex VDAs such as higher-dimensional techniques could be trained using DL in the future in order to utilize their approaches in real time.
Another advantage of the ORPG surrogates is portability. Modern DL frameworks and deployment frameworks simplify the importing and application of trained models to new datasets without the need for additional software installations and dependencies. In addition, the ORPG code is written exclusively for the WSR-88D network, with little ability to port algorithms to other types of radar (i.e., different vendors, frequencies, beamwidths, etc.). This flexibility provided by the U-Net allows for simpler application of an ORPG-like VDA to new radar types without the need for the significant data reformatting required by the traditional ORPG algorithm. Improved performance on other radar types was also gained by training with Nyquist velocities that were lower than that commonly set by WSR-88D operators. This enabled impressive performance on other unseen radars (TDWR; PX-1000) despite the fact that the model was not trained directly on these systems. The prospect of porting ORPG-like algorithms to other radar systems may provide benefits to weather radar networks outside of the United States where these algorithms are not currently available. Further performance improvements on non-WSR-88D data can potentially be enabled using transfer learning by fine-tuning the network developed here.
Note that the VDA is far from the only ORPG product that is computationally complex. As of Build 20, the ORPG maintains a suite of over 150 products, which must all be run in real time and keep up with the fastest scan rates. This plethora of algorithms may need to be run in a much more expeditious manner in the future with faster scan rates, such as those discussed in Kollias et al. (2022) and Palmer et al. (2022). Additionally, other radar systems may not have access to the hardware available for the ORPG. For example, research radars may want to implement WSR-88D algorithms in real time but do not have the computational capabilities.
Given the results of this work, a natural question to ask is how this approach can be leveraged for other weather radar product enhancements. A whole suite of Level-III radar products has been successfully developed using traditional signal, image, and video processing techniques, and these methods have their constraints and weaknesses that can lead to difficulty in both implementation and operations. Additionally, like any algorithm, many of the ORPG algorithms suffer from false alarms, missed detections, and/or relatively noisy interpretations of untruthed data [e.g., the hydrometeor classification algorithm (HCA) (Park et al. 2009)]. However, ORPG algorithms, in many cases, are considered state of the art, and are desirable tools to have available to other radar systems. One way to explore this possibility is to implement other ML-based surrogates to ORPG algorithms. ML-based versions of the ORPG algorithms could result in several advantages, ranging from faster compute times to (potentially) improved accuracy through continued fine-tuning of ML surrogates using additional datasets.
Overall, in this work, a DL approach was developed for two-dimensional velocity dealiasing of weather radar data. To construct a training dataset, archives of Level-II and Level-III WSR-88D radial velocity products were combined to create a large dataset of over 100 000 dealiased velocity PPIs. To create aliased/dealiased pair samples for training, these scans were artificially folded using a known Nyquist velocity.
A U-Net model was trained using this dataset to output the correct fold number at each point in an input velocity image. The U-Net model developed here utilized several convolutional layers of different sizes to ensure it properly learned both global and local structure in the input image. Once trained, the U-Net model generated similar results to the operational ORPG algorithm from which was trained based on a test set of held-out WSR-88D data. Encouragingly, the model also performed reasonably well on radars that were not included in the training dataset, including the TDWR and PX-1000 radars.
Future work will continue to explore potential DL applications in radar meteorology, including the development of other ORPG surrogate algorithms. These techniques have the potential to improve existing radar products and to enable ORPG product generation on non-WSR-88D radars.
Acknowledgments.
The authors thank Mike Donovan, Betty Bennett, and Mike Istok for helpful discussions about the WSR-88D ORPG. The authors also thank the anonymous reviewers for their valuable insights and suggestions. This material is based upon work supported by the Department of the Air Force under Air Force Contract FA8702-15-D-0001. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the Department of the Air Force.
Data availability statement.
Trained model weights, sample data, and code for running the U-Net VDA are available at GitHub (https://github.com/mit-ll/unet-vda). Source codes for ORPG are available online (https://www.weather.gov/code88d/). WSR-88D data used in this work are publicly available from AWS Open Data Registry (https://registry.opendata.aws/noaa-nexrad/) and Google Cloud Storage (https://cloud.google.com/storage/docs/public-datasets/nexrad). TDWR data are publicly available from the National Center for Environmental Information (https://www.ncei.noaa.gov/products/radar/terminal-doppler-weather-radar). The PX-1000 (https://arrc.ou.edu/radar_px1000.html) is maintained and operated by the Advanced Radar Research Center of the University of Oklahoma, and data can be requested by contacting data@arrc.ou.edu.
REFERENCES
Agrawal, S., L. Barrington, C. Bromberg, J. Burge, C. Gazen, and J. Hickey, 2019: Machine learning for precipitation nowcasting from radar images. arXiv, 1912.12132v1, https://doi.org/10.48550/arXiv.1912.12132.
Alford, A. A., M. I. Biggerstaff, and G. D. Carrie, 2019: Mobile ground-based SMART radar observations and wind retrievals during the landfall of Hurricane Harvey (2017). Geosci. Data J., 6, 205–213, https://doi.org/10.1002/gdj3.82.
Atkins, N. T., K. M. Butler, K. R. Flynn, and R. M. Wakimoto, 2014: An integrated damage, visual, and radar analysis of the 2013 Moore, Oklahoma, EF5 tornado. Bull. Amer. Meteor. Soc., 95, 1549–1561, https://doi.org/10.1175/BAMS-D-14-00033.1.
Biggerstaff, M. I., A. A. Alford, G. D. Carrie, and J. A. Stevenson, 2021: Hurricane Florence (2018): Long duration single- and dual-Doppler observations and wind retrievals during landfall. Geosci. Data J., 9, 273–287, https://doi.org/10.1002/gdj3.137.
Bisong, E., 2019: Tensorflow 2.0 and keras. Building Machine Learning and Deep Learning Models on Google Cloud Platform, Springer, 347–399.
Bluestein, H. B., M. M. French, R. L. Tanamachi, S. Frasier, K. Hardwick, F. Junyent, and A. L. Pazmany, 2007: Close-range observations of tornadoes in supercells made with a dual-polarization, X-band, mobile Doppler radar. Mon. Wea. Rev., 135, 1522–1543, https://doi.org/10.1175/MWR3349.1.
Brown, R. A., L. R. Lemon, and D. W. Burgess, 1978: Tornado detection by pulsed Doppler radar. Mon. Wea. Rev., 106, 29–38, https://doi.org/10.1175/1520-0493(1978)106<0029:TDBPDR>2.0.CO;2.
Burgess, D. W., and L. R. Lemon, 1990: Severe thunderstorm detection by radar. Radar in Meteorology, Springer, 619–647.
Burgess, D. W., M. A. Magsig, J. Wurman, D. C. Dowell, and Y. Richardson, 2002: Radar observations of the 3 May 1999 Oklahoma City tornado. Wea. Forecasting, 17, 456–471, https://doi.org/10.1175/1520-0434(2002)017<0456:ROOTMO>2.0.CO;2.
Burgess, D., and Coauthors, 2014: 20 May 2013 Moore, Oklahoma, tornado: Damage survey and analysis. Wea. Forecasting, 29, 1229–1237, https://doi.org/10.1175/WAF-D-14-00039.1.
Chandrasekar, V., 2020: AI in weather radars. 2020 IEEE Radar Conf. (RadarConf20), Florence, Italy, Institute of Electrical and Electronics Engineers, 1–3, https://doi.org/10.1109/RadarConf2043947.2020.9266442.
Chen, H., V. Chandrasekar, R. Cifelli, and P. Xie, 2020: A machine learning system for precipitation estimation using satellite and ground radar network observations. IEEE Trans. Geosci. Remote Sens., 58, 982–994, https://doi.org/10.1109/TGRS.2019.2942280.
Cheong, B. L., R. Kelley, R. D. Palmer, Y. Zhang, M. Yeary, and T.-Y. Yu, 2013: PX-1000: A solid-state polarimetric X-band weather radar and time-frequency multiplexed waveform for blind range mitigation. IEEE Trans. Instrum. Meas., 62, 3064–3072, https://doi.org/10.1109/TIM.2013.2270046.
Cho, J. Y. N., 2010: Signal processing algorithms for the Terminal Doppler Weather Radar: Build 2. MIT Lincoln Laboratory Tech. Rep., Project Rep. ATC-363, 92 pp.
Cho, J. Y. N., and M. E. Weber, 2010: Terminal Doppler Weather Radar enhancements. 2010 IEEE Radar Conf., Arlington, VA, Institute of Electrical and Electronics Engineers, 1245–1249, https://doi.org/10.1109/RADAR.2010.5494427.
Collis, S., and Coauthors, 2018: It’s time for color vision deficiency friendly color maps in the radar community. 10th European Conf. on Radar in Meteorology and Hydrology, Utrecht, Netherlands, Wageningen University, https://ntrs.nasa.gov/api/citations/20180004634/downloads/20180004634.pdf.
Doviak, R. J., and D. S. Zrnić, 1993: Doppler Radar and Weather Observations. Dover, 562 pp.
Eilts, M., and S. Smith, 1990: Efficient dealiasing of Doppler velocities using local environment constraints. J. Atmos. Oceanic Technol., 7, 118–128, https://doi.org/10.1175/1520-0426(1990)007<0118:EDODVU>2.0.CO;2.
Evans, J., and D. Turnbull, 1989: Development of an automated windshear detection system using Doppler weather radar. Proc. IEEE, 77, 1661–1673, https://doi.org/10.1109/5.47729.
Feldmann, M., C. N. James, M. Boscacci, D. Leuenberger, M. Gabella, U. Germann, D. Wolfensberger, and A. Berne, 2020: R2D2: A region-based recursive Doppler dealiasing algorithm for operational weather radar. J. Atmos. Oceanic Technol., 37, 2341–2356, https://doi.org/10.1175/JTECH-D-20-0054.1.
Haque, I. R. I., and J. Neubert, 2020: Deep learning approaches to biomedical image segmentation. Inf. Med. Unlocked, 18, 100297, https://doi.org/10.1016/j.imu.2020.100297.
He, G., G. Li, X. Zou, and P. S. Ray, 2012: A velocity dealiasing scheme for synthetic C-band data from China’s new generation weather radar system (CINRAD). J. Atmos. Oceanic Technol., 29, 1263–1274, https://doi.org/10.1175/JTECH-D-12-00010.1.
Helmus, J. J., and S. M. Collis, 2016: The Python ARM Radar Toolkit (Py-ART), a library for working with weather radar data in the Python programming language. J. Open Res. Software, 4, e25, https://doi.org/10.5334/jors.119.
Hilburn, K. A., I. Ebert-Uphoff, and S. D. Miller, 2021: Development and interpretation of a neural-network-based synthetic radar reflectivity estimator using GOES-R satellite observations. J. Appl. Meteor. Climatol., 60, 3–21, https://doi.org/10.1175/JAMC-D-20-0084.1.
Hong, Y., and J. J. Gourley, 2014: Radar Hydrology: Principles, Models, and Applications. CRC Press, 176 pp.
Husnoo, N., T. Darlington, S. Torres, and D. Warde, 2021: A neural network quality-control scheme for improved quantitative precipitation estimation accuracy on the U.K. weather radar network. J. Atmos. Oceanic Technol., 38, 1157–1172, https://doi.org/10.1175/JTECH-D-20-0120.1
Iglovikov, V., and A. Shvets, 2018: Ternausnet: U-Net with VGG11 encoder pre-trained on ImageNet for image segmentation. arXiv, 1801.05746v1, https://doi.org/10.48550/arXiv.1801.05746.
Ioffe, S., and C. Szegedy, 2015: Batch normalization: Accelerating deep network training by reducing internal covariate shift. Proc. 32nd Int. Conf. on Machine Learning, Lille, France, PMLR, 448–456, http://static.googleusercontent.com/media/research.google.com/en//pubs/archive/43442.pdf.
Isola, P., J.-Y. Zhu, T. Zhou, and A. A. Efros, 2017: Image-to-image translation with conditional adversarial networks. Proc. IEEE Conf. Computer Vision and Pattern Recognition, Honolulu, HI, Institute of Electrical and Electronics Engineers, 1125–1134, https://doi.org/10.1109/CVPR.2017.632.
James, C. N., and R. A. Houze, 2001: A real-time four-dimensional Doppler dealiasing scheme. J. Atmos. Oceanic Technol., 18, 1674–1683, https://doi.org/10.1175/1520-0426(2001)018<1674:ARTFDD>2.0.CO;2.
Jatau, P., V. Melnikov, and T.-Y. Yu, 2021: A machine learning approach for classifying bird and insect radar echoes with S-band polarimetric weather radar. J. Atmos. Oceanic Technol., 38, 1797–1812, https://doi.org/10.1175/JTECH-D-20-0180.1
Jing, Z., and G. Wiener, 1993: Two-dimensional dealiasing of Doppler velocities. J. Atmos. Oceanic Technol., 10, 798–808, https://doi.org/10.1175/1520-0426(1993)010<0798:TDDODV>2.0.CO;2.
Kepner, J., and Coauthors, 2017: LLGrid: Supercomputer for sensor processing. Contemporary High Performance Computing, J. S. Vetter, Ed., Chapman and Hall/CRC, 637–645.
Kingma, D. P., and J. Ba, 2014: Adam: A method for stochastic optimization. arXiv, 1412.6980v9, https://doi.org/10.48550/arXiv.1412.6980.
Kollias, P., and Coauthors, 2022: Science applications of phased array radars. Bull. Amer. Meteor. Soc., 103, E2370–E2390, https://doi.org/10.1175/BAMS-D-21-0173.1.
Krinitskiy, M., P. Verezemskaya, K. Grashchenkov, N. Tilinina, S. Gulev, and M. Lazzara, 2018: Deep convolutional neural networks capabilities for binary classification of polar mesocyclones in satellite mosaics. Atmosphere, 9, 426, https://doi.org/10.3390/atmos9110426.
Kristovich, D. A. R., N. F. Laird, and M. R. Hjelmfelt, 2003: Convective evolution across Lake Michigan during a widespread lake-effect snow event. Mon. Wea. Rev., 131, 643–655, https://doi.org/10.1175/1520-0493(2003)131<0643:CEALMD>2.0.CO;2.
Kurdzo, J. M., D. J. Bodine, B. L. Cheong, and R. D. Palmer, 2015: High-temporal resolution polarimetric X-band Doppler radar observations of the 20 May 2013 Moore, Oklahoma, tornado. Mon. Wea. Rev., 143, 2711–2735, https://doi.org/10.1175/MWR-D-14-00357.1.
Kurdzo, J. M., B. J. Bennett, M. S. Veillette, D. J. Smalley, E. R. Williams, and M. F. Donovan, 2017a: WSR-88D chaff detection and characterization using an optimized hydrometeor classification algorithm. 18th Conf. on Aviation, Range, and Aerospace Meteorology, Lexington, MA, Amer. Meteor. Soc., J2.5, https://ams.confex.com/ams/97Annual/webprogram/Paper301430.html.
Kurdzo, J. M., and Coauthors, 2017b: Observations of severe local storms and tornadoes with the atmospheric imaging radar. Bull. Amer. Meteor. Soc., 98, 915–935, https://doi.org/10.1175/BAMS-D-15-00266.1.
Louf, V., A. Protat, R. C. Jackson, S. M. Collis, and J. Helmus, 2020: UNRAVEL: A robust modular velocity dealiasing technique for Doppler radar. J. Atmos. Oceanic Technol., 37, 741–758, https://doi.org/10.1175/JTECH-D-19-0020.1.
Michelson, M., W. Shrader, and J. Wieler, 1990: Terminal Doppler weather radar. Microw. J., 33, 139–148.
Mitchell, E. D. W., S. V. Vasiloff, G. J. Stumpf, A. Witt, M. D. Eilts, J. T. Johnson, and K. W. Thomas, 1998: The National Severe Storms Laboratory tornado detection algorithm. Wea. Forecasting, 13, 352–366, https://doi.org/10.1175/1520-0434(1998)013<0352:TNSSLT>2.0.CO;2.
NWS, 2021: WSR-88D CODE: The Common Operations and Development Environment. NOAA, accessed November 2022, https://www.weather.gov/code88d/.
Ortega, K. L., J. M. Krause, and A. V. Ryzhkov, 2016: Polarimetric radar characteristics of melting hail. Part III: Validation of the algorithm for hail size discrimination. J. Appl. Meteor. Climatol., 55, 829–848, https://doi.org/10.1175/JAMC-D-15-0203.1.
Oye, R. C., K. Mueller, and S. Smith, 1995: Software for radar translation, editing, and interpolation. 27th Conf. on Radar Meteorology, Vail, CO, Amer. Meteor. Soc., 359–361.
Palmer, R., and Coauthors, 2022: A primer on phased array radar technology for the atmospheric sciences. Bull. Amer. Meteor. Soc., 103, E2391–E2416, https://doi.org/10.1175/BAMS-D-21-0172.1.
Park, H. S., A. V. Ryzhkov, D. S. Zrnić, and K.-E. Kim, 2009: The hydrometeor classification algorithm for the polarimetric WSR-88D: Description and application to an MCS. Wea. Forecasting, 24, 730–748, https://doi.org/10.1175/2008WAF2222205.1.
Pazmany, A. L., J. B. Mead, H. B. Bluestein, J. C. Snyder, and J. B. Houser, 2013: A mobile rapid-scanning X-band polarimetric (RaXPol) Doppler radar system. J. Atmos. Oceanic Technol., 30, 1398–1413, https://doi.org/10.1175/JTECH-D-12-00166.1.
Peng, X., Q. Li, and J. Jing, 2022: CNGAT: A graph neural network model for radar quantitative precipitation estimation. IEEE Trans. Geosci. Remote Sens., 60, 1–14, https://doi.org/10.1109/TGRS.2021.3120218.
Racah, E., C. Beckham, T. Maharaj, S. Kahou, Prabhat, and C. J. Pal, 2017: Extreme weather: A large-scale climate dataset for semi-supervised detection, localization, and understanding of extreme weather events. NIPS’17: Proc. 31st Int. Conf. on Neural Information Processing Systems, Long Beach, CA, Association for Computing Machinery, 3405–3416, https://dl.acm.org/doi/10.5555/3294996.3295099.
Ravuri, S., and Coauthors, 2021: Skillful precipitation nowcasting using deep generative models of radar. Nature, 597, 672–677, https://doi.org/10.1038/s41586-021-03854-z.
Ray, P. S., and C. Ziegler, 1977: De-aliasing first-moment Doppler estimates. J. Appl. Meteor., 16, 563–564, https://doi.org/10.1175/1520-0450(1977)016<0563:DAFMDE>2.0.CO;2.
Ronneberger, O., P. Fischer, and T. Brox, 2015: U-Net: Convolutional networks for biomedical image segmentation. Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, N. Navab et al., Eds., Lecture Notes in Computer Science, Vol. 9351, Springer, 234–241, https://doi.org/10.1007/978-3-319-24574-4_28.
Samsi, S., M. Jones, and M. M. Veillette, 2020: Compute, time and energy characterization of encoder-decoder networks with automatic mixed precision training. 2020 IEEE High Performance Extreme Computing Conf. (HPEC), Waltham, MA, Institute of Electrical and Electronics Engineers, 1–6, https://doi.org/10.1109/HPEC43674.2020.9286241.
Skolnik, M., 2002: Introduction to Radar Systems. 3rd ed. McGraw-Hill, 786 pp.
Stensrud, D. J., and Coauthors, 2009: Convective-scale warn-on-forecast system: A vision for 2020. Bull. Amer. Meteor. Soc., 90, 1487–1500, https://doi.org/10.1175/2009BAMS2795.1.
Stumpf, G. J., A. Witt, E. D. Mitchell, P. L. Spencer, J. T. Johnson, M. D. Eilts, K. W. Thomas, and D. W. Burgess, 1998: The National Severe Storms Laboratory mesocyclone detection algorithm for the WSR-88D. Wea. Forecasting, 13, 304–326, https://doi.org/10.1175/1520-0434(1998)013<0304:TNSSLM>2.0.CO;2.
Veillette, M. S., E. P. Hassey, C. J. Mattioli, H. Iskenderian, and P. M. Lamey, 2018: Creating synthetic radar imagery using convolutional neural networks. J. Atmos. Oceanic Technol., 35, 2323–2338, https://doi.org/10.1175/JTECH-D-18-0010.1.
Wang, C., G. Tang, and P. Gentine, 2021: PrecipGAN: Merging microwave and infrared data for satellite precipitation estimation using generative adversarial network. Geophys. Res. Lett., 48, e2020GL092032, https://doi.org/10.1029/2020GL092032.
Weber, M., and Coauthors, 2021: Towards the next generation operational meteorological radar. Bull. Amer. Meteor. Soc., 102, E1357–E1383, https://doi.org/10.1175/BAMS-D-20-0067.1.
Weyn, J. A., D. R. Durran, R. Caruana, and N. Cresswell-Clay, 2021: Sub-seasonal forecasting with a large ensemble of deep-learning weather prediction models. J. Adv. Model. Earth Syst., 13, e2021MS002502, https://doi.org/10.1029/2021MS002502.
Witt, A., R. A. Brown, and Z. Jing, 2009: Performance of a new velocity dealiasing algorithm for the WSR-88D. 34th Conf. on Radar Meteorology, Williamsburg, VA, Amer. Meteor. Soc., P4.8, https://ams.confex.com/ams/34Radar/techprogram/paper_155951.htm.
Wurman, J., J. M. Straka, and E. N. Rasmussen, 1996: Fine-scale Doppler radar observations of tornadoes. Science, 272, 1774–1777, https://doi.org/10.1126/science.272.5269.1774.
Xing, D., J. Hou, C. Huang, and W. Zhang, 2022: Spatiotemporal reconstruction of MODIS normalized difference snow index products using U-Net with partial convolutions. Remote Sens., 14, 1795, https://doi.org/10.3390/rs14081795.
Xingjian, S., Z. Chen, H. Wang, D.-Y. Yeung, W.-K. Wong, and W.-c. Woo, 2015: Convolutional LSTM network: A machine learning approach for precipitation nowcasting. NIPS’15: Proc. 28th Int. Conf. on Neural Information Processing Systems, Montreal, QC, Canada, Association for Computing Machinery, 802–810, https://dl.acm.org/doi/10.5555/2969239.2969329.
Xu, Q., K. Nai, L. Wei, P. Zhang, S. Liu, and D. Parrish, 2011: A VAD-based dealiasing method for radar velocity data quality control. J. Atmos. Oceanic Technol., 28, 50–62, https://doi.org/10.1175/2010JTECHA1444.1.
Yang, J., K. Zhao, G. Zhang, G. Chen, H. Huang, and H. Chen, 2019: A Bayesian hydrometeor classification algorithm for C-band polarimetric radar. Remote Sens., 11, 1884, https://doi.org/10.3390/rs11161884.
Yo, T., S. Su, J. Chu, C. Chang, and H. Kuo, 2021: A deep learning approach to radar-based QPE. Earth Space Sci., 8, e2020EA001340, https://doi.org/10.1029/2020EA001340.
Yu, T.-Y., Y. Wang, A. Shapiro, M. B. Yeary, D. S. Zrnić, and R. J. Doviak, 2007: Characterization of tornado spectral signatures using higher-order spectra. J. Atmos. Oceanic Technol., 24, 1997–2013, https://doi.org/10.1175/2007JTECHA934.1.
Zhang, J., and S. Wang, 2006: An automated 2D multipass Doppler radar velocity dealiasing scheme. J. Atmos. Oceanic Technol., 23, 1239–1248, https://doi.org/10.1175/JTECH1910.1.
Zittel, W. D., Z. Jing, and N. Langlieb, 2011: A two-dimensional velocity dealiasing algorithm for the WSR-88D. 65th Interdepartmental Hurricane Conf., Miami, FL, Radar Operations Center, https://www.roc.noaa.gov/WSR88d/PublicDocs/AppsDocs/BriefIHC_Presentation_rev1.pdf.