1. Introduction
For convection-permitting numerical weather prediction systems, assimilating remotely sensed observation networks (e.g., radar and cloudy radiance measurements) is required to depict mesoscale weather features accurately (e.g., Vukicevic et al. 2004; Stengel et al. 2009; Privé et al. 2013). It is well known that strongly nonlinear model dynamics and observation operators, however, can induce bias in Gaussian-based data assimilation methods that are commonly used for numerical weather prediction (e.g., Bocquet et al. 2010). Since ensemble Kalman filters (EnKFs; Evensen 1994; Houtekamer and Mitchell 1998; Evensen and van Leeuwen 2000) approximate prior densities using a Gaussian and solve a linear system of equations to adjust a sample of model states to fit the posterior mean and covariance, strongly nonlinear model dynamics and measurement operators can lead to bias, which impedes achieving accurate convection-permitting initial conditions for next-generation weather forecast models. This limitation is apparent for multiscale weather prediction systems that exhibit large uncertainty in smaller scales, or when observations are sensitive to cloud processes (e.g., Poterjoy et al. 2017; Poterjoy 2022a). As a result, most infrared satellite data assimilation studies mainly focus on clear-sky observations (e.g., Errico et al. 2007; Fabry and Sun 2010; Minamide and Zhang 2017; Honda et al. 2018). Therefore, developing new data assimilation methods that mitigate Gaussian assumptions is an active area of research.
One strategy, which has gained momentum in recent years, is to apply dimension-reduction procedures (viz., localization) to particle filters (PFs; Penny and Miyoshi 2016; Poterjoy and Anderson 2016; Poterjoy et al. 2017, 2019; Potthast et al. 2019). PFs avoid the parametric estimation of Bayesian posterior densities, thus providing great flexibility for solving a range of complex data assimilation problems. These methods, however, are more sensitive to sampling errors than EnKFs. As such, computational limitations pose a major obstacle, which has limited research examining the potential of PFs for operational weather prediction. Incorporating statistics from a large number of high-resolution ensemble members into the data assimilation step is one of the most effective ways to mitigate the effects of sampling errors, but this strategy is not often tractable.
Given the challenges discussed above, a natural progression is to combine PFs with methods that rely on parametric density estimates when appropriate. Several papers have proposed to hybridize PFs with EnKFs (Stordal et al. 2011; Frei and Kunsch 2013; Slivinski et al. 2015; Chustagulprom et al. 2016; Grooms and Robinson 2021) and with variational methods (Morzfeld et al. 2018). These methods are remarkably accurate for cases of “moderate nonlinearity,” which are characteristic of situations with a non-Gaussian priors but Gaussian-like posterior distribution (Metref et al. 2014; Morzfeld and Hodyss 2019; Grooms and Robinson 2021). For example, Frei and Kunsch (2013) introduced a procedure that makes a continuous transition between the ensemble and the particle filter update by factoring the likelihood. They choose a “splitting factor” to ensure that an effective ensemble size is maintained within a certain tolerance of a user-specified threshold. Based on this approach, Chustagulprom et al. (2016) developed a method to hybridize the general linear ensemble transform filter (LETF) framework and PFs, which can use observation-space localization and avoid linear assumptions for observation operators. Grooms and Robinson (2021) also introduced a filter that combines PFs with EnKFs, which is generally similar to Frei and Kunsch (2013) and Chustagulprom et al. (2016) in that it factors the likelihood. This method, like others, is effective for problems characterized by medium nonlinearity in model dynamics or measurement operators. In these papers, the value of the splitting factor is determined adaptively by the effective ensemble size. This choice is a heuristic one, thus motivating additional research into how to optimally combine PFs with EnKFs. For example, Nerger (2022) propose a method for estimating hybrid coefficients that is based not only the effective ensemble size but also the kurtosis and skewness of the ensemble. This method still requires the tuning of hyper-parameters, but generates more accurate filter estimates than using effective ensemble size alone. As mentioned in Chustagulprom et al. (2016), a more powerful and computationally feasible alternative is to adopt Kullback–Leibler divergence (KL divergence; Kullback and Leibler 1951) as a means of identifying proper choices of prior error distribution. This approach is one of the most frequently used objective functions to measure deviations from Gaussianity in forecast error distribution for weather models (Kondo and Miyoshi 2019; Li et al. 2019; Ruiz et al. 2021; Pimentel and Qranfal 2021). However, it is difficult to measure non-Gaussianity by the KL divergence for high-dimensional systems when the ensemble size is small or when a strange attractor makes numerical convergence and proper definition of the continuous limit complicated (Bocquet et al. 2010).
In this study, we introduce a novel approach to forming an adaptive PF–EnKF data assimilation methodology, which exploits the theoretical strengths of nonparametric (PF) and parametric (EnKF) filters. For this purpose, we use a recently proposed PF (Poterjoy 2022a,b), which introduces an iterative strategy for PFs that match posterior moments. For this PF, iterations improve the filter’s ability to draw samples from non-Gaussian posterior densities despite fitting a limited number of moments. The iterations follow from a factorization of particle weights, which also provide a straightforward means of combining PFs with EnKFs to reduce the impact of sampling errors. To achieve the adaptive mixed methodology at each data assimilation cycle, we repeat the iterative PF update while the prior sample distribution is non-Gaussian, and update with EnKF when Gaussian distributions for marginal quantities are detected. Here, we introduce a statistical hypothesis testing approach to determine when to make this transition between filter algorithms. Several papers on data assimilation use statistical hypothesis tests for normality to measure the difference between the prior or posterior distribution and the normal distribution (e.g., Bocquet et al. 2010; Poterjoy 2016). However, research of incorporating a normality test directly into assimilation processes is unexplored.
The current study first compares the power of several hypothesis tests by performing Monte Carlo simulations of data generated from choice distributions that are often used to characterize errors. Then we examine the newly developed mixed hybrid methodology employing the Shapiro–Wilk test (Shapiro and Wilk 1965), which has outstanding power among omnibus tests for detecting departures from normality (e.g., Srivastava and Hui 1987; Mendes and Pala 2003; Farrell et al. 2007; Villaseñor and González-Estrada 2009). Using the hypothesis testing approach allows us to accurately detect Gaussianity, even with small ensemble sizes, and explore possibilities other than effective ensemble size and the KL divergence for determining the splitting factor adaptively.
The manuscript is organized in the following manner. In section 2, we briefly review the four well-known normality tests and compare the power of the tests. Section 3 introduces a statistical hypothesis testing approach to forming an adaptive PF–EnKF hybrid. We discuss the results and findings of numerical experiments conducted using low-dimensional toy models in section 4. The last section discusses major findings from this study.
2. Power comparisons of four normality tests
The assumption of a normal distribution is often an underlying premise of many academic fields and studies, including data assimilation. When the assumption of normality is violated, interpretations and inferences may lack reliability and validity. There are three commonly used procedures for evaluating whether a random independent sample comes from a normal population: graphical methods (histograms, box plots, Q–Q plots), moment-based methods (skewness and kurtosis), and formal normality tests. In particular, a significant amount of normality tests have been proposed, and several studies have already compared their power (e.g., Dufour et al. 1998; Thadewald and Büning 2007; Razali and Wah 2011; Saculinggan and Balase 2013). In this section, we compare the power of four well-known formal tests of normality: Shapiro–Wilk test (SWT; Shapiro and Wilk 1965), Kolmogorov–Smirnov test (Kolmogorov 1933; Smirnov 1939), Lilliefors test (Lilliefors 1967), and Anderson–Darling test (Anderson and Darling 1954). The following subsections briefly review the four normality tests, and describe the simulation procedure.
a. Methodology for the four hypothesis tests
1) Shapiro–Wilk test
The null hypothesis of SWT is that the data originate from a normally distributed population. Small values of W lead to the rejection of the null hypothesis, where 0 ≤ W ≤ 1. The original SWT was limited to a sample size of 50 or less. Royston (1982) extended SWT to large samples and provided an approximation of the test statistic W, and Royston (1983) suggested a test for multivariate normality based on SWT. Royston (1992) arrived at an improved approximation to the weights, which allows SWT to effectively detect departures from multivariate normality for smaller sample sizes. Last, Royston (1995) introduced the FORTRAN algorithm AS R94, which is used in the current study. The algorithm includes a scaling process that sets the mean of the sample to zero and a centering process that sets the variance of the sample to one. Therefore, the algorithm allows the results to generalize not only to the standard normal distribution, but also to a normal distribution with a nonzero mean and nonunit variance.
2) Kolmogorov–Smirnov test
3) Lilliefors test
4) Anderson–Darling test
b. Simulation procedures
Monte Carlo simulations are the most commonly used approaches to compare and evaluate the accuracy of a hypothesis test in detecting the degree of contamination by outliers and the test power with respect to sample size. Following previous studies, we use Monte Carlo simulations to evaluate the power of SWT, Kolmogorov–Smirnov test, Lilliefors test, and Anderson–Darling test statistics in testing whether a random sample of n independent observations is obtained from a population with a normal N(μ, σ2) distribution. Several papers have already shown the superiority of SWT over the other tests (e.g., Mendes and Pala 2003; Razali and Wah 2011). Thus, the simulation in this section focuses on a larger selection of distributions than previous studies, which are motivated by the diverse shapes of error distributions found for geophysical models and observing systems, and aims to reconfirm the superiority of SWT over the other tests.
Classification of cases by skewness and kurtosis of the distribution.
Definitions of terminologies in a statistical test.
In Table 2, the probability of a type I error occurring is denoted by α and the probability of a type II error by β. The probability of both the type I and type II errors should be small. However, it is impossible to make both small because the risk ratios α and β are in a trade-off relationship. Generally speaking (in society), committing a type I error is often a more serious problem. Moreover, as discussed in detail in section 3, in the current study, we prefer erring on the side of a lower type I error in order to avoid the use of an EnKF when the distribution is truly non-Gaussian. This is because the PF update can provide adequate estimates for both non-Gaussian or Gaussian prior distributions; the same cannot be said about the EnKF. Therefore, the correct procedure for hypothesis testing is to determine the acceptable risk rate α in advance, and then select the hypothesis test method with the highest test power 1 − β among them. Hence, in the current study, we focus on the type I error for N(0, 1) and on the test power for the other distributions (cases B–E in Table 1); We will not look at the type II error and specificity in this study.
c. Results
Figure 1 shows the variation of the type I error (Fig. 1a) and test power (Figs. 1b–i) with the sample sizes n for the four tests for each distribution when α = 0.05. When the distribution is N(0, 1), all four tests generally achieved α = 0.05. In the case of the symmetric distributions (
The overall results show that in all cases the power of SWT is superior to the other tests for small sample sizes, which is the regime of interest for ensemble data assimilation applied to weather models. This result is generally consistent with the results of many other previous studies mentioned before. Therefore, this paper uses SWT to detect Gaussianity in the new hybrid method hereafter. Note that “failing to reject the null hypothesis” is not the same as “accepting the null hypothesis.” In such a case, it is still not exactly clear whether the null or alternative hypothesis is correct. For simplicity, this study interprets the null hypothesis to be that “the samples are from a population that follows a normal distribution.” If the null hypothesis of SWT is not rejected then the samples are assumed to come from a Gaussian distribution.
3. Implementation with the local particle filter
This section presents the mathematical framework for implementing the adaptive PF–EnKF hybrid method by embedding SWT, the most powerful statistical test in the previous section, into the local PF. For the purposes of the adaptive hybrid methodology, the current study uses the recently proposed PF by Poterjoy (2022b, hereafter P22), which introduces an iterative strategy. We briefly describe the parts of PF that are relevant to the implementation of the proposed method.
Poterjoy et al. (2019) introduced several filter stabilization strategies in the PF of Poterjoy (2016) to avoid particle degeneracy. In particular, regularization and tempering are effective methods when sampling errors are large, and the sample size is small. Regularization is equivalent to increasing the particle weights up to a power β by inflating the observation error variance. This regularization allows the particles to acquire a specific “effective sample size Neff,” and is particularly helpful in stabilizing the filter when all particles are far from an observation. The regularization provides temporary iterations for the local PF, which is a posterior tempering strategy. This iterative approach also improves the filter’s ability to sample from non-Gaussian posterior densities, even though it fits a limited number of moments. The iterations consist of a factorization of particle weights, thus providing a natural framework for combining the local PF with alternative filters to reduce the impact of sampling error. When provided with Gaussian likelihoods, a partial update performed by a PF can adjust particles to more closely resemble samples from a Gaussian, even if the prior exhibits a complex nn-Gaussian shape. The resulting intermediate update then makes the EnKF an appropriate choice for the remaining update (Grooms and Robinson 2021). P22 introduced the hybrid parameter κ, which is an Nx-dimensional vector that determines when to switch from a PF update to a parametric filter update. For the iterative PF, the hybrid parameter κ and “the target effective sample size
In the case that SWT does not detect Gaussianity during Nk iterations at the jth grid point, κj and ηj become 1 and 0, respectively, and no EnKF update is performed at the grid point. Similarly, if SWT detects Gaussianity in the first iteration at the jth grid point, then κj = 0 and ηj = 1, and no PF update is performed at the grid point. Thus, in situations where the posterior is clearly non-Gaussian, the filter can have the option of retaining the local PF. The hybrid approach aims to obtain an intermediate distribution that is closer to Gaussian than the prior distribution by the PF updates, and to make this intermediate distribution closer to Gaussian by the EnKF. In cases that we cannot obtain an intermediate distribution closer to Gaussian, we can perform the iterative PF updates alone, without using EnKF in the last step, which is the strength of the adaptive strategy. The advantage of using the EnKF in the last step if a Gaussian is encountered during iterations is purely due to it being a more robust choice when ensemble sizes are small (and the distribution is indeed Gaussian).
As in P22, the
Since the computational cost of SWT is not expensive, the adaptive approach introduced in the current study, which incorporates the statistical test into the local PF introduced by P22, is generally less computationally expensive than the iterative LPF. Under most circumstances, the hybrid requires fewer iterations, thus leading to a cost saving. Nevertheless, the PF introduced by P22 is computationally more costly than pure EnKF because of the use of regularization and tempering. For more information, please refer to Poterjoy (2022a) and P22.
Note that κ and η are uniquely specified for each observation-space prior variable as well. In this case, they are Ny-dimensional vectors and we again use SWT to determine when each element of observation-space forecast ensembles may follow a Gaussian distribution. The κ and η defined for the observation-space are used for the observation-space filter updates.
Algorithm 1 Adaptive mixed PF–EnKF update with SWT
- 1: function pf_enkf_hybrid
- 2: k = 1
- 3:
κ = 0.0
Vector with Nx dimensions - 4: κ_residual = 1.0 – κ
- 5:
while max(κ_residual) > 0 do
Tempering - 6: for j = 1: Nx do
- 7: if κ_residual (j) > 0 then
- 8:
⊳ Shapiro-Wilk Test - 9: if SWT_result = Gaussian then
- 10: ηj = κ_residual(j)
- 11: κ_residual(j) = 0.0
- 12: end if
- 13: end if
- 14: end for
- 15: (βk, κ_residual) ← Regularization(κ_residual)
- 16: for i = 1: Ny do
- 17:
xa ← The Local PF (
) The Local PF core - 18: xf ← xa
- 19: end for
- 20: k = k + 1
- 21: end while
- 22: xa ← EnKF_tempered(xf, yo, η) ⊳ EnKF as the last adjustment
- 23: end function
Algorithm 2 Serial EnSRF update with inflated
- 1: function EnKF_tempered(xf, yo, η)
- 2: for I = 1: Ny do
- 3:
- 4:
- 5:
- 6: for j = 1: Nx do
- 7:
- 8: end for
- 9:
- 10:
- 11:
- 12:
- 13:
- 14: xf ← xa
- 15: end for
- 16: end function
4. Numerical experiments with low-order models
This section explores the behavior of the newly developed method through numerical simulations. In the first experiment, we use a simple univariate problem to illustrate the difference of the adaptive hybrid method between iterative EnKF and bootstrap PF using tempering. In the second experiment, we use the 40-variable dynamic model of Lorenz (1996) to compare the advantages of the adaptive method over EnKF, the local PF, and hybrid PF–EnKF with fixed values of κ and η. These experiments use simulated measurements to target several scenarios, such as varying spatial density, highly nonlinear dynamics, mixed measurement operators, and unresolved model error. The last experiment uses an idealized kinematic vortex, which was used in Poterjoy (2022a) to replicate findings from real-data applications. The kinematic vortex model allows us to generate observations that emulate realistic observations for an application containing large spatial dependence across variables, while retaining great flexibility in our construction of data assimilation experiments. Among the common parameters used for idealized data assimilation applications, such as observation error variance, observation locations, and ensemble size, these experiments contain parameters that indirectly control the shape of the full multivariate prior, thus allowing for an analysis of the adaptive hybrid technique under controllable conditions.
a. Univariate application
Consider the example shown in Fig. 2, where 104 prior members are updated using an observation whose value is 4 and observation error standard deviation is set to σy = 0.8. Among the 104 prior members, three quarters are selected from N(−4, 1.22), while the rest are from N(3.5, 1.22), whose mean value is smaller than the observation. Therefore, the prior ensemble is a bimodal distribution. Using the same prior for all three filters, Figs. 2a–c show the posterior distribution after the first iteration. In all cases, we can see that each filter shifts the ensemble toward the observation; however, the EnKF–MDA inherits the bimodal distribution of the prior for the posterior distribution, while the IPF and IPF–EnKF correctly retain a single mode.1 The bimodal posterior distributions in the EnKF–MDA are not relieved by the completion of all iterations (Fig. 2j). In the IPF, after all the updates, the posterior pdf is relatively close to the likelihood of the observation, but exhibits negative skewness because numerous particles remain in the leftmost mode (Fig. 2k). In the adaptive IPF–EnKF case, SWT detected Gaussianity in the distribution of the ensembles after three iterations of the bootstrap PF (Fig. 2i), and then EnKF was performed using R inflated by 8/5, which is the inverse of the remaining observation error variance (Fig. 2l). As a result, the IPF–EnKF posterior is close to the IPF, indicating that the hybrid method correctly transitioned to the partial EnKF step once a Gaussian distribution was detected. Furthermore, we emphasize that the univariate application is presented for illustration only, as the IPF–EnKF is not expected to provide benefits over the IPF when the ensemble size is large.
b. 40-variable dynamical system
1) Experimental designs
Experiments include three forms of measurement operator H as in Kurosawa and Poterjoy (2021): “Linear Case,” “Nonlinear Case 1,” and “Nonlinear Case 2” use
Configuration of cycling data assimilation experiments.
All experiments in this section use an observation frequency of 6 h. Observations are assimilated over a 10-yr period, and root-mean-square errors (RMSEs) from the last 9 years are used to quantify the accuracy of the posterior analyses, ignoring the first year spin up period. In this set of experiments, we perform 100 parallel trials out of an abundance of caution. For localization, we use the fifth-order correlation function controlled by a radius of influence (ROI) given by Gaspari and Cohn (1999). For posterior inflation, the current study adopts the strategy known as relaxation to prior perturbation (RTPP; Zhang et al. 2004) after EnKF update. Similar to the α used in the relaxation method, for the local PF, we use a mixing parameter γ to maintain particle diversity during updates in (8). When the ensemble size is small, this parameter works to prevent filter divergence. γ is a scalar between 0 and 1, and acts to increase diversity in particles without modifying prior or posterior error variance. Each time the particles in state space are updated, the prior particles are mixed with the resampled particles (P22). The target Neff is fixed at
2) Results
We summarize the results for the “normal” setting in Fig. 3, and the mean values of κ in the adaptive experiments with this setting are shown later (see Fig. 8a). In Linear Case (Fig. 3a), when the number of members is small, the higher the ratio of the EnKF, or the closer the value of κ is to zero, the lower the RMSE. However, with 40 members, the experiment that performs partial local PF is optimal (κ = 0.3); after 40 members, the performance of the EnKF hardly improves as the number of members increases, and in the experiments with 300 members, pure EnKF shows the worst score. Here, since the sampling error decreases as the number of particles increases, it would seem that using more of the local PF update would give better results. However, for this particular model and a linear observation operator, this is not the case. Even with Ne = 300, the best performing experiments use a factorization that amounts to 70% of the EnKF increment being used. The experiment that determines the value of κ adaptively shows less optimal but suitable results for smaller ensemble sizes—with the added benefit of not needing to be tuned. As the number of members increases, however, the sample size for SWT increases, thus making the test more accurate. Increasing the ensemble size also increases the rejection rate of the null hypothesis, which is a desirable property. While the mean value of κ becomes larger as the ensemble size increases, the value converges slowly to 0.2 (Fig. 8a). The L96 priors remain close to Gaussian for most data assimilation cycles when using a sufficiently dense network of observations with linear measurement operators, thus leading us to conclude that SWT operates appropriately for this application.
Results from Nonlinear Case 1 are shown in Fig. 3b. With Ne = 20–40, the trend is the same as in the Linear Case: the experiments with a more significant percentage of the EnKF show better scores. However, this feature is maintained even with large ensemble sizes. This is because of the precision and frequency of the observations compared to ones in Linear Case, as described in Kurosawa and Poterjoy (2021). Since model variables are around the magnitude of O(10), the nonlinear operator
Figure 3c shows the mean RMSEs from experiments that use measurements simulated with Nonlinear Case 2. In experiments using this observation network, a situation occurs in which the nonlinearity in the application becomes much larger than the sampling error in the prior and posterior distributions estimated by the ensemble. Owing to the strong nonlinearity of the observations, experiments using mainly the Gaussian-based method struggle to provide an accurate RMSE. In particular, the pure EnKF diverges, even with Ne = 300. The mean value of κ in the adaptive experiment shows that most of the update is used for the local PF (Fig. 8a). This result occurs as the strongly nonlinear observation operator tends to induce skewness in prior distributions, and SWT frequently rejects the null hypothesis.
Based on the above results from the “normal” setting, results from “mixed” observation networks yield intuitive results (Fig. 4). For example, in the case where the observation operators
We summarize the results for the “data void” setting in Fig. 5. Mean RMSEs are uniformly higher than the “normal” setting with any observation operators despite using the smaller localization scale. Notably, in Nonlinear Case 2, several fixed experiments diverged, but the experiments with the appropriate blend of the EnKF and the local PF (κ = 0.3–1.0) are stable (Fig. 5c). The mean value of κ used in the adaptive experiment is close to 1 (Fig. 8b). We note that this experiment shows slight advantages over experiments that keep κ fixed near 1, thus underscoring the importance of allowing κ to change over space and time.
Last, results obtained from simulated “model error” experiments show elevated errors for all experiments, regardless of observation operators in Fig. 6. The presence of model errors means that the prior variance can be quite large, which leads to more frequent non-Gaussian prior distributions for L96. The hybrid strategies with a value of κ close to 0.5 show clear advantages in this regime. The parametric (Gaussian) assumption that follows the PF steps in hybrid configurations allows the filter to more easily adjust solutions for observations that lie outside the span of the ensemble. Hence, it shifts particles closer to observations in a manner that is not permitted by the PF—for variables that are detected to have Gaussian errors.
To investigate the behavior of SWT specification of κ for these simulations, we examine a sample time series of prior ensemble variance and estimated κ for the experiment using a linear measurement operator (Fig. 7). The plotted values come from the first variable of the L96 model in the first trial of the experiments with ensemble size Ne = 300. Because of an imperfect model, the prior variance fluctuates significantly over the entire period. When the prior distribution has a larger variance, the nonlinear model dynamics can more readily produce non-Gaussian priors, which SWT successfully detects. For this example, the only factor that can contribute to non-Gaussian priors is the nonlinear model itself as the measurement operator is linear. Hence, fluctuations of the ensemble variance and κ are highly correlated. In Fig. 7, the time- and space-average value of κ over the period is 0.2655, which is very close to the value 0.2766 in Fig. 8b (note that the value of κ in Fig. 7 is from the first variable in the L96 model, while in Fig. 8 is from the average of all variables in the model). As in the “data void” simulations, the experiments with adaptively estimated κ again show improvements over experiments with values of κ that are configured to use close to the average mean estimated κ, but fixed over space and time (Fig. 6a). In general, we find that choosing κ adaptively is beneficial in “model error” experiments, owing to its ability to maintain filter stability without rigorous tuning. The sporadic non-Gaussian priors produced by L96 in “model error” experiments introduce a major challenge that mimics the expected behavior of real weather systems.
Based on the above results, the statistical hypothesis testing approach yields adequate hybrid factor estimates in all situations we examined for this study. Moreover, the approach has significant value for more realistic applications, such as nonhomogeneous observation networks and unknown model process error. We expect similar benefits for geophysical problems that are characterized by a variety of dynamic instabilities as well. Furthermore, the proposed adaptive hybrid method avoids the need to tune heuristic parameters, such as the hybrid factor, which we find to be sensitive to observation operators, observation density, and model process uncertainty.
c. Idealized vortex model
In contrast to the low-dimensional applications used in the previous subsections, realistic atmospheric forecast models have several variables at each grid point, such as air temperature, winds, pressure, and specific humidity. These variables also exhibit large spatial error dependence with one another, which is not accounted for in adaptive choices for κ. As such, an observation of one variable must be used to update all collocated and nearby variables; for EnKFs, this step considers prior error covariance across each variable. Furthermore, extending κ to be the same dimension as the full state vector, rather than the grid dimension, would bring additional algorithmic complexity to the proposed hybrid filter. To address the problem of collocated variables for estimating κ, a natural choice is to perform the hypothesis test using all variables that are expected to be correlated with that variable at a grid point, i.e., by performing a test for multivariate normality. The numerical experiments performed in this section serve the purpose of illustrating the advantages of optimally adjusted κ estimated via SWT extended to the test for multivariate normality proposed by Royston (1983). Considering marginal PDFs in the test for multivariate normality is expected to provide a reasonable κ for data assimilation updates that account for correlations between collocated variables, which is an important practical feature of the proposed hybrid method in real weather applications, as the transition between the local PF and EnKF updates are decided across grid points rather than state variables. The method, however, still neglects dependence for variables at different grid points, which is one theoretical shortcoming.
The current section provides an illustrative comparison of the EnKF, the local PF, and hybrid updates using a low dimensional application that mimics a common challenge for filtering geophysical flow, namely, the problem of assimilating measurements for mesoscale weather features that are not well constrained by measurements.
1) Experimental designs
The current study generates vortices on a two-dimensional Cartesian region consisting of 91 × 91 equally spaced grid points. To generate a prior sample, we first designate a control state with the center of the vortex located at (
Observations are generated by uniformly selecting points from within the scan area of a hypothetical Doppler radar with a radius of 30 grids, placed at coordinates (iradar = 25, jradar = 25) in the lower-left corner of the domain. We produce each observation by projecting the truth state wind in the direction of the hypothetical radar beam pointing outward from the radar. In this experiment, the errors added to each observation are drawn from
As in the experiment with the L96 in section 4b, out of an abundance of caution, we perform 3000 parallel trials with unique sets of priors, true solutions, and observations in order to capture the range of plausible outcomes for this application. To perform each trial, for the truth state, we generate the center (
2) Results
This subsection discusses results obtained by performing single-cycle data assimilation using the described sets of observations and prior members. Figure 11 shows the posterior RMSEs for four experiments as examples: EnKF (κ = 0.0), the local PF (κ = 1.0), PF–EnKF with κ = 0.5, and PF–EnKF with adaptive κ estimation. For experiments that use σp = 0.0 for the prior, each data assimilation method shows low RMSEs that are visibly similar (first row of Fig. 11). This finding is expected because the small σp = 0.0 leads to the Gaussian assumption being valid (Poterjoy 2022a). However, as the value of σp increases, each experiment yields vast differences in the upper right portion of the domain where each filter must infer wind estimates from distant observations. This application is especially problematic for the EnKF, as linear updates do not properly capture nonlinear dependence in winds across the vortex (Poterjoy 2022a). While the local PF and PF–EnKF produce smaller mean RMSEs than the EnKF, we note that these errors also continue to decrease as the number of members increases, because of the decrease in sampling error. Furthermore, compared to the EnKF and local PF experiments, both hybrid experiments show more accurate results across all domains, which demonstrates that the hybrid PF–EnKF method is effective at shifting particles into an approximate Gaussian before applying the EnKF step—even for the highly non-Gaussian vortex application discussed in Poterjoy (2022a).
Comparing experiments with fixed and adaptive κ, Fig. 12 shows the grid points over the domain where a specified value of κ produced the smallest RMSEs for each choice of prior. For all settings, the adaptive estimate yields the smallest errors outside of the vortex, thus reflecting diversity in optimal κ in this region. When σp = 0, experiments with a value of κ fixed near 0.5 show the best results near the vortex center (Figs. 12a–c). We suspect this result occurs because the location of the prior and true vortex centers are the same for all prior distributions when σp = 0, but since R is drawn from N(RCTRL, 1), the winds exhibit bimodal behavior, which is controlled by parameter R in (14); recall, this parameter divides the domain into regions of zero and nonzero—but constant—vorticity. Since the region near the vortex center is characterized by the presence of both zeros and nonzeros, it is conceivable that the case of κ = 0.5, where both PF and EnKF can be used in a balanced manner, happens to be the most optimal. Therefore, a fixed value of κ can be identified via rigorous tuning, rather than resorting to hypothesis testing. When σp > 0 and a sufficiently large ensemble size is used (e.g., Ne = 300), SWT correctly identifies values for κ that outperform fixed values for κ over most of the domain (Figs. 12f,i,l). Prior vortices are no longer collocated as σp increases, so the region where fixed values for κ are optimal gradually extends outward from the center.
We also examine the mean value of κ (averaged over trials) in the experiments where κ is adaptively adjusted (Fig. 13). First, for the case of σp = 0.0, the area close to the center of the control vortex, where experiments with a value of κ fixed near 0.5 corresponds to the location where estimated κ is about 0.2–0.5 in the first row of Fig. 13. This indicates that the area has a low percentage of the local PF updates compared to the fixed experiment. For the cases of Ne = 40 and 100, the areas where the experiment using the adaptively adjusted κ is inferior in Fig. 12 are generally estimated to have a value of κ less than 0.5 (the first two columns of Fig. 13). However, in the case with Ne = 300, the values of κ in those locations are generally more than 0.5, and the difference from the fixed experiment is not significant (third column of Fig. 13). This may be because the larger sample size used in SWT leads to more frequent rejection of the null hypothesis, resulting in more iterations of the local PF. Furthermore, the experiment with estimated κ is generally more stable in the other areas, far away from the center, especially in the upper-right domains, where there are no observations.
As in the previous subsections, the kinematic vortex experiments illustrate the advantage of specifying κ adaptively versus keeping this parameter fixed. In this demonstration, however, κ is estimated using u and υ with SWT extended to handle multivariate normality. Hence, results from this section differ in that we successfully analyzed samples from multivariate probability distributions. The multivariate approach is more practical for data assimilation with real atmospheric models, when we need to consider correlation among collocated variables. Furthermore, as noted by Poterjoy (2022a), the non-Gaussian data assimilation problem constructed in this section has univariate marginal distributions that are close to Gaussian, but multivariate marginals for variables across grid points that are far from Gaussian. In terms of computing burden, high-resolution models, such as those used for weather forecasting, are constrained by ensemble size. As a result, detecting such characteristics with a limited number of ensembles to be utilized for operational models is extremely challenging. Nevertheless, we find the proposed SWT approach to be sufficient for identifying deviations from a multivariate normality for collocated winds, which shows added value over a rigorously tuned hybrid methodology that uses fixed specifications for κ.
5. Discussion and conclusions
The current study introduces a novel approach to forming an adaptive hybrid data assimilation method that mixes the theoretical strengths and flexibility of particle filters with Gaussian-based ensemble Kalman filters (EnKFs), which are more resilient to bias in sample-estimated prior uncertainty. For this purpose, we use a recently proposed PF by Poterjoy (2022b), which introduces a regularization and tempering methodology to improve filter performance when sampling error is large. The tempering step consists of a factorization of the particle weights, which provides a natural framework for combining local PFs with alternative filters to mitigate the effects of sampling error. In addition to identifying portions of the state space where a PF may provide more accurate marginal posterior estimates than an EnKF, the adaptive strategy can switch between filters partway through data assimilation steps. The latter property is beneficial when Gaussian assumptions are appropriate for posteriors but not for priors, which is common when likelihoods are Gaussian. In this case, partial updates performed by the PF can adjust the distribution of particles to more closely fit a Gaussian, which allows for a more effective use of EnKFs. To determine the timing of the transition between these filter updates, we use the Shapiro–Wilk test (SWT), which has excellent power among omnibus tests to detect deviations from normality. The use of SWT allows for accurate detection of Gaussianity even when the ensemble size is small. SWT also requires minimal computing time, thus permitting its use between PF iterations, which can be carried out until prior sample distributions for marginals at each grid point are detected to be Gaussian. Increasing the ensemble size also increases the rejection rate of the null hypothesis and leads to a smaller portion of updates being made by an EnKF, which is a desirable property.
To examine the performance of the adaptive hybrid, this study constructs numerous data assimilation experiments using a low-dimensional dynamical model, which is characterized by 40 equally spaced variables on a periodic domain. In general, the statistical hypothesis testing approach yields adequate estimates of the hybrid factor in all situations considered in this study. Given a homogeneous network of equally spaced observations, the adaptive formulations of the hybrid filter are as accurate as the rigorously tuned hybrid parameters. The adaptive approach also demonstrates clear advantages in experiments containing heterogeneous observation networks and unknown model process errors—in which case, the optimal choice of adaptive parameter varies temporally or across variables.
The study also examines practical challenges for adopting the new method for real Earth system models, which are characterized by multiple variables at common grid points and large error correlations through space; e.g., modern weather prediction models. The computational expense of such models limits the amount of tuning that can be performed for heuristic parameters used during data assimilation, which can be sensitive to observation operators, observation frequency, and model process uncertainty. Therefore, this study adopts an idealized kinematic vortex model to study the behavior of the adaptive hybrid. This model permits large error dependence across variables displaced over a two-dimensional domain, and contains two variables (zonal and meridional winds) at each grid point thus requiring a multivariate SWT to adaptively choose how to partition PF and EnKF updates. For this application, the hybrid factor is estimated using SWT extended to detect multivariate normality for ensembles of u and υ at each grid point. This approach allows the use of the appropriate factor to account for multivariate marginal distributions for updating the state variables, alongside observation-space priors. Specifying the hybrid factor for collocated variables also simplifies the algorithmic formulation of the adaptive methodology, as it only requires the factor to be specified for all grid points and observation-space priors used during data assimilation. The experiments reveal spatial patterns of adaptively chosen hybrid factors that result in large PF updates in portions of the state space where Gaussian assumptions are known to be incorrect, and are close to the values identified at each grid point from rigorously tuned experiments aimed at reducing posterior mean RMSEs. These results encourage further testing for real geophysical problems that are characterized by a variety of dynamic instabilities.
In summary, the proposed adaptive hybrid method performs well in idealized simulations that mimic data assimilation problems encountered for real geophysical modeling systems. Because the new strategy relies on statistical hypothesis testing, it becomes more stable when the ensemble size increases. The proposed method obviates the need for tuning a hybrid parameter that influences when an EnKF is preferred over PF, which can depend on a number of factors including the underlying model dynamics and observation network. This property of the method has theoretical benefits for real Earth system models where rigorous tuning of data assimilation parameters is not always feasible, and the shape of error distributions is flow dependent. Last, this study demonstrates how SWT can be extended to consider error dependence for collocated variables. Further research will explore the use of multivariate error dependence for variables across grid points, which may be needed for prior distributions that are characterized by strong nonlinear dependence for variables displaced geographically.
For the provided ensemble size, the last iteration of the IPF is an accurate estimate of the true Bayesian posterior.
Acknowledgments.
Funding for this work was provided by NOAA Grant NA20OAR4600281 and NSF/CAREER Award AGS1848363.
Data availability statement.
This paper uses numerous low-dimensional model simulations that can be easily replicated using software written for this study. The model code, compilation script, and the namelist settings are available at https://github.com/Kenta9638/MWR_2022.
REFERENCES
Acheson, D. J., 1990: Elementary Fluid Dynamics. Oxford University Press, 408 pp.
Althouse, L. A., W. B. Ware, and J. M. Ferron, 1998: Detecting departures from normality: A Monte Carlo simulation of a new omnibus test based on moments. Diversity and Citizenship in Multicultural Societies, San Diego, CA, American Educational Research Association, 33 pp., https://eric.ed.gov/?id=ED422385.
Anderson, T. W., and D. A. Darling, 1954: A test of goodness of fit. J. Amer. Stat. Assoc., 49, 765–769, https://doi.org/10.1080/01621459.1954.10501232.
Arshad, M., M. T. Rasool, and M. I. Ahmad, 2003: Anderson Darling and modified Anderson Darling tests for generalized Pareto distribution. J. Appl. Sci., 3, 85–88, https://doi.org/10.3923/jas.2003.85.88.
Bocquet, M., C. A. Pires, and L. Wu, 2010: Beyond Gaussian statistical modeling in geophysical data assimilation. Mon. Wea. Rev., 138, 2997–3023, https://doi.org/10.1175/2010MWR3164.1.
Chustagulprom, N., S. Reich, and M. Reinhardt, 2016: A hybrid ensemble transform particle filter for nonlinear and spatially extended dynamical systems. SIAM/ASA J. Uncertainty Quantif., 4, 592–608, https://doi.org/10.1137/15M1040967.
Cramér, H., 1928: On the composition of elementary errors. Scand. Actuar. J., 1928, 13–74, https://doi.org/10.1080/03461238.1928.10416862.
Dufour, J., A. Farhat, L. Gardiol, and L. Khalaf, 1998: Simulation-based finite sample normality tests in linear regressions. Econom. J., 1, C154–C173, https://doi.org/10.1111/1368-423X.11009.
Emerick, A. A., and A. C. Reynolds, 2012: History matching time-lapse seismic data using the ensemble Kalman filter with multiple data assimilations. Comput. Geosci., 16, 639–659, https://doi.org/10.1007/s10596-012-9275-5.
Errico, R. M., P. Bauer, and J.-F. Mahfouf, 2007: Issues regarding the assimilation of cloud and precipitation data. J. Atmos. Sci., 64, 3785–3798, https://doi.org/10.1175/2006JAS2044.1.
Evensen, G., 1994: Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. J. Geophys. Res., 99, 10 143–10 162, https://doi.org/10.1029/94JC00572.
Evensen, G., and P. J. van Leeuwen, 2000: An ensemble Kalman smoother for non-linear dynamics. Mon. Wea. Rev., 128, 1852–1867, https://doi.org/10.1175/1520-0493(2000)128<1852:AEKSFN>2.0.CO;2.
Fabry, F., and J. Sun, 2010: For how long should what data be assimilated for the mesoscale forecasting of convection and why? Part I: On the propagation of initial condition errors and their implications for data assimilation. Mon. Wea. Rev., 138, 242–255, https://doi.org/10.1175/2009MWR2883.1.
Farrell, P. J., M. Salibian-Barrera, and K. Naczk, 2007: On tests for multivariate normality and associated simulation studies. J. Stat. Comput. Simul., 77, 1065–1080, https://doi.org/10.1080/10629360600878449.
Frei, M., and H. R. Kunsch, 2013: Bridging the ensemble Kalman and particle filters. Biometrika, 100, 781–800, https://doi.org/10.1093/biomet/ast020.
Gaspari, G., and S. E. Cohn, 1999: Construction of correlation functions in two and three dimensions. Quart. J. Roy. Meteor. Soc., 125, 723–757, https://doi.org/10.1002/qj.49712555417.
Grooms, I. G., and G. Robinson, 2021: A hybrid particle-ensemble Kalman filter for problems with medium nonlinearity. PLOS ONE, 16, e0248266, https://doi.org/10.1371/journal.pone.0248266.
Honda, T., and Coauthors, 2018: Assimilating all-sky Himawari-8 satellite infrared radiances: A case of Typhoon Soudelor (2015). Mon. Wea. Rev., 146, 213–229, https://doi.org/10.1175/MWR-D-16-0357.1.
Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126, 796–811, https://doi.org/10.1175/1520-0493(1998)126<0796:DAUAEK>2.0.CO;2.
Kolmogorov, A. N., 1933: Sulla determinazione empirica di una legge di distribuzione. Giorn. Instit. Ital. Attu., 4, 83–91.
Kondo, K., and T. Miyoshi, 2019: Non-Gaussian statistics in global atmospheric dynamics: A study with a 10 240-member ensemble Kalman filter using an intermediate atmospheric general circulation model. Nonlinear Processes Geophys., 26, 211–225, https://doi.org/10.5194/npg-26-211-2019.
Kullback, S., and R. A. Leibler, 1951: On information and sufficiency. Ann. Math. Stat., 22, 79–86, https://doi.org/10.1214/aoms/1177729694.
Kurosawa, K., and J. Poterjoy, 2021: Data assimilation challenges posed by nonlinear operators: A comparative study of ensemble and variational filters and smoothers. Mon. Wea. Rev., 149, 2369–2389, https://doi.org/10.1175/MWR-D-20-0368.1.
Li, R., N. Magbool Jan, B. Huang, and V. Prasad, 2019: Constrained multimodal ensemble Kalman filter based on Kullback–Leibler (Kl) divergence. J. Process Control, 79, 16–28, https://doi.org/10.1016/j.jprocont.2019.03.012.
Lilliefors, H. W., 1967: On the Kolmogorov-Smirnov test for normality with mean and variance unknown. J. Amer. Stat. Assoc., 62, 399–402, https://doi.org/10.1080/01621459.1967.10482916.
Lorenz, E. N., 1996: Predictability: A problem partly solved. Proc. Seminar on Predictability, Shinfield Park, Reading, United Kingdom, ECMWF, 1–18, https://www.ecmwf.int/node/10829.
Lorenz, E. N., and K. A. Emanuel, 1998: Optimal sites for supplementary weather observations: Simulation with a small model. J. Atmos. Sci., 55, 399–414, https://doi.org/10.1175/1520-0469(1998)055<0399:OSFSWO>2.0.CO;2.
Mendes, M., and A. Pala, 2003: Type I error rate and power of three normality tests. Inf. Technol. J., 2, 135–139, https://doi.org/10.3923/itj.2003.135.139.
Metref, S., E. Cosme, C. Snyder, and P. Brasseur, 2014: A non-Gaussian analysis scheme using rank histograms for ensemble data assimilation. Nonlinear Processes Geophys., 21, 869–885, https://doi.org/10.5194/npg-21-869-2014.
Minamide, M., and F. Zhang, 2017: Adaptive observation error inflation for assimilating all-sky satellite radiance. Mon. Wea. Rev., 145, 1063–1081, https://doi.org/10.1175/MWR-D-16-0257.1.
Morzfeld, M., and D. Hodyss, 2019: Gaussian approximations in filters and smoothers for data assimilation. Tellus, 71A, 1–27, https://doi.org/10.1080/16000870.2019.1600344.
Morzfeld, M., D. Hodyss, and J. Poterjoy, 2018: Variational particle smoothers and their localization. Quart. J. Roy. Meteor. Soc., 144, 806–825, https://doi.org/10.1002/qj.3256.
Nerger, L., 2022: Data assimilation for nonlinear systems with a hybrid nonlinear Kalman ensemble transform filter. Quart. J. Roy. Meteor. Soc., 148, 620–640, https://doi.org/10.1002/qj.4221.
Penny, S. G., and T. Miyoshi, 2016: A local particle filter for high dimensional geophysical systems. Nonlinear Processes Geophys., 23, 391–405, https://doi.org/10.5194/npg-23-391-2016.
Pimentel, S., and Y. Qranfal, 2021: A data assimilation framework that uses the Kullback-Leibler divergence. PLOS ONE, 16, e0256584, https://doi.org/10.1371/journal.pone.0256584.
Poterjoy, J., 2016: A localized particle filter for high-dimensional nonlinear systems. Mon. Wea. Rev., 144, 59–76, https://doi.org/10.1175/MWR-D-15-0163.1.
Poterjoy, J., 2022a: Implications of multivariate non-Gaussian data assimilation for multi-scale weather prediction. Mon. Wea. Rev., 150, 1475–1493, https://doi.org/10.1175/MWR-D-21-0228.1.
Poterjoy, J., 2022b : Regularization and tempering for a moment-matching localized particle filter. Quart. J. Roy. Meteor. Soc., 148, 2631–2651, https://doi.org/10.1002/qj.4328.
Poterjoy, J., and J. L. Anderson, 2016: Efficient assimilation of simulated observations in a high-dimensional geophysical system using a localized particle filter. Mon. Wea. Rev., 144, 2007–2020, https://doi.org/10.1175/MWR-D-15-0322.1.
Poterjoy, J., R. A. Sobash, and J. L. Anderson, 2017: Convective-scale data assimilation for the Weather Research and Forecasting Model using the local particle filter. Mon. Wea. Rev., 145, 1897–1918, https://doi.org/10.1175/MWR-D-16-0298.1.
Poterjoy, J., L. J. Wicker, and M. Buehner, 2019: Progress toward the application of a localized particle filter for numerical weather prediction. Mon. Wea. Rev., 147, 1107–1126, https://doi.org/10.1175/MWR-D-17-0344.1.
Potthast, R., A. Walter, and A. Rhodin, 2019: A localized adaptive particle filter within an operational NWP framework. Mon. Wea. Rev., 147, 345–362, https://doi.org/10.1175/MWR-D-18-0028.1.
Privé, N. C., Y. Xie, J. S. Woollen, S. E. Koch, R. Atlas, and R. E. Hood, 2013: Evaluation of the Earth Systems Research Laboratory’s Global Observing System Simulation Experiment system. Tellus, 65A, 19011, https://doi.org/10.3402/tellusa.v65i0.19011.
Razali, N., and Y. Wah, 2011: Power comparison of Shapiro–Wilk, Kolmogorov–Smirnov, Lilliefors and Anderson Darling tests. J. Stat. Model. Anal., 2, 21–33.
Royston, J. P., 1982: An extension of Shapiro and Wilk’s W test for normality to large samples. J. Roy. Stat. Soc., 31, 115–124, https://doi.org/10.2307/2347973.
Royston, J. P., 1983: Some techniques for assessing multivarate normality based on the Shapiro–Wilk. J. Roy. Stat. Soc., 32, 121–133, https://doi.org/10.2307/2347291.
Royston, J. P., 1992: Approximating the Shapiro–Wilk W-test for non-normality. Stat. Comput., 2, 117–119, https://doi.org/10.1007/BF01891203.
Royston, J. P., 1995: Remark AS R94: A remark on algorithm AS 181: The W-test for normality. J. Roy. Stat. Soc., 44, 547–551, https://doi.org/10.2307/2986146.
Ruiz, J., G.-Y. Lien, K. Kondo, S. Otsuka, and T. Miyoshi, 2021: Reduced non-Gaussianity by 30 s rapid update in convective-scale numerical weather prediction. Nonlinear Processes Geophys., 28, 615–626, https://doi.org/10.5194/npg-28-615-2021.
Saculinggan, M., and E. A. Balase, 2013: Empirical power comparison of goodness of fit tests for normality in the presence of outliers. J. Phys. Conf. Ser., 435, 012041, https://doi.org/10.1088/1742-6596/435/1/012041.
Shapiro, S. S., and M. B. Wilk, 1965: An analysis of variance test for normality (complete samples). Biometrika, 52, 591–611, https://doi.org/10.2307/2333709.
Slivinski, L., E. Spiller, A. Apte, and B. Sandstede, 2015: A hybrid particle-ensemble Kalman filter for Lagrangian data assimilation. Mon. Wea. Rev., 143, 195–211, https://doi.org/10.1175/MWR-D-14-00051.1.
Smirnov, N. V., 1939: Estimate of deviation between empirical distribution functions in two independent samples. Bull. Moscow Univ., 2, 3616.
Srivastava, M., and T. Hui, 1987: On assessing multivariate normality based on Shapiro–Wilk W statistic. Stat. Probab. Lett., 5, 15–18, https://doi.org/10.1016/0167-7152(87)90019-8.
Stengel, M., P. Unden, M. Lindskog, P. Dahlgren, N. Gustafsson, and R. Bennartz, 2009: Assimilation of SEVIRI infrared radiances with HIRLAM 4D-Var. Quart. J. Roy. Meteor. Soc., 135, 2100–2109, https://doi.org/10.1002/qj.501.
Stordal, A. S., H. A. Karlsen, G. Nævdal, H. J. Skaug, and B. Vallès, 2011: Bridging the ensemble Kalman filter and particle filters: The adaptive Gaussian mixture filter. Comput. Geosci., 15, 293–305, https://doi.org/10.1007/s10596-010-9207-1.
Thadewald, T., and H. Büning, 2007: Jarque–Bera test and its competitors for testing normality—A power comparison. J. Appl. Stat., 34, 87–105, https://doi.org/10.1080/02664760600994539.
Villaseñor Alva, J. A., and E. González-Estrada, 2009: A generalization of Shapiro–Wilk’s test for multivariate normality. Commun. Stat. Theory Methods, 38, 1870–1883, https://doi.org/10.1080/03610920802474465.
Vukicevic, T., T. Greenwald, M. Zupanski, D. Zupanski, T. V. Haar, and A. S. Jones, 2004: Mesoscale cloud state estimation from visible and infrared satellite radiances. Mon. Wea. Rev., 132, 3066–3077, https://doi.org/10.1175/MWR2837.1.
Whitaker, J. S., and T. M. Hamill, 2002: Ensemble data assimilation without perturbed observations. Mon. Wea. Rev., 130, 1913–1924, https://doi.org/10.1175/1520-0493(2002)130<1913:EDAWPO>2.0.CO;2.
Zhang, F., C. Snyder, and J. Sun, 2004: Impacts of initial estimate and observation availability on convective-scale data assimilation with an ensemble Kalman filter. Mon. Wea. Rev., 132, 1238–1253, https://doi.org/10.1175/1520-0493(2004)132<1238:IOIEAO>2.0.CO;2.