Ensemble Kalman Filtering with One-Step-Ahead Smoothing

Naila F. Raboudi King Abdullah University of Science and Technology, Thuwal, Saudi Arabia

Search for other papers by Naila F. Raboudi in
Current site
Google Scholar
PubMed
Close
,
Boujemaa Ait-El-Fquih King Abdullah University of Science and Technology, Thuwal, Saudi Arabia

Search for other papers by Boujemaa Ait-El-Fquih in
Current site
Google Scholar
PubMed
Close
, and
Ibrahim Hoteit King Abdullah University of Science and Technology, Thuwal, Saudi Arabia

Search for other papers by Ibrahim Hoteit in
Current site
Google Scholar
PubMed
Close
Full access

Abstract

The ensemble Kalman filter (EnKF) is widely used for sequential data assimilation. It operates as a succession of forecast and analysis steps. In realistic large-scale applications, EnKFs are implemented with small ensembles and poorly known model error statistics. This limits their representativeness of the background error covariances and, thus, their performance. This work explores the efficiency of the one-step-ahead (OSA) smoothing formulation of the Bayesian filtering problem to enhance the data assimilation performance of EnKFs. Filtering with OSA smoothing introduces an updated step with future observations, conditioning the ensemble sampling with more information. This should provide an improved background ensemble in the analysis step, which may help to mitigate the suboptimal character of EnKF-based methods. Here, the authors demonstrate the efficiency of a stochastic EnKF with OSA smoothing for state estimation. They then introduce a deterministic-like EnKF-OSA based on the singular evolutive interpolated ensemble Kalman (SEIK) filter. The authors show that the proposed SEIK-OSA outperforms both SEIK, as it efficiently exploits the data twice, and the stochastic EnKF-OSA, as it avoids observational error undersampling. They present extensive assimilation results from numerical experiments conducted with the Lorenz-96 model to demonstrate SEIK-OSA’s capabilities.

© 2018 American Meteorological Society. For information regarding reuse of this content and general copyright information, consult the AMS Copyright Policy (www.ametsoc.org/PUBSReuseLicenses).

Corresponding author: Ibrahim Hoteit, ibrahim.hoteit@kaust.edu.sa

Abstract

The ensemble Kalman filter (EnKF) is widely used for sequential data assimilation. It operates as a succession of forecast and analysis steps. In realistic large-scale applications, EnKFs are implemented with small ensembles and poorly known model error statistics. This limits their representativeness of the background error covariances and, thus, their performance. This work explores the efficiency of the one-step-ahead (OSA) smoothing formulation of the Bayesian filtering problem to enhance the data assimilation performance of EnKFs. Filtering with OSA smoothing introduces an updated step with future observations, conditioning the ensemble sampling with more information. This should provide an improved background ensemble in the analysis step, which may help to mitigate the suboptimal character of EnKF-based methods. Here, the authors demonstrate the efficiency of a stochastic EnKF with OSA smoothing for state estimation. They then introduce a deterministic-like EnKF-OSA based on the singular evolutive interpolated ensemble Kalman (SEIK) filter. The authors show that the proposed SEIK-OSA outperforms both SEIK, as it efficiently exploits the data twice, and the stochastic EnKF-OSA, as it avoids observational error undersampling. They present extensive assimilation results from numerical experiments conducted with the Lorenz-96 model to demonstrate SEIK-OSA’s capabilities.

© 2018 American Meteorological Society. For information regarding reuse of this content and general copyright information, consult the AMS Copyright Policy (www.ametsoc.org/PUBSReuseLicenses).

Corresponding author: Ibrahim Hoteit, ibrahim.hoteit@kaust.edu.sa

1. Introduction

Data assimilation (DA) is the process by which available observations are combined with dynamical models to constrain model predictions and determine the best possible estimate of a system’s state (Ghil and Malanotte-Rizzoli 1991). DA methods are generally grouped into two categories: 1) variational methods that aspire to fit the model’s trajectory to available observations through adjustments of a well-chosen set of control parameters (Le Dimet and Talagrand 1986) and 2) sequential methods that follow a probabilistic framework to split the Bayesian estimation problem into cycles of alternating forecast steps, to advance the distribution of the system state with the model, and analysis steps, to compute the posterior distribution based on incoming observations (Todling 1999).

The ensemble Kalman filter (EnKF) was introduced by Evensen (1994) as a Monte Carlo implementation of the Kalman filter (KF) (Kalman 1960) to tackle large-scale, sequential, nonlinear DA problems. It approximates the distribution of the system state by a collection of random realizations, or ensemble members, whose mean is the state estimate and sample covariance is the associated error covariance (Hoteit et al. 2015). The EnKF integrates the (analysis) ensemble members forward with the model in the forecast step and then updates them in the analysis step with incoming observations using a Kalman-like (best linear unbiased estimation) correction. Because of its reasonable computational cost, robust performance, and nonintrusive formulation, the EnKF has seen widespread use in many geophysical applications (Reichle et al. 2002; Carton and Giese 2008; Anderson et al. 2009; Hoteit et al. 2013; Nævdal et al. 2005; Triantafyllou et al. 2013).

The EnKF perturbs observations with random noise sampled from the observational error distribution before assimilation, such that its error covariance asymptotically matches that of the KF (Burgers et al. 1998). Later, this scheme became known as the stochastic EnKF (Tippett et al. 2003). This perturbation, however, introduces an additional source of sampling error and often underestimates the analysis error covariance, especially when the filter is implemented with small ensembles (Whitaker and Hamill 2002; Hoteit et al. 2015). Several deterministic EnKFs were proposed to update the ensemble’s sample mean and covariance without perturbing the observations, exactly as in the KF (Tippett et al. 2003), and then to resample a new ensemble to proceed with the next forecast step. Among these, we cite the ensemble transform Kalman filter (ETKF) (Bishop et al. 2001), the ensemble adjustment Kalman filter (EAKF) (Anderson 2001), and the ensemble square root Kalman filter (EnSRF) (Whitaker and Hamill 2002), which performs the ensemble resampling implicitly using different deterministic square root transformations. This method of resampling contrasts with the explicit stochastic resampling of the singular evolutive interpolated Kalman (SEIK) filter (Pham et al. 1998; Hoteit et al. 2002), which is based on orthogonal random transformations.

Accurate estimates of the forecast error covariances, also known as background covariances, are crucial to EnKF-like filters. In real-world, large-scale applications, however, EnKFs are implemented with relatively small ensembles to limit computational cost and with poor model error statistics (Song et al. 2010). This limits the representativeness of EnKFs’ backgrounds and, thus, the filters’ performance. Various auxiliary techniques have been proposed to mitigate this background limitation, such as inflation (Anderson 2001), localization (Houtekamer and Mitchell 1998), hybrid (Hamill and Snyder 2000) and adaptive formulations (Song et al. 2010), and robust filtering (Luo and Hoteit 2011). Here, we follow a different approach that exploits information in (future) observations to compute improved backgrounds in EnKFs. The idea is to follow the one-step-ahead (OSA) smoothing formulation of the Bayesian filtering problem (Desbouvries et al. 2011), which first applies a smoothing step with the future observation before integrating the resulting “smoothed” state with the model to compute the background, all within a fully Bayesian-consistent framework. This approach has been recently used by Ait-El-Fquih et al. (2016) to derive a Bayesian framework for state-parameter estimation for the well-known dual EnKF (Moradkhani et al. 2005). In the resulting EnKF-OSA scheme, the observation is exploited twice: first to constrain the sampling of the forecast ensemble with the future observation, and next to update the resulting ensemble with the data. This additional OSA-smoothing step provides an improved background for the next analysis cycle and is, thus, expected to enhance EnKFs’ results. Indeed, when implemented under challenging situations, such as strong nonlinearities, sparse observation networks, or large assimilation windows, the ensembles’ perturbations estimated by standard EnKFs may not well represent the uncertainties of the prior state, thereby limiting the impact of the observation. Conditioning the forecast ensemble by the observations through a smoothing step should provide more accurate estimates of the background uncertainties for better exploitation of the observation in the analysis step.

A very similar idea, known as “running in place” (RIP), was proposed in Kalnay and Yang (2010) as a heuristic way to accelerate the spinup of the ETKF. It iteratively applies the two-stage update step based on criteria that ensure a reasonable fit to the data. Using the observations more than once in each assimilation window during spinup is justified by the need to maximize the initial extraction of information (Kalnay and Yang 2010). The RIP iterations may be interpreted as some kind of iterations on the OSA-smoothing step to deal with the nonlinear character of this step—a very common approach in the subsurface community (Emerick and Reynolds 2013; Luo et al. 2015; Ma et al. 2017; Chen and Oliver 2013). This method was successfully tested with a series of realistic models (Penny et al. 2013; Yang et al. 2013). By ignoring the underlying criteria, we found that RIP with one iteration is very similar to the proposed OSA-smoothing-based filtering scheme. This means that RIP with one iteration is Bayesian consistent and, thus, not heuristic, as has been suggested.

The goal of our work is twofold. Realizing the benefit of an improved background covariance for state estimation with the EnKF, we first demonstrate the relevance of the OSA-smoothing-based approach for state estimation. The performance of EnKF-like methods is known to level off after enough members are used (about the rank of the covariance matrices) (e.g., Hoteit et al. 2002; Evensen 2003). The idea here is to better exploit the observations in order to enhance the (approximate) ensemble background using future information. We then derive a new deterministic OSA-smoothing-like ensemble filter, the SEIK-OSA, which is suitable for data assimilation into large-scale oceanic and atmospheric models. This should improve the filtering results, compared to the EnKF-OSA, when the ensemble size is small, compared to the number of assimilated observations, a common scenario in geophysical applications (Hoteit et al. 2015). We demonstrate the efficiency of the SEIK-OSA with extensive numerical experiments with the strongly nonlinear Lorenz-96 (L96) model under various assimilation settings and scenarios.

The remainder of this paper is organized as follows. Section 2 describes the concept of the filtering problem with OSA smoothing and presents a (stochastic) EnKF-based solution for general nonlinear systems. The SEIK-OSA algorithm is derived in section 3. Results of numerical experiments with the Lorenz-96 model are then analyzed in section 4. The main conclusions are summarized and discussed in section 5.

2. OSA-smoothing-based filtering

a. Problem formulation

Consider the following discrete-time dynamical system:
e1
where is the dimensional system state at time , is the corresponding observation, is a (nonlinear) operator integrating the system state from time to , and is the observation operator that projects from the state space onto the observation space. Here, we assume that is linear to simplify the presentation of the filtering algorithms. The case of a nonlinear could be treated as usual in EnKFs; see, for example, Liu et al. (2016). The model and the observation process noise, and , respectively, are assumed to be Gaussian with zero means and covariances and , respectively; η and are also assumed to be independent, jointly independent, and independent of the initial state .

The filtering problem consists of estimating, at each time , the state from the previous observations . The posterior mean (PM), which minimizes the mean squared error (MSE), is commonly taken as a standard solution for this problem (Ait-El-Fquih and Hoteit 2016). Classical filtering schemes involve a forecast step to move from the analysis estimate at time to the analysis estimate at the next time . The OSA-smoothing formulation of the filtering problem is based on the idea that the aforementioned standard filtering path, which computes the analysis state at a given time from the analysis state at the previous time, is not unique (Desbouvries and Ait-El-Fquih 2008; Desbouvries et al. 2011). In a linear Gaussian estimation problem, the same analysis estimate can, indeed, be also obtained if one includes a smoothing step between the two successive analyses. This formulation introduces a new update step to the state, based on future observations (smoothing), without violating the Bayesian filtering framework. The resulting OSA-smoothing-based filters are said to be Bayesian consistent, as they are derived, under only the common Gaussian assumption, from a theoretically sound generic algorithm that involves two Bayesian updates with the same data [see Eqs. (21) and (24) in Ait-El-Fquih et al. (2016)]. Unlike iterative algorithms, using the data twice does not affect in any way the Bayesian character of the OSA-smoothing-based filtering schemes, as these data are used to update two different quantities, and . Fig. 1 outlines the difference between the classical KF algorithm and the KF with OSA smoothing (called hereafter KF-OSA). Both schemes provide the same state analysis estimate in the linear Gaussian case. The OSA-smoothing-based formulation is, however, expected to enhance the filter’s performance when the filter is not implemented under ideal conditions, with the future observation providing more information to compute the background statistics.

Fig. 1.
Fig. 1.

Diagram depicting the KF and KF-OSA algorithms, where the subscript n stands for time, and are the forecast and analysis estimates computed by KF, and , , , and are, respectively, the forecast, smoothed, pseudoforecast, and analysis estimates computed by KF-OSA.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Desbouvries and Ait-El-Fquih (2008) and Desbouvries et al. (2011) derived KF and particle filter (PF)-like algorithms based on the OSA-smoothing formulation [see also Pitt and Shephard (1999) and Luo et al. (2014), who introduced an auxiliary PF involving OSA-smoothing, but in a different way]. Lee and Farmer (2014) also followed a similar strategy and presented a number of algorithms for estimating both the state and system noise. More recently, EnKF-OSA schemes were also proposed by Ait-El-Fquih et al. (2016) and Gharamti et al. (2015) to derive a dual-filter-like algorithm for state-parameter estimation of large-scale nonlinear systems. It is important to realize that the OSA-smoothing-based schemes are not fixed-lag smoothers (see, e.g., Simon 2006; Anderson and Moore 1979). The main difference between an OSA-smoothing-based scheme and a fixed-lag scheme is in the computation of the analysis distribution. The former computes it from the OSA-smoothing distribution, while the fixed-lag scheme computes it from the forecast distribution, following exactly the standard filtering update. The smoothing distribution/estimate is the final goal of the fixed-lag algorithms, while in our OSA-smoothing-based scheme, this distribution is only used to compute the analysis distribution, which is our final solution. Below, we present the EnKF-OSA algorithm for state estimation based on Ait-El-Fquih et al. (2016).

b. The (stochastic) EnKF-OSA algorithm

Let , , and denote the forecast, analysis, and OSA-smoothed ensembles, respectively. For the analysis ensemble (and, similarly, for the forecast and smoothing ones), let denote its empirical mean and the ensemble perturbation matrix, with the ith column defined as . Starting from an analysis ensemble at time , the analysis ensemble at the next time is obtained following two steps:

1) Smoothing step

The forecast ensemble and the corresponding observation forecast ensemble are first computed as
e2
e3
where , and . These are then used along with the observation to smooth the previous analysis ensemble with the “future” observation as
e4
where the cross covariances and are computed from the corresponding ensembles as
e5
e6
Here, is the observation forecast ensemble perturbation matrix, with the ith column defined as , with the observation forecast ensemble mean.

2) Analysis step

The resulting smoothed ensemble is then integrated forward in time with the model to compute the pseudoforecast ensemble , which, in turn, is updated based on to obtain the analysis ensemble of interest at :
e7
e8
e9
with and , the Kalman gain , and
e10
e11
where is defined in a similar way as , but using the pseudoforecast members. To better explain the difference between the EnKF-OSA and the standard EnKF, we write how the analysis members are computed, starting from their counterparts at the previous time:
e12
e13
Here, we ignored the model noise term for simplicity. To compute the analysis members , the standard EnKF updates the forecast members resulting from the model integration of the previous analysis ensemble using one Kalman-like correction. In contrast, EnKF-OSA updates the pseudoforecast ensemble , which is computed by model integration of the smoothed ensemble using a second Kalman-like correction. This explains why the EnKF-OSA is expected to provide an improved background for the analysis step and, thus, enhanced state estimates.

3. A deterministic EnKF-OSA algorithm (SEIK-OSA)

To derive a deterministic EnKF-OSA, we first present the KF-OSA algorithm that was introduced by Desbouvries et al. (2011), based on which we then derive SEIK-OSA.

a. KF-OSA

In the linear Gaussian framework [i.e., ], the forecast, analysis, and smoothing probability density functions (pdfs) are Gaussian (Desbouvries et al. 2011). Without abuse of notation with respect to their EnKF counterparts, let , , and , respectively, denote the forecast, analysis, and OSA-smoothed PM estimates, and let , , and denote their corresponding error covariances. The KF-OSA algorithm can be summarized as follows.

1) Smoothing step

Forecast estimate and error covariance are first computed as in the classical KF:
e14
e15
Once the observation is available, a smoothing Kalman gain is computed to smooth the previous analysis estimate at time as follows:
e16
e17
e18

2) Analysis step

The smoothed state is integrated forward in time using the dynamical model to compute the pseudoforecast:
e19
The observation is then reexploited to update the pseudoforecast and compute the analysis mean and covariance as follows:
e20
e21
where
e22
e23
e24
In the EnKF-OSA algorithm presented above, the analysis step of the pseudoforecast ensemble is performed based on the analysis Kalman gain , estimated from the ensembles, while in the KF-OSA algorithm (Desbouvries et al. 2011), the analysis correction is performed using . Here, we show that in the linear Gaussian case, these two expressions are identical (see appendix for the proof), that is,
e25
In an ensemble setting, this alternative expression allows us to directly estimate the analysis Kalman gain from the filter ensembles and enables more flexible incorporation of the model error covariance, following the various approaches discussed above.

b. SEIK-OSA algorithm

We follow the formulation of the SEIK filter in Pham (2001) and Hoteit et al. (2002) to derive SEIK-OSA from the KF-OSA algorithm.

1) Smoothing step

As in the stochastic EnKF-OSA described above, the smoothing step of SEIK-OSA aims at computing an OSA-smoothed ensemble from an analysis ensemble , but following a deterministic formulation (Tippett et al. 2003). Accounting for model errors in the forecast step of the SEIK filter would continuously increase the rank of the forecast error covariances, thereby limiting the usefulness of the low-rank approximation (Hoteit et al. 2007). Various approaches have been proposed to account for model errors in ensemble-based techniques by, for example, inflating the background error covariance (Pham et al. 1998), projecting the model error onto the ensemble subspace (Pham 2001), or perturbing the forecast ensemble runs (Hoteit et al. 2007). For the sake of simplicity, we apply inflation here, keeping in mind the other available techniques. The forecast members and the associated error covariance matrix are then given by
e26
e27
Denote by the matrix whose columns are the forecast members; similar notations hold for the analysis , smoothing , and pseudoforecast ensembles . Following the formulation of the error covariances in SEIK (Pham 2001), we can factorize the forecast and analysis error covariances as
e28
e29
where , , , and is an full column-rank matrix with a zero-column sum. The smoothed estimate is computed using the smoothing Kalman gain given by Eq. (16). Given Eqs. (28), (29), and (Hoteit et al. 2008), with the tangent linear model, can be decomposed as
e30
e31
The smoothed state estimate and the associated error covariance can be respectively computed as
e32
e33
The OSA-smoothed ensemble is then sampled from these two moments using the resampling scheme of SEIK:
e34
where is a square root of the matrix satisfying , is an matrix with orthonormal columns and a zero-column sum, and denotes the ith row of . An efficient procedure to generate the random matrices, , is described by Pham (2001) and Hoteit et al. (2002).

2) Analysis step

The analysis step [Eqs. (20)(24)] of KF-OSA involves the model error covariance matrix and requires the computation of the matrices and . For the analysis Kalman gain, we use Eq. (25) and replace with , which can be directly estimated from the ensembles. This expression of the Kalman gain can be further written as
e35
e36
where is the pseudoforecast error, is the pseudoforecast error covariance, defined as , is the pseudoinnovation, and is the cross covariance between the observational error and the pseudoforecast error.
To factorize the analysis error covariance in a square root form, we express it directly in terms of the analysis error as . Using (20) leads to
e37
Typically, one should expect (and this is often observed in practice) the analysis error to be weakly correlated with the observational error and, likewise, the smoothing error with the observational error at the next step, . Integrating the smoothing error with a deterministic model with additive noise independent from the should result in a pseudoforecast error that is weakly correlated with , which justifies the assumption that the pseudoforecast error should be weakly correlated with the observational error (i.e., ). This was further confirmed in our numerical experiments. Neglecting this cross-covariance term in (36) allows us to decompose as
e38
and, thus, as
e39
e40
where , , and . We finally insert (38) into (20) to obtain the analysis update of the pseudoforecast state:
e41
The analysis members are then generated as
e42
where is a square root of the matrix , such that .

c. Summary of the SEIK-OSA algorithm

The SEIK-OSA algorithm can be summarized as follows, starting from a given analysis ensemble :

1) Smoothing step

  • The forecast ensemble is first computed by integrating with the model to the time of the next available observation.

  • The forecast members are then used with the incoming observation to smooth the previous analysis mean using (32).

  • The new smoothed members are then generated using (34).

2) Analysis Step

  • The pseudoforecast ensemble is computed by integrating the with the model.

  • The resulting with the current observation are used to compute the analysis mean, as in (41).

  • The new analysis members are sampled using (42).

The proposed SEIK-OSA scheme, therefore, involves two forecast steps with the model: one to compute the forecast ensemble and the other to compute the pseudoforecast ensemble. It further performs two update steps: one to smooth the previous analysis estimate and the corresponding error covariance matrix using the “future” observation and another to update the pseudoforecast estimate and its error covariance using the same observation. This is, therefore, roughly twice more computationally demanding than the standard SEIK. As we will demonstrate with the numerical experiments presented in the next section, however, exploiting the information in the observation more efficiently, enhances the ensemble filter’s performance, even when implemented with only half the ensemble of the classical filtering scheme—a situation in which both schemes require comparable computational cost.

4. Numerical experiments

a. Experimental setting

Numerical experiments are performed with the strongly nonlinear Lorenz-96 model (Lorenz and Emanuel 1998). The model simulates the time evolution of an atmospheric quantity based on a set of differential equations:
e43
where denotes the ith element of the state at time t. The nonlinear (quadratic) terms represent advection, and the linear term simulates dissipation. L96 obeys the energy conservation law and is sensitive to the initial conditions and external forcing. In its most common form, the system dimension is , and the forcing term F is set to 8. For this value of F, disturbances propagate from low to high indices (west to east), and the model exhibits chaotic behavior. Boundary conditions are periodic [i.e., , , and ].
L96 was discretized using the Runge–Kutta 4th-order scheme with a constant time step, (which corresponds to 6 h in real-world time). The trajectory of a reference run is taken as the “true” trajectory to which Gaussian noise of zero mean and unit variance is added to generate the observations. The mean of the initial ensemble is taken as the mean of the model trajectory over the first 5000 model time steps, and the initial ensemble is generated by perturbing it with Gaussian noise with zero mean and identity covariance. We implement all filters with the covariance inflation (Anderson and Anderson 1999) and localization techniques (Houtekamer and Mitchell 1998). We apply the standard local analysis approach by restricting the update of each grid point to only observations falling within some influence radius (Sakov and Bertino 2011). After a spinup period of roughly 20 days to remove any detrimental impact, the simulations are run for a period of 5 years in model time (i.e., 7300 model steps, unless otherwise mentioned). We evaluate the performance of the filters using the root-mean-square error (RMSE) between the reference states and the filters’ estimates , averaged over all variables and over the assimilation period of cycles:
e44
To reduce statistical fluctuations, we repeat each experiment independently 10 times, with a randomly generated initial ensemble and randomly generated observational errors. We took the average of the RMSEs over these multiple runs as the final result.

b. Results and discussion

To test the filters under realistic and challenging situations, we consider three observational scenarios: full (i.e., all model variables are observed), half (i.e., every second variable is observed), and quarter (i.e., every fourth variable is observed). In each scenario, a series of sensitivity experiments is performed with various ensemble sizes, spatial and temporal observation frequencies, and observational error variances. Results of these experiments are presented with two-dimensional plots of the RMSE as a function of the inflation factor and the localization radius. The localization support radii vary from 2 (strong localization) to 40 (weak localization) grid points. The inflation varies from 1 to 1.3 when no bias is introduced in the forecast model and from 1 to 2.5 otherwise.

1) On the approximation

The SEIK-OSA algorithm was derived assuming that . To evaluate the order of this term in comparison with the other terms in the Kalman gain (36), we implement EnKF-OSA using an ensemble size of 1000 (without localization and inflation) to eliminate the covariance undersampling noise and assimilate all model variables every four model steps. The results at day 365 shown in Fig. 2 suggest that is more than 10 orders of magnitude smaller than . Accordingly, is much smaller than . We found that in terms of RMSEs, neglecting this term or not in EnKF-OSA resulted in similar performances. With small ensembles, neglecting this term in the EnKF-OSA even improved its performance. This is because is undersampled in this case. Based on these results, we neglect this term in both EnKF-OSA and SEIK-OSA in all following experiments.

Fig. 2.
Fig. 2.

Plots of the covariance terms and in (36), as estimated by the EnKF-OSA algorithm with 1000 ensembles at day 365 while assimilating all model variables every four model steps.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

2) Sensitivity to the ensemble size

We first examine the sensitivity of the filters to the ensemble size. In realistic large-scale applications, one would be restricted to small ensembles to avoid prohibitive computational burden. We conduct the experiments using five different ensemble sizes: 10, 20, 40, 80, and 160. For the last two sizes, we implement the filters without localization and assess only the effect of inflation. We assimilate the observations every day, that is, every four model steps. We then investigate the three observational scenarios in which all, half, and one-quarter of the observations are assimilated. Figures 3, 4, and 5 present the performance of EnKF, EnKF-OSA, SEIK, and SEIK-OSA in terms of RMSE for , respectively. As expected, the deterministic EnKFs outperform the stochastic EnKFs when the number of observations is close to or larger than the ensemble’s size (Hoteit et al. 2015). As the ensemble size increases, the performances of SEIK and EnKF become more comparable, whether or not filtering is implemented based on OSA smoothing. We consider, for instance, the case in which half of the observations are assimilated. With only 10 members, SEIK achieves a lower minimum RMSE (0.84) than EnKF (1.06), corresponding to an improvement of 20%. Similarly, SEIK-OSA reaches an RMSE minimum of 0.7 versus 0.87 for EnKF-OSA, which corresponds also to an improvement of roughly 20%, while with 40 members, the differences become less pronounced.

Fig. 3.
Fig. 3.

Time-averaged RMSE as a function of the localization radius (x axis) and inflation factor (y axis). All filters were implemented with 10 members, and observations were assimilated from (top) all, (middle) half, and (bottom) one-quarter of the model variables at every fourth model time step (or 24 h in real time). A logarithmic color scale is used to emphasize the low RMSE values. The minimum-averaged RMSEs are indicated by asterisks, and their associated values are given in the title. White boxes indicate divergence of the filter.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Fig. 4.
Fig. 4.

As in Fig. 3, but for 20 ensemble members.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Fig. 5.
Fig. 5.

As in Fig. 3, but for 40 ensemble members.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

The OSA-smoothing-based filters clearly perform better than the standard filters; this becomes more and more pronounced when less data are assimilated and when the filters are implemented with small ensembles, as shown in Fig. 3, in which the filters used members. The minimum RMSEs achieved by SEIK-OSA are 0.38, 0.7, and 1.18 when, respectively, all, half, and one-quarter of the observations are assimilated, compared to 0.44, 0.84, and 1.52 with the standard SEIK. The stochastic EnKF-based versions exhibit similar behavior. Moreover, with OSA smoothing, the robustness of both EnKF and SEIK is clearly enhanced, mainly when fewer observations are assimilated (divergences are indicated in the figure by white boxes). With relatively larger ensembles, both EnKF-OSA and SEIK-OSA still outperform EnKF and SEIK in all observational scenarios, but the resulting improvements become less significant as the number of members increases (see Figs. 4 and 5 for which and 40, respectively). This is consistent with our expectation of the benefit of OSA smoothing, which should be more pronounced when the filter is not implemented in an ideal setting, highlighting how a better exploitation of the data could mitigate filtering deficiencies.

Figures 35 suggest that the overall performances of the different schemes are sensitive to the choice of the filtering parameters, such as inflation factor and localization radius. For instance, the SEIK-based algorithms seem to require less localization than the EnKF ones, mainly when the ensemble size is smaller than the number of observations, a case in which the stochastic perturbations in the EnKF may introduce noise in the filter error covariances. When the ensemble size is relatively large, compared to the number of observations (see, e.g., Fig. 5, bottom panels, where and ), SEIK and EnKF require similar localization radii. In general, SEIK-OSA seems to require less localization than SEIK, which is probably related to the improved ensemble resulting from its two-stage update. This is, however, true for the EnKF only when the ensemble size is larger than the number of observations. Although the OSA smoothing provides an improved background, the effect of undersampling becomes more pronounced because of its two update steps. Overall, SEIK-OSA is (generally) the least demanding in terms of localization, as it exploits the benefits of both the OSA-smoothing update to improve the background and the SEIK update to avoid the observations’ perturbations.

In terms of inflation, OSA-smoothing-based schemes seem to require larger inflations in general, compared to the standard schemes, and this becomes more pronounced as less data are assimilated. OSA smoothing tends to decrease the ensemble spread because of its two-stage update, thereby requiring large inflation to preserve enough spread for the following assimilation cycle.

To investigate more closely the impact of the additional smoothing step, we plot in Figs. 6 and 7 the time evolution of two model variables (observed variable 19 and nonobserved variable 20) over a 30-day period (i.e., between days 366 and 395) as estimated by SEIK and SEIK-OSA, respectively. The EnKFs exhibit similar behavior and are, thus, not shown. We present the box plots in which the whiskers are set using 0 and 95th percentiles. The dashed green line represents the filter estimate (the mean of the ensembles), the stars are the observations, and the circles are the true state. Plots are shown for the challenging case with only 10 ensemble members and assimilation of one-quarter of the observations. We use a localization radius of 4 and an inflation factor of 1.15. Overall, the forecast and analysis of SEIK-OSA are closer to the reference states than those of SEIK. This is more pronounced for the nonobserved variable. Consider, for example, the period ranging from day 382 to 387 in Fig. 7, where SEIK fails to accurately recover the reference states. SEIK-OSA, on the other hand, better tracks the true trajectory. This confirms our expectations related to the effect of the smoothing step on the pseudoforecast ensemble. From a probabilistic point of view, the forecast members are, indeed, samples from the forecast distribution , where and are, respectively, the analysis and forecast distributions, and is the state transition density represented by the dynamical model. On the other hand, the pseudoforecast ensemble can be sampled from the distribution . Thus, the forecast and the pseudoforecast ensembles are both computed following the same mechanism [i.e., by sampling from ], the only difference being that the pseudoforecast ensemble starts from a (smoothed) ensemble , which is already constrained by the observation. This conditioning on the observation explains the improved pseudoforecast uncertainties, compared with the original forecast uncertainties. This two-stage update also explains why SEIK-OSA tends to reduce the ensemble spread, compared with the classical SEIK. Exploiting the information more efficiently in the observations, thus, improves the representativeness of the background ensembles and reduces their spread. As a matter of fact, the averaged forecast ensemble spread for the observed variable 19 over the whole simulation period reduces from 1.95 with SEIK to 1.17 with SEIK-OSA, about a 40% reduction. This further explains why the OSA-smoothing-based schemes generally require larger inflation values than the standard schemes, mainly for the cases with sparse observations.

Fig. 6.
Fig. 6.

Time evolution between days 366 and 395 of the true state (circles), observations (stars), mean of analysis ensemble members (green dashed line), and boxplots of 10 ensemble members for the observed model variable 19, as resulting from (a) SEIK forecast, (b) SEIK analysis, (c) SEIK-OSA forecast, (d) SEIK-OSA analysis, (e) SEIK-OSA pseudoforecast, and (f) SEIK-OSA smoothing.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Fig. 7.
Fig. 7.

As in Fig. 6, but for the nonobserved model variable 20.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Figure 8 presents bar plots of the minimum RMSEs resulting from the best configuration of each filter, including also results from 80 and 160 members, as a function of the ensemble size, for the full, half, and one-quarter observation scenarios. Overall, the OSA-smoothing-based filters outperform the standard ones for all ensemble sizes. Furthermore, as expected, for large enough ensembles (), the stochastic variant, EnKF-OSA, outperforms SEIK-OSA, mainly when less data are assimilated (Lei et al. 2010; Hoteit et al. 2015). We also see that the OSA-smoothing-based filters not only outperform the corresponding standard schemes for the same ensemble size, but also when they are implemented with only half the number of ensemble members—that is, when the standard and OSA-smoothing-based filters require similar computational costs. In the same context, Fig. 9 summarizes the run times of EnKF-OSA and SEIK-OSA and of their standard counterparts for different ensemble sizes. We implement all the filters without inflation and localization and report the computing times required by each of them (in seconds) when all the data are assimilated every four model steps. One can see that, as expected, the computing times of EnKF-OSA and SEIK-OSA are roughly doubled, compared with those of EnKF and SEIK, respectively, for a given ensemble size. Besides, we clearly notice that the computing times of EnKF-OSA and SEIK-OSA are practically the same as those of EnKF and SEIK when implemented with twice the number of ensemble members ().

Fig. 8.
Fig. 8.

Minimum average RMSE for all tested filters as a function of ensemble size. (left) All, (middle) half, and (right) one-quarter of the observations were assimilated every four model steps.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Fig. 9.
Fig. 9.

Summary of the run times (s) of EnKF, EnKF-OSA, SEIK, and SEIK-OSA for different ensemble sizes. All filters were implemented without inflation and localization, and all data are assimilated every four model steps.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Table 1 reports the percentage of improvements computed using the minimum average RMSEs resulting from EnKF-OSA and SEIK-OSA when implemented with only half of the number of members of EnKF and SEIK, respectively. SEIK-OSA outperforms the standard SEIK for all tested scenarios, while EnKF-OSA improves upon the classical EnKF starting at , when the undersampling errors are less pronounced, which could be more severe in the two-stage update step of the EnKF-OSA.

Table 1.

Percentages of improvements, in terms of RMSE, computed using the minimum average RMSEs resulting from the OSA-smoothing-based filters and the standard filters using only half of the ensemble size. The notation means that the OSA-smoothing-based filter uses members, while the standard filter uses members. The (−) sign suggests that the standard filter with members produces more accurate estimates than the OSA-smoothing-based filter with members.

Table 1.

3) Sensitivity to the frequency of observations

In the second set of experiments, we fix the ensemble size to and vary the frequency at which the observations are assimilated. We tolerate an increase or decrease of 5 days in model time (i.e., 20 model steps) to ensure that the last model step coincides with the last filter step if it is not the case with the 5-yr simulation period. Figures 10 and 11 plot the results of these experiments when the data are assimilated every model step and every six model steps, respectively. These are compared with those in Fig. 4, in which the data are assimilated every four model steps.

Fig. 10.
Fig. 10.

As in Fig. 4, but with assimilation of observations every model time step.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Fig. 11.
Fig. 11.

As in Fig. 4, but with assimilation of observations every six model time steps.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

When the observations are assimilated frequently in time (every model step), filtering with or without OSA smoothing leads to similar results. With small assimilation windows, the system is well constrained by the data, and the filters are implemented under favorable conditions, with a large enough ensemble and weakly nonlinear conditions. With fewer observations assimilated in time, a situation in which the model becomes more nonlinear within the forecast period, the accuracy of the background approximation degrades, and the benefit of the OSA smoothing becomes clearly pronounced. The OSA-smoothing-based schemes are found to be more robust and provide more accurate estimates, compared with those of EnKF and SEIK (Figs. 4, 11). When the observations are assimilated every six model time steps, EnKF-OSA leads to about 15%, 20%, and 16% improvement over EnKF when all, half, and one-quarter of the observations are assimilated, respectively, while SEIK-OSA leads to 18%, 22%, and 21% improvement over SEIK for the same observational scenarios.

Figure 12 illustrates the evolution of the minimum RMSEs for different frequencies of assimilated observations, ranging between every model step and every 20 model steps. The percentages of improvements resulting from the OSA-smoothing-based schemes with respect to the standard ones are also reported. As stated above, the benefit of the OSA-smoothing step becomes pronounced only when the data are not assimilated very frequently; the largest relative improvements generally correspond to an assimilation frequency of six or eight model steps. This is, however, true only up to a certain frequency (2 to 12 or 16, depending on the observational density), beyond which the performance of the OSA-smoothing-based schemes starts to converge toward that of the standard ones. For large assimilation windows, the model becomes strongly nonlinear, and the linear cross correlations become irrelevant to conducting the smoothing step (Song et al. 2013). One ad hoc way that might improve the results with large assimilation windows would be to apply the smoothing on the model state between the previous and current analysis times, as suggested by (Yang et al. 2013).

Fig. 12.
Fig. 12.

(bottom) Minimum average RMSE for all tested filters as a function of the frequency of observations, defined as the ratio between the filter step (fstep) and the model step (mstep), and (top) percentages of relative improvements, respectively, brought by EnKF-OSA and SEIK-OSA with respect to EnKF and SEIK. (left) All, (middle) half, and (right) one-quarter of the observations were assimilated every four model steps using 20 ensemble members.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

4) Sensitivity to the variance of the observational errors

In this set of experiments, we fix the ensemble size to and assimilate the observations every four model steps. Three new values of the variance of the observational error are tested: a small , a relatively large , and a large . Figures 13 and 14 present the results of the experiments with and . Comparing these results with those of Fig. 4, where the observational error variance is set to 1, we can see that, as expected, the filters’ performance degrades as the noise level in the observations increases. The standard and OSA-smoothing-based filters exhibit comparable performances in the case of small observational errors. However, additional experiments that we conducted, in which we designed the background to be less accurately estimated, compared with that of Fig. 13 using smaller ensemble sizes and larger assimilation windows, suggest that the OSA-smoothing-based filters outperform the standard schemes. For instance, repeating the same experiment of Fig. 13 but assimilating the observations every eight model steps (instead of every four), we found that OSA-smoothing-based schemes significantly outperform the standard ones, especially when the observations are sparse. The results, indeed, suggest that when only one-quarter of the observations are assimilated, SEIK and EnKF, respectively, reach minimum RMSEs of 1.58 and 1.45, compared to 0.90 and 0.84 with SEIK-OSA and EnKF-OSA. One can also see from Figs. 13 and 14 that as the noise level increases, EnKF-OSA and SEIK-OSA outperform EnKF and SEIK, respectively, in all observational scenarios.

Fig. 13.
Fig. 13.

As in Fig. 4, but for an observational error variance equal to 0.1.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Fig. 14.
Fig. 14.

As in Fig. 4, but for an observational error variance equal to 2.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

Figure 15 plots the minimum RMSE values for all filtering schemes, considering the four tested values of observational error variances (0.1, 1, 2, and 4) along with the percentages of relative improvements resulting from the OSA-smoothing-based schemes with respect to the standard ones. Overall, OSA-smoothing-based filters always outperform the classical ones, particularly when the variance of the noise in the observations is larger than 1, suggesting better ability to filter out the noise in the data using the improved background. For instance, when the observational error variance is set to 2 (which is the case of Fig. 14), the estimates are improved by about 12%, 12%, and 8% when, respectively, all, half, and one-quarter of the observations are assimilated using EnKF-OSA, and 14%, 11%, and 9% using SEIK-OSA.

Fig. 15.
Fig. 15.

(bottom) Minimum average RMSE for all tested filters as a function of the observational error variance and (top) percentages of relative improvements, respectively, brought by EnKF-OSA and SEIK-OSA with respect to EnKF and SEIK. (left) All, (middle) half, and (right) one-quarter of the observations were assimilated every four model steps using 20 ensemble members.

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

5) Sensitivity to bias in the forecast model

We test the filters’ performance with a biased forecast model. The L96 model is integrated with in the filter’s forecast step, while the reference states are simulated using . Larger inflation-factor values, ranging from 1 to 2.5, are considered in this set of experiments. The ensemble size is set to 20, data are assimilated every four model steps, and the observational error variance is set to 1. As shown in Fig. 16, the results are consistent with those of the unbiased case (Fig. 4); that is, EnKF-OSA and SEIK-OSA still outperform EnKF and SEIK when the model is biased, and as the number of assimilated observations decreases, these improvements become more pronounced. Indeed, EnKF-OSA performs 3%, 10%, and 15% better, in terms of RMSE, than EnKF performs when, respectively, all, half, and one-quarter of the observations are assimilated, whereas SEIK-OSA is 4%, 10%, and 17% better than SEIK.

Fig. 16.
Fig. 16.

As in Fig. 4, but assimilation was carried out with a biased forcing .

Citation: Monthly Weather Review 146, 2; 10.1175/MWR-D-17-0175.1

To eliminate the ensemble undersampling error and focus on background limitations due to the bias, we conducted an experiment using a large ensemble of 200 members without localization. We found that the EnKF-OSA estimates are 10%, 12%, and 13% more accurate than those of the standard EnKF when, respectively, all, half, and one-quarter of the observations are assimilated, whereas SEIK-OSA estimates are 13%, 13%, and 14% more accurate than those of SEIK for the same scenarios. This suggests that exploiting more the observations may also help to mitigate bias in the forecast model.

5. Discussion and summary

We investigated the relevance of an ensemble filtering strategy that replaces the classical filtering path, which involves a forecast step between two successive analyses, with an alternative path that exploits more information in the observations. The proposed filtering scheme, which introduces one-step-ahead smoothing of the state between two successive analyses, is an ensemble implementation of a fully Bayesian filtering approach. The new smoothing step involves more information in the estimation process, which provides more reliable background. This is shown to improve the ensemble filter’s performance when the filter is implemented under imperfect conditions. The resulting filters bear strong resemblance to the running in place ETKF scheme, which was introduced as a heuristic approach.

Our goals were to evaluate the relevance of the OSA-smoothing strategy and to propose a new deterministic ensemble variant that is suitable for data assimilation into large-scale oceanic and atmospheric models. We constructed such a filter based on the SEIK filter, which we called SEIK-OSA.

Numerical experiments with the Lorenz-96 model demonstrated the relevance of the proposed OSA-smoothing strategy. Even though OSA-smoothing-based ensemble filters are computationally about twice more demanding than are standard ensemble filtering schemes for the same ensemble size, these schemes generally lead to more accurate estimates using only half the ensemble size of the standard schemes. The assimilation results also suggested that the proposed OSA-smoothing-based filters are clearly more efficient and more robust than the standard filters under all tested experimental settings and scenarios. This was more pronounced particularly when the filters were implemented in the challenging situations of low spatial and temporal data coverage, large sampling errors, noisy observations, and biased forecast model.

The proposed ensemble-OSA schemes are straightforward to implement and require only a few modifications to an existing ensemble filtering code. They were further found particularly efficient in situations depicting realistic scenarios. Future work will focus on testing the new OSA-smoothing schemes with large-scale oceanic and atmospheric data assimilation problems. Another future direction we will take is to consider using an OSA-smoothing-based strategy in the context of nonlinear, non-Gaussian filtering, such as the particle and Gaussian mixture filters.

Acknowledgments

This work is supported by King Abdullah University of Science and Technology Award CRG3-2156.

APPENDIX

Proof of (25)

By inserting (17) in (19), we obtain
ea1
where . The errors and , respectively, associated with the estimates and of and are given by
ea2
ea3
The cross covariance can thus be expressed as
ea4
The covariance can also be expressed as
ea5
We can then check that is invertible, as it can be expressed as
ea6
From (A4), (A5), and (A6), we readily obtain (25).

REFERENCES

  • Ait-El-Fquih, B., and I. Hoteit, 2016: A variational Bayesian multiple particle filtering scheme for large-dimensional systems. IEEE Trans. Signal Process., 64, 54095422, https://doi.org/10.1109/TSP.2016.2580524.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Ait-El-Fquih, B., M. El Gharamti, and I. Hoteit, 2016: A Bayesian consistent dual ensemble Kalman filter for state-parameter estimation in subsurface hydrology. Hydrol. Earth Syst. Sci., 20, 32893307, https://doi.org/10.5194/hess-20-3289-2016.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Anderson, B. D., and J. B. Moore, 1979: Optimal Filtering. Information and System Science Series, T. Kailath, Ed., Prentice-Hall, 367 pp.

  • Anderson, J. L., 2001: An ensemble adjustment Kalman filter for data assimilation. Mon. Wea. Rev., 129, 28842903, https://doi.org/10.1175/1520-0493(2001)129<2884:AEAKFF>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Anderson, J. L., and S. L. Anderson, 1999: A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts. Mon. Wea. Rev., 127, 27412758, https://doi.org/10.1175/1520-0493(1999)127<2741:AMCIOT>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Anderson, J. L., T. Hoar, K. Raeder, H. Liu, N. Collins, R. Torn, and A. Avellano, 2009: The Data Assimilation Research Testbed: A community facility. Bull. Amer. Meteor. Soc., 90, 12831296, https://doi.org/10.1175/2009BAMS2618.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Bishop, C. H., B. J. Etherton, and S. J. Majumdar, 2001: Adaptive sampling with the ensemble transform Kalman filter. Part I: Theoretical aspects. Mon. Wea. Rev., 129, 420436, https://doi.org/10.1175/1520-0493(2001)129<0420:ASWTET>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Burgers, G., P. J. van Leeuwen, and G. Evensen, 1998: Analysis scheme in the ensemble Kalman filter. Mon. Wea. Rev., 126, 17191724, https://doi.org/10.1175/1520-0493(1998)126<1719:ASITEK>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Carton, J. A., and B. S. Giese, 2008: A reanalysis of ocean climate using Simple Ocean Data Assimilation (SODA). Mon. Wea. Rev., 136, 29993017, https://doi.org/10.1175/2007MWR1978.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Chen, Y., and D. S. Oliver, 2013: Levenberg–Marquardt forms of the iterative ensemble smoother for efficient history matching and uncertainty quantification. Comput. Geosci., 17, 689703, https://doi.org/10.1007/s10596-013-9351-5.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Desbouvries, F., and B. Ait-El-Fquih, 2008: Direct, prediction-based and smoothing-based particle filter algorithms. Proc. Fourth World Conf. of the Int. Association for Statistical Computing, Yokohama, Japan, IASC, 384–393.

  • Desbouvries, F., Y. Petetin, and B. Ait-El-Fquih, 2011: Direct, prediction- and smoothing-based Kalman and particle filter algorithms. Signal Process., 91, 20642077, https://doi.org/10.1016/j.sigpro.2011.03.013.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Emerick, A. A., and A. C. Reynolds, 2013: Ensemble smoother with multiple data assimilation. Comput. Geosci., 55, 315, https://doi.org/10.1016/j.cageo.2012.03.011.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Evensen, G., 1994: Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. J. Geophys. Res., 99, 10 14310 162, https://doi.org/10.1029/94JC00572.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Evensen, G., 2003: The ensemble Kalman filter: Theoretical formulation and practical implementation. Ocean Dyn., 53, 343367, https://doi.org/10.1007/s10236-003-0036-9.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Gharamti, M., B. Ait-El-Fquih, and I. Hoteit, 2015: An iterative ensemble Kalman filter with one-step-ahead smoothing for state-parameters estimation of contaminant transport models. J. Hydrol., 527, 442457, https://doi.org/10.1016/j.jhydrol.2015.05.004.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Ghil, M., and P. Malanotte-Rizzoli, 1991: Data assimilation in meteorology and oceanography. Advances in Geophysics, Vol. 33, Academic Press, 141–266, https://doi.org/10.1016/S0065-2687(08)60442-2.

    • Crossref
    • Export Citation
  • Hamill, T. M., and C. Snyder, 2000: A hybrid ensemble Kalman filter–3D variational analysis scheme. Mon. Wea. Rev., 128, 29052919, https://doi.org/10.1175/1520-0493(2000)128<2905:AHEKFV>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., D. T. Pham, and J. Blum, 2002: A simplified reduced order Kalman filtering and application to altimetric data assimilation in tropical Pacific. J. Mar. Syst., 36, 101127, https://doi.org/10.1016/S0924-7963(02)00129-X.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., G. Triantafyllou, and G. Korres, 2007: Using low-rank ensemble Kalman filters for data assimilation with high dimensional imperfect models. J. Numer. Anal. Ind. Appl. Math., 2, 6778.

    • Search Google Scholar
    • Export Citation
  • Hoteit, I., D. T. Pham, G. Triantafyllou, and G. Korres, 2008: A new approximate solution of the optimal nonlinear filter for data assimilation in meteorology and oceanography. Mon. Wea. Rev., 136, 317334, https://doi.org/10.1175/2007MWR1927.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., T. Hoar, G. Gopalakrishnan, N. Collins, J. Anderson, B. Cornuelle, A. Köhl, and P. Heimbach, 2013: A MITgcm/DART ensemble analysis and prediction system with application to the Gulf of Mexico. Dyn. Atmos. Oceans, 63, 123, https://doi.org/10.1016/j.dynatmoce.2013.03.002.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., D.-T. Pham, M. Gharamti, and X. Luo, 2015: Mitigating observation perturbation sampling errors in the stochastic EnKF. Mon. Wea. Rev., 143, 29182936, https://doi.org/10.1175/MWR-D-14-00088.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126, 796811, https://doi.org/10.1175/1520-0493(1998)126<0796:DAUAEK>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Kalman, R., 1960: A new approach to linear filtering and prediction problems. J. Basic Eng., 82, 3545, https://doi.org/10.1115/1.3662552.

  • Kalnay, E., and S.-C. Yang, 2010: Accelerating the spin-up of ensemble Kalman filtering. Quart. J. Roy. Meteor. Soc., 136, 16441651, https://doi.org/10.1002/qj.652.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Le Dimet, F.-X., and O. Talagrand, 1986: Variational algorithms for analysis and assimilation of meteorological observations: Theoretical aspects. Tellus, 38A, 97110, https://doi.org/10.1111/j.1600-0870.1986.tb00459.x.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Lee, W., and C. Farmer, 2014: Data assimilation by conditioning of driving noise on future observations. IEEE Trans. Signal Process., 62, 38873896, https://doi.org/10.1109/TSP.2014.2330807.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Lei, J., P. Bickel, and C. Snyder, 2010: Comparison of ensemble Kalman filters under non-Gaussianity. Mon. Wea. Rev., 138, 12931306, https://doi.org/10.1175/2009MWR3133.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Liu, B., B. Ait-El-Fquih, and I. Hoteit, 2016: Efficient kernel-based ensemble Gaussian mixture filtering. Mon. Wea. Rev., 144, 781800, https://doi.org/10.1175/MWR-D-14-00292.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Lorenz, E. N., and K. A. Emanuel, 1998: Optimal sites for supplementary weather observations: Simulation with a small model. J. Atmos. Sci., 55, 399414, https://doi.org/10.1175/1520-0469(1998)055<0399:OSFSWO>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Luo, X., and I. Hoteit, 2011: Robust ensemble filtering and its relation to covariance inflation in the ensemble Kalman filter. Mon. Wea. Rev., 139, 39383953, https://doi.org/10.1175/MWR-D-10-05068.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Luo, X., R. J. Lorentzen, A. S. Stordal, and G. Nævdal, 2014: Toward an enhanced Bayesian estimation framework for multiphase flow soft-sensing. Inverse Probl., 30, 114012, https://doi.org/10.1088/0266-5611/30/11/114012.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Luo, X., A. S. Stordal, R. J. Lorentzen, and G. Nævdal, 2015: Iterative ensemble smoother as an approximate solution to a regularized minimum-average-cost problem: Theory and applications. Soc. Pet. Eng. J., 20, https://doi.org/10.2118/176023-PA.

    • Search Google Scholar
    • Export Citation
  • Ma, X., G. Hetz, X. Wang, L. Bi, D. Stern, and N. Hoda, 2017: A robust iterative ensemble smoother method for efficient history matching and uncertainty quantification. SPE Reservoir Simulation Conf., Montgomery, TX, Society of Petroleum Engineers, https://doi.org/10.2118/182693-MS.

    • Crossref
    • Export Citation
  • Moradkhani, H., S. Sorooshian, H. V. Gupta, and P. R. Houser, 2005: Dual state–parameter estimation of hydrological models using ensemble Kalman filter. Adv. Water Resour., 28, 135147, https://doi.org/10.1016/j.advwatres.2004.09.002.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Nævdal, G., L. Johnsen, S. Aanonsen, and E. Vefring, 2005: Reservoir monitoring and continuous model updating using ensemble Kalman filter. Soc. Pet. Eng. J., 10, 6674, https://doi.org/10.2118/84372-PA.

    • Search Google Scholar
    • Export Citation
  • Penny, S., E. Kalnay, J. Carton, B. Hunt, K. Ide, T. Miyoshi, and G. Chepurin, 2013: The local ensemble transform Kalman filter and the running-in-place algorithm applied to a global ocean general circulation model. Nonlinear Processes Geophys., 20, 10311046, https://doi.org/10.5194/npg-20-1031-2013.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Pham, D. T., 2001: Stochastic methods for sequential data assimilation in strongly nonlinear systems. Mon. Wea. Rev., 129, 11941207, https://doi.org/10.1175/1520-0493(2001)129<1194:SMFSDA>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Pham, D. T., J. Verron, and M. C. Roubaud, 1998: A singular evolutive extended Kalman filter for data assimilation in oceanography. J. Mar. Syst., 16, 323340, https://doi.org/10.1016/S0924-7963(97)00109-7.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Pitt, M. K., and N. Shephard, 1999: Filtering via simulation: Auxiliary particle filters. J. Amer. Stat. Assoc., 94, 590599, https://doi.org/10.1080/01621459.1999.10474153.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Reichle, R. H., D. B. McLaughlin, and D. Entekhabi, 2002: Hydrologic data assimilation with the ensemble Kalman filter. Mon. Wea. Rev., 130, 103114, https://doi.org/10.1175/1520-0493(2002)130<0103:HDAWTE>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Sakov, P., and L. Bertino, 2011: Relation between two common localisation methods for the EnKF. Comput. Geosci., 15, 225237, https://doi.org/10.1007/s10596-010-9202-6.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Simon, D., 2006: Optimal State Estimation: Kalman, H Infinity, and Nonlinear Approaches. John Wiley & Sons, 552 pp.

    • Crossref
    • Export Citation
  • Song, H., I. Hoteit, B. D. Cornuelle, and A. C. Subramanian, 2010: An adaptive approach to mitigate background covariance limitations in the ensemble Kalman filter. Mon. Wea. Rev., 138, 28252845, https://doi.org/10.1175/2010MWR2871.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Song, H., I. Hoteit, B. D. Cornuelle, X. Luo, and A. C. Subramanian, 2013: An adjoint-based adaptive ensemble Kalman filter. Mon. Wea. Rev., 141, 33433359, https://doi.org/10.1175/MWR-D-12-00244.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Tippett, M. K., J. L. Anderson, C. H. Bishop, T. M. Hamill, and J. S. Whitaker, 2003: Ensemble square root filters. Mon. Wea. Rev., 131, 14851490, https://doi.org/10.1175/1520-0493(2003)131<1485:ESRF>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Todling, R., 1999: Estimation theory and foundations of atmospheric data assimilation. DAO Office Note 1999-01, 187 pp.

  • Triantafyllou, G., I. Hoteit, X. Luo, K. Tsiaras, and G. Petihakis, 2013: Assessing a robust ensemble-based Kalman filter for efficient ecosystem data assimilation of the Cretan Sea. J. Mar. Syst., 125, 90100, https://doi.org/10.1016/j.jmarsys.2012.12.006.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Whitaker, J. S., and T. M. Hamill, 2002: Ensemble data assimilation without perturbed observations. Mon. Wea. Rev., 130, 19131924, https://doi.org/10.1175/1520-0493(2002)130<1913:EDAWPO>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Yang, S.-C., K.-J. Lin, T. Miyoshi, and E. Kalnay, 2013: Improving the spin-up of regional EnKF for typhoon assimilation and forecasting with Typhoon Sinlaku (2008). Tellus, 65A, 20804, https://doi.org/10.3402/tellusa.v65i0.20804.

    • Crossref
    • Search Google Scholar
    • Export Citation
Save
  • Ait-El-Fquih, B., and I. Hoteit, 2016: A variational Bayesian multiple particle filtering scheme for large-dimensional systems. IEEE Trans. Signal Process., 64, 54095422, https://doi.org/10.1109/TSP.2016.2580524.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Ait-El-Fquih, B., M. El Gharamti, and I. Hoteit, 2016: A Bayesian consistent dual ensemble Kalman filter for state-parameter estimation in subsurface hydrology. Hydrol. Earth Syst. Sci., 20, 32893307, https://doi.org/10.5194/hess-20-3289-2016.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Anderson, B. D., and J. B. Moore, 1979: Optimal Filtering. Information and System Science Series, T. Kailath, Ed., Prentice-Hall, 367 pp.

  • Anderson, J. L., 2001: An ensemble adjustment Kalman filter for data assimilation. Mon. Wea. Rev., 129, 28842903, https://doi.org/10.1175/1520-0493(2001)129<2884:AEAKFF>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Anderson, J. L., and S. L. Anderson, 1999: A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts. Mon. Wea. Rev., 127, 27412758, https://doi.org/10.1175/1520-0493(1999)127<2741:AMCIOT>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Anderson, J. L., T. Hoar, K. Raeder, H. Liu, N. Collins, R. Torn, and A. Avellano, 2009: The Data Assimilation Research Testbed: A community facility. Bull. Amer. Meteor. Soc., 90, 12831296, https://doi.org/10.1175/2009BAMS2618.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Bishop, C. H., B. J. Etherton, and S. J. Majumdar, 2001: Adaptive sampling with the ensemble transform Kalman filter. Part I: Theoretical aspects. Mon. Wea. Rev., 129, 420436, https://doi.org/10.1175/1520-0493(2001)129<0420:ASWTET>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Burgers, G., P. J. van Leeuwen, and G. Evensen, 1998: Analysis scheme in the ensemble Kalman filter. Mon. Wea. Rev., 126, 17191724, https://doi.org/10.1175/1520-0493(1998)126<1719:ASITEK>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Carton, J. A., and B. S. Giese, 2008: A reanalysis of ocean climate using Simple Ocean Data Assimilation (SODA). Mon. Wea. Rev., 136, 29993017, https://doi.org/10.1175/2007MWR1978.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Chen, Y., and D. S. Oliver, 2013: Levenberg–Marquardt forms of the iterative ensemble smoother for efficient history matching and uncertainty quantification. Comput. Geosci., 17, 689703, https://doi.org/10.1007/s10596-013-9351-5.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Desbouvries, F., and B. Ait-El-Fquih, 2008: Direct, prediction-based and smoothing-based particle filter algorithms. Proc. Fourth World Conf. of the Int. Association for Statistical Computing, Yokohama, Japan, IASC, 384–393.

  • Desbouvries, F., Y. Petetin, and B. Ait-El-Fquih, 2011: Direct, prediction- and smoothing-based Kalman and particle filter algorithms. Signal Process., 91, 20642077, https://doi.org/10.1016/j.sigpro.2011.03.013.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Emerick, A. A., and A. C. Reynolds, 2013: Ensemble smoother with multiple data assimilation. Comput. Geosci., 55, 315, https://doi.org/10.1016/j.cageo.2012.03.011.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Evensen, G., 1994: Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. J. Geophys. Res., 99, 10 14310 162, https://doi.org/10.1029/94JC00572.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Evensen, G., 2003: The ensemble Kalman filter: Theoretical formulation and practical implementation. Ocean Dyn., 53, 343367, https://doi.org/10.1007/s10236-003-0036-9.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Gharamti, M., B. Ait-El-Fquih, and I. Hoteit, 2015: An iterative ensemble Kalman filter with one-step-ahead smoothing for state-parameters estimation of contaminant transport models. J. Hydrol., 527, 442457, https://doi.org/10.1016/j.jhydrol.2015.05.004.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Ghil, M., and P. Malanotte-Rizzoli, 1991: Data assimilation in meteorology and oceanography. Advances in Geophysics, Vol. 33, Academic Press, 141–266, https://doi.org/10.1016/S0065-2687(08)60442-2.

    • Crossref
    • Export Citation
  • Hamill, T. M., and C. Snyder, 2000: A hybrid ensemble Kalman filter–3D variational analysis scheme. Mon. Wea. Rev., 128, 29052919, https://doi.org/10.1175/1520-0493(2000)128<2905:AHEKFV>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., D. T. Pham, and J. Blum, 2002: A simplified reduced order Kalman filtering and application to altimetric data assimilation in tropical Pacific. J. Mar. Syst., 36, 101127, https://doi.org/10.1016/S0924-7963(02)00129-X.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., G. Triantafyllou, and G. Korres, 2007: Using low-rank ensemble Kalman filters for data assimilation with high dimensional imperfect models. J. Numer. Anal. Ind. Appl. Math., 2, 6778.

    • Search Google Scholar
    • Export Citation
  • Hoteit, I., D. T. Pham, G. Triantafyllou, and G. Korres, 2008: A new approximate solution of the optimal nonlinear filter for data assimilation in meteorology and oceanography. Mon. Wea. Rev., 136, 317334, https://doi.org/10.1175/2007MWR1927.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., T. Hoar, G. Gopalakrishnan, N. Collins, J. Anderson, B. Cornuelle, A. Köhl, and P. Heimbach, 2013: A MITgcm/DART ensemble analysis and prediction system with application to the Gulf of Mexico. Dyn. Atmos. Oceans, 63, 123, https://doi.org/10.1016/j.dynatmoce.2013.03.002.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Hoteit, I., D.-T. Pham, M. Gharamti, and X. Luo, 2015: Mitigating observation perturbation sampling errors in the stochastic EnKF. Mon. Wea. Rev., 143, 29182936, https://doi.org/10.1175/MWR-D-14-00088.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126, 796811, https://doi.org/10.1175/1520-0493(1998)126<0796:DAUAEK>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Kalman, R., 1960: A new approach to linear filtering and prediction problems. J. Basic Eng., 82, 3545, https://doi.org/10.1115/1.3662552.

  • Kalnay, E., and S.-C. Yang, 2010: Accelerating the spin-up of ensemble Kalman filtering. Quart. J. Roy. Meteor. Soc., 136, 16441651, https://doi.org/10.1002/qj.652.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Le Dimet, F.-X., and O. Talagrand, 1986: Variational algorithms for analysis and assimilation of meteorological observations: Theoretical aspects. Tellus, 38A, 97110, https://doi.org/10.1111/j.1600-0870.1986.tb00459.x.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Lee, W., and C. Farmer, 2014: Data assimilation by conditioning of driving noise on future observations. IEEE Trans. Signal Process., 62, 38873896, https://doi.org/10.1109/TSP.2014.2330807.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Lei, J., P. Bickel, and C. Snyder, 2010: Comparison of ensemble Kalman filters under non-Gaussianity. Mon. Wea. Rev., 138, 12931306, https://doi.org/10.1175/2009MWR3133.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Liu, B., B. Ait-El-Fquih, and I. Hoteit, 2016: Efficient kernel-based ensemble Gaussian mixture filtering. Mon. Wea. Rev., 144, 781800, https://doi.org/10.1175/MWR-D-14-00292.1.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Lorenz, E. N., and K. A. Emanuel, 1998: Optimal sites for supplementary weather observations: Simulation with a small model. J. Atmos. Sci., 55, 399414, https://doi.org/10.1175/1520-0469(1998)055<0399:OSFSWO>2.0.CO;2.

    • Crossref
    • Search Google Scholar
    • Export Citation
  • Luo, X., and I. Hoteit, 2011: Robust ensemble filtering and its relation to covariance inflation in the ensemble Kalman filter. Mon. Wea. Rev., 139, 39383953, https://doi.org/10.1175/MWR-D-10-05068.1.

    • Crossref
    • Search Google Scholar
    • Export Citation