Sampling Errors in Ensemble Kalman Filtering. Part I: Theory

William Sacher McGill University, Montréal, Québec, Canada

Search for other papers by William Sacher in
Current site
Google Scholar
PubMed
Close
and
Peter Bartello McGill University, Montréal, Québec, Canada

Search for other papers by Peter Bartello in
Current site
Google Scholar
PubMed
Close
Full access

Abstract

This paper discusses the quality of the analysis given by the ensemble Kalman filter in a perfect model context when ensemble sizes are limited. The overall goal is to improve the theoretical understanding of the problem of systematic errors in the analysis variance due to the limited size of the ensemble, as well as the potential of the so-called double-ensemble Kalman filter, covariance inflation, and randomly perturbed analysis techniques to produce a stable analysis—that is to say, one not subject to filter divergence. This is achieved by expressing the error of the ensemble mean and the analysis error covariance matrix in terms of the sampling noise in the background error covariance matrix (owing to the finite ensemble estimation) and by comparing these errors for all methods. Theoretical predictions are confirmed with a simple scalar test case. In light of the analytical results obtained, the expression of the optimal covariance inflation factor is proposed in terms of the limited ensemble size and the Kalman gain.

Corresponding author address: William Sacher, Dept. of Atmospheric and Oceanic Sciences, McGill University, 805 Sherbrooke St. W., Montreal, QC H3A 2K6, Canada. Email: sacher@mail.mcgill.ca

Abstract

This paper discusses the quality of the analysis given by the ensemble Kalman filter in a perfect model context when ensemble sizes are limited. The overall goal is to improve the theoretical understanding of the problem of systematic errors in the analysis variance due to the limited size of the ensemble, as well as the potential of the so-called double-ensemble Kalman filter, covariance inflation, and randomly perturbed analysis techniques to produce a stable analysis—that is to say, one not subject to filter divergence. This is achieved by expressing the error of the ensemble mean and the analysis error covariance matrix in terms of the sampling noise in the background error covariance matrix (owing to the finite ensemble estimation) and by comparing these errors for all methods. Theoretical predictions are confirmed with a simple scalar test case. In light of the analytical results obtained, the expression of the optimal covariance inflation factor is proposed in terms of the limited ensemble size and the Kalman gain.

Corresponding author address: William Sacher, Dept. of Atmospheric and Oceanic Sciences, McGill University, 805 Sherbrooke St. W., Montreal, QC H3A 2K6, Canada. Email: sacher@mail.mcgill.ca

1. Introduction

This paper begins to address what is in our view the vital question facing those who employ the ensemble Kalman filter (EnKf; Evensen 1994): What are the systematic biases in the statistics on which it is based, resulting from a limited ensemble size? Ensemble forecasting is clearly an expensive proposition. The forecaster is faced with many choices to reduce the computational burden of the task. For example, is it better to have a small ensemble of high-resolution models (thereby potentially reducing model error at the risk of sampling error) or to have a large ensemble of lower-resolution models (thereby reducing sampling error at the risk of model error)? The answer depends on quite a number of things and in this study, the focus is much less ambitious. As a first step in the process, we attempt to quantify the sampling error resulting from the use of a relatively small (i.e., less than infinite) ensemble. Of course, we are not the first to be aware of these issues and our statistical error scaling arguments are employed to understand and improve previous attempts to deal with these biases, such as covariance inflation and the subdivision of the ensemble into smaller subensembles in order to remove “inbreeding” (see references below).

Blending of background and observation information is a necessary step in “data assimilation” (later referred to as DA), because neither model estimate nor observation is ever perfect. In this study, we consider a sequential (as opposed to variational) DA approach where the product of the analysis cycle is used as an initial state of a numerical prediction system. The basic tool in this approach is filtering; that is, all of the information is used up to and including the considered analysis time. We focus on one particular technique, the EnKf, and its numerical weather prediction (NWP) application. The EnKF has been the subject of studies in various meteorological and oceanographic contexts, many of them reviewed by Evensen (2007). However, to date, only the Canadian Meteorological Centre (CMC) has implemented it operationally (Houtekamer and Mitchell 2005). The Kalman filter (e.g., Gelb 1974), or its nonlinear extension the extended Kalman filter (EKF; e.g., Evensen 1992), are common examples of filtering methods. Nevertheless, it has been recognized that the EKF is impossible to implement in an operational NWP context, because of the computational burden (Evensen 1992; Keppenne 2000; Lawson and Hansen 2004). Moreover, unbounded error variance growth due to the linearization procedure has been observed (Evensen 1992). Alternatives to the EKF have been proposed, relying on the assumption that “the trajectory of the system state in a high-dimensional dynamical system typically lies on a smaller dimensional subspace of the entire phase space” (Farrell and Ioannou 2001) and, thus, can be called reduced-state filters (Lawson and Hansen 2004). One class of these employs Monte Carlo–based methods, where one generates an ensemble of N forecasts that samples the state space. They provide an estimation of the state’s probability density function (pdf), whose quality is dependent on the size of the ensemble. The generated pdf can naturally be used to initiate an ensemble prediction system, where one performs “stochastic dynamic prediction” (Epstein 1969). This latter approach is possible with the atmosphere’s deterministic laws because of the chaotic nature of the solutions. Monte Carlo filters can be divided into two subclasses: deterministic and stochastic, as stated by Tippett et al. (2003) or Lawson and Hansen (2004). In the stochastic filter (Houtekamer and Mitchell 1998; Burgers et al. 1998), observations are treated as random variables and thus perturbed to avoid an underestimation of the analysis variance (Burgers et al. 1998). This version has shown efficiency in various contexts (Houtekamer and Mitchell 1998; Hansen and Smith 2000; Reichle et al. 2002; Snyder and Zhang 2003). Deterministic formulations of Monte Carlo–based techniques, as reviewed by, for example, Tippett et al. (2003), do not require perturbed observations.

2. Sampling issues

Previous experiments dealing with realistic models have highlighted a few issues concerning EnKf’s operational application: (a) The limited number of members [usually no more than O(100)] can cause spurious covariances between widely separated locations, (b) the unavoidable presence of model errors of different types (presence of unresolved scales, imperfectly known forcing, etc.) can also cause spurious covariances, and (c) the presence of observational error of representation and errors in the statistical description of observations. In the current study, we only address issue a above. However, issue a (as well as issue b) can potentially lead to “underestimation” of the analysis variance. In some cases, this is known to be fatal to the EnKf: the analysis loses track of the truth, whereas its spread remains “too small.” As spread represents the uncertainty, the filter “believes” it performs better than it does in reality. This behavior is called filter divergence (e.g., Maybeck 1979, p. 338) and was observed by, for example, Houtekamer and Mitchell (1998) and Hamill et al. (2001) for the EnKf. We consider it important to clarify the words “too small” and “underestimation.” For this purpose, we recall that “the goal of the assimilation is to produce a random sample of the conditional probability distribution that is consistent with the truth while minimizing the rms error of the ensemble mean from the truth” (Anderson and Anderson 1999). Hence, the aim of Monte Carlo filtering is to satisfy the following two criteria:

  • (i) to produce a sample with an ensemble mean, commonly used as the state’s best estimate, having the smallest expected error, and

  • (ii) to produce a reliable ensemble of analysis, which is to say that the variations of the ensemble mean error, though not accessible, should be well represented by the ensemble of analyses given by the filter itself (in other words, the true state should be statistically indistinguishable from a randomly selected member of the ensemble).

In fact, criterion ii is a necessary condition for a perfect assimilation (Anderson and Anderson 1999). To be sure to satisfy it, a requirement is that the spread among the ensemble members remains representative of the difference between the ensemble mean and the true state, as the assimilation cycle proceeds (Anderson and Anderson 1999; Houtekamer and Mitchell 1998). We define the optimal analysis in the EnKf framework as the one obtained with an infinite number of members, and the ideal analysis as the one satisfying criteria i and ii when using a finite ensemble. Then, as only N members are used, and not an infinite number, the ideal value of the spread is not the one obtained with an infinite ensemble. It is rather represented by the mean squared error of the analysis mean calculated over the N ensemble members, which is itself different from the mean squared error (MSE) of an analysis using an ensemble of infinite size. Now, if the analysis spread calculated over the ensemble of size N under- or overestimates this ideal value, error statistics are not well represented (henceforth “misrepresented”), and criterion ii is not achieved. In fact, the analysis spread of a N-member ensemble systematically underestimates the MSE of the analysis mean calculated over these members.

To illustrate this, results of a twin-experiment simulation of one analysis cycle with the very simple case of a Gaussian process represented by a scalar are shown in Fig. 1. The curves show an average over 105 realizations. Forecasts and observations are taken from a normal distribution of zero mean and unit variance [N(0, 1)] using a pseudorandom number generator, for various ensemble sizes. One can see that the EnKf analysis error variance converges toward the optimal value [in this case σ2a = (1 − k) = 0.5] from below, that is, by negative error values. This result was also found by van Leeuwen (1999) in the multivariate case and by Whitaker and Hamill (2002) in a similar scalar case. On the other hand, the MSE of the ensemble mean converges toward the optimal value of the analysis error variance from above, with positive error values. This shows that, on average, the analysis generally does not satisfy criterion ii. Moreover, the analysis variance underestimates the optimal analysis variance. It subsequently underestimates the ideal value, which corresponds to the MSE of the ensemble mean.

Note that in experimental applications of the EnKf, underestimation of the analysis variance with respect to the optimal value is more often observed than overestimation, whereas it is known that an overestimate is “safer” than an underestimate (Maybeck 1979, p. 339; Daley 1991, section 4.9). Houtekamer and Mitchell (1998) claimed that besides sampling noise related to the limited size of the ensemble, “inbreeding” problems may come into play and also cause spread underestimation. This would refer to the fact that we produce the estimate of the analysis error variation with the same ensemble used to compute the analysis. This is another way to express that the gain and the analysis variance depend nonlinearly on the background error covariance matrix (Whitaker and Hamill 2002). Various techniques have been proposed to correct for inbreeding-sampling errors and stabilize the EnKf. They include covariance localization (Houtekamer and Mitchell 2001; Ott et al. 2004), a hybrid formulation of EnKf–three-dimensional variational data assimilation (3DVAR; Hamill and Snyder 2000), a double-ensemble Kalman filter (DEnKf) (Houtekamer and Mitchell 1998), and covariance inflation (Anderson and Anderson 1999; Hamill et al. 2001; Whitaker and Hamill 2002; Ott et al. 2004; Gillijns et al. 2005). Finally, Pham et al. (1998) used a “forgetting factor” to reduce the observation error covariance matrix 𝗥. We do not present “covariance localization” nor hybrid formulations here, though conclusions may be drawn from our analytic results. In this study, we want to improve the theoretical understanding of the problem of misrepresentation of the analysis variance of the EnKf, as well as investigate the potential of the DEnKf and covariance inflation techniques to produce stable analyses (i.e., not subject to filter divergence).

To quantitatively address the question, we need to compare representative measures of the spread and the error of the ensemble mean. The Gaussian hypothesis for the statistics leads one to naturally calculate the second-order moments of the pdf of the error of the ensemble mean and of the analysis error to verify whether the criteria are satisfied. We consider here an idealized context that intentionally does not include any localization procedure, though Houtekamer and Mitchell (2001) show that localization can be effective in forcing the filter to satisfy criterion ii. Moreover, signal contamination by errors present in an operational context (e.g., model or representation errors) is not considered either. Using the established results, we then propose a theoretically based covariance inflation technique. The technique presented, which modifies the analysis update, turns out to be easier to implement and less costly than the DEnKf, whose performance is also considered. Section 3 describes the formulation of the perturbed-observation EnKf used here. In section 4, we establish the analytical expressions for the average analysis error covariance matrix and the MSE of the ensemble mean for the EnKf. In section 5, similar analytic results are demonstrated for the DEnKf, and a new covariance inflation technique is described. For each analytical result, we give a simple application using the analysis of a Gaussian scalar process. This application, though very simplistic, gives a fairly good insight into the problem. Moreover, we will see in an upcoming second paper that our analytic results compare well to analyses made with a barotropic model.

3. Formulation of the EnKf

The Kalman filter is a “recursive state estimation technique” (Lawson and Hansen 2004). It consists of two steps: (i) a propagation step using the system dynamics to evolve the first two moments of the state’s pdf (characterizing completely a Gaussian distribution) and (ii) an analysis step using a solution derived from Bayes’s rule, giving “either a maximum likelihood estimate (e.g., Lorenc 1986) or a minimum error variance estimate (e.g., within estimation theory; Cohn 1997).” It can also be identified “as a recursive least squares problem properly weighted by the inverses of the relevant error covariance matrices (e.g., Wunsch 1996),” as reviewed by Lawson and Hansen (2004).

Conceptually, the EnKf differs from other Kalman filters only by the way it treats the propagation step. It uses a Monte Carlo method to generate an ensemble of N forecast members ψj, j1, N, that are independent realizations of the process, to sample the state space. Their evolutions between times tk−1 and tk are calculated by using the fully nonlinear model:
i1520-0493-136-8-3035-e1
where ψfj (respectively, ψaj) is the jth member of the forecast (respectively, analysis) ensemble of state vectors. Each member is a state vector of length n, the dimension of the state space. In the general case, a model error term should be present on the right-hand side of (1). In this study, we consider the model as perfect.
This ensemble can subsequently be used to estimate the evolution of any moment of the pdf with respect to time. Particularly,
i1520-0493-136-8-3035-e2
where the overbar stands for the ensemble-based estimate of the mean and ξfjf = ψfj. In addition, 𝗣 fN is an estimate (using N ensemble members) of the background error covariance matrix 𝗣 f = 𝗣 fN=∞, defined as
i1520-0493-136-8-3035-e3
where E{} stands for an expectation value. This expression is itself an estimate of the true background error covariance matrix 𝗣 = E{(ψfψt)(ψfψt)T}, where ψt is the true state. In the EnKf, one sees that the background error is approximated by the departure from the ensemble mean. To simplify the analytic expressions, the notation ξba = ψaψb is extensively used henceforth. Moreover, any matrix refers to 𝗫N, the approximation to 𝗫 = 𝗫N=∞ using N ensemble members. Finally, the true mean value of the process is noted ψμ = E{ψf}. In the stochastic version of the ensemble Kalman filter used in this paper, the analysis step at time tk consists of the following equations:
i1520-0493-136-8-3035-e4
i1520-0493-136-8-3035-e5
where 𝗥, an m × m matrix, is the observation error covariance matrix, m being the number of observations, and 𝗜 is the n × n identity matrix. Also, 𝗛, an m × n matrix, is an interpolation operator, which projects from model space onto observation space, and 𝗞N, an n × m matrix, is the so-called Kalman gain matrix. Here, ∀j1, N, and dj = d + δj, where δj is a random vector taken from N(0, 𝗥), and d is the actual measurement vector. Then, as in (2), we have for the analysis error covariance matrix:
i1520-0493-136-8-3035-e6
The ensemble of maximum likelihood (or minimum error variance) estimates is thus a linear combination of the two available fields, namely background and observation fields. Note that in this formulation, one keeps a second-order moment closure. This is not mandatory with Monte Carlo filters, because theoretically one has access to higher-order-moment estimates. Moreover, a linearized version 𝗛 of the generally nonlinear interpolation operator H has been used here.

This definition leads to an interpretation of the EnKf as a Monte Carlo method where the ensemble of model states evolves in state space, with the mean as the best estimate and the error variance as an estimate of the spread. Observations are similarly represented by another ensemble, generated by the addition of random noise sampled from the observational error distribution (Burgers et al. 1998). Hence, at measurement times, each ensemble member is updated using a different observation taken from this ensemble of perturbations. Here, 𝗞N is the Kalman gain evaluated over the ensemble. Note the nonlinear dependence in (5) of 𝗞N on 𝗣 fN.

4. On sampling errors in the EnKf

In this section we want to establish a general expression for the average analysis error covariance matrix and for the MSE of the ensemble mean in terms of the optimal values 𝗣a and 𝗞 (obtained for an ensemble of infinite size). The MSE is the second-order moment of the pdf of the difference between the ensemble mean and the truth and, therefore, is a measure of the analysis accuracy [as used in Houtekamer and Mitchell (1998) and Charron et al. (2006)]. We introduce here the new matrix ΔaN:
i1520-0493-136-8-3035-e7
where ψt is the unknown true state, assumed to be a freely evolving solution of the system. We will also use the equivalent definition for the forecast, using superscript f instead of a. As suggested by Houtekamer and Mitchell (1998), the EnKf satisfies criterion ii (i.e., produces a representative analysis ensemble of the actual error variation) if tr(ΔaN) ≈ tr(𝗣aN). We proceed to this comparison in order to discuss the performance of the EnKf.

a. Averaged analysis error covariance matrix

Following van Leeuwen (1999), let us now assume that the ensemble estimates are not too far off; hence, we express the statistical error with respect to the unknown true background covariance matrix 𝗣f as ϵ: 𝗣 fN = 𝗣 f + ϵ, with ‖ϵ‖ ≪ ‖𝗣f‖. We want to express the analysis error covariance matrix 𝗣aN as a series expansion in ϵ. Similarly, we suppose = 𝗥 + ρ with ‖ρ‖ ≪ ‖𝗥‖. We shall assume E{ϵ} = 0n and E{ρ} = 0m, where 0n and 0m are, respectively, the null matrix for the state and observation spaces.

Henceforth, in order to simplify the algebra, we define 𝗟N = 𝗜 − 𝗞N𝗛, as well as ΠaN = 𝗟N𝗣 fN. Also, Θ = (𝗛𝗣f𝗛T + 𝗥) and Φ is the inverse of Θ. Then, the optimal Kalman gain, given by (5) when taking an infinite ensemble, can be written 𝗞 = 𝗣f𝗛TΦ, and its transpose 𝗞T = (𝗣f𝗛TΦ)T = Φ𝗛𝗣f. We finally define the gainlike matrix κ = ϵ𝗛TΦ.

Substituing (4) and the mean of (4) into (6), one obtains (Burgers et al. 1998; Whitaker and Hamill 2002; Furrer and Bengtsson 2007)
i1520-0493-136-8-3035-e8
Hence, while the perturbed observations avoid systematic error covariance underestimation with respect to the optimal analysis, as shown by Burgers et al. (1998), they introduce sampling errors represented by the last three terms in (8). Then, since ‖ϵ‖ ≪ ‖𝗣f‖, we have ‖𝗛κ‖ ≪ ‖𝗜m‖ and orders bigger than 2 in 𝗛κ and/or ρ will be systematically discarded below. Using this approximation in the Taylor expansions (A1) and (A2), given in appendix A, leads to
i1520-0493-136-8-3035-e9
i1520-0493-136-8-3035-e10
where we have considered 𝗣a = 𝗟𝗣f as the optimal analysis error covariance matrix and ‖ ‖ is the matrix norm of interest.
One may calculate the expectation of (8), or equivalently a mean over many independent realizations of an analysis cycle using N ensemble members, denoted as 〈.〉. If we suppose 〈〉 = 0n,m, that is, that the observation and the forecast error are uncorrelated (a common assumption in DA), we are left with
i1520-0493-136-8-3035-e11
i1520-0493-136-8-3035-e12
that is to say, the sum of the optimal value and a negative term arising from the limited-size ensemble, implying that the analysis variance and the gain experience a systematic negative bias, as noted by van Leeuwen (1999) and Furrer and Bengtsson (2007). This result is consistent with the general observation that ensemble spread is too low and that the filter can ultimately diverge by putting too much weight on the background. Here, 〈κΘκT〉 = 〈ϵ𝗛TΦ𝗛ϵ〉 represents the variance of the error variance, weighted by Φ = Θ−1. For a Gaussian process, it can be shown (see appendix B) that
i1520-0493-136-8-3035-eq1
then,
i1520-0493-136-8-3035-e13
Note that for the straightforward case of a scalar, we have tr(𝗣a𝗟𝗞𝗛) + tr(𝗣a𝗟)tr(𝗞𝗛) = 2σ2a(1 − k)k, where σa and k are the optimal analysis variance and Kalman gain, respectively. Then, 〈σ2a(N)〉 ≈ σ2a[1 − 2(1 − k)k(N − 1)−1], which is plotted in Fig. 1 and matches with the simulation quite closely.

b. Mean squared error of the ensemble mean

We consider here a twin experiment in which one of the realizations, ψt, is arbitrarily chosen to be the truth, and an ensemble set of N realizations, ψj, j1, N, is used to calculate the statistical properties of the process. Taking the expectation of ΔaN = ξatξatT, assuming a Gaussian process, and using (9), and (C1) from appendix C, we get (details can be found in appendix D)
i1520-0493-136-8-3035-e14
which is the sum of three terms: (i) the optimal analysis variance 𝗣a, (ii) the sampling variance 𝗣a/N, and (iii) a systematic bias similar but opposite to the one in (12). The first two terms give the regular variance of the error variance term [as seen in (C1) for the background analysis variance], and the third one comes from the nonlinear dependence of 𝗞N on 𝗣 fN. Unlike for the analysis covariance in (12), we note that the MSE is converging toward the optimal value 𝗣a with a positive error. In absolute value, this error is on average larger than the error made on the analysis covariance by the additional factor 𝗣a/N. Taking the trace of (14) leads to
i1520-0493-136-8-3035-e15
When compared with (13), we see that criterion ii is generally not satisfied and that the average analysis variance underestimates the MSE of the ensemble mean. For a scalar process, we have 〈δ2a(N)〉 ≈ (1 + N−1)σ2a[1 + 2(1 − k)k(N − 1)−1], which again matches with the result of the simulation presented in Fig. 1.

Order 2 truncation is a very good approximation for 〈𝗣aN〉 even for small values of N. In this case, higher-order terms almost cancel each other out, though they are not negligible individually. Although it is almost not perceptible in the scalar case, this is not exactly true in the expression of 〈ΔaN〉, for which higher-order terms in ϵ may have to be added.

5. Proposed solutions: DEnKf and covariance inflation

The results of the previous section show the impossibility for the EnKf to satisfy criterion ii, and therefore that it is naturally subject to divergence for small ensembles. The double-ensemble Kalman filter (DEnKf; Houtekamer and Mitchell 1998), where the N-member ensemble is split into two N/2-member subsets is an attempt to fix this problem. The covariance information from one of the subsets is used in the data assimilation of the other subset. By using one subensemble to calculate the background error covariance needed to update the forecast of the other subensemble, the DEnKf is expected to remove the negative bias in the analysis error variance described in the last section, because the dominant term in (10), −𝗟κΘκT, responsible for the systematic negative bias in the analysis error variance described in the last section, will average to zero.

The authors further suggested that the ensemble of size N can be divided in l subensembles of sizes N/l, updating each N/l-member subset by using the other one [of size (l − 1)N/l] to calculate 𝗞N. As it still implies two ensembles, we will later refer to this as lDEnKf (l2, N), a filter using l independent subensembles. To attain criterion ii perfectly, Houtekamer and Mitchell (1998) also suggested the use of a NDEnKf, therefore updating each member with a gain matrix estimated from all other members of the ensemble (though they did not test this formulation). Note that a lDEnKf theoretically requires the calculation of l Kalman gain matrices and may therefore be more expensive than the single-ensemble version.

Alternatively, some authors correct directly for the spread misrepresentation by using the so-called covariance inflation technique, where the ensemble-based covariances are multiplied by a tunable factor r ≥ 1 (Anderson and Anderson 1999; Whitaker and Hamill 2002; Ott et al. 2004; Anderson 2001). As seen previously, in order to satisfy criterion ii, these methods should seek an enhanced analysis variance that is as close as possible to the ideal value (and not as close as possible to the optimal value), although it is necessarily suboptimal due to the use of an ensemble of limited size. We stress here that this inflation is a priori dependent on N as well as on 𝗞, as verified below.

In section 5a, we examine the ability of the DEnKf to produce an ideal analysis variance in a perfect model context by extending our theoretical analysis. In section 5b, this is employed to formulate covariance inflation and randomly perturbed analysis methods giving, on average, an analysis that satisfies criteria i and ii.

a. Sampling errors in the DEnKf

Here, we examine separately the quality of the partitioned and the merged analysis. This distinction only stands for the way the analysis is used once produced. The former considers the whole ensemble given by merging all of the subensembles, whereas the latter considers each subensemble separately. In a partitioned analysis, each subensemble is defined at the first analysis cycle and evolves independently. In a merged analysis, the l subensembles of size N/l are unified in a single one of size N at the end of each analysis, and then randomly redivided. We chose to do so because authors have used both types of presentations in previous studies. We treat the 2DEnKf fully, and infer the result for the general case of an lDEnKf.

1) Partitioned analysis

The equivalent of (8) for the analysis error covariance matrix of the first ensemble in the 2DEnKf is
i1520-0493-136-8-3035-e16
where 𝗣a1, 𝗣 f1, 𝗣 f2, 𝗞1, 𝗞2 are, respectively, the analysis, and the forecast error covariance matrices and Kalman gains calculated over the indicated subensemble, and 𝗜 is the m × m identity matrix. The j in ξfjf and δj, as well as the overbar refer to the first ensemble. Furthermore, 𝗟2 = 𝗜 − 𝗞2𝗛, Πa1 = 𝗟2𝗣 f1, and = 𝗥 + ρ1. An equivalent expression can be written for 𝗣a2. We follow here van Leeuwen (1999) in his attempt to write an equivalent of (12) for this case, as well as Houtekamer and Mitchell (1999), considering 𝗣 f1 = 𝗣 f + ϵ1 and 𝗣 f2 = 𝗣 f + ϵ2. Neglecting terms of orders higher than two in (A3) and (A1), we have the following expressions:
i1520-0493-136-8-3035-e17
i1520-0493-136-8-3035-e18
(see proof in appendix A). When comparing (18) with (10), we can see that the term κ2ΘκT1 may explain the inbreeding effect suggested by Houtekamer and Mitchell (1998). When assuming the two sets of members to be independent, 〈−κ2ΘκT1〉 = 0m and the average effect disappears, as was already predicted by van Leeuwen (1999). Van Leeuwen (1999) showed that the error term in (18) is partly due to the errors in the ensemble covariance in combination with those in the gain, and partly due to the nonlinearity of the gain itself. The first type of errors, represented by the term κ2ΘκT1, leads to inbreeding, while the latter represented by the term κ2ΘκT2 are of different sign and partly compensate the inbreeding effect. In the DEnKF the inbreeding part disappears, but the second effect is still present. Then, taking into account the other terms in (16), and averaging leads to (see appendix E)
i1520-0493-136-8-3035-e19
i1520-0493-136-8-3035-e20
Equivalent and symmetric expressions can be written for 〈𝗞1〉 and 〈𝗣a2〉. By comparing this expression with (12), one can see that the sign of the error with respect to the optimal analysis covariance matrix 𝗣a is reversed. Hence, by cutting the ensemble in two, one reverses the systematic bias (with respect to optimality) of the analysis error covariance matrix from negative to positive. On the other hand, the gain still has a systematic negative bias. For the 2DEnKf, each subensemble has N/2 members. Hence, using (B1), it follows that
i1520-0493-136-8-3035-e21
In the more general case of an lDEnKf, each subensemble is of size N/l. Equation (20) remains valid with tr(𝗟〈κ2ΘκT2〉𝗟T) = [l−1(l − 1)N − 1]−1[tr(𝗣a𝗟𝗞𝗛) + tr(𝗣a𝗟)tr(𝗞𝗛)], leading to
i1520-0493-136-8-3035-e22
with l2, N and p1, l. Then, the bigger l is, the closer to the optimal analysis variance Pa is.
Now, to evaluate the MSE 〈ΔaN〉 for the first subensemble, we note that the third term of (14) comes from the gain used to compute the analysis. With the 2DEnKf, it follows that
i1520-0493-136-8-3035-e23
i1520-0493-136-8-3035-e24
Comparing (15) and (24), we see that the analysis is less accurate than with the EnKf. This means that criterion i is not satisfied. Then, comparing (21) and (24), we see that criterion ii is generally not satisfied either, because the average analysis variance underestimates the MSE of the ensemble mean. These two results are consistent with the simulation results of Houtekamer and Mitchell (1998).
When generalized to the lDEnKf, l, p2, N × 1, l, (24) becomes
i1520-0493-136-8-3035-e25
Because of the term [l tr(𝗣a)]/N, the accuracy can be substantially degraded. Moreover, the discrepancy between (22) and (25) increases with l. Consequently, a partitioned analysis is not likely to take full advantage of a multiple division of the ensemble. The analysis error variance remains smaller than the mean squared error of the ensemble mean.

We stress here that for small values of N, the second-order truncation is still a good one for tr〈𝗣a1〉, but not as precise as for the EnKf. Indeed, higher-order terms will not compensate because those containing products of ϵ1 and ϵ2 average to zero, whereas the others do not.

2) Merged analysis

We define 𝗣aN(1,2) as the analysis error covariance matrix using the whole ensemble given by the 2DEnKf analysis (i.e., merging the two subensembles to get only one). It can be proven (see appendix F) that
i1520-0493-136-8-3035-e26
where the second and third terms look like the second term in (20), but weighted differently. Taking the trace leads to
i1520-0493-136-8-3035-e27
The generalized version for an lDEnKf has the following form:
i1520-0493-136-8-3035-e28
We now define Δa(1,2)N = ξatξaTt as the error of the mean calculated over the whole ensemble for the 2DEnKf. Considering = 0.5( + ), and taking the average, it follows after some calculation (details can be found in appendix G) that
i1520-0493-136-8-3035-e29
i1520-0493-136-8-3035-e30
Recalling (15), one notices that for N is not too small tr[〈Δa(1,2)N〉] ≈ tr(〈ΔaN〉). This means that cutting the ensemble in two has almost no effect on the precision of the filter when using the mean over the whole ensemble. This result can be generalized to the lDEnKf:
i1520-0493-136-8-3035-e31
By comparing with (28), we see that criterion ii can theoretically be verified if there exists an l2, N such that
i1520-0493-136-8-3035-e32
Nevertheless, the solution is problem dependent, and it is not clear how to find a general solution for the optimal value of l.

Figure 2 shows the results obtained with a scalar process and a 2DEnKf. One can see that in this case the analysis variance given by the merged analysis matches very well the error of the ensemble mean on average. This means that in this case (k = 0.5), 2 is the optimal value for l, a value effectively given by solving the scalar version of (32). On the other hand, the partitioned analysis (the first ensemble in this case) gives an analysis with a degraded accuracy, whose variance is still underestimating the MSE of the ensemble mean. However, as predicted above, it overestimates the optimal value. Furthermore, the experimental results match the theoretical curves quite closely, as in the case of the EnKf. This is a good indication of the accuracy of the approximations made. Higher values of l have also been tested, namely l = 4 and l = 8. The results shown in Figs. 3 and 4 are also quite consistent with the theory. Figure 5 shows the simulation for the extreme case of an NDEnKf. As predicted, the analysis variance underestimates the MSE of the ensemble mean. The error with respect to the optimal value is almost exactly equal and opposite to the one given by the EnKf.

We see that the use of a DEnKf is a very good alternative to the EnKf. It can give an analysis that satisfies criteria i and ii, which is impossible with the EnKf. Nevertheless, it appears that the improvement is not as great when one divides the ensemble in more than two in this case.

b. Optimal covariance inflation

To correct for systematic sampling errors in 𝗣aN and to satisfy criterion ii, some authors multiply the forecast or analysis covariances by an inflation factor r. To properly satisfy criterion ii, we have seen that this factor is necessarily dependent on the ensemble size. Instead of using a blindly chosen factor, we propose here two inflation techniques, namely inflation of the deviations from the background before analysis, and addition of random errors after analysis.

The analyses given by the two methods on average satisfy criteria i and ii. They do not lead to underestimated error covariances. We therefore expect these analyses to be less subject to the problem of filter divergence.

We follow here Hamill et al. (2001), inflating the deviations from the ensemble-mean first guess by a constant factor r for each member of the ensemble at the end of the propagation step, that is, = ψfj + (r − 1)(ψfj), implying + (r − 1)𝗟Nξfjf, which is equivalent to a multiplication of the background error covariance matrix 𝗣 fN by r2. This inflation technique does not affect the forecast and analysis ensemble means; therefore, = 〈ΔaN〉. The average of the new mean analysis covariance matrix is
i1520-0493-136-8-3035-eq2
Satisfying criterion ii requires tr() ≈ tr(); that is,
i1520-0493-136-8-3035-e33
which can be approximated by judiciously using the value 𝗣aN, 𝗟N, 𝗞N as well as an expression for 〈𝗟N𝗣aN〉 similar to (12). In the scalar case, rN takes the simple form
i1520-0493-136-8-3035-e34
This expression gives a “flow dependent” covariance inflation, which is a priori different for each analysis cycle. In Fig. 6a, we see that the application of (34) at k = ½ gives an analysis that satisfies criterion ii on average. Figure 6b shows that a constant value for r is optimal for one particular value of N. Figure 7 displays isolines of r given by (34). Typical inflation values of a few percent are observed for k ≤ 0.8. This is consistent with the values used by, for example, Anderson and Anderson (1999) and Whitaker and Hamill (2002).
Instead of using the inflation factor r, some authors (Gillijns et al. 2005; Corazza et al. 2002) suggest perturbing the analysis by adding a random vector, in the way that perturbed observations are added to increase the analysis variance in Burgers et al. (1998). Added vectors can be chosen such that criterion ii is satisfied. Then, we suggest
i1520-0493-136-8-3035-e35
where μ, υ are two independent random scalars of zero mean and unit variance. Then, we have = ; that is to say = 〈ΔaN〉. With this formulation, we have and therefore criterion ii is satisfied. Simulations applying (35) to the scalar problem (not shown) yield a similar quality as in Fig. 6. This formulation may be easier to implement than the covariance inflation factor because of the absence of additive matrix calculations.

6. Conclusions

Theoretical expressions have been established for the analysis error covariance matrices and the mean squared error of the ensemble mean given by various versions of a perturbed-observation EnKf. The performance of each version is examined by verifying whether it produces a sample with an ensemble mean having the smallest error (a requirement termed criterion i) and if the MSE of the ensemble mean is well represented by the spread drawn from the ensemble of analyses given by the filter (termed criterion ii). For the EnKf used with a finite ensemble, our results show that it never satisfies criterion ii fully. That is, the analysis spread among the ensemble members always underestimates the mean squared difference between the ensemble mean and the true state and the filter is therefore naturally subject to divergence. The possibility for the double-ensemble Kalman filter technique to properly correct this problem was studied. We showed that an lDEnKf (i.e., cutting the ensemble in l subensembles of size N/l) can potentially give adequate analyses when using the ensemble resulting from the merging of all the subensembles. Nevertheless, the best value of l is problem dependent, typically depending on the observation density and frequency. In addition, using an lDEnKf requires the computation of l Kalman gains, whose cost may be prohibitive. For the case of a partitioned lDEnKf, that is using each subensemble independently, we showed that the analysis never satisfies criterion ii, though performing relatively better than the EnKf for small values of l, but at the expense of a degradation of the accuracy of the ensemble mean (criterion i is not satisfied) whose level increases with l. To correct for the misrepresentation problems and satisfy criteria i and ii on average, we presented flow-dependent techniques such as an optimal covariance inflation, and a randomly perturbed analysis. They are motivated by the present theoretical analysis, which takes into account the dependence of criterion ii on the number of members and the Kalman gain. We expect these methods to be less subject to the problem of filter divergence. It should be stressed here that all of the calculations have been done in an idealized context, ignoring model or representativity errors. This can virtually invalidate some of our results and would affect our theoretical value for the inflation factor r. It is also important to keep in mind that the analytic results concern the means over a large number of realizations of analyses. For one given analysis, they might not be so relevant—especially if the variance of the squared error of the ensemble mean is relatively high. In NWP, as one works with state spaces of high dimensions, we are confident that the averages will have some significance. We should also stress that the whole theory developed here follows the traditional assumption of a Gaussian process. As a result, we may expect correction to r in a strongly nonlinear context. Nevertheless, we expect our results to better guide the user in the determination of the r factor in an operational context, previously chosen by trial and error (Anderson and Anderson 1999). We have performed simulations using a very nonlinear two-dimensional turbulence model that have led to very encouraging results regarding the capacity of these techniques to avoid filter divergence. They will be presented in the forthcoming Part II of this paper.

Acknowledgments

This work would not have been possible without the contributions of Peter Houtekamer and Herschel Mitchell, with whom we had hours of stimulating and fruitful discussions. We thank two anonymous reviewers for their very helpful and productive comments. We acknowledge financial support from the Canadian Foundation for Climate and Atmospheric Sciences (CFCAS) in the form of Grant GR-500B.

REFERENCES

  • Anderson, J. L., 2001: An ensemble adjustment Kalman filter for data assimilation. Mon. Wea. Rev., 129 , 28842903.

  • Anderson, J. L., and S. L. Anderson, 1999: A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts. Mon. Wea. Rev., 127 , 27412758.

    • Search Google Scholar
    • Export Citation
  • Burgers, G., P. J. van Leeuwen, and G. Evensen, 1998: Analysis scheme in the ensemble Kalman filter. Mon. Wea. Rev., 126 , 17191724.

  • Charron, M., P. L. Houtekamer, and P. Bartello, 2006: Assimilation with an ensemble Kalman filter of synthetic radial wind data in anisotropic turbulence: Perfect model experiments. Mon. Wea. Rev., 134 , 618637.

    • Search Google Scholar
    • Export Citation
  • Cohn, S., 1997: An introduction to estimation theory. J. Meteor. Soc. Japan, 75 , 257288.

  • Corazza, M., E. Kalnay, D. J. Patil, E. Ott, J. Yorke, B. R. Hunt, I. Szunyogh, and M. Cai, 2002: Use of the breeding technique in the estimation of the background error covariance matrix for a quasi-geostrophic model. Preprints, Symp. on Observations, Data Assimilation and Probabilistic Prediction, Orlando, FL, Amer. Meteor. Soc., 6.4. [Available online at http://ams.confex.com/ams/pdfpapers/28755.pdf.].

  • Daley, R., 1991: Atmospheric Data Analysis. Cambridge University Press, 457 pp.

  • Epstein, E., 1969: Stochastic dynamic prediction. Tellus, 21 , 739759.

  • Evensen, G., 1992: Using the extended Kalman filter with a multilayer quasi-geostrophic ocean model. J. Geophys. Res., 97 (C11) , 1790517924.

    • Search Google Scholar
    • Export Citation
  • Evensen, G., 1994: Inverse methods and data assimilation in nonlinear ocean models. Physica D, 77 , 108129.

  • Evensen, G., 2007: Data Assimilation: The Ensemble Kalman Filter. Springer, 280 pp.

  • Farrell, B., and P. Ioannou, 2001: State estimation using a reduced-order kalman filter. J. Atmos. Sci., 58 , 36663680.

  • Furrer, R., and T. Bengtsson, 2007: Estimation of high-dimensional prior and posterior covariance matrices in Kalman filter variants. J. Multivariate Anal., 98 , 227255.

    • Search Google Scholar
    • Export Citation
  • Gelb, A., 1974: Applied Optimal Estimation. The MIT Press, 382 pp.

  • Gillijns, S., D. S. Bernstein, and B. D. Moor, 2005: The reduced rank transform square root filter for data assimilation. Proc. 14th IFAC Symp. on System Identification (SYSID-2006), Newcastle, Australia, Int. Federation on Automatic Control, FrB2.5.

  • Hamill, T. M., and C. Snyder, 2000: A hybrid ensemble Kalman filter–3D variational analysis scheme. Mon. Wea. Rev., 128 , 29052919.

  • Hamill, T. M., J. S. Whitaker, and C. Snyder, 2001: Distance-dependent filtering of background error covariance estimates in an ensemble Kalman filter. Mon. Wea. Rev., 129 , 27762790.

    • Search Google Scholar
    • Export Citation
  • Hansen, J. A., and L. A. Smith, 2000: The role of operational constraints in selecting supplementary observations. J. Atmos. Sci., 57 , 28592871.

    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126 , 796811.

    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 1999: Reply. Mon. Wea. Rev., 127 , 13781379.

  • Houtekamer, P. L., and H. L. Mitchell, 2001: A sequential ensemble Kalman filter for atmospheric data assimilation. Mon. Wea. Rev., 129 , 123137.

    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 2005: Ensemble Kalman filtering. Quart. J. Roy. Meteor. Soc., 131 , 32693289.

  • Keppenne, C., 2000: Data assimilation into a primitive-equation model with a parallel ensemble Kalman filter. Mon. Wea. Rev., 128 , 19711981.

    • Search Google Scholar
    • Export Citation
  • Lawson, W. G., and J. A. Hansen, 2004: Implications of stochastic and deterministic filters as ensemble-based data assimilation methods in varying regimes of error growth. Mon. Wea. Rev., 132 , 19661981.

    • Search Google Scholar
    • Export Citation
  • Lorenc, A. C., 1986: Analysis methods for numerical weather prediction. Quart. J. Roy. Meteor. Soc., 112 , 11771194.

  • Maybeck, P. S., 1979: Stochastic Models, Estimation, and Control. Vol. 1. Mathematics in Science and Engineering, Vol. 141-1, Academic Press, 442 pp.

    • Search Google Scholar
    • Export Citation
  • Ott, E., and Coauthors, 2004: A local ensemble Kalman filter for atmospheric data assimilation. Tellus, 56A , 415428.

  • Pham, D. T., J. Verron, and M. C. Roubaud, 1998: A singular evolutive extended Kalman filter for data assimilation in oceanography. J. Mar. Syst., 16 , 323340.

    • Search Google Scholar
    • Export Citation
  • Reichle, R. H., D. B. McLaughlin, and D. Entekhabi, 2002: Hydrologic data assimilation with the ensemble Kalman filter. Mon. Wea. Rev., 130 , 103114.

    • Search Google Scholar
    • Export Citation
  • Snyder, C., and F. Zhang, 2003: Assimilation of simulated Doppler radar observations with an ensemble Kalman filter. Mon. Wea. Rev., 131 , 16631677.

    • Search Google Scholar
    • Export Citation
  • Tippett, M. K., J. L. Anderson, C. H. Bishop, T. M. Hamill, and J. Whitaker, 2003: Ensemble square root filters. Mon. Wea. Rev., 131 , 14851490.

    • Search Google Scholar
    • Export Citation
  • van Leeuwen, P. J., 1999: Comment on “Data assimilation using an ensemble Kalman filter technique”. Mon. Wea. Rev., 127 , 13741377.

    • Search Google Scholar
    • Export Citation
  • Whitaker, J. S., and T. M. Hamill, 2002: Ensemble data assimilation without perturbed observations. Mon. Wea. Rev., 130 , 19131924.

  • Wunsch, W., 1996: The Ocean Circulation Inverse Problem. Cambridge University Press, 458 pp.

APPENDIX A

Taylor Expansion of 𝗞N and aN in the EnKf

Substituting 𝗣 fN = 𝗣 f + ϵ in (5) leads to (𝗜m + 𝗛κ)N = (𝗣f + ϵ)𝗛TΦ (𝗜m + 𝗛κ)−1. We suppose ‖𝗛κ‖ ≪ ‖𝗜m‖, then ρ(𝗛κ) < 1 (where ρ is the spectral radius), and the Taylor expansion for (𝗜m + 𝗛κ)−1 is
i1520-0493-136-8-3035-eqa1
where ‖ ‖ is the matrix norm of interest. It leads to the following expressions for 𝗞N and ΠaN:
i1520-0493-136-8-3035-ea1
and
i1520-0493-136-8-3035-ea2
For the DEnKf, we have 𝗞2 = 𝗣 f2𝗛T(𝗛𝗣f2𝗛T + 𝗥)−1 and Πa1 = 𝗟2𝗣 f1, and then
i1520-0493-136-8-3035-ea3
Here, 𝗞2 (respectively, K1) has the same expression as 𝗞 N, replacing κ by κ2 (respectively, κ1). The expression for Πa2 = 𝗟1𝗣 f2 is obtained by symmetry.

APPENDIX B

Calculation of tr(〈𝗟κΘκT𝗟T〉) for a Gaussian Multivariate Process

We know (e.g., Maybeck 1979, p. 130) that var(σ2s2) = 2σ4(N − 1)−1 for a Gaussian scalar process, if σ2 and s2 are the true and the unbiased estimated variances. To generalize this result to the multivariate case, as was done in Furrer and Bengtsson (2007), we use the spectral decomposition of 𝗣f = ΓΛΓT, where Λ is an n × n diagonal matrix with terms λii1, n, and define ΛN = ΓT𝗣 fNΓ (in general not diagonal) as an unbiased estimator of Λ. For any matrix 𝗔 and 𝗕, we have
i1520-0493-136-8-3035-eqb1
We consider now (𝗫)i,j as the general term of 𝗫 and define (Γ)i,j = γi,j, (ΓT)i,j = γi,j, (𝗔)i,j = ai,j, (𝗕)i,j = bi,j, (𝗣 fN)i,j = , (𝗣 f)i,j = pi,j, (Λ)i,j = λi,j, (ΛN)i,j = , and (ΛNΓT𝗔Γ)ij(ΓT𝗕ΓΛN)kl = fi,j,k,l. This leads to
i1520-0493-136-8-3035-eqb2
where we have used = pq,ppr,s + (N − 1)−1(pq,spp,r + pq,rpp,s). This leads to
i1520-0493-136-8-3035-eqb3
and
i1520-0493-136-8-3035-eb1
where we have replaced 𝗔 by 𝗜 and 𝗕 by 𝗛TΦ𝗛, and have used the trace properties of a product of matrices.

APPENDIX C

Calculation of ΔfN for a Gaussian Process

We consider ΔfN = ξftξfTt, where ξft = ψt, and introduce ψμ = E{ψ} ≈ 〈ψ〉, the expectation value of the process (or the value obtained when the ensemble is of infinite size). Then, if ξμt = ψtψμ, and ξfμ = ψμ, it follows that ΔfN = (ξμt + ξfμ)(ξμt + ξfμ)T. Taking the average yields 〈ΔfN〉 = 〈ξμtξμTt〉 + 〈ξμtξfTμ〉 + 〈ξfμξμTt〉 + 〈ξfμξfTμ〉. Considering that the truth ψt is independent of each realization, and that its deviation with respect to the expectation value is independent of the deviation of each realization with respect to the expectation value, we have 〈ξμtξfTμ〉 = 〈ξfμξμTt〉 = 0n, and, finally, for a Gaussian process,
i1520-0493-136-8-3035-ec1

APPENDIX D

Taylor Expansion for ΔaN

Taking the mean of (4), and providing d − 𝗛 = 𝗛ξftδt, with δt = Hψtd, it follows that ΔaN = [ξft − 𝗞N(𝗛ξftδt)][ξft − 𝗞N(𝗛ξftδt)]T. Considering 〈ξftδTt〉 = 0nm leads to
i1520-0493-136-8-3035-eqd1
We assume that the background and observation errors are not correlated with the sampling error ϵ (e.g., 〈κδtδTt〉 = 〈κ〉〈δtδTt〉, 〈κξftξfTt〉 = 〈κ〉〈ΔfN〉). We recall that 〈κ〉 = 0m. We use (C1) and 〈δtδTt〉 = R(1 + N−1). Then, replacing 𝗞N by the expression found in (A1), and noticing that sums cancel out each other because of the identity 𝗞𝗥 = 𝗟𝗣f𝗛T, we finally have
i1520-0493-136-8-3035-ed1

APPENDIX E

Taylor Expansion for 𝗣aN in the Partitioned DEnKf Method

In (16), the last three terms average to zero. Then, using expressions for 𝗞1 and 𝗞2 given by (17) and using (𝗛κΘ)T = ΘκT𝗛T in the second term leads to
i1520-0493-136-8-3035-eqe1
Using this last expression, then (16) and (18), taking the mean over a large number of realizations, assuming that 〈〉 = 0nm, 〈κ1ΘκT2〉 = 〈ρ1〉 = 0m, and finally that 〈κρ〉 = 0m, one obtains the following for 𝗣a1:
i1520-0493-136-8-3035-eqe2

APPENDIX F

Taylor Expansion for 𝗣aN in the Merged DEnKf Method

Isolating members coming from each ensemble and taking ξa,pt as in appendix G yields 𝗣a(1,2)N = (N − 1)−1(Σ1 + Σ2), with
i1520-0493-136-8-3035-eqf1
Expression of, for example, Σ1 can be obtained by considering that ξa,pjaξa,lTja = ξa,mjaξa,mTja + ξa,ma,pξa,mTa,lξa,mjaξa,mTa,lξa,ma,pξa,mTja, and ΣN/2j=1ξa,1ja = 0. This gives
i1520-0493-136-8-3035-eqf2
Then, taking the average, using (20), (23), and 〈ξa,1tξa,2Tt〉 = 𝗣a + O(‖κ1,2Θκ1,2‖) (see appendix G) gives
i1520-0493-136-8-3035-eqf3
and a symmetric expression for 〈Σ2〉. Further, if we suppose (N − 1)−1κiΘκTi〉 = O(‖κiΘκTi‖), then
i1520-0493-136-8-3035-ef1
The generalized version for an lDEnKF has the following form:
i1520-0493-136-8-3035-ef2
where κp = ϵp𝗛T, ϵp is the error of a background error covariance matrix calculated over all the subensembles but the pth.

APPENDIX G

Taylor Expansion for ΔaN in the Merged DEnKf Method

Considering = 0.5( + ) and ξa,pt = ψt, we have
i1520-0493-136-8-3035-eqg1
The first two terms of this sum are given by (23). By using the expression of and , assuming observation and forecast error are uncorrelated, and 〈ξf,1μξf,2μ〉 = 0n, 〈ξjfμξkfμ〉 = 0n, ∀(j, k) ∈ 1, N2 and jk (ψμ being the expectation value of the process), it follows that
i1520-0493-136-8-3035-eqg2
Then, we use (19) and the symmetric expression for 𝗞1, as well as 〈ξμtξμTt〉 = 𝗣 f and 〈δtδTt〉 = 𝗥. We also assume that κ1 and κ2 are uncorrelated, and as is done in appendix D, we assume that the background and observation errors are not correlated with ϵ. It follows that 〈ξa,1tξa,2Tt〉 = 〈ξa,2tξa,1Tt〉 = 𝗣a + O(‖κ1,2ΘκT1,2‖); therefore,
i1520-0493-136-8-3035-eg1
Generalization to the lDEnKf leads to
i1520-0493-136-8-3035-eg2
where κp is as in appendix F.

Fig. 1.
Fig. 1.

Simulation of one analysis cycle with a Gaussian scalar, taken from N(0, 1). The observation ensemble is taken from N(0, 1) too; hence, the optimal gain K = 0.5, and the optimal analysis variance σ2a = 0.5. Triangles show the MSE of the ensemble mean, and squares show the average analysis ensemble variance. Means are performed over 100 000 realizations. For each experimental curve, the theoretical curves 〈σ2a(N)〉 = σ2a[1 − 2k(1 − k)(N − 1)−1] and 〈δ2a(N)〉 = σ2a(1 + N−1)[1 + 2k(1 − k)(N − 1)−1], the scalar versions of (13) and (15), are drawn in continuous lines.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Fig. 2.
Fig. 2.

Same simulation as in Fig. 1 but for a 2DEnKf. (a) Statistics for the merged analysis. The scalar versions of (27) and (30) are drawn in continuous lines. (b) Statistics for the analysis of the first ensemble. The scalar versions of (21) and (24) are drawn in continuous lines.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Fig. 3.
Fig. 3.

Same simulation as in Fig. 1 but for a 4DEnKf. (a) Statistics for the merged analysis. The scalar versions of (28) and (31) for l = 4 are drawn in continuous lines. (b) Statistics for the analysis of the first ensemble. The scalar versions of (22) and (25) for l = 4 are drawn in continuous lines.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Fig. 4.
Fig. 4.

Same simulation as in Fig. 1 but for an 8DEnKf. (a) Statistics for the merged analysis. The scalar versions of (28) and (31) for l = 8 are drawn in continuous lines. (b) Statistics for the analysis of the first ensemble. The scalar versions of (22) and (25) for l = 8 are drawn in continuous lines.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Fig. 5.
Fig. 5.

Same simulation as in Fig. 1 but for an NDEnKf, N being the number of ensemble members. The scalar versions of (28) and (31) for l = N are drawn in continuous lines.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Fig. 6.
Fig. 6.

Same simulation as in Fig. 1 but for a sampling-error-corrected EnKf, using an inflation factor. (a) The inflation coefficient rN, which depends on the number of ensemble members, has been used. (b) The constant value r = 1.035 has been used. The curves match closely the corresponding analytical curve 〈σ2a(N)〉 = σ2a[1.035 − 2k(1 − k)(N − 1)−1], represented by the continuous line.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Fig. 7.
Fig. 7.

Isolines of the scalar version of the inflation coefficient rN, as a function of the number of ensemble members and the gain k.

Citation: Monthly Weather Review 136, 8; 10.1175/2007MWR2323.1

Save
  • Anderson, J. L., 2001: An ensemble adjustment Kalman filter for data assimilation. Mon. Wea. Rev., 129 , 28842903.

  • Anderson, J. L., and S. L. Anderson, 1999: A Monte Carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts. Mon. Wea. Rev., 127 , 27412758.

    • Search Google Scholar
    • Export Citation
  • Burgers, G., P. J. van Leeuwen, and G. Evensen, 1998: Analysis scheme in the ensemble Kalman filter. Mon. Wea. Rev., 126 , 17191724.

  • Charron, M., P. L. Houtekamer, and P. Bartello, 2006: Assimilation with an ensemble Kalman filter of synthetic radial wind data in anisotropic turbulence: Perfect model experiments. Mon. Wea. Rev., 134 , 618637.

    • Search Google Scholar
    • Export Citation
  • Cohn, S., 1997: An introduction to estimation theory. J. Meteor. Soc. Japan, 75 , 257288.

  • Corazza, M., E. Kalnay, D. J. Patil, E. Ott, J. Yorke, B. R. Hunt, I. Szunyogh, and M. Cai, 2002: Use of the breeding technique in the estimation of the background error covariance matrix for a quasi-geostrophic model. Preprints, Symp. on Observations, Data Assimilation and Probabilistic Prediction, Orlando, FL, Amer. Meteor. Soc., 6.4. [Available online at http://ams.confex.com/ams/pdfpapers/28755.pdf.].

  • Daley, R., 1991: Atmospheric Data Analysis. Cambridge University Press, 457 pp.

  • Epstein, E., 1969: Stochastic dynamic prediction. Tellus, 21 , 739759.

  • Evensen, G., 1992: Using the extended Kalman filter with a multilayer quasi-geostrophic ocean model. J. Geophys. Res., 97 (C11) , 1790517924.

    • Search Google Scholar
    • Export Citation
  • Evensen, G., 1994: Inverse methods and data assimilation in nonlinear ocean models. Physica D, 77 , 108129.

  • Evensen, G., 2007: Data Assimilation: The Ensemble Kalman Filter. Springer, 280 pp.

  • Farrell, B., and P. Ioannou, 2001: State estimation using a reduced-order kalman filter. J. Atmos. Sci., 58 , 36663680.

  • Furrer, R., and T. Bengtsson, 2007: Estimation of high-dimensional prior and posterior covariance matrices in Kalman filter variants. J. Multivariate Anal., 98 , 227255.

    • Search Google Scholar
    • Export Citation
  • Gelb, A., 1974: Applied Optimal Estimation. The MIT Press, 382 pp.

  • Gillijns, S., D. S. Bernstein, and B. D. Moor, 2005: The reduced rank transform square root filter for data assimilation. Proc. 14th IFAC Symp. on System Identification (SYSID-2006), Newcastle, Australia, Int. Federation on Automatic Control, FrB2.5.

  • Hamill, T. M., and C. Snyder, 2000: A hybrid ensemble Kalman filter–3D variational analysis scheme. Mon. Wea. Rev., 128 , 29052919.

  • Hamill, T. M., J. S. Whitaker, and C. Snyder, 2001: Distance-dependent filtering of background error covariance estimates in an ensemble Kalman filter. Mon. Wea. Rev., 129 , 27762790.

    • Search Google Scholar
    • Export Citation
  • Hansen, J. A., and L. A. Smith, 2000: The role of operational constraints in selecting supplementary observations. J. Atmos. Sci., 57 , 28592871.

    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126 , 796811.

    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 1999: Reply. Mon. Wea. Rev., 127 , 13781379.

  • Houtekamer, P. L., and H. L. Mitchell, 2001: A sequential ensemble Kalman filter for atmospheric data assimilation. Mon. Wea. Rev., 129 , 123137.

    • Search Google Scholar
    • Export Citation
  • Houtekamer, P. L., and H. L. Mitchell, 2005: Ensemble Kalman filtering. Quart. J. Roy. Meteor. Soc., 131 , 32693289.

  • Keppenne, C., 2000: Data assimilation into a primitive-equation model with a parallel ensemble Kalman filter. Mon. Wea. Rev., 128 , 19711981.

    • Search Google Scholar
    • Export Citation
  • Lawson, W. G., and J. A. Hansen, 2004: Implications of stochastic and deterministic filters as ensemble-based data assimilation methods in varying regimes of error growth. Mon. Wea. Rev., 132 , 19661981.

    • Search Google Scholar
    • Export Citation
  • Lorenc, A. C., 1986: Analysis methods for numerical weather prediction. Quart. J. Roy. Meteor. Soc., 112 , 11771194.

  • Maybeck, P. S., 1979: Stochastic Models, Estimation, and Control. Vol. 1. Mathematics in Science and Engineering, Vol. 141-1, Academic Press, 442 pp.

    • Search Google Scholar
    • Export Citation
  • Ott, E., and Coauthors, 2004: A local ensemble Kalman filter for atmospheric data assimilation. Tellus, 56A , 415428.

  • Pham, D. T., J. Verron, and M. C. Roubaud, 1998: A singular evolutive extended Kalman filter for data assimilation in oceanography. J. Mar. Syst., 16 , 323340.

    • Search Google Scholar
    • Export Citation
  • Reichle, R. H., D. B. McLaughlin, and D. Entekhabi, 2002: Hydrologic data assimilation with the ensemble Kalman filter. Mon. Wea. Rev., 130 , 103114.

    • Search Google Scholar
    • Export Citation
  • Snyder, C., and F. Zhang, 2003: Assimilation of simulated Doppler radar observations with an ensemble Kalman filter. Mon. Wea. Rev., 131 , 16631677.

    • Search Google Scholar
    • Export Citation
  • Tippett, M. K., J. L. Anderson, C. H. Bishop, T. M. Hamill, and J. Whitaker, 2003: Ensemble square root filters. Mon. Wea. Rev., 131 , 14851490.

    • Search Google Scholar
    • Export Citation
  • van Leeuwen, P. J., 1999: Comment on “Data assimilation using an ensemble Kalman filter technique”. Mon. Wea. Rev., 127 , 13741377.

    • Search Google Scholar
    • Export Citation
  • Whitaker, J. S., and T. M. Hamill, 2002: Ensemble data assimilation without perturbed observations. Mon. Wea. Rev., 130 , 19131924.

  • Wunsch, W., 1996: The Ocean Circulation Inverse Problem. Cambridge University Press, 458 pp.

  • Fig. 1.

    Simulation of one analysis cycle with a Gaussian scalar, taken from N(0, 1). The observation ensemble is taken from N(0, 1) too; hence, the optimal gain K = 0.5, and the optimal analysis variance σ2a = 0.5. Triangles show the MSE of the ensemble mean, and squares show the average analysis ensemble variance. Means are performed over 100 000 realizations. For each experimental curve, the theoretical curves 〈σ2a(N)〉 = σ2a[1 − 2k(1 − k)(N − 1)−1] and 〈δ2a(N)〉 = σ2a(1 + N−1)[1 + 2k(1 − k)(N − 1)−1], the scalar versions of (13) and (15), are drawn in continuous lines.

  • Fig. 2.

    Same simulation as in Fig. 1 but for a 2DEnKf. (a) Statistics for the merged analysis. The scalar versions of (27) and (30) are drawn in continuous lines. (b) Statistics for the analysis of the first ensemble. The scalar versions of (21) and (24) are drawn in continuous lines.

  • Fig. 3.

    Same simulation as in Fig. 1 but for a 4DEnKf. (a) Statistics for the merged analysis. The scalar versions of (28) and (31) for l = 4 are drawn in continuous lines. (b) Statistics for the analysis of the first ensemble. The scalar versions of (22) and (25) for l = 4 are drawn in continuous lines.

  • Fig. 4.

    Same simulation as in Fig. 1 but for an 8DEnKf. (a) Statistics for the merged analysis. The scalar versions of (28) and (31) for l = 8 are drawn in continuous lines. (b) Statistics for the analysis of the first ensemble. The scalar versions of (22) and (25) for l = 8 are drawn in continuous lines.

  • Fig. 5.

    Same simulation as in Fig. 1 but for an NDEnKf, N being the number of ensemble members. The scalar versions of (28) and (31) for l = N are drawn in continuous lines.

  • Fig. 6.

    Same simulation as in Fig. 1 but for a sampling-error-corrected EnKf, using an inflation factor. (a) The inflation coefficient rN, which depends on the number of ensemble members, has been used. (b) The constant value r = 1.035 has been used. The curves match closely the corresponding analytical curve 〈σ2a(N)〉 = σ2a[1.035 − 2k(1 − k)(N − 1)−1], represented by the continuous line.

  • Fig. 7.

    Isolines of the scalar version of the inflation coefficient rN, as a function of the number of ensemble members and the gain k.

All Time Past Year Past 30 Days
Abstract Views 0 0 0
Full Text Views 2348 1885 79
PDF Downloads 456 89 8