1. Introduction
Data assimilation addresses the problem of producing useful analyses and forecasts given imperfect dynamical models and observations. The Kalman filter is the optimal data assimilation method for linear dynamics with additive, state-independent Gaussian model and observation errors (Cohn 1997). An attractive feature of the Kalman filter is its calculation of forecast and analysis error covariances, in addition to the forecasts and analyses themselves. In this way, the Kalman filter produces estimates of forecast and analysis uncertainty, consistent with the dynamics and prescribed model and observation error statistics. However, the error covariance calculation components of the Kalman filter are difficult to implement in realistic systems because of (i) their computational cost, (ii) the nonlinearity of the dynamics, and (iii) poorly characterized error sources.
The ensemble Kalman filter (EnKF), proposed by Evensen (1994), addresses the first two of these problems by using ensemble representations for the forecast and analysis error covariances. Ensemble size limits the number of degrees of freedom used to represent forecast and analysis errors, and Kalman filter error covariance calculations are practical for modest-sized ensembles. The EnKF algorithm begins with an analysis ensemble whose mean is the current state estimate or analysis and whose statistics reflect the analysis error. Applying the full nonlinear dynamics to each analysis ensemble member produces the forecast ensemble; tangent linear and adjoint models of the dynamics are not required. Statistics of the forecast ensemble represent forecast errors; in its simplest form, the EnKF only accounts for forecast error due to uncertain initial conditions, neglecting forecast error due to model deficiencies. The forecast ensemble mean and covariance are then used to assimilate observations and compute a new analysis ensemble with appropriate statistics, and the cycle is repeated. The new analysis ensemble can be formed either stochastically (Houtekamer and Mitchell 1998; Burgers et al. 1998) or deterministically (Bishop et al. 2001; Anderson 2001; Whitaker and Hamill 2002). Deterministic methods were developed to address the adaptive observational network design problem and to avoid sampling issues associated with the use of “perturbed observations” in stochastic analysis ensemble update methods.
The EnKF and other ensemble data assimilation methods belong to the family of square root filters (SRFs), and a purpose of this paper is to demonstrate that deterministic analysis ensemble updates are implementations of Kalman SRFs (Bierman 1977; Maybeck 1982; Heemink et al. 2001). An immediate benefit of this identification is a framework for understanding and comparing deterministic analysis ensemble update schemes (Bishop et al. 2001; Anderson 2001; Whitaker and Hamill 2002). SRFs, like ensemble representations of covariances, are not unique. We begin our discussion in section 2 with a presentation of the Kalman SRF; issues related to implementation of ensemble SRFs are presented in section 3; in section 4 we summarize our results.
2. The Kalman SRF
Kalman SRF algorithms, originally developed for space-navigation systems with limited computational word length, demonstrate superior numerical precision and stability compared to the standard Kalman filter algorithm (Bierman 1977; Maybeck 1982). SRFs by construction avoid loss of positive definiteness of the error covariance matrices. SRFs have been used in earth science data assimilation methods where error covariances are approximated by truncated eigenvector expansions (Verlaan and Heemink 1997).
The forecast and analysis error covariance matrices are symmetric positive-definite matrices and can be represented as
Covariance matrix square roots are closely related to ensemble representations. The sample covariance
3. Ensemble SRFs
a. Analysis ensemble
Beginning with the same forecast error covariance, the direct, serial, ETKF, and EAKF methods produce different analysis ensembles that span the same state-space subspace and have the same covariance. Higher-order statistical moments of the different models will be different, a relevant issue for nonlinear dynamics. The computation costs of the direct, ETKF, and EAKF methods are seen in Table 1 to scale comparably (see the appendix for details). There are differences in precise computational cost; for instance, the EAKF contains an additional singular value decomposition (SVD) calculation of the forecast with cost O(m3 + m2). The computational cost of the serial filter is less dependent on the rank of the forecast error covariance and more sensitive to the number of observations. This difference is important when techniques to account for model error and control filter divergence, as described in the next section, result in an effective forecast error covariance dimension m much larger than the dynamical forecast ensemble dimension.
b. Forecast error statistics
In the previous section we examined methods of forming the analysis ensemble given a matrix square root of the forecast error covariance. There are two fundamental problems associated with directly using the ensemble generated by (3). First, ensemble size is limited by the computational cost of applying the forecast model to each ensemble member. Small ensembles have few degrees of freedom available to represent errors and suffer from sampling error that further degrades forecast error covariance representation. Sampling error leads to loss of accuracy and underestimation of error covariances that can cause filter divergence. Techniques to deal with this problem are distance-dependent covariance filtering and covariance inflation (Whitaker and Hamill 2002). Covariance localization in the serial method consists of adding a Schur product to the definition of
4. Summary and discussion
Ensemble forecast/assimilation methods use low-rank ensemble representations of forecast and analysis error covariance matrices. These ensembles are scaled matrix square roots of the error covariance matrices, and so ensemble data assimilation methods can be viewed as square root filters (SRFs; Bierman 1977). After assimilation of observations, the analysis ensemble can be constructed stochastically or deterministically. Deterministic construction of the analysis ensemble eliminates one source of sampling error and leads to deterministic SRFs being more accurate than stochastic SRFs in some examples (Whitaker and Hamill 2002; Anderson 2001). SRFs are not unique since different ensembles can have the same covariance. This lack of uniqueness is illustrated in three recently proposed ensemble data assimilation methods that use the Kalman SRF method to update the analysis ensemble (Bishop et al. 2001; Anderson 2001; Whitaker and Hamill 2002). Identifying the methods as SRFs allows a clearer discussion and comparison of their different analysis ensemble updates.
Accounting for small ensemble size and model deficiencies remains a significant issue in ensemble data assimilation systems. Schur products can be used to filter ensemble covariances and effectively increase covariance rank (Houtekamer and Mitchell 1998, 2001; Hamill et al. 2001; Whitaker and Hamill 2002). Covariance inflation is one simple way of accounting for model error and stabilizing the filter (Hamill et al. 2001; Anderson 2001; Whitaker and Hamill 2002). Hybrid methods represent forecast error covariances with a combination of ensemble and parameterized correlation models (Hamill and Snyder 2000). Here we have shown deterministic methods of including model error into a square root or ensemble data assimilation system when the model error has large-scale representation and when the model error is represented by a correlation model. However, the primary difficulty remains obtaining estimates of model error.
The nonuniqueness of SRFs has been exploited in estimation theory to design filters with desirable computational and numerical properties. An open question is whether there are ensemble properties that would make a particular SRF implementation better than another, or if the only issue is computational cost. For instance, it may be possible to choose an analysis update scheme that preserves higher-order, non-Gaussian statistics of the forecast ensemble. This question can only be answered by detailed comparisons of different methods in a realistic setting where other details of the assimilation system such as modeling of systematic errors or data quality control may prove to be as important.
Acknowledgments
Comments and suggestions of two anonymous reviews improved the presentation of this work. The authors thank Ricardo Todling (NASA DAO) for his comments, suggestions, and corrections. IRI is supported by its sponsors and NOAA Office of Global Programs Grant NA07GP0213. Craig Bishop received support under ONR Project Element 0601153N, Project Number BE-033-0345, and also ONR Grant N00014-00-1-0106.
REFERENCES
Anderson, J. L., 2001: An ensemble adjustment filter for data assimilation. Mon. Wea. Rev., 129 , 2884–2903.
Andrews, A., 1968: A square root formulation of the Kalman covariance equations. AIAA J., 6 , 1165–1168.
Bierman, G. J., 1977: Factorization Methods for Discrete Sequential Estimation. Vol. 128, Mathematics in Science and Engineering, Academic Press, 241 pp.
Bishop, C. H., B. Etherton, and S. J. Majumdar, 2001: Adaptive sampling with the ensemble transform Kalman filter. Part I: Theoretical aspects. Mon. Wea. Rev., 129 , 420–436.
Burgers, G., P. J. van Leeuwen, and G. Evensen, 1998: On the analysis scheme in the ensemble Kalman filter. Mon. Wea. Rev., 126 , 1719–1724.
Cohn, S. E., 1997: An introduction to estimation theory. J. Meteor. Soc. Japan, 75 , 257–288.
Cohn, S. E., and R. Todling, 1996: Approximate data assimilation schemes for stable and unstable dynamics. J. Meteor. Soc. Japan, 74 , 63–75.
Cohn, S. E., A. M. da Silva, J. Guo, M. Sienkiewicz, and D. Lamich, 1998: Assessing the effect of data selection with the DAO Physical-space Statistical Analysis System. Mon. Wea. Rev., 126 , 2913–2926.
Dee, D. P., 1995: On-line estimation of error covariance parameters for atmospheric data assimilation. Mon. Wea. Rev., 123 , 1128–1145.
Evensen, G., 1994: Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. J. Geophys. Res., 99 , 1043–1062.
Golub, G. H., and C. F. Van Loan, 1996: Matrix Computations. 3d ed. The Johns Hopkins University Press, 694 pp.
Hamill, T. M., and C. Snyder, 2000: A hybrid ensemble Kalman filter—3D variational analysis scheme. Mon. Wea. Rev., 128 , 2905–2919.
Hamill, T. M., J. S. Whitaker, and C. Snyder, 2001: Distance-dependent filtering of background error covariance estimates in an ensemble Kalman filter. Mon. Wea. Rev., 129 , 2776–2790.
Heemink, A. W., M. Verlaan, and A. J. Segers, 2001: Variance reduced ensemble Kalman filtering. Mon. Wea. Rev., 129 , 1718–1728.
Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126 , 796–811.
Houtekamer, P. L., and H. L. Mitchell, 2001: A sequential ensemble Kalman filter for atmospheric data assimilation. Mon. Wea. Rev., 129 , 123–137.
Houtekamer, P. L., L. Lefaivre, J. Derome, H. Ritchie, and H. L. Mitchel, 1996: A system simulation approach to ensemble prediction. Mon. Wea. Rev., 124 , 1225–1242.
Maybeck, P. S., 1982: Stochastic Models, Estimation, and Control. Vol. 1. Academic Press, 423 pp.
Mitchell, H. L., and P. L. Houtekamer, 2000: An adaptive ensemble Kalman filter. Mon. Wea. Rev., 128 , 416–433.
Pham, D., 2001: Stochastic methods for sequential data assimilation in strongly nonlinear systems. Mon. Wea. Rev., 129 , 1194–1207.
Pham, D., J. Verron, and M. Roubaud, 1998: A singular evolutive extended Kalman filter for data assimilation in oceanography. J. Mar. Syst., 16 , 323–340.
Verlaan, M., and A. W. Heemink, 1997: Tidal flow forecasting using reduced rank square filters. Stochastic Hydrol. Hydraul., 11 , 349–368.
Whitaker, J., and T. M. Hamill, 2002: Ensemble data assimilation without perturbed observations. Mon. Wea. Rev., 130 , 1913–1924.
APPENDIX
Computational Costs
Here we detail the computational cost scalings summarized in Table 1. All the methods require applying the observation operator to the ensemble members to form 𝗛k
Direct method
- Solve (𝗛k
𝗣fk + 𝗥k)𝗬k = 𝗛k𝗛Tk for 𝗬k. If 𝗥−1 is available, the solution can be obtained using the Sherman–Morrison–Woodbury identity (Golub and Van Loan 1996),𝗭fk and only inverting m × m matrices. Cost: O(m3 + m2p). Form 𝗜 − (𝗛k
)T𝗬k. Cost: O(pm2).𝗭fk Compute matrix square root of the m × m matrix 𝗜 − (𝗛k
)T𝗬k. Cost: O(m3).𝗭fk Apply matrix square root to 𝗭f. Cost: O(m2n).
Total cost: O(m3 + m2p + m2n).
Serial method
For each observation:
Form 𝗗. Cost: O(m).
Form 𝗜 − β𝗩𝗩T and apply to
. Cost: O(nm).𝗭fk
Total cost: O(mp + mnp).
ETKF
Form 𝗭fT𝗛T𝗥−1𝗛𝗭f. Assume 𝗥−1 inexpensive to apply. Cost: O(m2p).
Compute eigenvalue decomposition of m × m matrix. Cost: O(m3).
Apply to 𝗭f. Cost: O(m2n).
Total cost: O(m2p + m3 + m2n).
EAKF
Cost in addition to ETKF:
Eigenvalue decomposition of 𝗣f (low rank). Cost: O(m2n + m3).
Form 𝗙T𝗭f. Cost: O(m2p).
Total cost: O(m2p + m3 + m2n).
Summary of analysis ensemble calculation computational cost as a function of forecast ensemble size m, number of observations p, and state dimension n
For instance, the columns of 𝗖k that span the (m − p)-dimensional null space of
The appearance of
International Research Institute for Climate Prediction Contribution Number IRI-PP/02/03.