1. Introduction
The fixed-lag Kalman smoother (FLKS) has been proposed by Cohn et al. (1994; CST94 hereafter) as an approach to perform retrospective data assimilation. The term retrospective data assimilation denotes a procedure to incorporate data observed well past each analysis time into each analysis, taking into account error propagation through dynamical effects. Since a goal of reanalysis efforts is to produce a long archive of best-possible analyses based on all available data, whereas current reanalysis projects (e.g., Gibson et al. 1997; Kalnay et al. 1996; Schubert and Rood 1995) incorporate only data observed up to and including each analysis time, retrospective data assimilation should be an ultimate goal of reanalysis efforts, as pointed out in CST94. Moreover, although retrospective data assimilation is studied in this article primarily as a means of improving analysis quality, it is foreseeable that such a procedure could also be adopted in numerical weather prediction to improve mid- to long-range forecasts, starting from a given retrospective analysis. The preliminary efforts of Gelaro et al. (1996) can be viewed as an approach to retrospective data assimilation with this purpose.
Cohn et al. (1994) gave a particular derivation of the optimal linear FLKS. In that work, it was pointed out that the same algorithm can be derived from the approach of “state enlargement,” or “state augmentation” as it is more commonly known, first suggested in the engineering literature by Willman (1969), to reduce the smoothing problem to a filtering problem. In the state augmentation approach, the state vector at each time is appended with the state vector at previous times when the desired smoother estimates are to be calculated. A Kalman filter (KF) problem can then be solved for the augmented system. The first derivation of a smoother algorithm via state augmentation was that of Biswas and Mahalanabis (1972) for the linear fixed-point smoothing problem. Subsequently, Moore (1973) derived a linear fixed-lag smoother via the same approach, which results in the same algorithm as that derived in CST94. Extension of the FLKS formulation to nonlinear systems can be achieved using the same technique of state augmentation, as indicated by Biswas and Mahalanabis (1973), for both the fixed-point and fixed-lag smoothing problems [see also Todling and Cohn (1996a) for an explicit derivation of the extended fixed-lag Kalman smoother]. The utility of state augmentation is that the resulting smoothers are often computationally less demanding than those arising from some other approaches (e.g., Sage and Melsa 1970, section 9.5). For instance, smoothers based on state augmentation avoid inversion of the filter error covariance matrices and of the tangent linear propagator (e.g., Ménard and Daley 1996; see also the appendix of the present article). These inversions are also avoided by an earlier smoother algorithm due to Bryson and Frazier (1963), which can be shown to reduce to the FLKS algorithm of CST94 for the case of linear systems. Algebraic equivalence between smoothers obtained by state augmentation and by methods such as maximum likelihood (Sage and Ewing 1970; Sage 1970) or conditional expectation (Leondes et al. 1970) exists in most cases. The interested reader is referred to Meditch (1973) and Kailath (1975) for detailed reviews of the literature on linear and nonlinear smoothing. The distinction among different types of smoothing problems, and the connection between fixed-interval smoothing and four-dimensional variational (4DVAR) analysis, is drawn in Cohn (1997).
Brute-force implementation of the (extended) FLKS to build an operational retrospective data assimilation system (RDAS) is not possible for the same reasons that a brute-force (extended) KF-based data assimilation system would be impractical: computational requirements are excessive, and knowledge of the requisite error statistics is lacking. Therefore, approximations not only must be employed but cannot be escaped from. Thus, in this article, we develop and evaluate the performance of potentially implementable approximate schemes. To provide an exact evaluation, we choose a barotropically unstable linear shallow water model as a test bed for this investigation. All of the approximate schemes evaluated here have relatively simple nonlinear equivalents.
In the sequel, we first briefly review, in section 2, the linear FLKS of CST94 and outline the performance evaluation technique employed to study the behavior of linear suboptimal filter and smoother algorithms. Section 3 gives a summary of the suboptimal filters and smoothers evaluated subsequently in section 4, in the context of the linear shallow water model. We draw conclusions in section 5.
2. Review and performance evaluation equations


The subscript notation utilized here is common in estimation theory, and is particularly important when considering smoothing problems. Specifically, the forecast vector
The FLKS update equation (3) is only applicable for l ⩽ k. If the (fixed) total number of lags is L, meaning that (3) is to be applied in general for l = 1, 2, . . . , L, then for k = 0, 1, . . . , L − 1, the condition l ⩽ k is not satisfied for all l. Therefore, the update (3) is applied only for l = 1, 2, . . . , min(k, L), which is a restricted range of l when k = 0, 1, . . . , L − 1. In the language of estimation theory, this restriction corresponds to computing fixed-point smoother results for all k up to k = L − 1, after which the fixed-lag smoother starts operating. This is an initialization procedure for the fixed-lag smoother (e.g., Gelb 1974, 173–176). In practice, because the FLKS algorithm employed by CST94 already has the structure of a fixed-point algorithm, this procedure simply amounts to controlling the ending points of certain loop statements in a computer code.






3. Summary of suboptimal filters and smoothers
We now summarize the suboptimal schemes to be evaluated here in the context of the linear shallow water model of the next section. The following are the suboptimal schemes considered in this article for the filter portion of the fixed-lag smoother algorithm (see Cohn and Todling 1996, CT96 hereafter; Todling et al. 1996;Todling and Cohn 1996a,b).
a. Constant forecast error covariance filter (CCF)
b. Partial singular-value decomposition filter (PSF)
c. Partial eigendecomposition filter (PEF)
d. Reduced resolution filter (RRF)
It should be pointed out that the approach of reduced resolution filtering is very general, falling in the broad category of order reduction schemes commonly known in estimation theory. In this regard, the PSF scheme described above can be seen as a reduced-order approximation with the matrix
The following are the suboptimal schemes considered here for the retrospective analysis portion of the fixed-lag smoother algorithm.
e. Partial singular-value decomposition retrospective analysis (PSRA/PSRA2)
In this category, there are at least two possibilities. The PSRA scheme extends the PSF approximation of the filter gain (1c) to the retrospective analysis gains (4), whereas the PSRA2 scheme extends the PEF approximation to the retrospective analysis gains.
f. Reduced resolution retrospective analysis (RRRA)
Many other suboptimal schemes have been proposed for filtering, particularly in the atmospheric data assimilation literature (Todling and Cohn 1994, and references therein). Since fixed-lag smoothing can always be regarded as filtering for an augmented-state system (e.g., Todling and Cohn 1996a), in principle all of these suboptimal strategies carry over to the fixed-lag smoothing problem. In this article we choose to concentrate only on the approximations presented above.
It is possible to construct approximate RDASs by combining different strategies for approximating the filter and the retrospective analysis portions of the RDAS. For instance, one could choose to approximate both portions equivalently—that is, with two similar schemes like the RRF and RRRA at the same resolution; or one could choose to approximate the filter and calculate the retrospective analysis portion exactly—that is, to approximate (6) and use (7); one could also build hybrid approximations in which the filter and the retrospective analysis employ different strategies. In any case, since our formulation of the fixed-lag smoother is based on the filter, whenever the filter is approximated the smoother becomes suboptimal. The converse is not true, in the sense that if the filter is kept exact and the retrospective analysis equations are approximated—if we use (6) and approximate (7)—only the smoother becomes suboptimal, but not the filter. This, however, may not be a very useful approach, since the major computational requirements are associated with the filter equation (6).
4. Results for a shallow water model
To evaluate the performance of the suboptimal schemes described above, we use the barotropically unstable model of CT96, a shallow water model linearized about a meridionally dependent squared-hyperbolic secant jet (Bickley jet; Haltiner and Williams 1980, 175). We refer the reader to Fig. 1 of CT96 for the shape, extent, and strength of the jet. The model domain is shown in Fig. 1 here. The assimilation experiments employ the observing network of CT96: 33 radiosonde stations observing winds and heights every 12 h and distributed outside the strongest part of the jet. The tick marks in the figure indicate the 25 × 16 model grid. In the experiments referring to a trailing error covariance matrix we construct it, exactly as in CT96, using the slow eigenmodes of the autonomous unstable dynamics of our shallow water model.
Before evaluating the performance of a few suboptimal RDASs, we discuss results obtained for the optimal FLKS. The performance of the optimal filter and fixed-lag smoother can be seen in Fig. 2, which shows the domain-averaged expected root-mean-square (erms) analysis error in the total energy as a function of time. This quantity is calculated as a weighted average of the analysis error variances of the three variables of the model, which are extracted from the main diagonal of the actual analysis error covariance matrix. The top curve in the figure corresponds to the filter analysis every 12 h, whereas successive retrospective analysis results are given by successively lower curves, which refer to analyses including data 12, 24, 36, and 48 h ahead in time—that is, lags l = 1, 2, 3, and 4. The filter curve is the same as that in Fig. 2 of CT96 (shown, here, only up to 5 days). The most relevant results are those for the transient part of the assimilation period, before the filter and smoother begin to approach steady state. Incorporating new data into past analyses reduces the corresponding past analysis errors considerably. The largest impact is on the initial analysis, which would not be the case if a significant amount of model error were present.
Further illustration of the behavior of the optimal FLKS is given in Fig. 3, where we display maps of the analysis error standard deviation in the height field at t = 0.5 days. The panels are for the filter analysis errors (Fig. 3a), and for the retrospective analysis errors for lags l = 1 (Figs. 3b) and l = 4 (Fig. 3c). Thus, in Figs. 3b,c and the analysis errors are reduced by incorporating data 12 and 48 h ahead of the current analysis time (t = 0.5 days), respectively. We see not only the overall decrease in error levels from Fig. 3a to Fig. 3c, as expected from Fig. 2, but also that within each panel errors are largest in the central band of the domain, where there are no observations and where the jet is strongest. Furthermore, the error maximum in the center of the domain moves westward and diminishes as more data are incorporated into the analysis through the smoothing procedure (from Figs. 3a to 3c). This property of the FLKS of propagating and reducing errors in the direction opposite of the flow has already been observed in the experiments of CST94 and Ménard and Daley (1996).
We now study the behavior of suboptimal RDASs. We start with schemes that approximate the filter and retrospective analysis portions similarly. In this category, we investigate the behavior of the RRF and RRRA corresponding to expressions (12) and (16), respectively, as well as the behavior of the PSF and PSRA corresponding to expressions (10) and (13), respectively.
The results of Todling and Cohn (1996b) showed that the RRF described above, with resolutions 13 × 16 and 13 × 12, provides good filter performance in our shallow water model context. This was attributed mainly to the fact that at these resolutions the barotropically unstable jet is fairly well resolved. As a matter of fact, the meridional jet is fully resolved at resolution 13 × 16. In Fig. 4 we show results of the performance evaluation for the RRF and RRRA algorithms at these resolutions (Fig. 4a for 13 × 16; Fig. 4b for 13 × 12). No trailing error covariance models are taken into account here, that is,
The results of the RRF can be improved by modeling the trailing error covariance matrix
Along similar lines, we investigate the performance of an RDAS using the PSF algorithm for the filter portion, and the PSRA algorithm for the retrospective analysis portion. From the experiments of CT96, we know that using the first 54 singular modes of the 12-h propagator of the linear shallow water model—those with singular values greater than or equal to one—is enough to produce a stable suboptimal filter. Moreover, we learned in CT96 that adaptively tuning a modeled trailing error covariance matrix
Figure 5 shows performance results for the PSF–PSRA suboptimal RDAS when the first 54 modes are used for both approximations (out of a total of 325 slow modes). The filter results, when compared to the optimal results of Fig. 2, are once again quite good—the reader is encouraged to compare the top curve of Fig. 5 with the curve labeled S54 in Fig. 11 of CT96; results now are better due to the adaptively tuned trailing error covariance matrix. The PSRA results, on the other hand, are not nearly as good as those for the optimal smoother (Fig. 2), with little difference among results for lag l = 1 and those for higher lags l = 2, 3, and 4 in Fig. 5. The next two experiments demonstrate that this poor smoother performance can be attributed mostly to neglecting the trailing forecast–retrospective analysis error cross-covariance matrix
To investigate the PSRA scheme further, we compare performance results between two RDASs using the KF for the filter portion, with the retrospective analysis portion given by either the PSRA scheme or the PSRA2 scheme, both with 54 singular modes retained. Thus, the suboptimality in these two RDASs is solely in the retrospective analysis portion. Figure 6 shows the erms errors in the total energy for these two cases: Fig. 6a corresponds to the RDAS using the PSRA scheme and Fig. 6b corresponds to the RDAS using the PSRA2 algorithm. The filter curves in both panels (topmost curves) are identical to one another, as well as to the filter curve in Fig. 2 for the optimal FLKS case. Comparison between the lower curves in the two panels of Fig. 6 shows the superiority of the PSRA2 scheme: beyond lag l = 1 little is gained in the PSRA scheme, even when based on the KF (cf. Fig. 6a with Fig. 5), whereas successively higher lags do have a significant impact in the PSRA2 scheme (Fig. 6b). The poor performance of the PSRA scheme indicates that its neglected trailing part contains a large amount of cross-(co)variance when retaining just the 54 singular modes of the propagator with singular values larger than or equal to one. The PSRA2 scheme with 54 modes, on the other hand, captures most of the cross-(co)variance, as comparison of Fig. 6b with the optimal result in Fig. 2 indicates. We conclude from these experiments that the trailing cross-covariance matrix is more significant in some approximate retrospective analysis schemes than in others. Moreover, the singular modes of the propagated filter analysis–retrospective analysis error cross-covariance matrix (employed in the PSRA2 scheme), rather than the singular modes of the propagator itself (employed in the PSRA scheme), contain most of the information relevant for retrospective analysis. This distinction between the PSRA and PSRA2 schemes is completely analogous to the distinction between the PSF and PEF schemes drawn in CT96, where somewhat better performance of the PEF scheme over the PSF scheme was demonstrated. This distinction is even more pronounced in the retrospective analysis context, as seen in Fig. 6.
The good performance of the PSRA2 scheme when combined with the KF suggests evaluating the performance of two hybrid RDASs. Figure 7a shows results for the combined PSF–PSRA2 algorithm, with 54 modes, and Fig. 7b shows results for the combined PEF–PSRA2 algorithm, with 54 modes as well. Both the PSF and PEF filtering strategies include an adaptive tuning procedure for the modeled trailing error covariance matrices
We evaluate next the performance of schemes that approximate only the filter portion and carry out the retrospective analysis calculations exactly. We start with an RDAS in which the adaptive CCF scheme is used for the filter part. Figure 8 shows the evolution of the actual erms errors up to day 5 (same as Fig. 3 of Todling et al. 1996). While the performance of the CCF scheme (top curve) is worse than that seen in Fig. 2 for the optimal KF, it is significantly worse only beyond day one; adaptive tuning of more than a single parameter would likely improve this filter result. As a consequence of suboptimality of the CCF scheme, the performance of the CCF-based retrospective analyses shown in Fig. 8 is also suboptimal. However, a comparison between Figs. 2 and 8 indicates that retrospective analysis based on a suboptimal filter can be viewed as a way of improving suboptimal filter performance toward optimal filter performance. For instance, notice that by day 2.5, the lag-2 suboptimal retrospective analysis of Fig. 8 has about the same error level as that of the optimal filter analysis of Fig. 2.
We also see in Fig. 8 that the retrospective analysis at 12 h for lag-4 is worse than the retrospective analyses for smaller lags. This is an indication that we can expect to extract only so much information from the data when using the rather crude approximate forecast error covariance matrix of the CCF. Even though the correct dynamics is still used to propagate filter information from 60 h back to 12 h, the approximate forecast error covariance used by the CCF is largely different from the true forecast error covariance at 60 h, and we expect that the retrospective analysis results can actually degrade during the transient period, the more lags we include in the retrospective scheme.
When comparing the RDAS using the CCF scheme (Fig. 8) with the RDASs using the RRF–RRRA of Fig. 4 and the PSF–PSRA of Fig. 5, we see that the performance of the CCF scheme itself is not much different than that of the RRF with 13 × 12 resolution and that of the PSF with 54 modes (top curve in each figure). The performance of the CCF-based retrospective analysis, however, exceeds that of the 13 × 12 RRF-based RRRA scheme and the 54-mode PSF-based PSRA scheme, for every lag, beyond the initial transient assimilation period. During the transient period, the RRF–RRRA algorithm shows better performance, for high lags, than either the CCF-based retrospective analysis algorithm or the PSF-based PSRA algorithm.
Analogously to Fig. 3, we show in Fig. 9 maps of the actual height analysis error standard deviation at day 0.5, for the experiment of Fig. 8. The panels are arranged as before: (a) filter analysis; (b) lag l = 1 retrospective analysis; and (c) lag l = 4 retrospective analysis. Comparing panels (a) and (b) with the corresponding panels in Fig. 3, we see that the CCF scheme and the resulting lag l = 1 retrospective analysis perform remarkably well. However, the retrospective analysis for lag l = 4 (Fig. 9c) is not significantly better than for lag l = 1 (Fig. 9b), as one might expect from Fig. 8 at day 0.5, and in fact compares poorly with the optimal case (Fig. 3c), particularly over the data-void central band.
Finally, we examine the performance of the more sophisticated PSF and PEF suboptimal filters and the corresponding suboptimal RDASs, using the exact retrospective analysis formulas. In both cases we retain only 54 leading modes and we adaptively tune the trailing error covariance matrices as before. In Fig. 10a the top curve refers to the performance of the PSF, whereas that in Fig. 10b refers to the performance of the PEF. The PSF result is identical to that displayed in Fig. 5 since the filter here retains the same number of modes as before. A comparison of the PSF-based retrospective analyses of Fig. 10a, which use the exact retrospective analysis formulation, and the PSF-based retrospective analyses of Fig. 5, where this formulation was approximated by the PSRA algorithm, shows clearly the superior performance of the exact formulation. The PSF-based RDAS (Fig. 10a) performance is similar to, and the PEF-based RDAS (Fig. 10b) performance is superior to, that of the CCF-based RDAS of Fig. 8. The RDAS using the PEF (Fig. 10b) presents very good long-term performance, with its results being fairly close to those of the optimal FLKS in Fig. 2, and only slightly inferior to those of the 13 × 16 RRF–RRRA scheme of Fig. 4a.
In Fig. 11 we show maps of the actual height analysis error standard deviations for the experiment using the PSF of Fig. 10a. Performance relative to the optimal case (Fig. 3) tends to worsen with increasing lag number, particularly over the data-void central band. Compared to the maps of Fig. 9, however, there is improvement in the analyses over specific regions of the domain. In particular, the lag l = 4 retrospective analysis in Fig. 11c shows a considerable error reduction beyond that of Fig. 9c over the central part of the domain and the Atlantic Ocean.
5. Conclusions
In this article we evaluated the performance of approximate (suboptimal) RDASs based on the FLKS formulation of Cohn et al. (1994). This formulation has several practical advantages over more commonly known smoother formulations. In particular, it avoids a number of large matrix inversions. This formulation also separates naturally into a filter portion and a retrospective analysis portion, enabling a variety of suboptimal implementations. Model error is incorporated implicitly in the retrospective analysis portion, because the filter portion is based directly on the Kalman filter, which already takes model error explicitly into account. Thus, a version of the retrospective analysis portion could be implemented operationally and remain unchanged while improvements in the filter portion, such as accounting for model error, take place.
For linear dynamics and observing systems, performance evaluation equations for approximate RDASs based on the FLKS formulation follow directly from the approach of state augmentation and the usual performance evaluation equations for linear filters utilizing general gain matrices. In this way, we examined the performance of a variety of suboptimal RDASs for a barotropically unstable shallow water model. We concentrated on evaluating the performance obtained when using approximate expressions for the error covariance propagation in the filtering portion of the RDAS, as well as for the error cross-covariance propagation in the retrospective analysis portion of the RDAS. Our experiments indicate that successful retrospective data assimilation schemes can be designed by approximating either the filter portion alone or by approximating both the filter and retrospective analysis portions simultaneously. An important conclusion from these experiments is that a few lags of suboptimal retrospective analysis may accomplish the performance of an optimal filter analysis. Sophisticated approximate filters that take dynamics of error covariances into account present the best suboptimal retrospective analysis performance. It would be interesting to verify how well these conclusions hold in more realistic data assimilation environments.
Acknowledgments
It is a pleasure to thank Dick P. Dee for many discussions throughout the course of this work. The numerical results were obtained on the Cray C90 through cooperation of the NASA Center for Computational Sciences at the Goddard Space Flight Center. This research was supported by a fellowship from the Universities Space Research Association (RT) and by the NASA EOS Interdisciplinary Project on Data Assimilation (SEC and NSS).
REFERENCES
Biswas, K. K., and A. K. Mahalanabis, 1972: An approach to fixed-point smoothing problems. IEEE Trans. Aerosp. Electron. Syst.,8, 676–682.
——, and ——, 1973: Suboptimal algorithms for nonlinear smoothing. IEEE Trans. Aerosp. Electron. Syst.,9, 529–534.
Bryson, A. E., and M. Frazier, 1963: Smoothing for linear and nonlinear dynamic systems. TDR 63–119, 12 pp. [Available from Wright–Patterson Air Force Base, OH 45433.].
Campbell, S. L., and C. D. Meyer Jr., 1991: Generalized Inverses of Linear Transformations. Dover Publications, 272 pp.
Cane, M. A., A. Kaplan, R. N. Miller, B. Tang, E. C. Hackert, and A. J. Busalacchi, 1996: Mapping tropical Pacific sea level: Data assimilation via a reduced state space Kalman filter. J. Geophys. Res.,101, 22 599–22 617.
Cohn, S. E., 1997: An introduction to estimation theory. J. Meteor. Soc. Japan,75, 257–288.
——, and R. Todling, 1996: Approximate Kalman filters for stable and unstable dynamics. J. Meteor. Soc. Japan,74, 63–75.
——, N. S. Sivakumaran, and R. Todling, 1994: A fixed-lag Kalman smoother for retrospective data assimilation. Mon. Wea. Rev.,122, 2838–2867.
Dee, D. P., 1995: On-line estimation of error covariance parameters for atmospheric data assimilation. Mon. Wea. Rev.,123, 1128–1196.
Fukumori, I., and P. Malanotte-Rizzoli, 1995: An approximate Kalman filter for ocean data assimilation: An example with an idealized Gulf Stream model. J. Geophys. Res. Oceans,100, 6777–6793.
Gelaro, R., E. Klinker, and F. Rabier, 1996: Real and near-real time corrections to forecast initial conditions using adjoint methods:A feasibility study. Preprints, 11th Conf. on Numerical Weather Prediction, Norfolk, VA, Amer. Meteor. Soc., J58–J60.
Gelb, A., Ed., 1974: Applied Optimal Estimation. The MIT Press, 374 pp.
Gibson, J. K., P. Kallberg, S. Uppala, A. Hernandez, A. Nomura, and E. Serrano, 1997: ERA description. ECMWF Re-Analysis Project Rep., Ser. 1, ECMWF, Reading, United Kingdom, 72 pp.
Haltiner, G. J., and R. T. Williams, 1980: Numerical Prediction and Dynamic Meteorology. John Wiley and Sons, 477 pp.
Jazwinski, A. H., 1970: Stochastic Processes and Filtering Theory. Academic Press, 376 pp.
Kailath, T., 1975: Supplement to “A survey of data smoothing for linear and nonlinear dynamic systems.” Automatica,11, 109–111.
Kalnay, E., and Coauthors, 1996: The NCEP/NCAR 40-year reanalysis project. Bull. Amer. Meteor. Soc.,77, 437–471.
Leondes, C. T., J. B. Peller, and E. B. Stear, 1970: Nonlinear smoothing theory. IEEE Trans. Syst. Sci. Cybernetics,6, 63–71.
Meditch, J. S., 1969: Stochastic Linear Estimation and Control. McGraw-Hill, 394 pp.
——, 1973: A survey of data smoothing for linear and nonlinear dynamic systems. Automatica,9, 151–162.
Ménard, R., and R. Daley, 1996: The application of Kalman smoother theory to the estimation of 4DVAR error statistics. Tellus,48A, 221–237.
Moore, J. B., 1973: Discrete-time fixed-lag smoothing algorithms. Automatica,9, 163–173.
Pu, Z., E. Kalnay, and J. Sela, 1997: Sensitivity of forecast error to initial conditions with a quasi-inverse linear model. Mon. Wea. Rev.,125, 2479–2503.
Sage, A. P., 1970: Maximum a posteriori filtering and smoothing algorithms. Int. J. Control,11, 641–658.
——, and W. S. Ewing, 1970: On filtering and smoothing for non-linear state estimation. Int. J. Control,11, 1–18.
——, and J. L. Melsa, 1970: Estimation Theory with Applications to Communications and Control. McGraw-Hill, 529 pp.
Schubert, S. D., and R. B. Rood, 1995: Proceedings of the workshop on the GEOS-1 five-year assimilation. NASA Tech. Memo. 104606, Vol. 7, 201 pp. [Available online from http://dao.gsfc.nasa.gov/subpages/tech-reports.html.].
Todling, R., 1995: Computational aspects of Kalman filtering and smoothing for atmospheric data assimilation. Numerical Simulations in the Environmental and Earth Sciences, Cambridge University Press, 191–202.
——, and S. E. Cohn, 1994: Suboptimal schemes for atmospheric data assimilation based on the Kalman filter. Mon. Wea. Rev.,122, 2530–2557.
——, and ——, 1996a: Some strategies for Kalman filtering and smoothing. Proc. ECMWF Seminar on Data Assimilation, Reading, United Kingdom, ECMWF, 91–111. [Available online from http://dao.gsfc.nasa.gov/DAO_people/todling.].
——, and ——, 1996b: Two reduced resolution filter approaches to data assimilation. Proc. Ninth Brazilian Meteorological Congress, São José dos Campos, Brazil, Braz. Meteor. Soc., 1069–1072. [Available online from http://dao.gsfc.nasa.gov/DAO_people/todling.].
——, N. S. Sivakumaran, and S. E. Cohn, 1996: Some strategies for retrospective data assimilation: Approximate fixed-lag Kalman smoothers. Preprints, 11th Conf. on Numerical Weather Prediction, Norfolk, VA, Amer. Meteor. Soc., 238–240.
Verlaan, M., and A. W. Heemink, 1995: Reduced rank square root filters for large scale data assimilation problems. Proc. Int. Symp. on Assimilation of Observations in Meteorology and Oceanography, Vol. 1, Tokyo, Japan, WMO, 247–252.
Wang, Z., K. K. Droegemeier, L. White, and I. M. Navon, 1997: Application of a Newton adjoint algorithm to the 3-D ARPS storm scale model using simulated data. Mon. Wea. Rev.,125, 2460–2478.
Willman, W. W., 1969: On the linear smoothing problem. IEEE Trans. Autom. Control,14, 116–117.
Wunsch, C., 1996: The Ocean Circulation Inverse Problems. Cambridge University Press, 442 pp.
APPENDIX
Implicit Account of Model Error in Retrospective Analysis






For retrospective data assimilation purposes, expression (A2) presents no advantage over (3), particularly due to the appearance of the inverses of the adjoint propagator and of the forecast and analysis error covariance matrices in the gain matrices
We emphasize that in the formulation of the FLKS employed in CST94, the retrospective analysis (3) is a system driven exclusively by appropriately weighted innovations (cf. Moore 1973):




One might consider using (A8) as an approximation for the retrospective analysis gains (A1) in the case when model error is present. Although the assumption of invertibility of the propagator is extremely stringent for atmospheric data assimilation, a quasi-inverse approximation similar to that of Pu et al. (1997) and Wang et al. (1997) could be employed. Performance evaluation experiments, like those in section 4 of the present article, have been conducted using both of these approximations. Results indicate that the quasi-inverse approximation performs well when (A8) is used, for the perfect model case; the use of (A8) together with the quasi-inverse propagator approximation in the presence of model error, however, does not perform very well in general (results not shown here).

Model domain and observational network composed of 33 radiosonde stations observing winds and heights every 12 h (same as Fig. 2 of CT96).
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

Model domain and observational network composed of 33 radiosonde stations observing winds and heights every 12 h (same as Fig. 2 of CT96).
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
Model domain and observational network composed of 33 radiosonde stations observing winds and heights every 12 h (same as Fig. 2 of CT96).
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

The erms analysis error in total energy for the Kalman filter (upper curve) and fixed-lag Kalman smoother (lower curves).
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

The erms analysis error in total energy for the Kalman filter (upper curve) and fixed-lag Kalman smoother (lower curves).
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
The erms analysis error in total energy for the Kalman filter (upper curve) and fixed-lag Kalman smoother (lower curves).
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

Analysis error standard deviation in the height field at time t = 0.5 days. (a) The filter analysis; (b) and (c) the retrospective analyses with lags l = 1 and 4, respectively. Contour interval is 1 m.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

Analysis error standard deviation in the height field at time t = 0.5 days. (a) The filter analysis; (b) and (c) the retrospective analyses with lags l = 1 and 4, respectively. Contour interval is 1 m.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
Analysis error standard deviation in the height field at time t = 0.5 days. (a) The filter analysis; (b) and (c) the retrospective analyses with lags l = 1 and 4, respectively. Contour interval is 1 m.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an approximate RDAS using the RRF and RRRA schemes for resolutions: (a) 13 × 16 and (b) 13 × 12.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an approximate RDAS using the RRF and RRRA schemes for resolutions: (a) 13 × 16 and (b) 13 × 12.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 2 but for an approximate RDAS using the RRF and RRRA schemes for resolutions: (a) 13 × 16 and (b) 13 × 12.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an approximate RDAS using the PSF and PSRA schemes simultaneously, both with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an approximate RDAS using the PSF and PSRA schemes simultaneously, both with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 2 but for an approximate RDAS using the PSF and PSRA schemes simultaneously, both with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an RDAS using the KF and approximate retrospective analysis schemes: (a) PSRA and (b) PSRA2, with singular 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an RDAS using the KF and approximate retrospective analysis schemes: (a) PSRA and (b) PSRA2, with singular 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 2 but for an RDAS using the KF and approximate retrospective analysis schemes: (a) PSRA and (b) PSRA2, with singular 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an RDAS using the PSRA2 scheme for the retrospective analysis portion, and either the (a) PSF or (b) PEF for the filter portion, all with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for an RDAS using the PSRA2 scheme for the retrospective analysis portion, and either the (a) PSF or (b) PEF for the filter portion, all with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 2 but for an RDAS using the PSRA2 scheme for the retrospective analysis portion, and either the (a) PSF or (b) PEF for the filter portion, all with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for the adaptive CCF scheme and exact retrospective analysis equations.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but for the adaptive CCF scheme and exact retrospective analysis equations.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 2 but for the adaptive CCF scheme and exact retrospective analysis equations.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 3 but using the CCF-based RDAS of Fig. 8.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 3 but using the CCF-based RDAS of Fig. 8.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 3 but using the CCF-based RDAS of Fig. 8.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but using (a) PSF and (b) PEF, both with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 2 but using (a) PSF and (b) PEF, both with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 2 but using (a) PSF and (b) PEF, both with 54 modes.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 3 but using the PSF-based RDAS of Fig. 10a.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2

As in Fig. 3 but using the PSF-based RDAS of Fig. 10a.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2
As in Fig. 3 but using the PSF-based RDAS of Fig. 10a.
Citation: Monthly Weather Review 126, 8; 10.1175/1520-0493(1998)126<2274:SSFRDA>2.0.CO;2