1. Introduction
In data assimilation, the process of detecting and accounting for observation errors that are statistical outliers is called quality control (QC; e.g., Daley 1991). An operational numerical weather prediction system may employ multiple layers of QC. For instance, observations with implausible values are usually rejected even before they enter the data assimilation process. We refer to the algorithms used for such rejection decisions as offline QC algorithms. The fact that an observation passes the offline QC procedures does not guarantee that it is not a statistical outlier, however. For instance, an error in a highly accurate observation can be a statistical outlier when the error has a large representativeness error component. Such errors have to be dealt with by the data assimilation algorithm. We refer to the QC procedures that are part of the data assimilation algorithms as online QC algorithms.
Online QC algorithms detect observation errors that are statistical outliers by examining the difference between the observation and the prediction of the observation by the background. This difference is called the innovation. For instance, a simple online QC can be implemented by rejecting the observations for which the absolute value of the innovation is larger than a prescribed threshold. Another approach, which is more desirable from a theoretical point of view, is to employ robust statistics in the formulation of the state-update step of the data assimilation scheme (e.g., Huber 1981; Hampel 1968; Maronna et al. 2006). In particular, the presumed probability distribution of the observation errors can be modified such that the update step can anticipate errors that would be considered statistical outliers if the observation errors were strictly Gaussian. The practical challenge posed by this approach is to find a modification of the prescribed probability distribution function, which leads to a data assimilation algorithm that can be implemented in practice.
An operational online QC algorithm using robust observation error statistics (Anderson and Järvinen 1999) was first introduced by the European Centre for Medium-Range Weather Forecasts (ECMWF). The general idea of this approach was to define the probability distribution of the observation errors as the sum of two probability distributions: a normal distribution representing the “normal” observation errors and another distribution representing the “gross” observation errors. This approach was originally proposed as an offline QC procedure by Ingleby and Lorenc (1993), but the variational framework made its integration into the data assimilation scheme possible. The formulation of the algorithm by Anderson and Järvinen (1999) became known as variational QC (Var-QC). In the latest operational version of Var-QC, called the Huber norm QC (Tavolato and Isaksen 2010), the probability of medium and large observation errors decreases linearly making it faster than a Gaussian distribution but slower than a uniform distribution.
A wide variety of robust filtering schemes has been proposed in the mathematical statistics literature in the past decades. In particular, Meinhold and Singpurwalla (1989) replaced the normality assumption with fat-tailed distributions such as the t distribution, whereas Naveau et al. (2005) considered a skewed version of the normal distribution. West (1981, 1983, 1984) suggested a method for robust sequential approximate Bayesian estimation. Fahrmeir and Kaufmann (1991) and Fahrmeir and Kunstler (1999) offered posterior mode estimation and penalized likelihood smoothing in robust state-space models. Kassam and Poor (1985) discussed the minimax approach for the design of robust filters for signal processing. Schick and Mitter (1994) derived a first-order approximation for the conditional prior distribution of the state. Ershov and Liptser (1978), Stockinger and Dutter (1987), Martin and Raftery (1987), Birmiwal and Shen (1993), and Birmiwal and Papantoni-Kazakos (1994) also proposed robust filtering schemes that were resistant to outliers.
Recently, Ruckdeschel (2010) proposed a robust Kalman filter in the setting of time-discrete linear Euclidean state-space models with an extension to hidden Markov models, which is optimal in the sense of minimax mean-squared errors. He used the Huberization method but investigated its performance only on a one-dimensional linear system. Luo and Hoteit (2011) employed the H∞ filter to make ensemble Kalman filters (EnKF) robust enough to gross background errors. The H∞ filter minimizes the maximum of a cost function different from the minimum variance used in the Kalman filter. They demonstrated their approach on both a one-dimensional linear and a multidimensional nonlinear model. Calvet et al. (2012) introduced an impact function that quantified the sensitivity of the state distribution and proposed a filter with a bounded impact function.
EnKFs have been successfully implemented in highly complex operational prediction models in the atmospheric and oceanic sciences. They are Monte Carlo approximations of the traditional Kalman filter (KF; Kalman 1960) and use ensembles of forecasts to estimate the mean and covariance of the presumed normal distribution of the background. Similar to KF, EnKFs are not robust enough to gross errors in the estimate of the background mean or the observation (e.g., Schlee et al. 1967). The main goal of this paper is to design an EnKF scheme that is robust to observation errors that are statistical outliers. Harlim and Hunt (2007) and Luo and Hoteit (2011) made EnKF robust to unexpectedly large background errors. Here, we propose to make EnKF robust to gross observation errors by Huberization, a procedure that can be implemented on any EnKF scheme.
The rest of the paper is organized as follows. Section 2 first illustrates the effects of gross observation errors on the performance of EnKF; then, it describes our proposed approach to cope with such errors. Section 3 demonstrates the effectiveness of our approach for a one-dimensional linear system, while section 4 shows the results for the 40-variable Lorenz model. Finally, section 5 summarizes the main results of the paper.
2. A robust ensemble Kalman filter
a. Ensemble Kalman filters



























The components of the vector of differences
b. The effects of observation outliers















Plot of the true states (solid line) and the traditional EnKF (dashed line) as a function of time t for a one-dimensional linear system with (top) additive outliers ξt = 5 and (bottom) innovations outliers with α = 0.2 and kt = 25. The occurrences of outliers are marked with open circles.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

Plot of the true states (solid line) and the traditional EnKF (dashed line) as a function of time t for a one-dimensional linear system with (top) additive outliers ξt = 5 and (bottom) innovations outliers with α = 0.2 and kt = 25. The occurrences of outliers are marked with open circles.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
Plot of the true states (solid line) and the traditional EnKF (dashed line) as a function of time t for a one-dimensional linear system with (top) additive outliers ξt = 5 and (bottom) innovations outliers with α = 0.2 and kt = 25. The occurrences of outliers are marked with open circles.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
c. A robust ensemble Kalman filter
The detrimental effect of the outliers on the EnKF state estimate can be reduced by decreasing the magnitude of those components of the innovation vector that have unusually large absolute values. This can be done by defining an upper bound for the allowable absolute value of the innovations. When the magnitude of an innovation is found to be larger than the prescribed upper bound, the magnitude of the innovation can be clipped at the upper bound. To be precise, the innovation δy is left unchanged if −c < δy < c for some c > 0 and clipped at −c if δy < −c and at c if δy > c. This componentwise clipping of the innovation is called Huberization, and the tunable parameter c is called the clipping height.









A simple alternative to Huberization for handling observation error outliers is to discard the suspect observations from the data assimilation process. In fact, this is the online QC approach that has been employed by EnKF algorithms in weather prediction models (e.g., Szunyogh et al. 2008). In the simple numerical examples given here, we discard the observation if |δy| > c for a prescribed c. In these applications, the prescribed smallest magnitude of the innovation that triggers a rejection of the observation depends on the magnitude of the ensemble-based estimate of the background error variance at the observation location (the related entry of
d. Choosing parameter c
The tunable parameter of both strategies to handle the outlier observation errors, which were described in section 2c, is the p-dimensional vector c ∈



















The important issues in selecting the clipping height c are the computational complexity of the sample covariance matrices. First, a small ensemble size may produce inaccurate estimates of the covariance matrices (Whitaker and Hamill 2002). Another is that doing the Monte Carlo integration method to choose the clipping height c for all time steps is time consuming. To increase the accuracy of the covariance matrices and save computation time, we may use
3. A one-dimensional linear system


We investigate the performance of the robust ensemble Kalman filter (REnKF) for this system using 20-member ensembles and a variance inflation factor of 1.1. A limit of the sample variance of the ensembles
To see the impact of additive outliers, we suppose that the additive outliers with ξt = 8 are present in the data at t = 31, 32, and 33. Figure 2 shows the boxplots of the bias versus efficiency δ. Figure 3 shows the boxplots of the bias versus radius r. As the clipping value c decreases, that is, as the efficiency decreases or as the radius increases, the bias of the robust estimators shrinks, whereas the error variance decreases to a point but then increases again. The chosen clipping heights are in the range where the error variance keeps increasing. The bias of the Huberizing filter decreases to zero slower than that of the discarding filter, but the error variance increases slower than that of the discarding filter. The bias starts to recover from t = 34 when the outliers disappear. Figure 4 shows the trajectories of the true state, the traditional ensemble Kalman filter, and two robust ensemble Kalman filters with efficiencies δ = 0.99 and 0.7. Both robust ensemble Kalman filters have smaller jumps in the state estimation errors at the times of the outliers than the traditional ensemble Kalman filter has. At efficiency δ = 0.99, the discarding filter removes the jump entirely, coinciding with a bias of zero, but at efficiency δ = 0.7, its estimation is inaccurate, coinciding with the big error variance shown in Fig. 2. Figure 5 shows the trajectories of the true state, the traditional ensemble Kalman filter, and the two robust ensemble Kalman filters with efficiencies δ = 0.99 and 0.7 and radii r = 0.0001 and 0.01. For r = 0.01, the discarding filter is more precise than the Huberizing filter at t = 31, 32, and 33, but it is more imprecise in the absence of outliers from t = 10 to 20. It agrees that the larger the radius, the smaller bias and the larger error variance shown in Fig. 3.

Bias vs efficiency for the EnKF and two REnKFs for a one-dimensional linear system for t = 30, 31, 32, 33, 34, and 35. The additive outliers with ξt = 8 occur at times t = 31, 32, and 33.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

Bias vs efficiency for the EnKF and two REnKFs for a one-dimensional linear system for t = 30, 31, 32, 33, 34, and 35. The additive outliers with ξt = 8 occur at times t = 31, 32, and 33.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
Bias vs efficiency for the EnKF and two REnKFs for a one-dimensional linear system for t = 30, 31, 32, 33, 34, and 35. The additive outliers with ξt = 8 occur at times t = 31, 32, and 33.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 2, but for bias vs radius.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 2, but for bias vs radius.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 2, but for bias vs radius.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

The true states, EnKF, and two REnKFs with efficiencies δ = 0.99 and 0.7 for a one-dimensional linear system. The additive outliers with ξt = 8 occur at times t = 31, 32, and 33.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

The true states, EnKF, and two REnKFs with efficiencies δ = 0.99 and 0.7 for a one-dimensional linear system. The additive outliers with ξt = 8 occur at times t = 31, 32, and 33.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
The true states, EnKF, and two REnKFs with efficiencies δ = 0.99 and 0.7 for a one-dimensional linear system. The additive outliers with ξt = 8 occur at times t = 31, 32, and 33.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 4, but for radii r = 0.0001 and 0.01 rather than efficiencies.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 4, but for radii r = 0.0001 and 0.01 rather than efficiencies.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 4, but for radii r = 0.0001 and 0.01 rather than efficiencies.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
To examine the effect of innovations outliers, we suppose that the innovation outliers with kt = 25 occur at time t = 31, 32, and 33. Figure 6 shows the boxplots of the bias versus efficiency δ. Figure 7 shows the boxplots of the bias versus radius r. The bias stays at zero for all filters because the innovations outliers are set to have zero means. In terms of the error variance, the robust ensemble Kalman filters have increasing error variance as the efficiency δ decreases or as the radius r increases. At t = 31, 32, and 33, the traditional ensemble Kalman filter has the largest error variance. The efficiency δ gives a smaller error variance than the radius r gives because it has a larger clipping value compared to the radius and as such does not clip much. The Huberization is better than getting rid of observations in terms of the error variance. At times with no outliers, the robust ensemble Kalman filter, however, has a larger error variance than the traditional ensemble Kalman filter.

As in Fig. 2, but with innovations outliers with kt = 25 occurring at times t = 31, 32, and 33 with probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 2, but with innovations outliers with kt = 25 occurring at times t = 31, 32, and 33 with probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 2, but with innovations outliers with kt = 25 occurring at times t = 31, 32, and 33 with probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 3, but with innovations outliers with kt = 25 occurring at times t = 31, 32, and 33 with a probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 3, but with innovations outliers with kt = 25 occurring at times t = 31, 32, and 33 with a probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 3, but with innovations outliers with kt = 25 occurring at times t = 31, 32, and 33 with a probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
4. A multidimensional nonlinear system
a. The Lorenz model







b. Choice of the clipping height for the Lorenz model
We discuss how to choose the clipping height c and investigate the behavior of the robust ensemble Kalman filter for the Lorenz model. We use the average of the sample background covariance matrix

The average sample ensemble covariances between variable 21 and other variables using 10 000 ensemble members from t = 101 to 300.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

The average sample ensemble covariances between variable 21 and other variables using 10 000 ensemble members from t = 101 to 300.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
The average sample ensemble covariances between variable 21 and other variables using 10 000 ensemble members from t = 101 to 300.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
Since the dynamics of the model, distribution of the observations, and observation error statistics are homogenous, all components of the clipping height vector c have similar values. The radii r = 0.0001, 0.0005, 0.001, 0.01, and 0.05 respectively correspond to the clipping heights 3.30, 2.89, 2.7, 2, and 1.44. The efficiencies δ = 0.9999, 0.999, 0.99, 0.985, and 0.98 respectively correspond to the clipping heights 2.45, 1.62, 0.55, 0.32, and 0.16 when we Huberize observations, and they respectively correspond to the clipping heights 3.8, 3, 1.8, 1.43, and 1.06 when we discard observations. We use 200 replications for graphical representations with boxplots.
c. The effects of outliers




Bias vs efficiency of the EnKF and two REnKFs for variable 11 of the Lorenz model for t = 70–78. The additive outliers ξt = 10 occur at variables 11, 12, and 13 at times t = 71, 72, and 73.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

Bias vs efficiency of the EnKF and two REnKFs for variable 11 of the Lorenz model for t = 70–78. The additive outliers ξt = 10 occur at variables 11, 12, and 13 at times t = 71, 72, and 73.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
Bias vs efficiency of the EnKF and two REnKFs for variable 11 of the Lorenz model for t = 70–78. The additive outliers ξt = 10 occur at variables 11, 12, and 13 at times t = 71, 72, and 73.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 9, but for bias vs radius rather than efficiencies.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 9, but for bias vs radius rather than efficiencies.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 9, but for bias vs radius rather than efficiencies.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

The true values, EnKF, and two REnKFs with efficiency δ = 0.9999 and 0.98 for variable 11 of the Lorenz model. The additive outliers with ξt = 10 occur at variables 11, 12, and 13 at times t = 71, 72, and 73.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

The true values, EnKF, and two REnKFs with efficiency δ = 0.9999 and 0.98 for variable 11 of the Lorenz model. The additive outliers with ξt = 10 occur at variables 11, 12, and 13 at times t = 71, 72, and 73.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
The true values, EnKF, and two REnKFs with efficiency δ = 0.9999 and 0.98 for variable 11 of the Lorenz model. The additive outliers with ξt = 10 occur at variables 11, 12, and 13 at times t = 71, 72, and 73.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig.11, but for r = 0.0001 and 0.05.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig.11, but for r = 0.0001 and 0.05.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig.11, but for r = 0.0001 and 0.05.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
To investigate the effect of innovations outliers, we assume that the observation error comes from white noise with extreme variance at variables 11, 12, and 13 at t = 71, 72, and 73. Figure 13 shows the boxplots of the bias versus efficiency in the presence of innovations outliers with kt = 100 in the Lorenz model. Figure 14 shows the boxplots of the bias versus radius in the presence of the same innovations outliers. For r > 0 and δ < 1, the bias stays at zero but the error variance decreases to a certain point and then increases again as the clipping height decreases, and at t = 70 when no outliers occur, both robust ensemble Kalman filters experience a loss of accuracy.

As in Fig. 9, but for innovations outliers with kt = 100 occur at variables 11, 12, and 13 at times t = 71, 72, and 73 with a probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 9, but for innovations outliers with kt = 100 occur at variables 11, 12, and 13 at times t = 71, 72, and 73 with a probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 9, but for innovations outliers with kt = 100 occur at variables 11, 12, and 13 at times t = 71, 72, and 73 with a probability of contamination α = 0.2.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 13, but for bias vs radius.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1

As in Fig. 13, but for bias vs radius.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
As in Fig. 13, but for bias vs radius.
Citation: Monthly Weather Review 141, 12; 10.1175/MWR-D-13-00091.1
5. Discussion
We proposed a robust ensemble Kalman filter for the robust estimation of the state of a spatiotemporal dynamical system in the presence of observational outliers. We applied this robust ensemble Kalman filter to a one-dimensional linear system and a multidimensional nonlinear system. Using this filtering technique, which is based on the Huberization method, the negative effects of the outliers on the state estimates can be greatly reduced. The clipping values were selected using the efficiency and radius criteria. We compared the results of the robust ensemble Kalman filter with those from the classical ensemble Kalman filter. We also compared the robust ensemble Kalman filter based on the Huberization filter, which pulls the outliers back to c or −c, and the robust ensemble Kalman filter, which discards outliers. We found that compared to the conventional EnKF, the robust ensemble Kalman filter reduced the bias in the state estimates at the expense of increasing the error variance. The increase of the error variance differed depending on the filtering method. The Huberization filter was found to perform better than the discarding filter for the examples given in the paper, which may be because the model we used gives the true state. The robust ensemble Kalman filter is efficient with simple models, and we plan to test it in realistic ocean and atmospheric systems.
Finding the proper clipping values for a data assimilation system that assimilates many types of observations using a complex model is expected to be a labor intensive process. There is no reason to believe, however, that the process would be more challenging or would require more work than determining the parameters of the quality-control procedures currently used in operational numerical weather prediction. In fact, the parameters used in the current operational systems should provide invaluable information about the gross errors in the different types of observations, which could be used as guidance for the selection of the clipping values.
Acknowledgments
This work was supported in part by Award KUS-C1-016-04 made by King Abdullah University of Science and Technology (KAUST). Mikyoung Jun's research was also partially supported by NSF Grants DMS-0906532 and DMS-1208421. Istvan Szunyogh acknowledges the support from ONR Grant N000141210785.
REFERENCES
Anderson, E., and H. Järvinen, 1999: Variational quality control. Quart. J. Roy. Meteor. Soc., 125, 697–722.
Birmiwal, K., and J. Shen, 1993: Optimal robust filtering. Stat. Decis., 11, 101–119.
Birmiwal, K., and P. Papantoni-Kazakos, 1994: Outlier resistant prediction for stationary processes. Stat. Decis., 12, 395–427.
Burgers, G., P. J. van Leeuwen, and G. Evensen, 1998: Analysis scheme in the ensemble Kalman filter. Mon. Wea. Rev., 126, 1719–1724.
Calvet, L. E., V. Czellar, and E. Ronchetti, cited 2012: Robust filtering. [Available online at http://ssrn.com/abstract=2123477.]
Daley, R., 1991: Atmospheric Data Analysis. Cambridge Atmospheric and Space Science Series, Cambridge University Press, 457 pp.
Ershov, A. A., and R. S. Liptser, 1978: Robust Kalman filter in discrete time. IEEE Trans. Autom. Remote Control, 39, 359–367.
Fahrmeir, L., and H. Kaufmann, 1991: On Kalman filtering, posterior mode estimation and Fisher scoring in dynamic exponential family regression. Metrika, 38, 37–60.
Fahrmeir, L., and R. Kunstler, 1999: Penalized likelihood smoothing in robust state space models. Metrika, 49, 173–191.
Fox, A. J., 1972: Outliers in time series. J. Roy. Stat. Soc., B34, 350–363.
Genton, M. G., 2003: Breakdown-point for spatially and temporally correlated observations. Developments in Robust Statistics, R. Dutter et al., Eds., Springer, 148–159.
Genton, M. G., and A. Lucas, 2003: Comprehensive definitions of breakdown-points for independent and dependent observations. J. Roy. Stat. Soc., B65, 81–94.
Genton, M. G., and A. Lucas, 2005: Discussion of “Breakdown and groups” by L. Davies and U. Gather. Ann. Stat., 33, 988–993.
Hampel, F. R., 1968: Contributions to the theory of robust estimation. Ph.D. thesis, University of California.
Harlim, J., and B. R. Hunt, 2007: A non-Gaussian ensemble filter for assimilating infrequent noisy observations. Tellus, 59A, 225–237.
Houtekamer, P. L., and H. L. Mitchell, 1998: Data assimilation using an ensemble Kalman filter technique. Mon. Wea. Rev., 126, 796–811.
Huber, P. J., 1981: Robust Statistics. Wiley, 308 pp.
Ingleby, N. B., and A. C. Lorenc, 1993: Bayesian quality control using multivariate normal distributions. Quart. J. Roy. Meteor. Soc., 119, 1195–1225.
Kalman, R. E., 1960: A new approach to linear filtering and prediction problems. J. Basic Eng., 82, 34–45.
Kassam, S. A., and H. V. Poor, 1985: Robust techniques for signal processing: A survey. Proc. IEEE, 73, 433–481.
Lorenz, E. N., and K. A. Emanuel, 1998: Optimal sites for supplementary weather observations: Simulation with a small model. J. Atmos. Sci., 55, 399–414.
Luo, X., and I. Hoteit, 2011: Robust ensemble filtering and its relation to covariance inflation in the ensemble Kalman filter. Mon. Wea. Rev., 139, 3938–3953.
Maronna, A., R. D. Martin, and V. J. Yohai, 2006: Robust Statistics: Theory and Methods. Wiley, 436 pp.
Martin, R. D., and A. E. Raftery, 1987: Robustness, computation and non-Euclidean models. J. Amer. Stat. Assoc., 82, 1044–1050.
Meinhold, R. J., and N. D. Singpurwalla, 1983: Understanding the Kalman filter. Amer. Stat., 37, 123–127.
Meinhold, R. J., and N. D. Singpurwalla, 1989: Robustification of Kalman filter models. J. Amer. Stat. Assoc., 84, 479–486.
Naveau, P., M. G. Genton, and X. Shen, 2005: A skewed Kalman filter. J. Multivariate Anal., 95, 382–400.
Ruckdeschel, P., 2010: Optimally robust Kalman filtering. Berichte des Fraunhofer ITWM 185, 53 pp.
Schick, I. C., and S. K. Mitter, 1994: Robust recursive estimation in the presence of heavy-tailed observation noise. Ann. Stat., 22, 1045–1080.
Schlee, F. H., C. J. Standish, and N. F. Toda, 1967: Divergence in the Kalman filter. Amer. Inst. Aeronaut. Astronaut. J., 5, 1114–1120.
Stockinger, N., and R. Dutter, 1987: Robust time series analysis: A survey. Kybernetika, 23, 3–88.
Szunyogh, I., E. J. Kostelich, G. Gyarmati, E. Kalnay, B. R. Hunt, E. Ott, E. Satterfield, and J. A. Yorke, 2008: A local ensemble transform Kalman filter data assimilation system for the NCEP global model. Tellus, 60, 113–130.
Tavolato, C., and L. Isaksen, 2010: Huber norm quality control in the IFS. ECMWF Newsletter, No. 122, ECMWF, Reading, United Kingdom, 27–31.
Tukey, J. W., 1970: Exploratory Data Analysis. Vol. 1. Addison-Wesley, 688 pp.
West, M., 1981: Robust sequential approximate Bayesian estimation. J. Roy. Stat. Soc., B43, 157–166.
West, M., 1983: Generalized linear models: Scale parameters, outlier accommodation and prior distributions. Bayesian Stat., 2, 531–558.
West, M., 1984: Outlier models and prior distributions in Bayesian linear regression. J. Roy. Stat. Soc., B46, 431–439.
Whitaker, J. S., and T. M. Hamill, 2002: Ensemble data assimilation without perturbed observations. Mon. Wea. Rev., 130, 1913–1924.