1. Introduction
In numerical weather prediction (NWP), necessity often arises to evaluate sensitivity of forecasts with respect to changes in initial conditions. A natural measure of such sensitivity is a gradient of some interesting aspects of forecasts with respect to initial conditions. This gradient-based sensitivity is called adjoint sensitivity and can be quantified by running adjoint models associated with numerical models from which forecasts are produced (Errico and Vukicevic 1992; Langland et al. 1995). Since adjoint models are essential in any four-dimensional variational (4DVAR) data assimilation systems, adjoint sensitivity can be estimated relatively easily if we have access to a 4DVAR system by utilizing the driving model and its adjoint code therein.
However, adjoint models are not always available in practice due to their high maintenance costs when NWP models are frequently updated with more sophisticated components. In such cases, adjoint sensitivity can still be estimated from ensembles of analyses and forecasts (Ancell and Hakim 2007; Hakim and Torn 2008), which are always available in any ensemble prediction system. This ensemble-based estimation of adjoint sensitivity can be reinterpreted as taking multiple linear regression coefficients between forecasts and initial conditions as a measure of sensitivity. Regression coefficients in this case are determined by two covariances: analysis error covariances
The regression-based estimation of adjoint sensitivity requires inversion of
The use of ensemble sensitivity has expanded beyond its original derivation as a convenient tool or an intermediate step in quantifying the impact of targeted observations and it has later been used independently in many sensitivity studies (Torn and Hakim 2008; Garcies and Homar 2009; Torn 2010; Chang et al. 2013; Hanley et al. 2013; Ito and Wu 2013; Bednarczyk and Ancell 2015; Brown and Hakim 2015; Hill et al. 2016; Yokota et al. 2016). The extensive use of ensemble sensitivity in sensitivity analysis can be attributed to its simplicity in formulation where the otherwise expensive inversion of
Along this vein, in order to get rid of the controversial diagonality assumption, Hacker and Lei (2015) came back to the understanding of the ensemble-based estimation of adjoint sensitivity as multiple linear regression coefficients. They avoided the inversion of
It is expected that multivariate ensemble sensitivity should be different from univariate ensemble sensitivity since the former does not rely on the diagonal approximation, as claimed and demonstrated in Hacker and Lei (2015) and Ren et al. (2019). Curiously, however, in their numerical experiments, Hacker and Lei (2015) found that changes of forecasts yielded by univariate and multivariate ensemble sensitivity with respect to one standard deviation changes of initial conditions are identical when localization is not applied (see Fig. 1a in their paper). The authors did not give any adequate explanations for such a coincidence. Note that localization is an ad hoc technique to mitigate the impact of sampling errors on sample covariances. Thus, in the limited case when we can access a large number of ensemble members so that localization can be turned off, those results imply an interesting dilemma: despite the difference in foundations of univariate and multivariate ensemble sensitivity, they have the same effect on changes of forecasts. In other words, we return to multivariate ensemble sensitivity to avoid the diagonality assumption, and finally we end up with univariate ensemble sensitivity again. How can we explain the effective work of univariate ensemble sensitivity given its controversial diagonality assumption?
In this paper, we show that it is not accidental that univariate and multivariate ensemble sensitivity yield the same changes of forecasts. Underlying this dilemma is a theory that establishes a robust foundation for ensemble sensitivity. In short, ensemble sensitivity is equivalent to an effective adjoint sensitivity that yields the same forecast response as adjoint sensitivity without a need of calculating perturbations at correlated points. This is justified by the proof that the same forecast response is obtained for any analysis error covariances. By showing that ensemble sensitivity is a rigorous concept, we can resolve confusion regarding the use of ensemble sensitivity in practice. Furthermore, we prove that ensemble sensitivities to all elements of the initial state form the most sensitive initial perturbation which maximizes changes of forecasts among all initial perturbations with the same magnitudes—a fact that, to the best knowledge of the authors, has been hitherto unrecognized in the literature.
This paper is organized as follows. Section 2 prepares a mathematical background for deploying a robust foundation of ensemble sensitivity. In section 3, we revisit the formerly proposed definition of sensitivity as regression coefficients and seek a new definition that also takes into account probability distributions of initial perturbations. Section 4 shows how ensemble sensitivity follows naturally from the established framework. Another important role of the proposed ensemble sensitivity is examined in section 5 in which ensemble sensitivity is proved to form the most sensitive analysis perturbation. Finally, section 6 summarizes the main findings of this study.
2. Background on ensemble sensitivity
Due to a limited number of ensemble members,
Thus, (7) shows that we can calculate adjoint sensitivity by using either the primal form (
3. Standardized ensemble sensitivity
Without using the diagonality assumption due to its potential confusion, we will define ensemble sensitivity through the forecast response considering both linear transformations
4. A unified theory for ensemble sensitivity
There exist many probability distributions of Δx0 that have the mean 0 and the covariance
As we have seen above, the ensemble sensitivity follows naturally from the definition of standardized ensemble sensitivity (11) and the multivariate normal distribution of Δx0. The controversial diagonality assumption is irrelevant in construction of ensemble sensitivity. This diagonality assumption has been the source of confusion in understanding and applying ensemble sensitivity. First, given the inadequacy of neglecting spatial and interelement correlation in analysis perturbations, ensemble sensitivity has often been considered as a rough approximation of adjoint sensitivity, and the validity of its use has been only justified by its simplicity. Second, since the diagonality assumption is equivalent to assuming that all elements Δx0i are independent, ensemble sensitivity has often been regarded to be univariate in nature, which has led to a common conception that sensitivity as represented by ensemble sensitivity is overestimated because significant contributions from other elements are ignored. The mathematical arguments in this section point out that ensemble sensitivity is multivariate in nature and is not just an approximation to adjoint sensitivity but rather is a distinct measure of sensitivity with a robust mathematical foundation.
In practice, both vectors s (standardized ensemble sensitivity) and std(ΔJ)s (normalized ensemble sensitivity) can be used as a measure of sensitivity. As shown by (16) si are simply correlations between the forecast response and initial conditions. As a result, it is relatively easy to estimate si and localization can even be imposed on s. Furthermore, because si are dimensionless unlike std(ΔJ)si, which are dimensioned, they have an advantage of allowing us to compare sensitivities of different forecast metrics. However, since std(ΔJ)si give changes of the forecast response when Δx0i = std(Δx0i), if forecast metrics are of the same type so that all std(ΔJ)si have the same unit, e.g., rainfall at different areas, it is better to use std(ΔJ)si. This is because two forecast metrics with the same correlations si = cor(Δx0i, ΔJ) will show different responses if their standard deviations std(ΔJ) are different.
Equation (16) points out a surprising fact: standardized ensemble sensitivities are simply cross correlation coefficients cor(Δx0, ΔJ) regardless of the form of
5. Ensemble sensitivity as the most sensitive perturbation
In this section, we show the relevance of ensemble sensitivity in another important problem related to the sensitivity problem. Instead of estimating how each state element will change the forecast response, we now examine which analysis perturbation among all possible ones with the same magnitudes will yield the largest change in the forecast response. Due to collaborative effect among different elements, this most sensitive analysis perturbation is not expected to be any individual analysis perturbations. As we shall see, intriguingly, ensemble sensitivity is closely involved in this optimization problem.
It is natural to continue with a question on the second most sensitive perturbation by taking the optimization problem further. Thus, we will find the analysis perturbation Δx0 that maximizes the norm of ΔJ subject to the two constraints
6. Discussion and conclusions
In ensemble sensitivity analysis, adjoint sensitivity is estimated from the regression coefficients that are obtained by regressing changes in a forecast response on initial perturbations. Ancell and Hakim (2007) pointed out the existence of a new kind of sensitivity in the ensemble context that they called ensemble sensitivity, and ensemble sensitivity has been proved to be a very useful sensitivity measure in practice. While being proved successful, ensemble sensitivity leaves some confusion in its interpretation because its derivation involves an ad hoc operation that replaces the analysis error covariance
-
why is univariate ensemble sensitivity so effective in practice despite its (apparent) reliance on the diagonal approximation?
-
why multivariate and univariate ensemble sensitivity result in equivalent sensitivity?
-
what are potential applications of ensemble sensitivity to optimal initial conditions that affect the forecast metric?
The key to resolving the three questions above is to distinguish two kinds of sensitivity: (i) the gradient
To overcome these deficits and inconveniences, we approach ensemble sensitivity in the sense of (ii) above that accounts for contributions to ΔJ not only from the gradient of forecast response but also from the initial perturbation Δx0. We have developed a mathematical background to precisely define such a sensitivity measure, which we term standardized ensemble sensitivity, and showed that it leads to the formerly proposed ensemble sensitivity without making any ad hoc assumption like the diagonal approximation, making it a sound and appealing choice as a sensitivity measure. Thus, the mathematical results provide rigorous explanations to the three questions raised above. First, it is found that the same change of the forecast response is obtained regardless of whether
It is worth emphasizing that the theory points out standardized ensemble sensitivity as an important and rigorous concept rather than ensemble sensitivity. This is because the vector containing standardized ensemble sensitivities carries in itself three important quantities at the same time: 1) standardized changes of the forecast response with one standard deviation changes of individual state variables; 2) correlations between the forecast response and individual state variables; and 3) the most sensitive analysis perturbation, which, for a quadratic response function, coincides with the leading ensemble singular vector. These different theoretical aspects provide a comprehensive picture of ensemble sensitivity.
The theoretical framework developed in this study clarifies the difference between the conventional, gradient-based sensitivity measure and the newly proposed impact-based sensitivity measure. While the former approaches the sensitivity problem from a dynamical systems perspective, the latter is better suited for assessment of sensitivity from a probabilistic perspective of predictability and causality. Future practitioners of ensemble sensitivity analysis are advised to clarify which aspect of sensitivity they intend to investigate before initiating analysis. We hope the theoretical elucidation given in this manuscript will be helpful in this regard.
Acknowledgments.
This work was supported by the Ministry of Education, Culture, Sports, Science and Technology (MEXT) through the Program for Promoting Researches on the Supercomputer Fugaku JPMXP1020351142 “Large Ensemble Atmospheric and Environmental Prediction for Disaster Prevention and Mitigation” (hp200128, hp210166, hp220167), Foundation of River and basin Integrated Communications (FRICS), and JST Moonshot R&D project (Grant JPMJMS2281). Constructive comments from Dr. Brian Ancell have allowed us to significantly improve the manuscript by correctly reframing the discussion. Comments from two other anonymous reviewers are also acknowledged.
Data availability statement.
No datasets were generated or analyzed during the current study.
REFERENCES
Ancell, B. C., and G. J. Hakim, 2007: Comparing adjoint- and ensemble-sensitivity analysis with applications to observation targeting. Mon. Wea. Rev., 135, 4117–4134, https://doi.org/10.1175/2007MWR1904.1.
Ancell, B. C., and A. A. Coleman, 2022: New perspectives on ensemble sensitivity analysis with applications to a climatology of severe convection. Bull. Amer. Meteor. Soc., 103, E511–E530, https://doi.org/10.1175/BAMS-D-20-0321.1.
Bednarczyk, C. N., and B. C. Ancell, 2015: Ensemble sensitivity analysis applied to a southern plains convective event. Mon. Wea. Rev., 143, 230–249, https://doi.org/10.1175/MWR-D-13-00321.1.
Bishop, C. M., 2006: Pattern Recognition and Machine Learning. Springer, 738 pp.
Brown, B. R., and G. J. Hakim, 2015: Sensitivity of intensifying Atlantic hurricanes to vortex structure. Quart. J. Roy. Meteor. Soc., 141, 2538–2551, https://doi.org/10.1002/qj.2540.
Chang, E. K. M., M. Zheng, and K. Raeder, 2013: Medium-range ensemble sensitivity analysis of two extreme Pacific extratropical cyclones. Mon. Wea. Rev., 141, 211–231, https://doi.org/10.1175/MWR-D-11-00304.1.
Errico, R. M., and T. Vukicevic, 1992: Sensitivity analysis using an adjoint of the PSU–NCAR mesoscale model. Mon. Wea. Rev., 120, 1644–1660, https://doi.org/10.1175/1520-0493(1992)120<1644:SAUAAO>2.0.CO;2.
Garcies, L., and V. Homar, 2009: Ensemble sensitivities of the real atmosphere: Application to Mediterranean intense cyclones. Tellus, 61A, 394–406, https://doi.org/10.1111/j.1600-0870.2009.00392.x.
Hacker, J. P., and L. Lei, 2015: Multivariate ensemble sensitivity with localization. Mon. Wea. Rev., 143, 2013–2027, https://doi.org/10.1175/MWR-D-14-00309.1.
Hakim, G. J., and R. D. Torn, 2008: Ensemble synoptic analysis. Synoptic-Dynamic Meteorology and Weather Analysis and Forecasting: A Tribute to Fred Sanders, Meteor. Monogr., No. 33, Amer. Meteor. Soc., 147–162, https://doi.org/10.1175/0065-9401-33.55.147.
Hanley, K. E., D. J. Kirshbaum, N. M. Roberts, and G. Leoncini, 2013: Sensitivities of a squall line over central Europe in a convective-scale ensemble. Mon. Wea. Rev., 141, 112–133, https://doi.org/10.1175/MWR-D-12-00013.1.
Hill, A. J., C. C. Weiss, and B. C. Ancell, 2016: Ensemble sensitivity analysis for mesoscale forecasts of dryline convection initiation. Mon. Wea. Rev., 144, 4161–4182, https://doi.org/10.1175/MWR-D-15-0338.1.
Ito, K., and C.-C. Wu, 2013: Typhoon-position-oriented sensitivity analysis. Part I: Theory and verification. J. Atmos. Sci., 70, 2525–2546, https://doi.org/10.1175/JAS-D-12-0301.1.
Langland, R. H., R. L. Elsberry, and R. M. Errico, 1995: Evaluation of physical processes in an idealized extratropical cyclone using adjoint sensitivity. Quart. J. Roy. Meteor. Soc., 121, 1349–1386, https://doi.org/10.1002/qj.49712152608.
Limpert, G. L., and A. L. Houston, 2018: Ensemble sensitivity analysis for targeted observations of supercell thunderstorms. Mon. Wea. Rev., 146, 1705–1721, https://doi.org/10.1175/MWR-D-17-0029.1.
Ren, S., L. Lei, Z.-M. Tan, and Y. Zhang, 2019: Multivariate ensemble sensitivity analysis for Super Typhoon Haiyan (2013). Mon. Wea. Rev., 147, 3467–3480, https://doi.org/10.1175/MWR-D-19-0074.1.
Saltelli, A., M. Ratto, T. Andres, F. Campolongo, J. Cariboni, D. Gatelli, M. Saisana, and S. Tarantola, 2008: Global Sensitivity Analysis: The Primer. John Wiley and Sons, 304 pp.
Torn, R. D., 2010: Ensemble-based sensitivity analysis applied to African easterly waves. Wea. Forecasting, 25, 61–78, https://doi.org/10.1175/2009WAF2222255.1.
Torn, R. D., and G. J. Hakim, 2008: Ensemble-based sensitivity analysis. Mon. Wea. Rev., 136, 663–677, https://doi.org/10.1175/2007MWR2132.1.
Torn, R. D., and G. J. Hakim, 2009: Initial-condition sensitivity of western Pacific extratropical transitions determined using ensemble-based sensitivity analysis. Mon. Wea. Rev., 137, 3388–3406, https://doi.org/10.1175/2009MWR2879.1.
Yokota, S., H. Seko, M. Kunii, H. Yamauchi, and H. Niino, 2016: The tornadic supercell on the Kanto Plain on 6 May 2012: Polarimetric radar and surface data assimilation with EnKF and ensemble-based sensitivity analysis. Mon. Wea. Rev., 144, 3133–3157, https://doi.org/10.1175/MWR-D-15-0365.1.