1. Introduction
The generalized inversion (GI) is the most flexible tool for combining information on the state of the ocean provided by physical models and observations (Egbert and Bennett 1996; Kivman 1997a). The approach explicitly takes into account uncertainties inherent in both model equations and measurements; this is achieved by the proper weighting of corresponding cost terms. An assumption that the model equations are perfect has been widely used in data assimilation. Though accounting for the model errors allows one to improve accuracy of the prediction (see, e.g., Evensen 1997; Gong et al. 1998), the question of what the model errors are remains an open scientific issue (Courtier 1997). The main goal of this paper is to show that as soon as we combine information from different sources into a unique solution, methods of the information theory are applicable and may be of help in selecting a number of error statistics parameters.
Though there have been some attempts to estimate systematic model errors (Dee and da Silva 1998; Griffith and Nichols 1996), in most studies, the errors in the model equations and initial and boundary conditions have been taken to be unbiased and Gaussian distributed. That is, in the linear case, the mean E(ψ) is assumed to be given by the model solution satisfying the initial and boundary data exactly (Bennett 1992). The initial and boundary conditions must be obtained from observations. Thus, the data obtained at the initial time and/or at the boundary of the domain are considered to be equal to the mean of the prior E(ψ), while the data assimilated inside the domain do not possess that exclusive property. It is a rather general situation that there are not enough observations at the intial moment or at the boundary; then, we should introduce bogus data to get a well-posed problem (Bennett and Miller 1991). We work with the two datasets that get different interpretations: the data at the initial time and/or at the boundary of the domain are considered to be equal to the mean E(ψ), while the data assimilated inside the domain are not. Partitioning the data into the two subsets is arbitrary to some extent; thus, essential asymmetry in the interpretation of these two sets is introduced.
Having few observations, we cannot estimate everything and should restrict our attention to tuning a certain number of statistical parameters, such as error variances and error decorrelation scales, while keeping other inputs fixed. Two methods have been widely used in data assimilation community, namely, maximum likelihood (ML; Dee 1995) and general cross-validation (GCV; Egbert et al. 1994). As the statistical theory suggests (Cramer 1954), if the true prior PDF is indeed of an assumed functional shape, then ML is the most powerful tool for estimating parameters of the prior PDF from the data in the sense of minimizing variances of the estimates. However, if the assumptions about p(ψ) are violated, this approach may be less effective than some other tuning criteria (Wahba 1985). Gao (1994) demonstrated that in a particular example, the ML substantially overestimated a parameter of smoothing (error decorrelation length scale) in comparison to the GCV. When even the mean E(ψ) is not known, the ML estimates of the statistical parameters are suspect.
Another approach to building the prior PDF may be based on the principle of maximum entropy (PME). The PME and Bayesian viewpoints complement each other (Jaynes 1988a). Indeed, the former assesses the prior probabilities in the most cautious way in the sense that the PDF obtained assumes nothing beyond what was actually known and used as input. Once the prior PDF is chosen, the Bayes method updates it when new data become available. It is worth emphasizing that probabilities derived from the PME should be considered as“degrees of belief” rather than “relative frequencies” (Jaynes 1988b); it has been precisely the interpretation of probabilities used by Tarantola (1987) in the statistical inverse problems theory. Thacker and Holloway (1990) pointed out that the PME might be of help in choosing the prior in oceanographic inverse problems.
We outline the fundamentals of the PME in section 2. Then, we show that in the linear case, any choice of the prior generates a measure in the space of possible outcomes, which may be viewed as a generalization of the probability measures. Thus, one is able to use the PME to select a certain number of statistical parameters (see section 3 for the theory and section 4 for a numerical example). Any method for tuning the statistical parameters substantially complicates data assimilation and is very difficult to employ in realistic large-dimensional problems. The maximum data cost criterion (MDC), which is easier to implement, is shown to be an approximation to the PME. Though the MDC established on heuristic grounds has worked well (Kivman 1997c; Kurapov and Kivman 1999), its statistical background is clarified only now (see section 5).
The cost functional generally includes certain smoothing terms, which are introduced to provide well-posedness. It was demonstrated by Gong et al. (1998) that they could be as important as the weight controlling the trade-off between the model and the data. We will show in section 6 how to tune a smoothing parameter in the scope of the PME and the MDC.
2. Maximum entropy formalism
The concept of entropy originated about a century ago within the context of a fundamental but specific physical problem; it has recently found widespread application in dynamical systems theory, ergodicity theory, inverse problems, communication theory, numerical analysis, theory of functions, decision theory, etc., that is, to problems staying far from the classical thermodynamics. Thanks to Shannon (1948), all these applications were made possible only after recognizing the fact that entropy has much more wider content than had been previously thought. To be concise, Shannon proved that any measure of uncertainty in a probability distribution is proportional to the Boltzmann–Gibbs entropy if that measure is
(i)ii a continuous function of probabilities;
(ii)i a decreasing function of the number of possible events in the case when they are equally probable;and
(iii) consistent, that is, if there is more than one way of evaluating its value, every possible way yields the same result.
Thus, the PME is in fact a refined version of Bernoulli’s principle of insufficient reason for assigning probabilities to possible events in the light of all prior knowledge (see Thacker and Holloway 1990, and references therein). It should be emphasized that the PME is an extremely conservative procedure in the sense that its application yields a probability distribution that is as uncertain or “spread” as allowed by properties we wish the distribution to have and that possesses a minimal information content among all distributions satisfying the constraints. In other words, the maximum entropy distribution is free from any information additional to that we actually knew and used as input.
3. An entropy approach to selecting priors in weak constraint data assimilation
Calculating the weights am and ad from the error variances with (6), (7), (11) is a tough routine (Urban 1996). We will not be able to determine the weights unless both σd and σm are known. Luckily, we are rarely interested in the whole PDF and would be pleased to obtain a reliable estimate of its maximizer or the mean. To calculate it, we only need the ratio w = ad/am. Surprisingly, if there is no model solution fitting the data exactly (i.e., the model and the data are in contradiction), the PME makes it possible to proceed a bit further.
We emphasize that the operators Mm and Md depend not only on weights am, ad and the operators A and L, but also on the specific form of the inner products ( · )r and ( · )d. Involved in the definitions of the operators A∗ and L∗, they reflect our a priori assumptions about the data and model error covariances, Cd and Cm. In the standard practice of data assimilation, we know the approximate Cd and have few ideas about what Cm looks like. Hence, we might try to maximize H(M) as a guide to selecting the priors. However, dimD ≪ dimR as a rule, and maximizing the entropy (25) will not determine Cm uniquely. In this instance, if we could estimate the relative weight w and the decorrelation length of the model errors, that would be quite a satisfactory result.
4. Case study: Weak constraint data assimilation into a 1D model for tides in Tatarsky Strait
a. Model
We would like the generalized inverse of our 1D model to reproduce features essential for weak constraint data assimilation into 2D and 3D models. One of them is ill-posedness of the problem in the continuum limit (see section 6). To allow for it, the dynamical errors are supposed to be uncorrelated. The same are assumed for the data errors. If we penalized the mean-square norm of the residuals referred to the model equations (27), (28), we would obtain ζ, q ∈
b. Maximum entropy and the quality of the solution
Inverse solutions have been computed at four different grids containing N = 20, 38, 75, or 149 nodes. At each grid, a series of calculations have been made to obtain the entropy H as a function of the weight ratio w. Along with it, we have estimated root-mean-square (rms) errors of the solution with respect to the data constraining the solution (rmsas) and to the data left for verification (rmsυ) (Fig. 2).
In principle, we wish to obtain a solution that would have as small rmsas and rmsυ as possible. By increasing the data weight w, we can achieve an arbitrary small value of the former. However, the latter does not generally diminish with w → ∞ and usually has a positive lower bound. Closeness between rmsas and rmsυ points to the fact that the inverse solution experiences minor sensitivity to the choice of the subset of the data taken from the whole set of observations and assimilated (presuming the data are all of the same quality). Alternatively, large differences between rmsas and rmsυ mean that the inverse solution varies notably depending on whether it is constrained by certain data; in the latter case, the inverse solution will be far from the truth. Therefore, in speaking about the quality of the solution, we impose two basic requirements. First, the solution should yield rmsυ as small as possible. Second, we would want rmsas to be close to rmsυ if the two datasets are of the same accuracy.
Let us check the quality of the solution corresponding to the weight w = wme at which it yields the maximum entropy. At each “coarse” grid (grids 20, 38, and 75), we will be quite satisfied if rmsυ is close to its minimum (see Fig. 2). Underdetermining the weight leads to a much bigger value. The second quality criterion is also fulfilled at w = wme: rmsυ is close to rmsas (Table 1). It is worth noting that the mean data amplitude is about 47 cm. Keeping in mind the crudeness of the model, the approximate 20% relative error seems satisfactory.
Overestimating w does not increase rmsυ significantly. However, rmsas becomes close to zero, and thus the inverse solution is sensitive to the input information. The other limit case (w → 0 or am → ∞) corresponds to the so-called strong constraint data assimilation. This approach has been widely used in the oceanographic inverse modeling and is based on the assumption that the model equations are error free. An observation made in several studies (see Gong et al. 1998) was that the weak constraint estimates (w > 0) were better if compared with the strong constraint solutions. Figure 2 clearly indicates the advantage of weak constraint data assimilation.
At the very fine resolution (grid 149), the situation is not as good as that at the coarser grids. The PME overestimates the data weight, which leads to a large discrepancy between rmsas and rmsυ. In the case study, the entropy peak value is attained at grid 38, and the maximum entropy value decreases when a finer or coarser grid is taken (Table 1). Numerical discretization acts as a regularizor to the problem, which is ill-posed in the continuum limit (see section 6). Consequently, following the PME we are to adopt grid 38 as optimal. Numerical results support our choice: rmsυ(wme) obtained at grid 149 is not improved in comparison with the value obtained at grid 75 (see Table 1). Thus, choosing the grid much finer than that having the biggest entropy does not improve the accuracy of the inverse solution.
5. Maximum data cost criterion
To compute the entropy, one needs a calculation of eigenvalues of the operator Md or Mm, which is not feasible to obtain for large N. We would like to find an approximation to the entropy approach and to avoid explicit calculation of the entropy. It is easy to check the multiple measurements example discussed in section 2 so that the maximum-entropy measure dα yields the weight maximizing the cost a1(ỹ − y1)2 when a2 is fixed, and vice versa. Does the same fact hold for operator-valued measures appearing in multidimensional spaces? A relevant theorem is proven in appendix B. The MDC coincides with the PME under the assumption that the shape of the error covariance matrices allows the entropy to attain its upper bound (26) at some w. Fixing Cm and Cd a priori, we cannot guarantee that the upper bound of the entropy will be attained. However, we may still select w in accordance with the MDC criterion, considering it as an approximation to the PME.
Let us return to the case study. The MDC predicts weights wmdc = 15 and wmdc = 5 at grids 20 and 38, correspondingly. Though these weights are underestimated in comparison with the maximum-entropy weights wme = 45 and wme = 22, the MDC solutions are still of good quality (see Fig. 2). At grid 75, the MDC weight is already an order of magnitude less than the PME weight (wmdc = 1 against wme = 10). As we know from above, a sort of smoothing should be applied at this resolution. Further grid refinement leads to splitting the extremum of the data cost curve. Hence, the necessary condition for the MDC to coincide with the PME is violated (see appendix B).
6. Smoothing
What lies behind difficulties rising at the fine resolution? Since the model errors are assumed to be uncorrelated, the Hilbert structure in Ψ, introduced by the inner product (18), does not guarantee smoothness of ψ in the continuum limit. Although the inverse solution is bounded at any discrete grid, it is very sensitive to the data noise and to the spatial resolution if the grid becomes too fine. As an example, if the data are values of model variables at isolated points, the inverse solution has pathological behavior at the data sites with holes and spikes in their vicinity, and it does not converge to a continuous function as numerical lattice tends to zero (Bennett and McIntosh 1982). Kivman (1997b) has demonstrated that ill-posedness is not just a mathematical technicality; refining the grid can in fact degrade the inverse solution compared with that obtained at a coarser numerical lattice. In principle, well-posedness may be restored by reducing the relative data weight with grid refinement. However, this scheme relies on the assumption that the continuous model equations are error free, which is difficult to judge.
Smoothing at the fine resolution allows us to account for the model uncertainties that are not caused by discretization. Several strategies have been put forward to smooth the solution to the generalized inverse. Bennett and McIntosh (1982) used spatially variable dynamical weights with suitable singularities at the data sites. Provost and Salmon (1986) augmented the cost functional by a penalty on higher derivatives of the solution. Dynamical error covariances with nonzero off-diagonal elements were adopted in Egbert et al. (1994). Instead of penalizing the L2 norm of the model equation residuals, Kivman (1996) imposed a penalty on higher norms.
The theory of partial differential equations says that if the inverse solution satisfies an elliptic equation, it is continuous everywhere except, possibly, at the data sites. Hence, smoothing is necessary only in the data neighborhood, not in the whole computational domain. Based on this fact, we propose to “spread” the data in order to regularize the problem. The procedure consists in the insertion of bogus data of the same value as the original datum in the vicinity of each data point. It is possible to show that the inverse solution converges to a smooth function in the continuum limit if the radius of spreading is fixed.
From the physical standpoint, the decorrelation length of the field of interst is nonzero, and thus the values that the field takes in the vicinity of the data site is close to what is actually observed. Thus, inserting such bogus data around the measurement location within an area of strong correlations is plausible. Furthermore, we have more ideas about the decorrelation length of physical fields than about that of the model errors.
Because the number of observations M is always finite, the norm generated by (18) becomes only a seminorm at some resolution in 2D and 3D applications. That is, 〈ψ, ψ〉 = 0 does not necessarily mean that ψ = 0. Consequently, we will have to close the inverse problem. When we consider an open system exchanging the energy with its surroundings, the number Mo of numerical grid nodes at the open boundary may exceed the amount of the data there, and maximizing (19) becomes an ill-posed problem, even if M > Mo. A standard regularization routine reduces to spreading the data along the open boundary (interpolating between available data). Here, we propose to apply this regularization methodology to the interior observations.
How can we embed a procedure for tuning the smoothing parameter into the MDC? We may try to determine a threshold spatial resolution where the general inverse starts working badly and additional regularization is already necessary. Smoothing makes the inverse solution more tolerant to the inputs. However, the solution must be insensitive to a smoothing parameter. The optimal weight wmdc estimated by the MDC can play the role of a sensitivity measure of the inverse solution to the spreading length. If the data cost curve changes dramatically after the data are spread to the closest neighbors of the actual data locations, the spreading is not necessary at all.
Does the sensitivity study point to the same threshold resolution as that estimated by the PME (N = 75)? Yes, it does. As Fig. 3 depicts, spreading the data over adjacent nodes results in a dramatic increase of the MDC weight for N = 20. At grid N = 38, spreading the data causes the appearance of the second maximum. For N = 75, the MDC weight does not change if we spread the data just over two neighboring nodes, while further spreading drastically affects the MDC weight. An interesting point is that the optimal spreading length obtained at the two finer grids (N = 75, 149) remains the same in physical units. Thus, the smoothing length scale may be tuned at the coarser grid and used to obtain a solution at the finer resolution.
7. Summary
The principle of maximum entropy (PME) is a rigorous scheme for estimating weights and smoothing parameters involved in weak constraint data assimilation. Though the operator-valued measure generated by the prior seems to be an extravagant mathematical object, measures of that type arise naturally in the quantum probability theory. They appear in there instead of classical probability distributions because of the incomplete observability of quantum systems (Kholevo 1972). We face quite the same situation in data assimilation. The data are not enough to estimate even the mean E(ψ), let alone the full statistics. In this instance, we invoke a hypothesis about the distribution of the model errors, which are unobservable; thus, the data space and the space of the model errors are generally nonisomorphic. Only in the simplest example of multiple measurements of one variable do we have Ψ = D = R; thus the classical probability measure dα can be used.
Application of the PME to weak constraint data assimilation in the 1D tide model has allowed us to obtain the inverse solution of good quality. In addition, the PME has been able to point at the threshold spatial resolution where the generalized inverse already requires smoothing.
The PME is difficult to employ in large-dimensional problems. In this instance, the maximum data cost criterion (MDC) is a feasible alternative to the PME. The MDC was used earlier as a heuristic rule for estimating the relative data weight in weak constraint data assimilation (Kivman 1997c; Kurapov and Kivman 1999). Connection between the MDC and the PME sheds light on reasons for efficiency of the former. Namely, the MDC can be viewed as an approximation to the PME when the curve ψ(w) ⊂ Ψ is close to a straight segment. It is also important that the MDC is computationally simpler than the cross-validation. Moreover, if the data are from different sources (say, those from pressure gauges and current meters; in situ and satellite altimetry measurements), the method can be applied to multiple weights tuning (Kurapov and Kivman 1999).
In the case study, the MDC underestimated the data weight in comparison with the maximum entropy value. Nevertheless, the inverse solution was fairly good. We should admit that the MDC still requires large computational efforts because the data cost value can be determined only after the inverse solution was obtained. Luckily, the weight can be and should be estimated at the coarse spatial resolution where no smoothing is required. Then we transfer it onto a finer grid.
The method also points to the fine grids where smoothing is necessary. We spread the data to provide well posedness. The spreading length obtained by the sensitivity criterion is in agreement with the PME value. It is worth noting that this smoothing parameter corresponds to the decorrelation length of the tidal fields and, consequently, is easier to be evaluated a priori than the decorrelation length of the model errors. In our numerical example, the quality of the inverse solution did not appear to degrade with grid refinement, even if the smoothing routine was not applied. A reason may lie in that the data are of high quality and stay close to each other if compared with the model length scale. To check this hypothesis, we have unrealistically shallowed the channel and observed that smoothing was actually necessary.
Acknowledgments
This material is based upon the work supported by the U.S. Civilian Research & Development Foundation under Award RGI-245.
REFERENCES
Adams, R. A., 1975: Sobolev Spaces. Academic Press, 268 pp.
Bennett, A. F., 1992: Inverse Methods in Physical Oceanography. Cambridge University Press, 346 pp.
——, and P. C. McIntosh, 1982: Open ocean modeling as an inverse problem: Tidal theory. J. Phys. Oceanogr.,12, 1004–1018.
——, and R. N. Miller, 1991: Weighting initial conditions in variational data assimilation. Mon. Wea. Rev.,119, 1098–1102.
Boltzmann, L., 1964: Lectures on Gas Theory. Cambridge University Press, 490 pp. [First published as Vorlesungen über Gastheorie, Barth, 1896.].
Courtier, P., 1997: Dual formulation of four-dimensional variational assimilation. Quart. J. Roy. Meter. Soc.,123, 2449–2461.
Cramer, H., 1954: Mathematical Methods of Statistics. 6th ed. Princeton University Press, 575 pp.
Davies, E. B., and J. T. Lewis, 1970: An operational approach to quantum probability. Commun. Math. Phys.,17, 239–260.
Dee, D. P., 1995: On-line estimation of error covariance parameters for atmospheric data assimilation. Mon. Wea. Rev.,123, 1128–1145.
——, and A. M. da Silva, 1998: Data assimilation in the presence of forecast bias. Quart. J. Roy. Meteor. Soc.,124, 269–295.
Egbert, G. D., and A. F. Bennett, 1996: Data assimilation methods for ocean tides. Modern Approaches to Data Assimilation in Ocean Modeling, P. Malanotte-Rizzoli, Ed., Elsevier Press, 147–179.
——, ——, and M. G. G. Foreman, 1994: TOPEX/POSEIDON tides estimated using a global inverse model. J. Geophys. Res.,99, 24 821–24 852.
Evensen, G., 1997: Advanced data assimilation for strongly nonlinear dynamics. Mon. Wea. Rev.,125, 1342–1354.
Gao, F., 1994: Fitting smoothing splines to data from multiple sources with unknown relative weights. Commun. Stat. Theor. Methods,23, 1665–1698.
Ghil, M., and P. Malanotte-Rizzoli, 1991: Data assimilation in meteorology and oceanography. Advances in Geophysics, Vol. 33, Academic Press, 141–266.
Gibbs, J. W., 1902: Elementary Principles in Statistical Mechanics. Yale University Press, 207 pp.
Gong, J., G. Wahba, D. Johnson, and J. Tribbia, 1998: Adaptive tuning of numerical weather prediction models: Simultaneous estimation of weighting, smoothing, and physical parameters. Mon. Wea. Rev.,126, 210–231.
Griffith, A. K., and N. K. Nichols, 1996: Accounting for model error in data assimilation using the adjoint methods. Proc. Second Int. SIAM Workshop on Computational Differentiation, Santa Fe, NM, SIAM, 195–204.
Jaynes, E. T., 1988a: The relation of Bayesian and maximum entropy methods. Maximum-Entropy and Bayesian Methods in Science and Engineering, G. J. Erickson and C. R. Smith, Eds., Kluwer, Reidel, 25–29.
——, 1988b: How does the brain do plausible reasoning? Maximum-Entropy and Bayesian Methods in Science and Engineering, G. J. Erickson and C. R. Smith, Eds., Kluwer, Reidel, 1–24.
Kholevo, A. C., 1972: Statistical problems in quantum physics. Proc. Second Japan–USSR Symp. on Probability Theory, Kyoto, Japan, 22–40.
Kivman, G. A., 1996: Assimilating sea-level data into a global tidal model by means of the generalized method of residuals (in Russian). Okeanologia,36, 835–841.
——, 1997a: Assimilating data into open ocean models. Surv. Geophys.,18, 621–645.
——, 1997b: Weak constraint data assimilation for tides in the the Arctic Ocean. Progress in Oceanography, Vol. 40, Pergamon Press, 179–196.
——, 1997c: Weak constraint data assimilation for tides in the the Arctic Ocean. Proc. Workshop on Data Assimilation in Large-scale Models, Delft, Netherlands, Delft University of Technology, 53–55.
Kurapov, A. L., and G. A. Kivman, 1999: Data assimilation in a finite element model of M2 tides in the Barents Sea (in Russian). Okeanology,39, 306–313.
Provost, C., and R. Salmon, 1986: A variational method for inverting hydrographic data. J. Mar. Res.,44, 1–34.
Shannon, C. E., 1948: A mathematical theory of communication. Bell Syst. Tech. J.,27, 379–423, 623–655.
Tarantola, A., 1987: Inverse Problem Theory: Methods for Data Fitting and Model Parameter Estimation. Elsevier, 613 pp.
Thacker, W. C., and G. Holloway, 1990: Inverse problems and the principle of maximum entropy. Int. Symp. on the Assimilation of Observations in Meteorology and Oceanography, Clermont-Ferrand, France, 6–11.
Toulany, B., and C. Garrett, 1984: Geostrophic control of fluctuating barotropic flow through straits. J. Phys. Oceanogr.,14, 649–655.
Urban, B., 1996: Retrieval of atmospheric thermodynamical parameters using satellite measurements with a maximum entropy method. Inverse Probl.,12, 779–796.
van Leeuwen, P. J., and G. Evensen, 1996: Data assimilation and inverse methods in terms of a probabilistic formulation. Mon. Wea. Rev.,124, 2898–2913.
Wahba, G., 1985: A comparison of GCV and GML for choosing the smoothing parameter in the generalized spline smoothing problem. Ann. Stat.,13, 1378–1402.
Yosida, K., 1980: Functional Analysis. Springer-Verlag, 500 pp.
APPENDIX A
Properties of Md and Mm
APPENDIX B
Equivalence between the MDC and the PME in a Special Case
Theorem. If H(
Let us notice that if υl is an eigenfunction corresponding to λl = 1 for ad > 0, then Lυl = 0. This implies that neither 〈υl, υl〉 nor 〈Δ, υl〉 depend on ad. Hence, the second multiplier at the right-hand side of (B13) does not depend on ad. The least upper bound of the first multiplier is equal to 0.25, and it is reached when λj = 0.5 for some j.
It follows from (B3) that eigenvalues λm are monotonic functions of ad. According to the hypothesis of the theorem, the entropy attains its least upper bound. Thus, all nonzero and nonunity eigenvalues are equal to 0.5 when ad =
Two points are worth noting in this connection. First, if the entropy attains its upper bound, all nonzero and nonunity eigenvalues of Md are equal to each other, and thus the inverse solutions corresponding to different data weights all lie on a straight segment in the state space Ψ. Second, it is seen from the proof that except for uniqueness of the extreme point, the theorem remains valid for the case of several data weights referred to different multiple data terms (Kurapov and Kivman 1999).
Tatarsky Strait. Dots indicate measurement locations
Citation: Journal of Atmospheric and Oceanic Technology 18, 2; 10.1175/1520-0426(2001)018<0266:AEATTW>2.0.CO;2
The entropy (solid line), the data cost (dashed line), rmsas (dotted line), and rmsυ (dashed–dotted line) as functions of the data weight w obtained on different grids: (a) 20 nodes, (b) 38 nodes, (c) 75 nodes, (d) 149 nodes. The units for rmsas and rmsυ are in centimeters
Citation: Journal of Atmospheric and Oceanic Technology 18, 2; 10.1175/1520-0426(2001)018<0266:AEATTW>2.0.CO;2
The data cost for different spreading lengths. (a)–(d) as in Fig. 2. Dashed line, no spreading;solid line, optimal spreading: for 3 nodes in (b) and (c), for 7 nodes in (d); dashed–dotted line, spreading is overestimated when made: for 3 nodes in (a), 5 in (b) and (c), and 9 in (d)
Citation: Journal of Atmospheric and Oceanic Technology 18, 2; 10.1175/1520-0426(2001)018<0266:AEATTW>2.0.CO;2
Comparison of the PME and the MDC.