Skip to main content

Skewness of the Prior Through Position Errors and Its Impact on Data Assimilation

  • Chapter
  • First Online:

Abstract

Uncertainty in the position of a feature is a ubiquitous influence on data assimilation (DA) in geophysical applications. This chapter explores the properties of distributions arising from the uncertainty of the location of a flow feature. It is shown that distributions arising from phase uncertainty have surprisingly complex, non-Gaussian characteristics. These non-Gaussian characteristics are explored from an ensemble DA perspective in which the skewness (third-moment) is shown to be a significant contributor to the state-estimates obtained through Bayesian state estimation. Idealized examples, as well as an example in a real tropical cyclone using a state-of-the-art numerical weather prediction model, will be shown.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    A chi-square distribution with one-degree of freedom is constructed by squaring each random draw from a Gaussian distribution.

References

  • Anderson JL, Anderson SL (1999) A monte carlo implementation of the nonlinear filtering problem to produce ensemble assimilations and forecasts. Mon Weather Rev 127:2741–2758

    Article  Google Scholar 

  • Anderson J, Hoar T, Raeder K, Liu H, Collins N, Torn R, Avellano A (2009) The data assimilation research testbed: a community facility. Bull Amer Meteor Soc 90:1283–1296

    Article  Google Scholar 

  • Bishop CH, Etherton BJ, Majumdar SJ (2001) Adaptive sampling with the ensemble transform Kalman filter. Part I: Theoretical aspects. Mon Weather Rev 129:420–436

    Article  Google Scholar 

  • Burgers G, Van Leeuwen PJ, Evensen G (1998) Analysis scheme in the ensemble Kalman filter. Mon Weather Rev 126:1719–1724

    Article  Google Scholar 

  • Chen Y, Snyder C (2007) Assimilating vortex position with an ensemble Kalman filter. Mon Weather Rev 135:1828–1845

    Article  Google Scholar 

  • Dowell DC, Wicker LJ, Snyder C (2011) Ensemble Kalman filter assimilation of radar observations of the 8 May 2003 Oklahoma City supercell: influences of reflectivity observations on storm-scale analyses. Mon Weather Rev 139:272–294

    Article  Google Scholar 

  • Doyle JD, Jin Y, Hodur R, Chen S, Jin H, Moskaitis J, Reinecke A, Black P, Cummings J, Hendricks E, Holt T, Liou C, Peng M, Reynolds C, Sashegyi K, Schmidt J, Wang S (2012) Real time tropical cyclone prediction using COAMPS-TC. In: Chun-Chieh Wu, Jianping Gan (eds) Advances in geosciences, vol 28. World Scientific, Singapore, pp 15–28

    Google Scholar 

  • Evensen G (1994) Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. J Geophys Res 99:143–162

    Article  Google Scholar 

  • Fletcher SJ, Zupanski M (2006) A data assimilation method for log-normally distributed observational errors. Q J R Meteorol Soc 132:2505–2519

    Article  Google Scholar 

  • Hacker JP, Snyder C, Ha S-Y, Pocernich M (2011) Linear and non-linear response to parameter variations in a mesoscale model. Tellus A 63:429–444

    Article  Google Scholar 

  • Hamill TM, Whitaker JS, Fiorino M, Benjamin SG (2011) Global ensemble predictions of 2009’s tropical cyclones initialized with an ensemble Kalman filter. Mon Weather Rev 139:668–688

    Article  Google Scholar 

  • Hodyss D (2011) Ensemble state estimation for nonlinear systems using polynomial expansions in the innovation. Mon Weather Rev 139:3571–3588

    Article  Google Scholar 

  • Hodyss D (2012) Accounting for skewness in ensemble data assimilation. Mon Weather Rev doi:10.1175/MWR-D-11-00198.1 140, 2346–2358

    Google Scholar 

  • Houtekamer PL, Mitchell HL (1998) Data assimilation using an ensemble Kalman filter technique. Mon Weather Rev 126:796–811

    Article  Google Scholar 

  • Houtekamer PL, Mitchell HL (2001) A sequential ensemble Kalman filter for atmospheric data assimilation. Mon Weather Rev 129:123–137

    Article  Google Scholar 

  • Houtekamer PL, Mitchell HL, Pellerin G, Buehner M, Charron M, Spacek L, Hansen B (2005) Atmospheric data assimilation with an ensemble Kalman filter: results with real observations. Mon Weather Rev 133:604–620

    Article  Google Scholar 

  • Jazwinski AH (1998) Stochastic processes and filtering theory. Academic, New York, 376pp

    Google Scholar 

  • Julier SJ, Uhlmann JK (1997) New extension of the Kalman filter to nonlinear systems. Proc of SPIE 3068:182–193

    Article  Google Scholar 

  • Julier SJ, Uhlmann JK (2004) Unscented filtering and nonlinear estimation. Proc of the IEEE 92:401–422

    Article  Google Scholar 

  • Kalman RE (1960) A new approach to linear filtering and prediction problems. J Basic Eng 82:35–45

    Article  Google Scholar 

  • Kalman RE, Bucy RS (1961) New results in linear filtering and prediction theory. Trans of the ASME Series D J of Basic Eng 83:95–107

    Article  Google Scholar 

  • Khare SP, Anderson JL, Hoar TJ, Nychka D (2008) An investigation into the application of an ensemble Kalman smoother to high-dimensional geophysical systems. Tellus A 60:97–112

    Article  Google Scholar 

  • Kushner HJ (1967) Approximations to optimal nonlinear filters. IEEE Trans Auto Control AC-12:546–556

    Article  Google Scholar 

  • Lawson GW, Hansen JA (2005) Alignment error models and ensemble-based data assimilation. Mon Weather Rev 133:1687–1709

    Article  Google Scholar 

  • Meng Z, Zhang F (2008) Tests of an ensemble Kalman filter for mesoscale and regional-scale data assimilation. Part IV: Comparison with 3DVAR in a month-long experiment. Mon Weather Rev 136:3671–3682

    Article  Google Scholar 

  • Snyder C, Bengtsson T, Anderson J (2008) Obstacles to high-dimensional particle filtering. Mon Weather Rev 136:4629–4640

    Article  Google Scholar 

  • Szunyogh I, Kostelich EJ, Gyarmati G, Kalnay E, Hunt BR, Ott E, Satterfield E, Yorke JA (2008) A local ensemble transform Kalman filter data assimilation system for the NCEP global model. Tellus A 60:113–130

    Google Scholar 

  • Torn RD, Hakim GJ (2008) Performance characteristics of a pseudo-operational ensemble Kalman filter. Mon Weather Rev 136:3947–3963

    Article  Google Scholar 

  • Tippet MK, Anderson JL, Bishop CH, Hamill TM, Whitaker JS (2003) Ensemble square root filters. Mon Weather Rev 131:1485–1490

    Article  Google Scholar 

  • van Leeuwen PJ (2009) Particle filtering in geophysical systems. Mon Weather Rev 137:4089–4114

    Article  Google Scholar 

  • Whitaker JS, Hamill TM, Wei X, Song Y, Toth Z (2008) Ensemble data assimilation with the NCEP global forecast system. Mon Weather Rev 136:463–482

    Article  Google Scholar 

  • Zupanski M (2005) Maximum likelihood ensemble filter: theoretical aspects. Mon Weather Rev 133:1710–1726

    Article  Google Scholar 

Download references

Acknowledgements

We gratefully acknowledge support from the Office of Naval Research through PE-0601153N.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Daniel Hodyss .

Editor information

Editors and Affiliations

Appendices

Appendix 1: Matrix and Notation Definitions

The prior error covariance matrix is extended to include the higher moments:

$$\mathbf{\hat{P}}_{f} = \frac{{\mathbf{\hat{Z}\hat{Z}}}^{T}} {K - 1} = \left [\begin{array}{cc} \mathbf{P}_{f} & \mathbf{T}_{f} \\ \mathbf{T}_{f}^{T}&\mathbf{F}_{f} -\mathbf{p}_{f}\mathbf{p}_{f}^{T}\\ \end{array} \right ],$$
(7.44)

whose square root form may be approximated with an ensemble as

$$\mathbf{\hat{Z}} = \left [\begin{array}{c@{\quad }c@{\quad }c@{\quad }c} \boldsymbol{\epsilon }_{1} \quad & \boldsymbol{\epsilon }_{2} \quad &\cdots \quad & \boldsymbol{\epsilon }_{K} \\ \boldsymbol{\epsilon }_{1} \otimes \boldsymbol{\epsilon }_{1} -\mathbf{p}_{f}\quad &\boldsymbol{\epsilon }_{2} \otimes \boldsymbol{\epsilon }_{2} -\mathbf{p}_{f}\quad &\cdots \quad &\boldsymbol{\epsilon }_{K} \otimes \boldsymbol{\epsilon }_{K} -\mathbf{p}_{f}\\ \quad \end{array} \right ]$$
(7.45)

and the vectorized covariance matrix, p f  = vec(P f ), is an N 2-vector constructed from the concatenation of the N columns of P f and whose organization follows that of the Kronecker product “ ⊗ ”, and K is the ensemble size.

The extended observation operator takes the following form:

$$\mathbf{\hat{H}} = \left [\begin{array}{c} \mathbf{H}\\ \mathbf{H} _{2}\\ \end{array} \right ] = \left [\begin{array}{c} \mathbf{H}\\ \mathbf{H} \otimes \mathbf{H}\\ \end{array} \right ].$$
(7.46)

Note that for nonlinear observation operators one would not use a linearized form of the operator. Instead, the correct procedure is to operate the nonlinear observation operator on each member of the ensemble and then perform linear or nonlinear regression on this new distribution of predicted prior observations against the state variables needing update (Houtekamer and Mitchell 2001).

The covariance matrices in the extended state-space takes the form:

$$\displaystyle\begin{array}{rcl} & & \left \langle \boldsymbol{\epsilon }_{f}\mathbf{{\hat{v}}^{{\prime}T}}\right \rangle = \left [\begin{array}{ccc} \mathbf{P}_{ f}{\mathbf{H}}^{T}&\mathbf{T}_{ f}\mathbf{H}_{2}^{T}&\cdots \\ \end{array} \right ],\end{array}$$
(7.47)
$$\displaystyle\begin{array}{rcl} & & \left \langle \mathbf{{\hat{v}}^{{\prime}}{\hat{v}}^{{\prime}T}}\right \rangle = \left \langle {\mathbf{\hat{v}\hat{v}}}^{T}\right \rangle -\left \langle \mathbf{\hat{v}}\right \rangle \left \langle {\mathbf{\hat{v}}}^{T}\right \rangle ,\end{array}$$
(7.48)
$$\displaystyle\begin{array}{rcl} & & \left \langle {\mathbf{\hat{v}\hat{v}}}^{T}\right \rangle = \left [\begin{array}{ccc} \mathbf{HP}_{f}{\mathbf{H}}^{T} + \mathbf{R}& \mathbf{HT}_{ f}\mathbf{H}_{2}^{T} &\cdots \\ \mathbf{H}_{2}\mathbf{T}_{f}^{T}{\mathbf{H}}^{T} &\mathbf{H}_{2}\mathbf{F}_{f}\mathbf{H}_{2}^{T} + \mathbf{A} + \mathbf{B} + \mathbf{C} + \mathbf{R}_{4} & \cdots \\ \vdots & \vdots & \ddots\\ \end{array} \right ]\end{array}$$
(7.49)
$$\displaystyle\begin{array}{rcl} & & \left \langle \mathbf{\hat{v}}\right \rangle \left \langle {\mathbf{\hat{v}}}^{T}\right \rangle = \left [\begin{array}{ccc} \mathbf{0}& \mathbf{0} &\cdots \\ \mathbf{0}&\left \langle {\mathbf{v}}^{2}\right \rangle \left \langle {\mathbf{v}}^{2T}\right \rangle &\cdots \\ \vdots & \vdots & \ddots\\ \end{array} \right ],\end{array}$$
(7.50)

and 0 is the p ×p zero matrix. The matrices in (7.17) are defined as:

$$\displaystyle\begin{array}{rcl} \mathbf{P}_{f} = \left \langle \boldsymbol{\epsilon }_{f}\boldsymbol{\epsilon }_{f}^{T}\right \rangle ,& &\end{array}$$
(7.51a)
$$\displaystyle\begin{array}{rcl} \mathbf{T}_{f} = \left \langle \boldsymbol{\epsilon }_{f}\boldsymbol{\epsilon }_{f}^{2T}\right \rangle ,& &\end{array}$$
(7.51b)
$$\displaystyle\begin{array}{rcl} \mathbf{F}_{f} = \left \langle \boldsymbol{\epsilon }_{f}^{2}\boldsymbol{\epsilon }_{ f}^{2T}\right \rangle ,& &\end{array}$$
(7.51c)
$$\displaystyle\begin{array}{rcl} \mathbf{R} = \left \langle \boldsymbol{\epsilon }_{o}\boldsymbol{\epsilon }_{o}^{T}\right \rangle ,& &\end{array}$$
(7.52a)
$$\displaystyle\begin{array}{rcl} \mathbf{R}_{4}& = \left \langle \boldsymbol{\epsilon }_{o}^{2}\boldsymbol{\epsilon }_{o}^{2T}\right \rangle ,&\end{array}$$
(7.52b)
$$\displaystyle\begin{array}{rcl} & & \mathbf{A} = \left \langle \boldsymbol{\epsilon }_{o}^{2}\boldsymbol{\epsilon }_{ f}^{2T}\right \rangle \mathbf{H}_{ 2}^{T} + \mathbf{H}_{ 2}\left \langle \boldsymbol{\epsilon }_{f}^{2}\boldsymbol{\epsilon }_{ o}^{2T}\right \rangle ,\end{array}$$
(7.53)
$$\displaystyle\begin{array}{rcl} & & \mathbf{B} = \mathbf{R} \otimes \mathbf{HP}_{f}{\mathbf{H}}^{T} + \mathbf{HP}_{ f}{\mathbf{H}}^{T} \otimes \mathbf{R},\end{array}$$
(7.54)
$$\displaystyle\begin{array}{rcl} & & \mathbf{C} = \left \langle \left (\boldsymbol{\epsilon }_{o}\boldsymbol{\epsilon }_{f}^{T}{\mathbf{H}}^{T}\right ) \otimes \left (\mathbf{H}\boldsymbol{\epsilon }_{ f}\boldsymbol{\epsilon }_{o}^{T}\right )\right \rangle + \left \langle \left (\mathbf{H}\boldsymbol{\epsilon }_{ f}\boldsymbol{\epsilon }_{o}^{T}\right ) \otimes \left (\boldsymbol{\epsilon }_{ o}\boldsymbol{\epsilon }_{f}^{T}{\mathbf{H}}^{T}\right )\right \rangle ,\end{array}$$
(7.55)

and H 2 = H ⊗ H is the matrix operator that takes an N 2 vector into the p 2 predictor space and copious use of the identity, \((\mathbf{H}\boldsymbol{\epsilon }_{f}) \otimes (\mathbf{H}\boldsymbol{\epsilon }_{f}) = (\mathbf{H} \otimes \mathbf{H})(\boldsymbol{\epsilon }_{f} \otimes \boldsymbol{\epsilon }_{f})\), has been made.

In (7.51a), P f is a square matrix listing the necessary second moments of the prior distribution. In (7.51b), T f is a rectangular matrix listing the necessary third moments of the prior distribution. In (7.51c), F f is a square matrix listing the necessary fourth moments of the prior distribution. In (7.52a), R 4 is a square matrix listing the necessary fourth moments of the observation likelihood. Note that even when the observation error covariance matrix, R, is diagonal R 4 is not diagonal. The matrices A, B, andC are sparse, square matrices that represent various combinations of observation error covariances and forecast error covariances.

The covariance matrix of squared innovations is

$$\displaystyle\begin{array}{rcl} \boldsymbol{\Pi } = \mathbf{H}_{2}\mathbf{F}_{f}\mathbf{H}_{2}^{T} + \mathbf{A} + \mathbf{B} + \mathbf{C} + \mathbf{R}_{ 4} -\mathbf{H}_{2}\mathbf{T}_{f}^{T}{\mathbf{H}}^{T}{\left (\mathbf{HP}_{ f}{\mathbf{H}}^{T} + \mathbf{R}\right )}^{-1}\mathbf{HT}_{ f}\mathbf{H}_{2}^{T} -\left \langle {\mathbf{v}}^{2}\right \rangle \left \langle {\mathbf{v}}^{2T}\right \rangle.& &\end{array}$$
(7.56)

Appendix 2: Non-Gaussian Phase Uncertainty from Variable Shear Flows

To isolate the effects of phase uncertainty we focus on the one-dimensional advection equation:

$$\frac{\partial p} {\partial t} + c\left (x\right )\frac{\partial p} {\partial x} = 0,$$
(7.57)

where p = p(x, t) is the wavefield being advected at the speed c(x). We attach to (7.57) an unbounded domain in x as well as a localized initial condition such as the functions (7.42), which was our prototype for the surface pressure field of members of a prior distribution of tropical cyclones. By the method of characteristics we know that the solutions to (7.57) move away from their initial location according to,

$$\frac{\mathit{dX}} {\mathit{dt}} = c(X),$$
(7.58)

where X = X(t) is the location of, say, the minimum central pressure of the function (7.42). We may immediately note that even though equation (7.57) is linear in the amplitude of the disturbance, equation (7.58) may be non-linear if the function c(x) is non-linear. Hence, the motion of the location of the minimum central pressure may evolve non-linearly and therefore even if the initial distribution of phase uncertainty is Gaussian it still may evolve into a non-Gaussian phase distribution owing to the non-linearity in (7.58). Two examples follow:

  1. (1)

    Linear Shear

In the case where the shear is a linear increasing function of distance from the origin, i.e.

$$c(x) = c_{0} \frac{x} {L},$$
(7.59)

where c 0 is a characteristic phase speed and L is characteristic length scale. By inserting (7.59) into (7.58) and solving finds

$$X(t) = X_{0}\exp \left (\frac{c_{0}t} {L} \right ),$$
(7.60)

where X 0 is the initial location of the minimum central pressure. In Sect. 7.3 the parameter X 0 was denoted as \(\varphi\) and was normally distributed. Notice that if the location of the minimum central pressure is normally distributed with mean \(\bar{x}\) and variance σ 2, then at a later time, t, the phase distribution will be normal with

$$X \sim N\left (\bar{x}\exp \left (\frac{c_{0}t} {L} \right ){,\sigma }^{2}\exp \left (\frac{2c_{0}t} {L} \right )\right ).$$
(7.61)

Hence, linear shear produces disturbances that move away from their initial location exponentially with time. Nevertheless, disturbances in linear shear preserve the Gaussian character of their initial phase uncertainty.

  1. (2)

    Quadratic Shear

In the case where the shear is a quadratic function of distance from the origin, i.e.

$$c(x) = c_{0}{\left ( \frac{x} {L}\right )}^{2},$$
(7.62)

the solution to (7.58) is

$$\displaystyle\begin{array}{rcl} X(t) = \frac{L} { \frac{L} {X_{0}} -\frac{c_{0}t} {L} }.& &\end{array}$$
(7.63)

In this case, the phase locations of the disturbances move away from their initial locations faster than exponential with time. This can be seen by the fact that disturbances in linear shear flow approach infinity, i.e. X → , as t → , but disturbances in quadratic shear flow reach infinity in finite-time \([t_{\infty } = {L}^{2}/(c_{0}X_{0})]\). Note that in the limit of small time, i.e., \(t \ll {L}^{2}/(c_{0}X_{0})\) then equation (7.63) behaves approximately as

$$X(t) \approx X_{0} + \frac{c_{0}t} {{L}^{2}} X_{0}^{2},$$
(7.64)

which is obviously non-Gaussian even when X 0 is normally distributed and becomes increasingly non-Gaussian as time goes on.

Hence, the structure of the shear flow the disturbances are being advected within will determine whether the resulting phase distribution will be Gaussian or non-Gaussian at some time later. Moreover, even though (7.57) is an equation linear in the amplitude of the disturbance, its characteristics maybe be non-linear, which could lead to non-Gaussian distributions. This implies that it is not sufficient to simply note that the physical system (7.57) is linear in amplitude in order to assess whether a Kalman filter will be optimal or not. The correct condition is that both the model and its characteristics must be linear and the initial distribution one draws from must also be Gaussian. Given the severity of these conditions it would appear that it is unlikely that the phase distributions of actual flow features found in nature would always maintain a normally distributed character. Rather, it would seem more likely that the atmosphere would evolve through time and flow configurations in which a Gaussian initial phase distribution would be altered to have non-Gaussian characteristics.

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this chapter

Cite this chapter

Hodyss, D., Reinecke, A. (2013). Skewness of the Prior Through Position Errors and Its Impact on Data Assimilation. In: Park, S., Xu, L. (eds) Data Assimilation for Atmospheric, Oceanic and Hydrologic Applications (Vol. II). Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35088-7_7

Download citation

Publish with us

Policies and ethics