Abstract
The sequential Gaussian algorithm is widespread to simulate Gaussian random fields. In practice, the determination of the successive conditional distributions only accounts for the information available in a moving neighborhood centered on the target location, which provokes a loss of accuracy with respect to a unique neighborhood implementation. In order to reduce this loss of accuracy, iterative methods for solving large kriging systems of equations are used to improve the determination of the conditional distributions, taking the results obtained in a moving neighborhood as a first approximation. Numerical experiments are presented to show the proposed strategies and the improvements in the reproduction of the correlation structure of the simulated random field.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Axelsson O (1985) A survey of preconditioned iterative methods for linear systems of algebraic equations. BIT Numer Math 25:166–187
Barret R (1994) Templates for the solution of linear systems: building blocks for iterative methods. SIAM publications, vol 117. SIAM, Philadelphia
Burden R, Faires D (2010) Numerical analysis, 9th edn. Brooks/Cole, Boston, 872 pp
Chilès JP, Delfiner P (1999) Geostatistics: modeling spatial uncertainty. Wiley, New York, 695 pp
Datta BN (2010) Numerical linear algebra and applications, 2nd edn. Society for Industrial and Applied Mathematics, Philadelphia, 530 pp
Deutsch CV, Journel AG (1998) GSLIB: geostatistical software library and user’s guide, 2nd edn. Oxford University Press, New York, 369 pp
Emery X (2004) Testing the correctness of the sequential algorithm for simulating Gaussian random fields. Stoch Environ Res Risk Assess 18(6):401–413
Emery X, Peláez M (2011) Assessing the accuracy of sequential Gaussian simulation and cosimulation. Comput Geosci 15(4):673–689
Golub GH, Van Loan CF (1996) Matrix computations, 3rd edn. Johns Hopkins University Press, Baltimore, 694 pp
Gómez-Hernández JJ, Cassiraga EF (1994) Theory and practice of sequential simulation. In: Armstrong M, Dowd PA (eds) Geostatistical simulations. Kluwer Academic, Dordrecht, pp 111–124
Hestenes MR, Stiefel E (1952) Methods of conjugate gradients for solving linear systems. J Res Natl Bur Stand 49(6):409–436
Saad Y (1981) Krylov subspace methods for solving large unsymmetric linear systems. Math Comput 37:105–126
Saad Y (2003) Iterative methods for sparse linear systems, 2nd edn. SIAM, Philadelphia, 567 pp
Stoer J, Burlisch R (2002) Introduction to numerical analysis, 3rd edn. Springer, New York, 732 pp
Tran TT (1994) Improving variogram reproduction on dense simulation grids. Comput Geosci 20(7–8):1161–1168
Van der Vorst HA (1992) Bi-CGSTAB: a fast a smoothly converging variant of Bi-CG for the solution of nonsymmetric linear systems. SIAM J Sci Comput 13:631–644
Acknowledgements
This research was funded by the Chilean program MECESUP UCN0711 and the FONDECYT project 11100029.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix: Analysis of Computational Cost
Appendix: Analysis of Computational Cost
In this appendix we assess the computational cost of sequential simulation using a moving neighborhood of fixed size k, and the additional cost of solving the kriging system with an iterative method in each step of sequential simulation.
1.1 A.1 Solving Kriging Systems with a Moving Neighborhood
If we sequentially simulate a Gaussian random vector with n components using a moving neighborhood containing k variables, one has to solve one kriging system of size 1×1, one system of size 2×2,… , one system of size (k−1)×(k−1) and (n−k) systems of size k×k. To solve a linear system of equations of size p×p by means of the Gaussian elimination (direct method), the number of floating point operations is 2p 3/3 [3]. Accordingly, the computational cost in floating point operations to solve all the kriging systems in the sequential simulation with a neighborhood of k variables is O(nk 3):
If we consider the ideal case of a unique neighborhood implementation (k=n−1), the number of floating point operations is O(n 4):
1.2 A.2 Solving Kriging Systems with Iterative Methods
The computational cost to calculate the kriging weights and that are used to construct initial approximations (to apply the iterative methods), using a moving neighborhood of size k and Gaussian elimination, is given by (9). To this cost, one has to add the cost of the iterative method from component (k+2) to component n of the simulated vector.
1.2.1 A.2.1 Gauss-Seidel Method
To solve a linear system of equations of size p×p with the Gauss-Seidel iterative method, the number of floating point operations is 2sp 2, where s is the number of iterations performed [3]. Therefore, the number of floating point operations for the (n−k−1) systems where we apply the Gauss-Seidel method is:
With respect to the classical implementation (9), the additional cost is O(sn 3).
1.2.2 A.2.2 Conjugate Gradient Method
The cost to solve a linear system of equations of size p×p with Conjugate Gradient method is s(6p+3+2u p ), where u p is the number of nonzero elements of the left-hand side matrix of the system of equations [9]. Accordingly, the number of floating point operations for solving the last (n−k−1) kriging systems with the Conjugate Gradient method is:
The computational order of this additional is O(sn 2) when the matrix of the system has many zero entries.
1.3 A.3 CPU Time on a Numerical Example
To illustrate the gain in CPU time when using an iterative method (here, the Conjugate Gradient) rather than the direct method (Gaussian elimination) to solve the successive kriging systems, we have run an experiment on a regular 2D grid with 100×100 nodes, which are visited through a random sequence using multiple grids [15]. Table 5 shows the standardized Frobenius norm of \(C_{\tilde{Y}}-C_{Y}\) and the CPU time for direct sequential Gaussian simulation (SGS) and for the proposed algorithm (SGS + CG), using the same set of conditioning variables to construct the successive conditional distributions (specifically, the variables located within a circle of radius 30 centered on the target node). The initial approximation for the Conjugate Gradient method is a vector of zeros. It is seen that the accuracy is comparable for both approaches (SGS and SGS+CG), insofar as the standardized Frobenius norms are practically the same, but there are considerable differences in CPU times. These differences increase for larger simulation grids and for larger sizes of the neighborhood used to select the conditioning variables.
Rights and permissions
Copyright information
© 2012 Springer Science+Business Media Dordrecht
About this chapter
Cite this chapter
Arroyo, D., Emery, X., Peláez, M. (2012). Sequential Simulation with Iterative Methods. In: Abrahamsen, P., Hauge, R., Kolbjørnsen, O. (eds) Geostatistics Oslo 2012. Quantitative Geology and Geostatistics, vol 17. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-4153-9_1
Download citation
DOI: https://doi.org/10.1007/978-94-007-4153-9_1
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-007-4152-2
Online ISBN: 978-94-007-4153-9
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)