Abstract
To predict the response of a system with unknown parameters, a common route is to quantify the parameters using test data and propagate the results through a computational model of the system. Activities in this process may include model calibration and/or model validation. Data uncertainty has a significant effect on model calibration and model validation, and therefore affects the response prediction. Data uncertainty includes the uncertainty regarding the amount of data and numerical values of data. Although its effect can be qualitatively observed by trying different data sets and visually comparing the response predictions, a quantitative methodology assessing the contributions of these two types of data uncertainty to the uncertainty in the response prediction is necessary in order to solve test resource allocation problems. In this paper, a novel computational technique based on pseudo-random numbers is proposed to efficiently quantify the uncertainty in the data value of each type of test. Then the method of auxiliary variable based on the probability integral transform theorem is applied to build a deterministic function so that variance-based global sensitivity analysis can be conducted. The resultant global sensitivity indices quantify the contribution of data value uncertainty of each type of test to the uncertainty in the response prediction. Thus a methodology for robust test resource allocation is proposed, i.e., quantifying the number of each type of tests so that the response predictions using different data set are consistent.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Rajashekhar MR, Ellingwood BR (1993) A new look at the response surface approach. Struct Saf 12:205–220
Rasmussen CE, Williams CKI (2006) Gaussian Processes for Machine Learning. MIT Press
Ghanem R, Spanos PD (1990) Polynomial chaos in stochastic finite elements. J Appl Mech 57(1):197–202, no 89
Liepe J, Filippi S, Komorowski M, Stumpf MPH (2013) Maximizing the information content of experiments in systems biology. PLoS Comput Biol 9(1). e1002888
Sankararaman S, McLemore K, Mahadevan S, Bradford SC, Peterson LD (2013) Test resource allocation in hierarchical systems using Bayesian networks. AIAA J 51(3):537–550
Mullins J, Li C, Mahadevan S (2014) Optimal selection of calibration and validation test samples under uncertainty. In IMAC XXXII. Orlando, FL
O’Hagan A, Stevens JW (2003) Assessing and comparing costs: how robust are the bootstrap and methods based on asymptotic normality? Health Econ 12(1):33–49
Saltelli A, Ratto M, Andres T, Campolongo F, Cariboni J, Gatelli D, Saisana M, Tarantola S (2008) Global sensitivity analysis: the primer. John Wiley & Sons
Sobol IM (2001) Global sensitivity indices for nonlinear mathematical models and their Monte Carlo estimates. Math Comput Simul 55(1–3):271–280
Matsumoto M, Nishimura T (1998) Mersenne twister: a 623-dimensionally equidistributed uniform pseudo-random number generator. ACM Trans Model Comput Simul 8(1):3–30
L’ecuyer P (1999) Good parameters and implementations for combined multiple recursive random number generators. Oper Res 47(1):159–164
Matteis A, Pagnutti S (1993) Long-range correlation analysis of the Wichmann-Hill random number generator. Stat Comput 3(2):67–70
Sankararaman S, Mahadevan S (2013) Separating the contributions of variability and parameter uncertainty in probability distributions. Reliab Eng Syst Saf 112:187–199
Schrijver A (1998) Theory of linear and integer programming. John Wiley & Sons
Red-Horse JR, Paez TL (2008) Sandia National Laboratories validation workshop: structural dynamics application. Comput Meth Appl Mech Eng 197(29–32):2578–2584
Li C, Mahadevan S (2014) Uncertainty quantification and output prediction in multi-level problems. In: 16th AIAA Non-Deterministic Approaches Conference, American Institute of Aeronautics and Astronautics
Acknowledgements
The research in this paper is partially supported by funds from Sandia National Laboratories through contract no. BG-7732 (Technical Monitor: Dr. Angel Urbina). This support is gratefully acknowledged. The authors also appreciate valuable discussions with Shankar Sankararaman (NASA Ames) and Joshua Mullins (Sandia National Laboratories).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 The Society for Experimental Mechanics, Inc.
About this paper
Cite this paper
Li, C., Mahadevan, S. (2015). Sensitivity Analysis for Test Resource Allocation. In: Atamturktur, H., Moaveni, B., Papadimitriou, C., Schoenherr, T. (eds) Model Validation and Uncertainty Quantification, Volume 3. Conference Proceedings of the Society for Experimental Mechanics Series. Springer, Cham. https://doi.org/10.1007/978-3-319-15224-0_14
Download citation
DOI: https://doi.org/10.1007/978-3-319-15224-0_14
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-15223-3
Online ISBN: 978-3-319-15224-0
eBook Packages: EngineeringEngineering (R0)