|

Efficient Probabilistic Analysis Methods for Complex
Numerical Models with Large Numbers of Random Variables, 20-9190
Printer Friendly Version
Principal Investigators
Sitakanta Mohanty
Michael P. Enright
Y-T (Justin) Wu (consultant)
Inclusive Dates: 04/01/00 - 05/01/02
Background - Physics-based probabilistic
reliability/risk analysis of engineered and natural systems is emerging as a less
expensive alternative to field and laboratory test-based methods. New
challenges exist, however, because highly complicated physics-based models are
computationally intensive and require a large number of variables to conduct
accurate system-level analyses. The Total-System Performance Assessment
code developed with assistance from the Center for Nuclear Waste Regulatory
Analyses (a division of the Southwest Research Institute) to evaluate a
high-level radioactive waste (HLW) disposal facility and the codes for
reliability-based motor vehicle design involving hundreds of input variables are
typical examples. These problems cannot be solved efficiently and accurately by
the existing methods that have been successfully used for problems with a small
number of variables.
Approach - The objective of the project is to develop
and demonstrate a fast and accurate method to tackle risk/reliability problems
involving computationally intensive and numerically complex models with large
numbers of parameters. The hybrid method combines the sampling approach that explores the parameter space and the
advanced reliability methods that focus the analysis at the tail of the
probability distribution. Probabilistic sensitivity measures that can
most effectively identify the most important random variables are being
developed. The method is being extended to combine the low-probability,
high-consequence scenarios with the most likely scenarios to obtain a single
performance measure. An efficient reliability analysis method and an efficient
reliability-based design optimization method are being developed. The
demonstration examples include nuclear waste management, oil and gas exploration
and transportation, and automotive applications.
Accomplishments - This research made significant
advances in developing a methodology for probabilistic analysis of complex
systems characterized by time-consuming analyses and a large number of random
variables. The following is a summary of the key accomplishments.
- Developed and formalized an efficient sampling-based screening procedure and
demonstrated through a multi-linear response function example that (i) the
ranking is more easily identified at the tails of the distribution; (ii) the
mean-sensitivity (response CDF with respect to mean) has a generally better
discriminating power in identifying influential variables than the
sigma-sensitivity (response CDF with respect to standard deviation),
particularly when the CDF is close to the median (i.e., CDF = 0.5) where
Ss is not
useful; and (iii) the deviation at the tail end is not due to the nonlinearity
of the performance function.
- Developed confidence interval or acceptance limit-based criteria that most
definitely groups and simplifies the variables that are not important with
specified confidence level.
- Developed and demonstrated five new sensitivity measures:
(i) two of these
(i.e., sample mean-mean and sample mean-sigma sensitivities) are particularly
suitable for conducting analyses consistent with the United States high-level
waste disposal regulatory criterion; (ii) another two are average sensitivity
measures developed to facilitate comparison of the CDF sensitivity method with
the results from other methods; and (iii) a mean-sigma-based sensitivity measure
that combines the power of mean and sigma sensitivity measures proposed in the
past.
- Combined the low-probability high-consequence scenarios with the most likely
scenarios to obtain a single performance measure. The approach is based on the
probabilistic sensitivity and failure cost ratio associated with each random
variable and provides for visualization of not only the most influential
variables, but also the scenarios that contribute most to this influence.
- Developed a methodology for efficient reliability analysis of problems with
large numbers of random variables in which an "approximate response
function" is constructed based on the important variables identified by the
sampling-based screening procedure.
- Developed and demonstrated an approximate method to include the remaining
(unimportant) variables identified by the sampling-based screening procedure
(the remaining variables are grouped into a single term and added to the
approximate response function for reliability computations).
- Investigated efficient reliability-based design optimization
method for problems with large numbers of random variables and identified
computational improvements for design problems in which the reliability
constraints are inactive.
- Developed an integrated oil and gas production system model that combined
reservoir depletion drive model, multiphase flow wellbore pressure losses,
production economic model, and the CDF sensitivity method to identify parameters
that define the economic viability of producing from a given reservoir.

(a)

(b)
Important variables identified by (a) and (b)
sensitivity measures for an example problem with 246 random variables. The
figure clearly differentiates between significant and insignificant variables
with a 95% accuracy or acceptance level. (Note: LHS - Latin Hypercube Sampling)
2002 Program
SwRI
Home
|