Uncertainty and Sensitivity Analysis Methods for Improving Design Robustness and Reliability

advertisement
Uncertainty and Sensitivity Analysis Methods for
Improving Design Robustness and Reliability
by
Qinxian He
Bachelor of Science in Engineering, Duke University, 2008
Master of Science, Massachusetts Institute of Technology, 2010
Submitted to the Department of Aeronautics and Astronautics
in partial fulfillment of the requirements for the degree of
Doctor of Philosophy
at the
MASSACHUSETTS INSTITUTE OF TECHNOLOGY
June 2014
c Massachusetts Institute of Technology 2014. All rights reserved.
Author . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Department of Aeronautics and Astronautics
May 7, 2014
Certified by . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Karen E. Willcox
Professor of Aeronautics and Astronautics
Thesis Supervisor
Certified by . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Ian A. Waitz
Jerome C. Hunsaker Professor of Aeronautics and Astronautics
Committee Member
Certified by . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Douglas L. Allaire
Assistant Professor of Mechanical Engineering, Texas A&M University
Committee Member
Accepted by . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Paulo C. Lozano
Associate Professor of Aeronautics and Astronautics
Chair, Graduate Program Committee
2
Uncertainty and Sensitivity Analysis Methods for
Improving Design Robustness and Reliability
by
Qinxian He
Submitted to the Department of Aeronautics and Astronautics
on May 7, 2014, in partial fulfillment of the
requirements for the degree of
Doctor of Philosophy
Abstract
Engineering systems of the modern day are increasingly complex, often involving numerous components, countless mathematical models, and large, globally-distributed
design teams. These features all contribute uncertainty to the system design process
that, if not properly managed, can escalate into risks that seriously jeopardize the
design program. In fact, recent history is replete with examples of major design setbacks due to failure to recognize and reduce risks associated with performance, cost,
and schedule as they emerge during the design process.
The objective of this thesis is to develop methods that help quantify, understand,
and mitigate the effects of uncertainty in the design of engineering systems. The
design process is viewed as a stochastic estimation problem in which the level of
uncertainty in the design parameters and quantities of interest is characterized probabilistically, and updated through successive iterations as new information becomes
available. Proposed quantitative measures of complexity and risk can be used in the
design context to rigorously estimate uncertainty, and have direct implications for
system robustness and reliability. New local sensitivity analysis techniques facilitate
the approximation of complexity and risk in the quantities of interest resulting from
modifications in the mean or variance of the design parameters. A novel complexitybased sensitivity analysis method enables the apportionment of output uncertainty
into contributions not only due to the variance of input factors and their interactions,
but also due to properties of the underlying probability distributions such as intrinsic
extent and non-Gaussianity. Furthermore, uncertainty and sensitivity information
are combined to identify specific strategies for uncertainty mitigation and visualize
tradeoffs between available options. These approaches are integrated with design
budgets to guide decisions regarding the allocation of resources toward improving
system robustness and reliability.
3
The methods developed in this work are applicable to a wide variety of engineering
systems. In this thesis, they are demonstrated on a real-world aviation case study to
assess the net cost-benefit of a set of aircraft noise stringency options. This study
reveals that uncertainties in the scientific inputs of the noise monetization model are
overshadowed by those in the scenario inputs, and identifies policy implementation
cost as the largest driver of uncertainty in the system.
Thesis Supervisor: Karen E. Willcox
Title: Professor of Aeronautics and Astronautics
Committee Member: Ian A. Waitz
Title: Jerome C. Hunsaker Professor of Aeronautics and Astronautics
Committee Member: Douglas L. Allaire
Title: Assistant Professor of Mechanical Engineering, Texas A&M University
4
Acknowledgments
The journey toward a PhD is one that cannot be undertaken alone. It is only with help
from countless people along the way that I have made it thus far, and I am eternally
grateful for their support. First and foremost, I would like to thank my advisor,
Professor Karen Willcox, for her guidance, mentorship, and encouragement over the
past four years. It has truly been a privilege to work with Karen, and I am greatly
indebted to her for everything she has taught me about research, life, and making
it all balance together. I also extend thanks to my committee members, Dean Ian
Waitz and Professor Doug Allaire, for lending their knowledge and support to every
aspect of this research, and for providing invaluable feedback and guidance along the
way. I would also like to acknowledge my thesis readers, Professor John Deyst and
Professor Olivier de Weck for contributing insightful comments and suggestions that
have helped to improve this thesis.
To my labmates in the ACDL, past and present, thank you so much for sharing
this journey with me. It has been an honor to work with and learn from all of you.
I would like to especially recognize Leo Ng, Andrew March, Sergio Amaral, Chad
Lieberman, Giulia Pantalone, Harriet Li, Laura Mainini, Rémi Lam, Marc Lecerf,
Tiangang Cui, Patrick Blonigan, Alessio Spantini, Eric Dow, Hemant Chaurasia, and
Xun Huan for making my stay in Building 37 such a memorable experience. I am also
grateful to the larger AeroAstro community for providing a welcoming and dynamic
environment in which to learn and grow. I have made so many friends in AeroAstro
and will always fondly remember the good times we’ve shared. A special thank you
to the ladies of the Women’s Graduate Association of Aeronautics and Astronautics
(WGA3 ) — Sunny Wicks, Sameera Ponda, Farah Alibay, Whitney Lohmeyer, Abhi
Butchibabu, Sathya Silva, and Carla Perez Martinez — it has been such a joy getting
to know you and I can’t wait to see the amazing things you will continue to accomplish
in the future. I also owe a great deal of thanks to Philip Wolfe. From a technical
standpoint, thank you for helping me set up and run the CAEP/9 noise stringency
case study on such short notice; from a personal perspective, thank you for your
5
friendship over the past 10 years — I have really appreciated our project meetings,
commiseration lunches, and random adventures both at Duke and at MIT. Finally, I
would like to acknowledge the AeroAstro staff members who work tirelessly to make
life easier and more enjoyable for graduate students. Jean Sofronas, Joyce Light,
Meghan Pepin, Marie Stuppard, Beth Marois, Bill Litant, Robin Palazzolo, and Sue
Whitehead — thank you for being so generous with your time, help, smiles, and words
of encouragement.
Outside of research, I have been extremely fortunate to have the support of a
diverse group of friends who help me keep things in perspective. Matt and Amanda
Edwards, Jim Abshire, David Rosen, Emily Kim, Audrey Fan, Lori Ling, Yvonne
Yamanaka, Tiffany Chen, Tatyana Shatova, and Robin Chisnell are fellow graduate
students in the Boston area, and dear friends who have accompanied me on this
journey. Thank you for the potluck dinners, ski trips, hiking outings, gym workouts,
and runs along the Charles River — you have done so much to lift my spirits and
bring joy to my life. Another huge source of strength comes from the Sidney-Pacific
Graduate Community, without a doubt the best graduate dormitory in the world.
Living at SP was one of the highlights of my time at MIT, and has taught me so
much about friendship, leadership, and the spirit of community. In particular, I wish
to thank the former Housemaster team — Professor Roger Mark and Mrs. Dottie
Mark, and Professor Annette Kim and Dr. Roland Tang — thank you for welcoming
me into your home and making sure that I am always well-fed. To my friends from SP
— Amy Bilton, Mirna Slim, Ahmed Helal, Jit Hin Tan, Brian Spatocco, George Lan,
Po-Ru Loh, Kendall Nowocin, George Chen, Boris Braverman, Fabián Kozynski, and
many, many others — thank you for sharing your kindness and talents, for inspiring
me to set ambitious goals, and for showing me that there are few problems in life that
can’t be solved through teamwork and more bacon.
Finally, I would like to thank my family for their steadfast love and support. My
parents, Jie Chen and Helin He, left behind friends, family, and successful careers in
China to immigrate to the United States in pursuit of better opportunities. Without
their sacrifice, I would not be where I am today. I credit them for instilling in me the
6
importance of a good education, the determination to overcome whatever challenges
may arise, and the confidence to forge my own path. Last but not least, I want to
thank Tim Curran for being my co-pilot on this incredible journey. Thank you for
always believing in me, for teaching me to be patient with myself and others, and for
giving me the strength to push through the tough times. I look forward to starting
the next chapter of our lives together.
The work presented in this thesis was funded in part by the International Design Center at the Singapore University of Technology and Design, by the DARPA
META program through AFRL Contract Number FA8650-10-C-7083 and Vanderbilt
University Contract Number VU-DSR #21807-S7, by the National Science Foundation Graduate Research Fellowship Program, and by the Zonta International Amelia
Earhart Fellowship Program.
7
Contents
1 Introduction
19
1.1
Motivation for Uncertainty Quantification . . . . . . . . . . . . . . .
20
1.2
Terminology and Scope . . . . . . . . . . . . . . . . . . . . . . . . . .
21
1.2.1
System Design Process . . . . . . . . . . . . . . . . . . . . . .
21
1.2.2
Characterizing Uncertainty . . . . . . . . . . . . . . . . . . . .
24
1.2.3
Some Common Probability Distributions . . . . . . . . . . . .
26
Designing for Robustness and Reliability . . . . . . . . . . . . . . . .
28
1.3.1
Defining Robustness and Reliability . . . . . . . . . . . . . . .
29
1.3.2
Background and Current Practices . . . . . . . . . . . . . . .
30
1.3.3
Opportunities for Intellectual Contribution . . . . . . . . . . .
33
1.4
Thesis Objectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
33
1.5
Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
34
1.3
2 Quantifying Complexity and Risk in System Design
2.1
2.2
2.3
37
Background and Literature Review . . . . . . . . . . . . . . . . . . .
37
2.1.1
Quantitative Complexity Metrics . . . . . . . . . . . . . . . .
38
2.1.2
Differential Entropy . . . . . . . . . . . . . . . . . . . . . . . .
40
A Proposed Definition of Complexity . . . . . . . . . . . . . . . . . .
41
2.2.1
Exponential Entropy as a Complexity Metric . . . . . . . . . .
42
2.2.2
Interpreting Complexity . . . . . . . . . . . . . . . . . . . . .
44
2.2.3
Numerical Complexity Estimation . . . . . . . . . . . . . . . .
46
Related Quantities . . . . . . . . . . . . . . . . . . . . . . . . . . . .
47
9
2.3.1
Variance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
49
2.3.2
Kullback-Leibler Divergence and Cross Entropy . . . . . . . .
51
2.3.3
Entropy Power . . . . . . . . . . . . . . . . . . . . . . . . . .
52
2.4
Estimating Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
53
2.5
Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . .
54
3 Sensitivity Analysis
3.1
3.2
3.3
3.4
3.5
55
Background and Literature Review . . . . . . . . . . . . . . . . . . .
56
3.1.1
Variance-Based Global Sensitivity Analysis . . . . . . . . . . .
56
3.1.2
Vary-All-But-One Analysis . . . . . . . . . . . . . . . . . . . .
59
3.1.3
Distributional Sensitivity Analysis . . . . . . . . . . . . . . . .
60
3.1.4
Relative Entropy-Based Sensitivity Analysis . . . . . . . . . .
61
Extending Distributional Sensitivity Analysis to Incorporate Changes
in Distribution Family . . . . . . . . . . . . . . . . . . . . . . . . . .
62
3.2.1
Cases with No Change in Distribution Family . . . . . . . . .
65
3.2.2
Cases Involving Truncated Extent for X o . . . . . . . . . . . .
67
3.2.3
Cases Involving Resampling . . . . . . . . . . . . . . . . . . .
71
Local Sensitivity Analysis . . . . . . . . . . . . . . . . . . . . . . . .
75
3.3.1
Local Sensitivity to QOI Mean and Standard Deviation . . . .
76
3.3.2
Relationship to Variance-Based Sensitivity Indices . . . . . . .
77
3.3.3
Local Sensitivity to Design Parameter Mean and Standard Deviation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
79
Entropy Power-Based Sensitivity Analysis . . . . . . . . . . . . . . .
81
3.4.1
Entropy Power Decomposition . . . . . . . . . . . . . . . . . .
81
3.4.2
The Effect of Distribution Shape . . . . . . . . . . . . . . . .
85
Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . .
88
4 Application to Engineering System Design
91
4.1
R-C Circuit Example . . . . . . . . . . . . . . . . . . . . . . . . . . .
91
4.2
Sensitivity Analysis of the R-C Circuit . . . . . . . . . . . . . . . . .
94
4.2.1
94
Identifying the Drivers of Uncertainty . . . . . . . . . . . . . .
10
4.2.2
4.3
4.4
Understanding the Impacts of Modeling Assumptions . . . . .
Design Budgets and Resource Allocation . . . . . . . . . . . . . . . . 101
4.3.1
Visualizing Tradeoffs . . . . . . . . . . . . . . . . . . . . . . . 102
4.3.2
Cost and Uncertainty Budgets . . . . . . . . . . . . . . . . . . 104
4.3.3
Identifying and Evaluating Design Alternatives . . . . . . . . 107
Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
5 Application to a Real-World Aviation Policy Analysis
5.1
5.2
5.3
5.4
94
115
Background and Problem Overview . . . . . . . . . . . . . . . . . . . 115
5.1.1
Aviation Noise Impacts . . . . . . . . . . . . . . . . . . . . . . 115
5.1.2
CAEP/9 Noise Stringency Options . . . . . . . . . . . . . . . 117
APMT-Impacts Noise Module . . . . . . . . . . . . . . . . . . . . . . 119
5.2.1
Model Inputs . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
5.2.2
Estimating Monetary Impacts . . . . . . . . . . . . . . . . . . 125
Results and Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . 127
5.3.1
Comparison of Stringency Options and Discount Rates . . . . 127
5.3.2
Global and Distributional Sensitivity Analysis . . . . . . . . . 128
5.3.3
Visualizing Tradeoffs . . . . . . . . . . . . . . . . . . . . . . . 132
Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6 Conclusions and Future Work
137
6.1
Thesis Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
6.2
Thesis Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
6.3
Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
A Bimodal Distribution with Uniform Peaks
141
A.1 Specifying Distribution Parameters . . . . . . . . . . . . . . . . . . . 142
A.2 Options for Variance Reduction . . . . . . . . . . . . . . . . . . . . . 143
A.3 Switching to a Bimodal Uniform Distribution . . . . . . . . . . . . . 145
B Derivation of Local Sensitivity Analysis Results
147
B.1 Perturbations in Mean . . . . . . . . . . . . . . . . . . . . . . . . . . 147
11
B.1.1 Effect on Complexity . . . . . . . . . . . . . . . . . . . . . . . 148
B.1.2 Effect on Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
B.2 Perturbations in Standard Deviation . . . . . . . . . . . . . . . . . . 149
B.2.1 Effect on Complexity . . . . . . . . . . . . . . . . . . . . . . . 150
B.2.2 Effect on Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
C Derivation of the Entropy Power Decomposition
153
D CAEP/9 Analysis Airports
157
Bibliography
161
12
List of Figures
1-1 The effect of complexity on system development time in the aerospace,
automobile, and integrated circuit industries [1, Figure 1] . . . . . . .
20
1-2 Schematic of the proposed system design process . . . . . . . . . . . .
22
1-3 Examples of uniform, triangular, Gaussian, and bimodal uniform probability densities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
26
1-4 Schematics of design for robustness and reliability. Red and blue probability densities represent initial and updated estimates of fY (y), respectively. Dashed line indicates the location of the requirement r.
Shaded portions denote regions of failure in which the requirement is
not satisfied. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
29
2-1 Probability density of a standard Gaussian random variable (left) and
a uniform random variable of equal complexity (right) . . . . . . . . .
45
2-2 Probability density of a multivariate Gaussian random variable with
µ = (0, 0) and Σ = I2 (top left), and uniform random variables of equal
complexity in R2 and R (top right and bottom, respectively) . . . . .
45
2-3 Numerical approximation of h(Y ) using k bins of equal size . . . . . .
47
2-4 Comparison of variance, differential entropy, and complexity between
a bimodal uniform distribution and its equivalent Gaussian distribution 50
2-5 Probability of failure associated with the requirement r . . . . . . . .
53
3-1 Apportionment of output variance in GSA [8, Figure 3-1] . . . . . . .
56
13
3-2 Examples of reasonable uniform, triangular, and Gaussian distributions. In each figure, the solid red line denotes fX o (x), the dashed blue
lines represent various possibilities for fX 0 (x), and the dotted black
lines show κfX o (x), which is greater than or equal to the corresponding density fX 0 (x) for all x.
. . . . . . . . . . . . . . . . . . . . . . .
65
3-3 Examples of reasonable uniform, triangular, and Gaussian distributions for cases involving a truncated extent for X o
. . . . . . . . . .
69
3-4 Examples of reasonable uniform, triangular, and Gaussian distributions for cases involving resampling. The left column shows instances
where δ is too large, and resampling is required. The right column
shows instances where δ is sufficiently small, and AR sampling can be
used to generate samples of X 0 from samples of X o .
. . . . . . . . .
73
3-5 The relative locations of r and µY greatly impact the change in risk
associated with a decrease in σY . Moving from the red probability
density to the blue, P (Y < r) decreases if µY − r > 0, and increases if
µY − r < 0. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
77
3-6 Examples of Y = X1 + X2 with increase, decrease, and no change in
Gaussianity between the design parameters and QOI . . . . . . . . .
87
3-7 Sensitivity indices Si , ηi , and ζi for three examples of Y = X1 + X2 .
For each factor, Si equals the product of ηi and ζi . . . . . . . . . . . .
88
4-1 R-C high-pass filter circuit . . . . . . . . . . . . . . . . . . . . . . . .
92
4-2 Bode diagrams for magnitude (top) and phase (bottom). The dashed
black line indicates the required cutoff frequency of 300 Hz. . . . . . .
92
4-3 Histogram of cutoff frequency for Case 0 generated using 10,000 MC
samples. The dashed black line indicates the required cutoff frequency
of 300 Hz. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
93
4-4 Global and distributional sensitivity analysis results for Case 0 . . . .
94
4-5 Distributional sensitivity analysis results with changes to distribution
family . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
14
95
4-6 Examples of uniform (U), triangular (T), Gaussian (G), and bimodal
uniform (BU) distributions for X2 , each with a variance of 0.30 µF2 .
96
4-7 Examples of uniform (U), triangular (T), Gaussian (G), and bimodal
uniform (BU) distributions for X2 supported on the interval [3.76, 5.64] µF 98
4-8 Histogram of Y resulting from various distributions for X2 considered
in Case 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
99
4-9 Main effect and entropy power-based sensitivity indices for Case 2 . . 100
4-10 Contours for variance, complexity, and risk corresponding to reductions
in factor variance, generated from 10,000 MC simulations (solid colored
lines) or approximated using distributional or local sensitivity analysis
results (dashed black lines) . . . . . . . . . . . . . . . . . . . . . . . . 103
4-11 Notional cost associated with a 100(1 − δ)% reduction in the variance
of X1 or X2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
4-12 Contours for variance, complexity, and risk corresponding to reductions
in factor variance (solid colored lines) overlaid with contours for cost
of implementation (dashed green lines) . . . . . . . . . . . . . . . . . 106
4-13 Contours for risk (solid blue lines) overlaid with contours for cost
(dashed green lines). Shaded region denotes area in which all budget constraints are satisfied. . . . . . . . . . . . . . . . . . . . . . . . 107
4-14 Uncertainty contours for variations in resistance and capacitance, generated from 10,000 MC simulations (solid colored lines) or approximated using entropy power-based sensitivity analysis results (dashed
black lines) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108
4-15 Uncertainty contours for variations in resistance and resistor tolerance 108
4-16 Uncertainty contours for variations in capacitance and capacitor tolerance109
4-17 Uncertainty contours for variations in resistor and capacitor tolerance
109
4-18 Tradeoffs in complexity and risk for various design options. Dashed
green lines bound region where both complexity and risk constraints
are satisfied. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
15
5-1 Joint and marginal distributions for the regression parameters obtained
using 10,000 bootstrapped samples . . . . . . . . . . . . . . . . . . . 122
5-2 Schematic of APMT-Impacts Noise Module [142, adapted from Figure 4]126
5-3 Comparison of mean estimates for Y in Stringencies 1–3 and discount
rates between 3%–9%. Error bars denote 10th , 90th percentile values.
127
5-4 Main effect sensitivity indices for Stringency 1, DR = 3%, with IC
modeled probabilistically in (a) and as a deterministic value in (b) . . 129
5-5 Adjusted main effect sensitivity indices for Stringency 1, DR = 3%,
with IC modeled probabilistically in (a) and as a deterministic value
in (b) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
5-6 Histogram of Y under various IC assumptions for Stringency 1, DR =
3% . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
5-7 Uncertainty contours associated with changes in the mean and variance
of BNL and IC, generated from 10,000 MC simulations (solid colored
lines) or approximated using sensitivity analysis results (dashed black
lines) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
5-8 Uncertainty contours associated with changes in the mean and variance
of CU and IC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
5-9 Uncertainty contours associated with changes in the mean and variance
of BNL and CU . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
A-1 Bimodal uniform distribution with peak bounds aL , bL , aR , and bR ,
and peak heights fL and fR . . . . . . . . . . . . . . . . . . . . . . . 141
A-2 Examples of distributions for X 0 obtained using Algorithms 1–3 (solid
o
lines), based on an initial distribution X BU (dashed green line) and
δ = 0.5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
A-3 Examples of switching from a uniform, triangular, or Gaussian distribution to a bimodal uniform distribution. In each figure, the solid
red line denotes fX o (x), the dashed blue line represents fX 0 (x), and
the dotted black line shows κfX o (x), which is greater than or equal to
fX 0 (x) for all x. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
16
List of Tables
2.1
Range, mean, variance, differential entropy, exponential entropy, and
entropy power for distributions that are uniform, triangular, Gaussian,
or bimodal with uniform peaks
2.2
. . . . . . . . . . . . . . . . . . . . .
The properties of variance, differential entropy, exponential entropy,
and entropy power under constant scaling or shift . . . . . . . . . . .
3.1
48
Ratio of δ =
var (X 0 )
var (X o )
49
for changes in factor distribution between the
uniform, triangular, and Gaussian families. Rows represent the initial
distribution for X o and columns correspond to the updated distribution
for X 0 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
64
3.2
Local sensitivity analysis results for complexity and risk . . . . . . . .
76
4.1
Change in uncertainty estimates for Cases 1-T, 1-G, and 1-BU, as
compared to Case 0 . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.2
Change in uncertainty estimates for Cases 2-T, 2-G, and 2-BU, as
compared to Case 0 . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4.3
96
98
Best achievable uncertainty mitigation results given individual budgets
for cost and uncertainty. Entries in red denote budget violations. . . . 107
4.4
Local sensitivity predictions for the change in mean and standard deviation (SD) of each factor required to reduce risk to 10% . . . . . . . 112
4.5
Uncertainty and cost estimates associated with various design options.
Entries in red denote budget violations. . . . . . . . . . . . . . . . . . 113
17
5.1
Inputs to CAEP/9 noise stringency case study. The definitions for
BNL, CU, RP, SL, and IGR correspond to the midrange lens of the
APMT-Impacts Noise Module, whereas IC and DR are used in the
post-processing of impacts estimates to compute net cost-benefit. . . 124
5.2
Comparison of mean and uncertainty estimates of Y for Stringencies 1–
3 and discount rates between 3%–9%. All monetary values (mean, 10th
and 90th percentiles, standard deviation, and complexity) are listed in
2006 USD (billions). . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
5.3
Main effect sensitivity indices for Stringency 1 at various discount rates 129
D.1 CAEP/9 US airports and income levels . . . . . . . . . . . . . . . . . 160
18
Chapter 1
Introduction
Over the years, engineering systems have become increasingly complex, with astronomical growth in the number of components and their interactions. With this rise
in complexity comes a host of new challenges, such as the adequacy of mathematical
models to predict system behavior, the expense and time to conduct experimentation
and testing, and the management of large, globally-distributed design teams. These
issues all contribute uncertainty to the design process, which can have serious implications for system robustness and reliability and potentially disastrous repercussions
for program outcome.
To address some of these issues, this thesis presents a stochastic process model
to describe the design of complex systems, in which uncertainty in various parameters and quantities of interest is characterized probabilistically, and updated through
successive design iterations as new estimates become available. Incorporated in the
model are methods to quantify system complexity and risk, and reduce them through
the allocation of resources to enhance robustness and reliability. The goal of this
approach is to enable the rigorous quantification and management of uncertainty,
thereby serving to help mitigate technical and programmatic risk.
Section 1.1 describes the motivation for uncertainty quantification, followed by
Section 1.2, which defines the scope of the research and the terminology used to
describe it. Section 1.3 gives an overview of current methodologies used in design
for robustness and reliability, and identifies several potential areas for intellectual
19
contribution. Sections 1.4 and 1.5 outline the main research objectives and provide a
roadmap for the remainder of the thesis.
1.1
Motivation for Uncertainty Quantification
In the aerospace industry, the challenges associated with complexity are particularly
daunting. A recent study by the Defense Advanced Research Projects Agency shows
that as aerospace systems have become more complex, their associated costs and
development times have also reached unsustainable levels (Figure 1-1) [1].1 Until this
unmitigated growth can be contained, it poses a looming threat to the continued
viability and competitiveness of the industry.
Figure 1-1: The effect of complexity on system development time in the aerospace,
automobile, and integrated circuit industries [1, Figure 1]
To understand the potentially devastating effects of complexity in its various manifestations, one needs only to look to recent history, which is replete with examples
of major design setbacks due to failure to recognize and reduce performance, cost,
1
The complexity metric used in the study is the number of parts plus the source lines of code.
For a discussion of various measures of complexity and their applicability, see Section 2.1.
20
and schedule risks as they emerge during the design process. A notable example is
the Hubble Space Telescope which, when first launched, failed its optical resolution
requirement by an order of magnitude. A Shuttle repair mission, costing billions of
additional dollars, was required to remedy the problem [2]. An investigation later
uncovered that data revealing optical system errors were available during the fabrication process, but were not recognized and fully examined; in fact, crucial error
indicators were disregarded and key verification steps omitted. The V-22 Osprey tiltrotor aircraft is another example: over the course of its 25-year development cycle,
the program was fraught with safety, reliability, and affordability challenges, resulting in numerous flight test crashes with concomitant loss of crew and passenger lives
[3]. More recently, the Boeing 787 Dreamliner transport aircraft program has experienced a number of major prototype subsystem test failures, causing budget overruns
of billions of dollars and service introduction delays of about three years. Many of
the program’s problems were attributed to Boeing’s aggressive strategy to outsource
component design and assembly, which created heavy program management burdens
and led to unforeseen challenges during vehicle integration [4].
In these cases and numerous others, the design program was unaware of the mounting risks in the system, and was surprised by one or more unfortunate outcomes. Although these examples are extreme, they are suggestive that improved system design
practices are required in order to identify and address performance, cost, and schedule
risks as they emerge. Tackling the problem of complexity — and more broadly — of
uncertainty in the design of engineering systems is an enormous yet urgent task that
requires innovation and collaboration on both sides of industry and academia.
1.2
1.2.1
Terminology and Scope
System Design Process
Figure 1-2 shows a schematic of the system design process proposed in this thesis. It
outlines the key steps in the design of engineering systems, by which is meant col-
21
Set
Targets
Define
Parameters
𝑋1
Requirements,
Constraints
𝑋2
𝑋3
Perform
Quantify
Uncertainty Sensitivity
Analysis
Evaluate
Quantities
of Interest
Generate
Designs
Models,
Governing
Equations,
Analysis
Tools, etc.
𝑌1
0.35
0.3
0.25
0.2
𝑌2
0.15
0.1
0.05
0
-0.05
-1
-0.5
0
0.5
1
1.5
2
2.5
3
QOI
Design
Feedback
Variance, Complexity,
and Risk
Uncertainty
Feedback
Robustness and Reliability
Resource
Allocation
Figure 1-2: Schematic of the proposed system design process
lections of interrelated elements that interact with one another to achieve a common
purpose [5]. A system can be a physical entity, such as a vehicle or consumer product,
or an abstract concept, such as a software program or procedure. Whatever its form,
a system must satisfy a set of targets in order to be deemed successful. These targets
typically refer to functional requirements that dictate the system’s performance, but
may also include additional constraints that stipulate conditions such as budget or
time.
From the system targets, a set of design parameters can be enumerated which
are used to characterize the system. Design parameters are variables that relate
the requirements in the functional domain to aspects of the design in the physical
domain that can be manipulated to satisfy those requirements [6]. They can include
quantities that define the system itself, such as individual component dimensions, as
well as data describing the procedures and other elements by which the system can
be manufactured, operated, and maintained [7]. In this thesis, we assume that all
design parameters within a system are independent, and denote them using the m × 1
22
vector x = [x1 , x2 , . . . , xm ]T , where xi is the ith design parameter among a total of m
design parameters.
In addition to the design parameters, there are also quantities that are used to
characterize aspects of the system’s performance that are of interest to the designer.
We refer to them generically as quantities of interest (QOI), and represent them using
the n × 1 vector y = [y1 , y2 , . . . , yn ]T , where yj is the j th QOI out of a total of n.
Quantities of interest are typically evaluated indirectly as functions of the design
parameters using available models and analysis tools. For example, in aerospace
engineering, relevant QOI can include the gross weight of an aircraft, which can be
computed from design parameters regarding vehicle geometry and material selection,
or the lift-to-drag ratio, which can be estimated from the external geometry using
results from computational fluid dynamics and wind tunnel testing. For simplicity,
we will assume that there already exist models and tools (may be black-box) with
which to evaluate the QOI of a system from the design parameters. We express this
relationship as:
y = g(x),
(1.1)
where g : Rm → Rn denotes the mapping from design parameters to QOI. For the j th
QOI yj , this mapping is given by:
yj = gj (x),
(1.2)
where gj : Rm → R1 , and j = 1, 2, . . . , n.
With regard to analysis models and tools used to aid system design, we will also
use the terms factor and output. Consistent with the definitions set forth in Ref. [8],
a factor refers to an external input to a model, whereas output is a model result of
interest. While these terms are similar to design parameter and quantity of interest
in the context of a system, they are not equivalent. A factor can refer to any model
input, and need not correspond to a parameter that can be manipulated by the
designer. Similarly, the outputs of a model are not necessarily limited to the QOI of
a system.
23
1.2.2
Characterizing Uncertainty
For most realistic engineering systems, the design process is made more difficult by
the presence of non-deterministic features that contribute to stochasticity in the outcome, which comprise the uncertainty associated with the QOI [9, 10]. In this thesis, uncertainty quantification refers to “the science of identifying, quantifying, and
reducing uncertainties associated with models, numerical algorithms, experiments,
and predicted outcomes or quantities of interest” [11]. We focus specifically on the
development of methods to quantify uncertainty during the intermediate stages of
system design, where requirements, design parameters, and QOI have already been
defined, but no feasible design has yet been realized. Uncertainty pertaining to the
formulation of functional requirements, development of analysis tools, manufacturing,
deployment, maintenance, and organizational structure will not be discussed.
Uncertainty can come in many flavors; for example, it can be broadly categorized
as epistemic, owing to insufficient or imperfect knowledge, or aleatory, which arises
from natural randomness and is therefore irreducible [10–13]. Furthermore, there
are several types of uncertainty associated with the use of computer-based models in
system design [14]:
• Parameter uncertainty, resulting from not knowing the true values of the inputs
to a model;
• Parametric variability, relating to unspecified conditions in the model inputs or
parameters;
• Residual variability, due to the inherently unpredictable or stochastic nature of
the system;
• Observation error, referring to uncertainty associated with actual observations
or measurements;
• Model inadequacy, owing to the insufficiency of any model to exactly predict
reality;
• Code uncertainty, arising from not knowing the output of a model given a
particular set of inputs until the code is run.
24
The framework presented in this thesis is intended to be general enough to describe
any of aforementioned types of uncertainty, provided that they can be described
probabilistically. However, we note that although both epistemic and aleatory sources
of uncertainty can be characterized within this framework, only epistemic uncertainty
can be reduced through additional research and improved knowledge. Thus, while
it may be possible to study the effects of aleatory sources of uncertainty, it is not
appropriate to allocate resources toward their reduction.
To characterize the propagation of uncertainty, we employ continuous random
variables to represent design parameters2 and quantities of interest [10, 18, 19], and
model the time evolution of design as a stochastic process whose outcome is governed
by some probability distribution. We simulate the behavior of the system using
Monte Carlo (MC) sampling, which generates numerous QOI estimates from which
a probability density can be estimated.3 We use X = [X1 , X2 , . . . , Xm ]T and Y =
[Y1 , Y2 , . . . , Yn ]T to denote vectors of random variables that correspond to each entry in
x and y, respectively. The probability densities associated with Xi and Yj are given by
fXi (xi ) and fYj (yj ), respectively. In general, these densities can be of arbitrary form,
and closed-form expressions may not be available. The stochastic representations of
(1.1) and (1.2) can be written as:
Y = g(X),
(1.3)
Yj = gj (X).
(1.4)
Once the QOI have been estimated, we use selected metrics to quantitatively
describe their levels of uncertainty. This information is compared to the system’s
2
The process of assigning design parameters to specific probability distributions will not be discussed in this thesis. Instead, the reader is referred to Refs. [15–17], which detail several methods
addressing this topic. They include density estimation techniques to leverage historical data, elicitation of expert opinion to supplement empirical evidence, and calibration and aggregation procedures
to combine information from multiple sources.
3
Monte Carlo simulation is computationally expensive, and may be intractable for some systems.
In practice, computational expense can be reduced by using quasi-random sampling techniques
[20, 21], reducing the dimensionality of the input space [22–24], or employing less-expensive surrogate models to approximate system response [25–29]. In this thesis, we will assume that forward
simulation using MC sampling is computationally feasible.
25
uncertainty targets through the uncertainty feedback loop. We also perform sensitivity
analysis to identify the key drivers of uncertainty, enumerate uncertainty reduction
strategies, and evaluate tradeoffs in the context of resource allocation decisions for
uncertainty mitigation, thus enhancing system robustness and/or reliability.
1.2.3
Some Common Probability Distributions
Throughout this thesis, we will focus on four particular types of probability distributions for characterizing uncertainty in the design parameters: uniform, triangular,
𝑓𝑋 𝑥
𝑓𝑋 𝑥
Gaussian, and bimodal with uniform peaks, which are shown in Figure 1-3.
𝑎
𝑏
𝑥
𝑎
𝑥
𝑐
𝑏
𝑓𝑋 𝑥
𝑓𝑋 𝑥
(a) Uniform distribution with lower bound (b) Triangular distribution with lower
a and upper bound b
bound a, upper bound b, and mode c
𝑓𝐿
𝜎
𝜇
𝑥
𝑎𝐿
𝑓𝑅
𝑏𝐿
𝑥
𝑎𝑅
𝑏𝑅
(c) Gaussian distribution with mean µ and (d) Bimodal distribution composed of
standard deviation σ
uniform peaks with bounds aL , bL , aR ,
and bR , and peak heights fL and fR
Figure 1-3: Examples of uniform, triangular, Gaussian, and bimodal uniform probability densities
26
A uniform distribution is specified by two parameters, a and b, which correspond
to the lower and upper bound of permissible values, respectively. All values on the
interval [a, b] are equally likely, and therefore a uniform distribution is often assumed
when no knowledge exists about a parameter except for its limits. The probability
density of a uniform random variable X is given by:

1



fX (x) = b − a


0
if a ≤ x ≤ b,
(1.5)
otherwise.
Similar to the uniform distribution, a triangular distribution is also parameterized
by a lower bound a and an upper bound b. There is also an additional parameter, c,
which represents the mode. A triangular distribution is appropriate when information
exists about the bounds of a particular quantity, as well as the most likely value. The
corresponding probability density is:

2(x − a)




(b − a)(c − a)



2(b − x)
fX (x) =


(b − a)(b − c)




0
if a ≤ x ≤ c,
if c < x ≤ b,
(1.6)
otherwise.
Unlike uniform and triangular distributions, a Gaussian distribution does not have
finite bounds. It is often used when there is strong confidence in the most likely value
[17]. A Gaussian distribution is parameterized by its mean µ and standard deviation
σ. Its probability density is nonzero everywhere along the real line:
(x − µ)2
fX (x) = √
exp −
.
2σ 2
2πσ 2
1
(1.7)
In practice, it is often unrealistic to work with random variables with infinite support.
Instead, it may be necessary to approximate a Gaussian distribution with a truncated
form [30].
Bimodal distributions are used to characterize quantities consisting of two distinct
27
sets of features or behaviors. They can arise in a variety of science and engineering
contexts: for example, in materials science, to characterize the size and properties
of polymer particles [31–33]; in medicine, to explain disease epidemiology within and
between populations [34, 35]; in human factors, to analyze individual and group communication patterns [36, 37]; in earth science, to describe the occurrence of natural
phenomena such as rainfall and earthquakes [38–41]; and in electrical engineering, to
relate disparities in circuit component lifetime to distinct modes of failure [42, 43].
For simplicity, we will restrict our consideration to bimodal distributions composed
of two uniform peaks, henceforth termed “bimodal uniform,” as shown in Figure 13(d). We use [aL , bL ] and [aR , bR ] to denote the lower and upper bounds of the left
and right peaks, respectively, and fL and fR to designate the peak heights. It takes
five parameters to uniquely specify a distribution of this type; therefore, given any
five of {aL , bL , fL , aR , bR , fR }, the sixth can be directly computed from the condition
that:
fL (bL − aL ) + fR (bR − aR ) = 1.
(1.8)
The probability density of a bimodal uniform distribution is given by:



fL



fX (x) = fR




0
1.3
if aL ≤ x ≤ bL ,
if aR ≤ x ≤ bR ,
(1.9)
otherwise.
Designing for Robustness and Reliability
Uncertainty in engineering design is closely related to the robustness and reliability of
the resulting systems. In this section, we provide definitions for these terms, discuss
current practices in design for robustness and reliability, as well as identify several
opportunities for intellectual contribution.
28
1.3.1
Defining Robustness and Reliability
According to Knoll and Vogel, robustness is “the property of systems that enables
them to survive unforeseen or unusual circumstances” [44]. That is to say, to improve
a system’s robustness is to make it less susceptible to exhibit unexpected behavior in
the presence of a wide range of stochastic elements [45, 46].
Reliability, on the other hand, describes a system’s “probability of success in
satisfying some performance criterion” [9], and is closely tied to system safety [45].4
It is also related to the definition of risk that we will introduce in Section 2.4, which
poses risk and reliability as complementary quantities.
Refinement: Change Std Dev
Redesign: Shift Mean
0.8
0.6
0.6
𝑓𝑌 𝑦
Constant
Scaling
0.4
0.2
𝑃 𝑌<𝑟
0
-0.2
-4
f X(x)
f X(x)
𝑓𝑌 𝑦
0.8
0.4
𝑃 𝑌<𝑟
0.2
𝑦
𝑟
-3
Constant Shift
0
𝑦
𝑟
-0.2
(a)-2 Design
for
robustness
(b)
-1
0
1
2
3
4
5
-4
-3
-2 Design
-1
0 for 1reliability
2
3
Redesign
and
Refinement
x
4
5
x
𝑓𝑌 𝑦
0.8
Constant
Shift &
Scaling
f X(x)
0.6
0.4
0.2
0
𝑃 𝑌<𝑟
𝑦
𝑟
-0.2
-4
(c) Design
and
-3
-2 for
-1 robustness
0
1
2
x
reliability
3
4
5
Figure 1-4: Schematics of design for robustness and reliability. Red and blue probability densities represent initial and updated estimates of fY (y), respectively. Dashed
line indicates the location of the requirement r. Shaded portions denote regions of
failure in which the requirement is not satisfied.
4
Alternate definitions of reliability also ascribe an element of time. For example, reliability is
sometimes described as the probability of failure within a given interval of time [47], or the expected
life of a product or process, typically measured by the mean time to failure or the mean time between
failures [46]. In this thesis, we will not address the temporal aspect of reliability.
29
Graphically, we illustrate the distinction between design for robustness and design
for reliability in Figure 1-4, in which our estimate of a QOI Y is updated from the
red probability density to the blue. The dashed line y = r represents a performance
criterion below which the design is deemed unacceptable; the area of the shaded region
corresponds to P (Y < r), the probability of failure.5 Figure 1-4(a) shows a constant
scaling of fY (y) with no change in the mean value, whereas Figure 1-4(b) depicts
a constant shift (horizontal translation) of fY (y) with no change in the standard
deviation. Both activities have direct implications for the uncertainty associated with
Y . In practice, improvements in robustness and reliability can occur concurrently,
as depicted in Figure 1-4(c). In the following chapters, however, we will treat them
as distinct activities in order to investigate how robustness and reliability are related
to two proposed measures of uncertainty — complexity and risk — as well as how
sensitivity analysis can be used to predict changes in these quantities.
1.3.2
Background and Current Practices
The origins of robust design, also known as parameter design, can be traced to postWorld War II Japan, where it was pioneered by Genichi Taguchi as a way to improve
design efficiency and stimulate industries crippled by the war [48]. The aim of the
so-called “Taguchi Methods” is to reduce the sensitivity of products and processes
to various noise factors, such as manufacturing variability, environmental conditions,
and degradation over time [49, 50]. To achieve this goal, Taguchi applied design of
experiment techniques [51, 52], in particular the orthogonal array, to systematically
vary the levels of the control factors (e.g., the design parameters) and noise factors,
quantify the effect in the output, and compute a signal-to-noise ratio. In order to
improve system robustness, the designer seeks to determine the parameter values
that maximize the signal-to-noise ratio, thus minimizing the system’s sensitivity to
variability. Finally, a loss function is used to estimate the expected decrement in
quality due to deviation of the system response from a target value.
5
The probability of failure can also be defined as P (Y > r) in the case that the QOI must not
exceed r, noting that P (Y > r) = 1 − P (Y < r).
30
More recently, the robust design literature has expanded to include techniques
such as the combined array format [53, 54], response modeling [53–55], and adaptive
one-factor-at-a-time experimentation [56, 57]. Whereas traditional Taguchi methods
evaluate the influence of control factors and noise factors separately, the combined
array format merges the two sets in a single experimental design, thereby reducing the
number of required experiments. In response modeling, the output of the system is
modeled and used as the basis for design optimization, rather than the loss function.
One advantage of this approach is that it allows the designer to model the mean
and variance of the output separately, and thus gain an improved understanding
of underlying system behaviors [55, 58]. Finally, in adaptive one-factor-at-a-time
experimentation, the designer begins with a baseline set of factor levels, then varies
each factor in turn to a level that has not yet been tested while holding all other
factors constant. Along the way, the designer also seeks to optimize the system
response by keeping a factor at its new level if the response improves, or reverting to
its previous level if the response worsens. Although this approach does not guarantee
optimal factor settings, it has been shown in certain situations to be more effective
than traditional fractional factorial experimentation for improving design robustness
[56, 57, 59, 60].
Another design philosophy used in many industrial sectors today is what is known
as “Six Sigma.” First introduced by Motorola, Inc. in the mid-1980s, Six Sigma is
a set of strategies and techniques aimed to minimize risk by reducing the sources of
variability in a system, not just the system’s sensitivity to said variability [46, 61].
Qualitatively, to design for Six Sigma is to ensure that the mean system performance
far exceeds the design requirement such that failures are extremely rare. In this
approach, uncertainty in the QOI is represented using a Gaussian distribution with
mean µ and standard deviation σ (or “sigma”), which is subject to a failure criterion
r. Based on experimental testing, the number of sigmas that a particular design
achieves is estimated as the standard score (also known as the z value or z score) of
31
r with respect to the observed µ and σ, expressed by:
z=
r−µ
.
σ
(1.10)
From (1.10), the standard Gaussian cumulative distribution function Φ(z) can be
used to compute the reliability of the system with respect to the QOI. Quantitatively,
Six Sigma is defined by a 99.99966% probability of success, or 3.4 defects per million.6
Note that because Six Sigma seeks to maximize the probability of success through
variance reduction, it has implications for both system robustness and reliability (i.e.,
see Figure 1-4(a)).
In the design for reliability context, a common approach to minimize risk is to
ensure that a system’s nominal performance deviates from a critical requirement by a
specified amount. This is typically done by applying a multiplicative factor of safety
or an additive margin of safety (usually expressed as a percentage) to the requirement,
thus injecting conservatism into the design [5].
Another class of well-known methods are the First- and Second-Order Reliability
Methods (FORM and SORM, respectively). Originally used in civil engineering to
assess the reliability of structures under combined loads, FORM linearize a system
about the current design point in order to estimate the mean and standard deviation
of the resulting QOI [9, 63, 64]. As in Six Sigma, all random variables are assumed to
be Gaussian and uncorrelated, a z score (termed the “reliability index” in FORM) is
computed from the estimates of µ and σ, and the system is assigned a probability of
failure according to 1 − Φ(z). The goal of FORM is to employ various optimization
techniques to modify the design in such a way as to maximize the reliability index.
A similar approach is adopted in SORM, except that a second-order approximation
of the system is used [65, 66].
6
Confusingly, the 3.4 defects per million benchmark does not actually correspond to six standard
deviations of a Gaussian distribution. In fact, a failure probability of 3.4 × 10−6 corresponds to
1 − Φ(4.5), or “4.5 sigma.” By convention, the extra 1.5 sigma shift is introduced to account for
degradation in reliability over time [62]. The actual six sigma failure probability is 1 − Φ(6) =
9.9 × 10−10 , or approximately 1 defect per billion.
32
1.3.3
Opportunities for Intellectual Contribution
In view of current practices in designing for robustness and reliability, this thesis
proposes to make intellectual contributions in three main areas.
First, many existing methods use variance as the sole measure of uncertainty.
Although variance has many advantages as an uncertainty metric (see Section 2.3.1),
we will show in the following chapters that it also has several limitations, and should
be supplemented with additional measures such as complexity and risk in order to
provide a richer description of system robustness and reliability.
Second, a common theme in existing methods is the use of Gaussian distributions to represent uncertainty in a system’s design parameters and QOI. As we will
demonstrate, this simplifying assumption can lead to under- or overestimates of various uncertainty measures, as well as erroneous prioritization of efforts for uncertainty
mitigation. In this thesis, we will investigate the influence of distribution shape in
system design as it relates to uncertainty and sensitivity estimates.
Third, current methods for designing for robustness and reliability typically emphasize variance reduction, but do not provide specific details on how it is to be
achieved. Thus, a potential opportunity lies in using uncertainty and sensitivity information to identify specific strategies to mitigate uncertainty. We propose to do
this by evaluating the tradeoffs between available design options with respect to robustness and reliability, and by incorporating cost and uncertainty budgets to aid
decisions regarding resource allocation.
1.4
Thesis Objectives
The primary goal of this thesis is to quantify and understand the effects of uncertainty
in the design of engineering systems in order to guide decisions aimed to improve
system robustness and reliability. Specifically, the thesis objectives are:
1. To define and apply an information entropy-based complexity metric to estimate
uncertainty in the design of engineering systems.
33
2. To develop a sensitivity analysis methodology that identifies the key contributors to uncertainty, evaluates tradeoffs between various design alternatives, and
informs decisions regarding the allocation of resources for uncertainty mitigation.
3. To apply uncertainty quantification and sensitivity analysis methods to an engineering case study to demonstrate their utility for enhancing system robustness
and reliability.
1.5
Thesis Outline
This thesis is composed of six chapters. The structure and content of the remaining
chapters are outlined below.
• Chapter 2 gives an overview of various metrics used to describe uncertainty,
and proposes complexity and risk as two possible uncertainty measures in the
context of design for robustness and reliability. Definitions are presented for
each of these terms, as well as a discussion of how they can be applied to
characterize uncertainty in various quantities of interest.
• Chapter 3 provides a discussion of existing sensitivity analysis methods and
how they can be used to identify the key drivers of uncertainty. New local
sensitivity analysis techniques are introduced, which can be used to predict
changes in complexity and risk associated with a reduction in variance. In
addition, a novel approach is presented that decomposes output uncertainty
into contributions due to the intrinsic extent and non-Gaussianity of the input
factors and their interactions.
• Chapter 4 describes how uncertainty and sensitivity information can be used
in conjunction with design budgets to identify specific uncertainty mitigation
strategies, visualize tradeoffs between available options, and guide the allocation
of resources aimed at improving system robustness and reliability.
34
• Chapter 5 presents a case study in which the methods developed in this thesis
are applied to assess the net cost-benefit of a set of real-world aviation noise
stringency options and support the design of cost-effective policy implementation responses.
• Chapter 6 summarizes the key contributions of this thesis and highlights some
areas of future work.
35
Chapter 2
Quantifying Complexity and Risk
in System Design
There are many possible ways to quantify uncertainty in a system. In this chapter,
we propose complexity and risk as two specific measures of uncertainty, which can
be used to supplement variance and convey system robustness and reliability. Section 2.1 presents a brief overview of the literature on complexity in system design, with
a particular focus on Shannon’s differential entropy. Section 2.2 proposes a new metric for complexity based on Campbell’s exponential entropy, presents the associated
qualitative and quantitative definitions, and provides some possible interpretations
of complexity. Section 2.3 discusses how this complexity metric relates to several
other quantities that can be used to characterize uncertainty. An analogous set of
definitions is also proposed for quantifying risk, which is presented in Section 2.4.
2.1
Background and Literature Review
Complexity in system design is an elusive concept for which many definitions have
been proposed. Early work in the field of complexity science by Warren Weaver
posited complexity as the nebulous middle ground between order and chaos, a region
in which problems require “dealing simultaneously with a sizeable number of factors
37
which are interrelated into an organic whole” [67]. Another interpretation of this
idea considers a set of “phase transitions” during which the fundamental features of
a system undergo drastic changes [68]. As an illustrative example, consider the phase
transitions of water [69]. On one end of the spectrum, water is frozen into a simple
lattice of molecules whose structure and behavior are straightforward to understand.
At the other extreme, water in gaseous form consists of millions of molecules vibrating
at random, and the study of such a system requires methods of statistical mechanics
or probability theory [70]. In between the two lies the complex liquid state, wherein
water molecules behave in a manner neither orderly nor chaotic, but at once enigmatic
and mesmerizing, which has captured the imagination of fluid dynamicists past and
present.
Although the above example makes the idea of complexity relatable to a large
audience, the debate over its definition still persists. When asked for the meaning of
complexity, Seth Lloyd once replied, “I can’t define it for you, but I know it when I
see it” [69]. While complexity may be in the eye of the beholder, many researchers
agree that there are several properties that complex systems tend to share [71–73]:
• They consist of many parts.
• There are many interactions among the parts.
• The parts in combination produce synergistic effects that are not easily predicted and may often be novel, unexpected, even surprising.
• They are difficult to model and to understand.
2.1.1
Quantitative Complexity Metrics
In addition to qualitative descriptions of complexity, there have also been many attempts to explain complexity using quantitative measures. These metrics can be classified into three general categories: structure-based, process-based, and informationbased.
Structure-based metrics quantify the complexity associated with the physical representation of a system [74]. They typically involve counting strategies: in software
38
engineering, the source lines of code can be used to describe a computer program [75];
in mechanical design, analogous measures include the number of parts [76], functions
[77], or core technologies [78] embodied in a product. Although appealing in their
simplicity, these counting metrics may be susceptible to different interpretations of
what constitutes a distinct component — depending on the level of abstraction, a
component may be as high-level as an entire subsystem, or as basic as the nuts and
bolts holding it together. More sophisticated structure-based metrics also attempt to
address the issue of component interactions. For example, McCabe proposed the idea
of cyclomatic complexity in software engineering, which uses graph theory to determine the number of control paths through a module [79]. Numerous others have also
recommended approaches to estimate system complexity by characterizing the number, extent, and nature of component interactions, which govern the interconnectivity
and solvability of a system [73, 80–82]. Overall, structure-based complexity metrics
are usually easy to understand and to implement, but they may not be meaningful
except in the later stages of design, after most design decision have been made, and
the system is well-characterized [83].
A second class of complexity metrics quantifies system uncertainty in terms of
processes required to realize the system. Metrics in this category can measure, for
example, the number of basic operations required to solve a problem (computational
complexity) [84, 85], the compactness of an algorithm needed to specify a message
(algorithmic complexity or Kolmogorov complexity) [86–88], or the amount of effort
necessary to design, modify, manufacture, or assemble a product [6, 74, 83, 89].
A third possible interpretation of complexity is related to a system’s information
content, often described using the notion of entropy. Like complexity, entropy is
also an abstruse concept.1 Early work by thermodynamicists Carnot and Clausius
sought to use entropy to describe a system’s tendency towards energy dissipation;
later, Gibbs and Boltzmann formalized this notion by giving it a statistical basis,
conceptualizing entropy as a measure of the uncertainty in the macroscopic state of
1
The Hungarian-American mathematician John von Neumann once famously said “nobody knows
what entropy really is, so in a debate you will always have the advantage” [90].
39
a system associated with the microstates that are accessible during the course of
the system’s thermal fluctuations. For a system with probability p(l) of accessing
microstate l, the Gibbs entropy S is defined as:
S = −kB
X
p(l) log p(l) ,
(2.1)
l
where kB = 1.38065 × 10−23 J/K is the Boltzmann constant.
From its origins in thermodynamics and statistical mechanics, entropy has since
been adapted to describe randomness and disorder in many other contexts, including
information theory [91], economics [92], and sociology [93–95]. Most relevant to this
thesis is the concept of information entropy, which was originally proposed by Claude
Shannon to study lossless compression schemes for communication systems [91]. It is
discussed in more detail in the following section; in Section 2.2.1, we will introduce a
complexity metric that is based on information entropy.
2.1.2
Differential Entropy
Shannon’s information entropy is a well-known measure of uncertainty, and represents
the amount of information required on average to describe a random variable. For
a discrete random variable Y that can assume any of k values y (1) , y (2) , . . . , y (k) , the
Shannon entropy is defined in terms of the probability mass function pY (y):
H(Y ) = −
k
X
pY (y (l) ) log pY (y (l) ).
(2.2)
l
The quantity H(Y ) is always nonnegative. In the limiting case, if Y is deterministic with a value of y ∗ , then pY (y) = 1 for y = y ∗ , and zero elsewhere, which
results in H(Y ) = 0. The units of H(Y ) depend on the base of the logarithm used in
(2.2); base 2 and base e are the most common, resulting in units of “bits” or “nats,”
respectively. In this thesis, we will deal exclusively with the natural logarithm. Comparing (2.1) and (2.2), it is easy to see that information entropy has an intuitive and
appealing analogy to thermodynamic entropy as a measure of a system’s tendency
40
toward disorder [96].
In the design of complex systems, design parameters and quantities of interest
typically can vary over a range of permissible values. In order to avoid restricting
these quantities to take only discrete values, we model them as continuous random
variables with associated probability densities. For a continuous random variable Y ,
the analogy to Shannon entropy is differential entropy, which is defined in terms of
fY (y) over the support set Y:
Z
h(Y ) = −
fY (y) log fY (y) dy.
(2.3)
Y
Unlike Shannon entropy, differential entropy can take on values anywhere along
the real line. For example, let Y be a uniform random variable on the interval [a, b].
Using (2.3) with fY (y) = (b − a)−1 for y ∈ [a, b], the differential entropy is computed
to be h(Y ) = log(b − a). Clearly, h(Y ) > 0 if b − a > 1, h(Y ) = 0 if b − a = 1,
and h(Y ) → −∞ as b − a → 0. This result comes about because whereas Shannon
entropy is an absolute measure of randomness in a thermodynamic sense, differential
entropy measures randomness relative to a reference coordinate system [88, 91].
2.2
A Proposed Definition of Complexity
In this thesis, we use complexity as a measure of uncertainty in a system. Qualitatively, we define complexity as the potential of a system to exhibit unexpected
behavior in the quantities of interest, whether detrimental or not [97]. This definition
is consistent with many of the theoretical formulations of complexity discussed in Section 2.1, such as the notions of emergent behavior, nonlinear component interactions,
and information content. It is also in line with the view that reducing complexity
translates to minimizing a system’s information content and maximizing its likelihood
of satisfying all requirements [74, 98].
41
2.2.1
Exponential Entropy as a Complexity Metric
Quantitatively, we use exponential entropy as our complexity metric. Exponential entropy was first proposed by L. Lorne Campbell to measure the extent of a probability
distribution [99]. For a continuous random variable Y , the simplest measure of extent is the range, ν(Y ), defined as the length of the interval on which the probability
density is nonzero:
Z
dy.
ν(Y ) =
(2.4)
Y
In the case where Y is a uniform random variable on the interval [a, b], ν(Y ) = b−a
as expected. One disadvantage of range, however, is that it gives equal weight to
all possible values of y ∈ Y, and consequently can be infinite, as is the case for
probability distributions such as Gaussian or exponential. In these cases, range is
no longer a useful measure of uncertainty; instead, it is desirable to formulate a
measure of “intrinsic extent.” Campbell put forward such a quantity in exponential
entropy, which is closely related to Shannon’s differential entropy [99]. The derivation
of exponential entropy is given below.2
First, consider Mt (Y ), the generalized mean of order t of fY (y) over Y [100]:
 Z !1/t
t

1



fY (y) dy

Y fY (y)
Mt (Y ) =
Z



1

exp
log
fY (y) dy
fY (y)
Y
if t 6= 0,
(2.5)
if t = 0.
From (2.5), it is easy to see that ν(Y ) = M1 (Y ); that is, range is the generalized
mean of order one. Furthermore, when t = 0, M0 (Y ) is the geometric mean of fY (y)
over Y [101].
For arbitrary orders s and t, where s < t, the relationship between Ms (Y ) and
2
Instead of working with fY (y) and Y, Campbell’s derivation uses the Radon-Nikodym derivative
of a probability space (Ω, A, P ), where Ω is the sample space, A is a σ-algebra, and P is a probability
measure. Thus, it is more general than our presentation, which applies to probability densities
defined over real numbers, in that it also extends to probability measures defined over arbitrary
sets.
42
Mt (Y ) can be described by the generalized mean inequality [100]:
Ms (Y ) ≤ Mt (Y ),
(2.6)
where equality holds if and only if Y is a uniform random variable.3 Setting s = 0
and t = 1, we can rewrite (2.6) to obtain a general relationship between M0 (Y ) and
ν(Y ):
M0 (Y ) ≤ M1 (Y ),
(2.7)
M0 (Y ) ≤ ν(Y ).
(2.8)
Campbell terms the quantity M0 (Y ) the intrinsic extent of Y , which is always
less than or equal to the range ν(Y ) (i.e., the “true” extent). Another interpretation
of M0 (Y ) is as “the equivalent side length of the smallest set that contains most of
the probability” [88]. That is to say, whereas ν(Y ) gives the length of the interval
(potentially infinite) that contains all possible values y for which Y = y, M0 (Y )
represents the (finite) extent of Y once the various values of y have been weighted by
their probability of occurrence. In the case where Y is a uniform random variable,
M0 (Y ) and ν(Y ) are equivalent.
In addition, M0 (Y ) is also closely related to Shannon’s differential entropy h(Y ).
To see this, consider the natural logarithm of M0 (Y ):
Z
1
fY (y) dy
log
log M0 (Y ) = log exp
fY (y)
Y
Z
1
=
fY (y) log
dy
fY (y)
Y
Z
= − fY (y) log fY (y) dy
(2.9)
(2.10)
(2.11)
Y
= h(Y ).
(2.12)
From (2.12), it follows that M0 (Y ) is the exponential of the differential entropy of Y ,
3
If Y is a uniform random variable, Mt (Y ) = ν(Y ) = b − a for all t.
43
hence the term “exponential entropy”:
M0 (Y ) = exp[h(Y )].
(2.13)
Using exponential entropy as our metric for complexity, a QOI represented by Y thus
has a complexity given by:
C(Y ) = exp[h(Y )].
2.2.2
(2.14)
Interpreting Complexity
As exponential entropy is a measure of intrinsic extent, it has the same units as the
original random variable. We note that it is also consistent with our proposed qualitative definition of complexity, as well as fits with the notion of robustness introduced
in Section 1.3. In the limiting case, a deterministic system devoid of any potential to
exhibit unexpected behavior in the QOI has no complexity, and thus its exponential
entropy is zero.
From (2.8), (2.13), and (2.14), we have that C(Y ) ≤ ν(Y ), with equality in the case
where Y is a uniform random variable. This relationship implies that of all probability
distributions of a given range, the uniform distribution is the one with the maximum
complexity. This result is consistent with the principle of maximum entropy, which
states that if only partial information is available about a system, the probability
distribution which best represents the current state of knowledge is the one with
largest information entropy [102, 103]. More specifically, if the only knowledge about
a system is the range [a, b] of permissible values, the largest differential entropy (and
hence the greatest exponential entropy) is achieved by assuming a uniform random
variable on [a, b] [104].
Another interpretation of exponential entropy is as the volume of the support
set Y, which is always nonnegative [88]. For an arbitrary continuous random variable Y , the complexity of the associated quantity of interest is equivalent to that
of a uniform distribution for which C(Y ) is the range. Figure 2-1 illustrates this
statement for a standard Gaussian random variable Y ∼ N (0, 1). In this case,
44
Y ~ N( = 0,  = 1)
Y ~ U[0, 4.13]
0.5
0.5
𝐶 𝑌 =
0.4
2𝜋𝑒𝜎 2
0.4
𝐶 𝑌 = 4.13
0.3
fY(y)
fY(y)
0.3
0.2
0.2

Range = 4.13
0.1
0.1
0
-3
-2
-1
0
1
2
0
-1
3
0
1
2
3
4
5
y
y
Figure 2-1: Probability density of a standard Gaussian random variable (left) and a
uniform random variable of equal complexity (right)
Y ~ N( = [0, 0],  = I )
Y ~ U([0, 4.13], [0, 4.13])
2
0.2
0.15
fY(y1, y2)
fY(y1, y2)
0.2
0.1
0.05
𝐶 𝑌 =
0
2
2
0
y
-2
2
0
-2
y
2𝜋𝑒
𝑛
Σ
0.15
Area = 17.08
0.1
0.05
0
4
𝐶 𝑌 = 17.08
𝐶 𝑌 = 4.13 2
4
2
0
y
1
2
2
0
y
1
Y ~ U[0, 17.08]
0.2
fY(y)
0.15
0.1
0.05
Range = 17.08
0
0
5
10
15
y
Figure 2-2: Probability density of a multivariate Gaussian random variable with
µ = (0, 0) and Σ = I2 (top left), and uniform random variables of equal complexity
in R2 and R (top right and bottom, respectively)
C(Y ) =
√
2πeσ 2 = 4.13, which suggests that Y is equivalent in complexity to a
uniform random variable with a range of 4.13. This interpretation also holds in
higher dimensions. For example, if Y is instead a multivariate Gaussian random vari-
45
able with mean µ ∈ Rd and covariance Σ ∈ Rd×d , then C(Y ) represents the volume
of a d-dimensional hypercube that captures the intrinsic extent of Y , and there is
a corresponding uniform random variable in Rd with the same complexity. Alternatively, C(Y ) can be also be represented as the range of a uniform random variable in
R. Figure 2-2 illustrates this example for a standard multivariate Gaussian random
variable with d = 2.
2.2.3
Numerical Complexity Estimation
In realistic complex systems, estimates of the QOI are often generated from the
outputs of numerous model evaluations, such as through Monte Carlo simulation.
The resulting distribution for Y can typically be of any arbitrary form; that is, no
closed-form expression is available for fY (y). This precludes computing C(Y ) through
the exact evaluation of h(Y ) using (2.3), requiring instead numerical integration [105].
Assuming that fY (y) is Riemann-integrable, h(Y ) can be estimated by discretizing
fY (y) into k bins of size ∆(l) , where l = 1, ..., k. Letting fY (y (l) ) denote the probability
density of Y evaluated in the lth bin, and h(Y ∆ ) represent the estimate of h(Y )
computed using numerical approximation, we have [88]:
∆
h(Y ) = −
=−
=−
k
X
l=1
k
X
l=1
k
X
fY (y (l) )∆(l) log[fY (y (l) )∆(l) ]
fY (y (l) )∆(l)
fY (y )∆ log
∆(l)
(l)
(l)
(l)
(l)
(l)
(2.15)
(l)
(2.16)
fY (y )∆ log[fY (y )∆ ] +
l=1
k
X
fY (y (l) )∆(l) log ∆(l) .
(2.17)
l=1
If we select bins of equal size (∆
(l)
= ∆ ∀ l), then
k
X
fY (y (l) )∆ ≈ 1, which gives:
l=1
∆
h(Y ) = −
k
X
fY (y (l) )∆ log[fY (y (l) )∆] + log ∆.
l=1
46
(2.18)
Note that h(Y ∆ ) → h(Y ) as ∆ → 0. Figure 2-3 provides an illustration for evaluating
h(Y ∆ ) using this discretization procedure.
𝑓𝑌 𝑦
𝑓𝑌 𝑦 𝑙
∆
𝑦
bin 𝑙
Figure 2-3: Numerical approximation of h(Y ) using k bins of equal size
2.3
Related Quantities
In this section, we examine the relationship between exponential entropy and several other quantities that can be used to describe uncertainty. As its connection to
differential entropy and range have already been addressed in previous sections, we
will focus on how exponential entropy is related to variance and entropy power. We
also introduce the Kullback-Leibler divergence, which will play a key role in entropy
power-based sensitivity analysis, to be discussed in Chapter 3.
Table 2.1 lists the expressions for several uncertainty metrics for the four types
of distributions introduced in Section 1.2.3. Table 2.2 shows how the various metrics
are affected by the transformations shown in Figures 1-4(a) and 1-4(b), where a QOI
Y is updated to a new estimate Y 0 through scaling or shifting by a constant amount.4
4
In Figure 1-4(a), fY (y) is scaled by a constant α with no change in µY . To represent this update
using Y 0 = αY requires that µY = 0. Instead, the action depicted in Figure 1-4(a) is more accurately
described by Y 0 = α(Y − µY ). However, the results presented in Table 2.2 for multiplicative scaling
by α hold in general, and do not restrict Y to be zero mean.
47
b−a
Uniform
Y ∼ U[a, b]
b−a
Triangular
Y ∼ T (a, b, c)
Infinite
Gaussian
Y ∼ N (µY , σY )
(bL − aL ) + (bR − aR )
Bimodal Uniform
Y ∼ BU(aL , bL , aR , bR , fL , fR )
Range
ν(Y )
fL
fR 2
2
(b2 − aL2 ) + (bR
− aR
)
2 L
2
a+b
2
fL
L
fR
2πe
−2fL (bL −a ) −2fR (bR −aR )
fL
−fL (bL −aL ) −fR (bR −aR )
fR
−fL (bL − aL ) log fL − fR (bR − aR ) log fR
µY
Mean
E [Y ]
(b − a)2
12
σY2
log(2πeσY2 )
2
q
2πeσY2
a+b+c
3
Variance
var (Y )
log(b − a)
a2 + b2 + c2 − ab − ac − bc
18
1
b−a
+ log
2
2
√
(b − a) e
2
fL
fR 3
3
(b3 − aL3 ) + (bR
− aR
) − (E [Y ])2
3 L
3
Diff. Entropy
h(Y )
b−a
σY2
Exp. Entropy
C(Y )
(b − a)2
2πe
(b − a)2
8π
Entropy Power
N (Y )
Table 2.1: Range, mean, variance, differential entropy, exponential entropy, and entropy power for distributions that are
uniform, triangular, Gaussian, or bimodal with uniform peaks
48
Constant Scaling
Y 0 = αY, α > 0
Constant Shift
Y 0 = β + Y, β ∈ R
Variance
var (Y 0 )
α2 var (Y )
var (Y )
Diff. Entropy
h(Y 0 )
h(Y ) + log α
h(Y )
Exp. Entropy
C(Y 0 )
αC(Y )
C(Y )
Entropy Power
N (Y 0 )
α2 N (Y )
N (Y )
Table 2.2: The properties of variance, differential entropy, exponential entropy, and
entropy power under constant scaling or shift
2.3.1
Variance
Variance is one of the most widely used metrics for characterizing uncertainty in
system design. It provides a measure of the dispersion of a random variable about
its mean value [106]. For a continuous random variable Y , variance is denoted by
var (Y ) or σY2 , and can be expressed in integral form as:
var (Y ) =
σY2
Z
=
ZY
=
(y − µY )2 fY (y) dy
(2.19)
y 2 fY (y) dy − µ2Y ,
(2.20)
Y
where µY denotes the mean (also known as the expected value or first moment) of Y ,
given by:
Z
E [Y ] = µY =
yfY (y) dy.
(2.21)
Y
There are several advantages to using variance for quantifying uncertainty. For
most distributions, variance can easily be estimated from sample realizations of the
random variable, without explicit knowledge of the probability density.5 Furthermore,
5
It is important to note, however, that variance is not always well-defined. For example, the
variance of a Cauchy distribution, which has heavy tails, is undefined. However, its differential
49
there is a rich body of literature in variance-based sensitivity analysis, which provides
rigorous methods to apportion a system’s output variability into contributions from
the inputs (see Section 3.1.1).
Although variance and differential entropy can both be used to describe a system’s
uncertainty or complexity, in general there is no explicit relationship between the two
quantities [107, 108]. Consider the example shown in Figure 2-4: let Y be a bimodal
random variable consisting of two symmetric uniform peaks, such that fY (y) = 0.5
for −2 ≤ y ≤ −1 and 1 ≤ y ≤ 2, and zero elsewhere. We define Y G to be its corresponding “equivalent Gaussian distribution” — that is, a Gaussian random variable
with the same mean and variance, parameterized by Y G ∼ N (µY , σY ).6 Intuitively,
we might expect that the complexity of Y is less than that of Y G , as the former can
take on values in one of only two groups, whereas the latter has infinite support. For
this case, var (Y ) = var (Y G ) whereas C(Y ) < C(Y G ), and thus exponential entropy
more accurately captures the limited range of possible outcomes in Y as compared to
Y G.
0.6
Probability Density
0.5
fY(y)
Y
YG
var (·)
2.33
2.33
h(·)
0.69
1.84
C(·)
2.00
6.31
0.4
0.3
0.2
fYG(y)
0.1
0
−4
−2
0
2
4
y
Figure 2-4: Comparison of variance, differential entropy, and complexity between a
bimodal uniform distribution and its equivalent Gaussian distribution
Based on exponential entropy, the bimodal uniform distribution has the same
complexity as a uniform random variable whose range is equal to two. This illustrates
another key feature of our proposed complexity metric — unlike variance, which
entropy is given by log(γ) + log(4π), where γ > 0 is the shape parameter, and thus its exponential
entropy is well-defined. This represents one potential advantage of our proposed complexity metric,
in that it can be estimated for any probability density using the procedure outlined in Section 2.2.3.
6
The notion of an equivalent Gaussian distribution will be explored in more detail in Section 3.4,
where it is used to characterize the non-Gaussianity of the factors and outputs of a model.
50
measures dispersion about the mean, exponential entropy reflects the intrinsic extent
of a random variable irrespective of the expected value.
2.3.2
Kullback-Leibler Divergence and Cross Entropy
Another quantity relevant to our discussion of complexity is the Kullback-Leibler
(K-L) divergence, which is a measure of the difference between two probability distributions [109]. For continuous random variables Y1 and Y2 , the K-L divergence from
Y1 to Y2 is defined as:
Z
∞
DKL (Y1 ||Y2 ) =
fY1 (y) log
Z−∞
∞
fY1 (y)
dy
fY2 (y)
(2.22)
Z
∞
fY1 (y) log fY1 (y) dy −
=
−∞
fY1 (y) log fY2 (y) dy
(2.23)
−∞
= −h(Y1 ) + h(Y1 , Y2 ).
(2.24)
As shown in (2.24), the K-L divergence is equal to the sum of two terms: the
negative of the differential entropy h(Y1 ) and the cross entropy between Y1 and Y2 ,
defined as:
Z
∞
h(Y1 , Y2 ) = −
fY1 (y) log fY2 (y) dy.
(2.25)
−∞
Note that neither K-L divergence nor cross entropy is a true metric as they are
not symmetric quantities — that is, in general DKL (Y1 ||Y2 ) 6= DKL (Y2 ||Y1 ) and
h(Y1 , Y2 ) 6= h(Y2 , Y1 ). Because of this, K-L divergence is also commonly referred to as
the relative entropy of one random variable with respect to another, which is always
non-negative.7 In the limiting case where Y1 = Y2 , we have that h(Y1 , Y2 ) = h(Y1 )
and h(Y2 , Y1 ) = h(Y2 ), and thus DKL (Y1 ||Y2 ) = DKL (Y2 ||Y1 ) = 0.
7
In (2.22), Y2 is designated as the reference random variable. If fY2 (y) = 1 over the support set
Y2 (e.g., a uniform random variable with an extent of one), then DKL (Y1 ||Y2 ) equals −h(Y1 ). This
lends some intuition about the statement that differential entropy is a measure of randomness with
respect to a fixed reference [88, 91].
51
2.3.3
Entropy Power
In his seminal work on information theory, Shannon proposed not only the use of
differential entropy h(Y ) to quantify uncertainty in a random variable, but also introduced a derived quantity N (Y ), which he termed entropy power [91]:
N (Y ) =
exp[2h(Y )]
.
2πe
(2.26)
Entropy power is so named because in the context of communication systems, it
represents the power in a white noise signal that is limited to the same band as the
original signal and having the same differential entropy. As Shannon writes, “since
white noise has the maximum entropy for a given power, the entropy power of any
noise is less than or equal to its actual power” [91]. Due to the 2πe normalizing
constant in (2.26), entropy power also has the interpretation of the “variance of a
Gaussian random variable with the same differential entropy, [which is] maximum
and equal to the variance when the random variable is Gaussian” [110].8 That is to
say, for an arbitrary random variable Y , N (Y ) ≤ var (Y ), with equality in the case
where Y is a Gaussian random variable (see Table 2.1).
Additionally, entropy power is always non-negative, and is proportional to the
square of exponential entropy. Rewriting (2.26), we obtain:
exp[h(Y )] exp[h(Y )]
2πe
C(Y )2
=
.
2πe
N (Y ) =
(2.27)
Whereas C(Y ) is a measure of intrinsic extent and has the same units as Y , N (Y )
exhibits properties of variance and has the same units as Y 2 (see Table 2.2). This
provides a very appealing analogy for the relationship between standard deviation
and variance to that between exponential entropy and entropy power.
8
This is consistent with the principle of maximum entropy, which states that if we known nothing
about a system’s behavior except for the first and second moments, the distribution that maximizes
differential entropy (and therefore entropy power) given our limited knowledge is the Gaussian [104].
52
2.4
Estimating Risk
Similar to complexity, the concept of risk in system design is also difficult to formalize. In engineering systems, the risk associated with a particular event is often
characterized by two features: the probability of the event’s occurrence (likelihood)
and the severity of the outcome should the event occur (consequence) [5, 111]. In
one common risk management approach, these attributes are multiplied together to
generate a quantitative measure of risk associated with the event, which is then used
to set design priorities accordingly [112].
In this thesis, we focus only on the first component of risk, namely, the probability
of a system incurring an undesirable outcome, such as a technical failure or the
exceedance of a cost constraint. Quantitatively, we compute risk as the probability of
failing to satisfy a particular criterion [9, 113]. Note that in this definition, risk is the
complement of reliability, which instead describes the probability of success. Suppose
that a QOI is subject to the requirement that it must be greater than or equal to a
specified value r. In this case, risk corresponds to the probability that the random
variable Y takes on a value less than r, given by:
Z
r
fY (y) dy.
P (Y < r) =
(2.28)
−∞
This notion of risk is illustrated in Figure 2-5, where P (Y < r) corresponds to
the area of the shaded region in which the requirement is not met. Conversely,
P (Y > r) = 1 − P (Y < r).
𝑓𝑌 𝑦
𝑃 𝑌<𝑟
𝑦
𝑟
Figure 2-5: Probability of failure associated with the requirement r
53
2.5
Chapter Summary
In this chapter, we introduced complexity and risk as possible measures of uncertainty in a system with respect to a quantity of interest, and provided definitions
for these terms. In particular, the use of exponential entropy as a complexity metric
was described in detail, and placed in the context of information-based measures of
uncertainty.
In the following chapter, we will explore the topic of sensitivity analysis, which
is used to explain how the various factors of a model contribute to uncertainty in
the output, and how this information can be leveraged to reduce uncertainty in the
context of improving robustness and reliability.
54
Chapter 3
Sensitivity Analysis
The focus of this chapter is sensitivity analysis in the context of engineering system
design. Within the framework of Figure 1-2, sensitivity analysis takes place after
we have quantified uncertainty using metrics such as variance, complexity, and risk.
Sensitivity analysis allows us to better understand the effects of uncertainty in a
system in order to make well-informed decisions aimed at uncertainty reduction. For
example, it can be used to study how “variation in the output of a model ... can be
apportioned, qualitatively or quantitatively, to different sources of variation, and how
the given model depends upon the information fed into it” [114]. This information
can help to answer questions such as “What are the key factors that contribute
to variability in model outputs?” and “On which factors should research aimed at
reducing output variability focus?” [8].
Section 3.1 provides a brief overview of the sensitivity analysis literature. New
contributions of this thesis are presented in Sections 3.2 through 3.4. Specifically,
Section 3.2 extends existing variance-based sensitivity analysis methods to permit
changes to distribution family and examine the effect of input distribution shape
on output variance. Section 3.3 details how to predict local changes in complexity
and risk due to perturbations in mean and standard deviation, which have direct
implications for system robustness and reliability. Section 3.4 describes a sensitivity
analysis approach based on entropy power, in which uncertainty is decomposed into
contributions due to the intrinsic extent and non-Gaussianity of each factor. These
55
sections set the stage for Chapter 4, which explores how sensitivity information can be
combined with design budgets to guide decisions regarding the allocation of resources
for uncertainty mitigation.
3.1
Background and Literature Review
There are several key types of sensitivity analysis: factor screening, which identifies
the influential factors in a particular system; global sensitivity analysis (GSA), which
apportions the uncertainty in a model output over the entire response range into
contributions from the input factors; distributional sensitivity analysis (DSA), which
extends GSA to consider scenarios in which only a portion of a factor’s variance is
reduced; and regional sensitivity analysis, which is similar to GSA and DSA but
focuses instead on a partial response region — for example, the tail region of a
distribution in which some design requirement is not satisfied. The sections below
discuss several methods that are most relevant to the current thesis: variance-based
GSA and DSA, and relative entropy-based sensitivity analysis.
3.1.1
Variance-Based Global Sensitivity Analysis
Global sensitivity analysis uses variance as a measure of uncertainty, and seeks to
apportion a model’s output variance into contributions from each of the factors, as
well as their interactions. This notion is illustrated in Figure 3-1.
Figure 3-1: Apportionment of output variance in GSA [8, Figure 3-1]
There are two main methods to perform GSA: the Fourier amplitude sensitivity
test (FAST), which uses Fourier series expansions of the output function to repre-
56
sent conditional variances [115–117], and the Sobol’ method, which utilizes Monte
Carlo simulation [118–121]. The Sobol’ method is based on high-dimensional model
representation (HDMR), which decomposes a function g(X) into the following sum:
X
g(X) = g0 +
gi (Xi ) +
i
X
gij (Xi , Xj ) + . . . + g12...m (X1 , X2 , . . . , Xm ),
(3.1)
i<j
where g0 is a constant, gi (Xi ) is a function of Xi only, gij (Xi , Xj ) is a function of Xi
and Xj only, etc. Although (3.1) is not a unique representation of g(X), it can be
made unique by enforcing the following constraints:
Z
1
gi1 ,...,is (Xi1 , . . . , Xis ) dxk = 0,
∀
0
k = i1 , . . . , is ,
(3.2)
s = 1, . . . , m.
In the above equation, the indices i1 , . . . , is represent all sets of s integers that satisfy
1 ≤ i1 < . . . < is ≤ m. That is, for s = 1, the constraint given by (3.2) applies to
all sub-functions gi (Xi ) in (3.1); for s = 2, the constraint applies to all sub-functions
gij (Xi , Xj ) with i < j in (3.1), etc. Furthermore, in (3.2) we have defined all factors
Xi1 , . . . , Xis on the interval [0, 1]; this is merely for simplicity of presentation, and
not a requirement of the Sobol’ method. With the sub-functions gi1 ,...,is (Xi1 , . . . , Xis )
now uniquely specified, we refer to (3.1) as the analysis of variances high-dimensional
model representation (ANOVA-HDMR) of g(X).
In the ANOVA-HDMR, all sub-functions gi1 ,...,is (Xi1 , . . . , Xis ) are orthogonal and
zero-mean, and g0 equals the mean value of g(X). Assuming that g(X) and therefore
all comprising sub-functions are square-integrable, the variance of g(X) is given by:
Z
D=
g(X)2 dx − g02 ,
(3.3)
and the partial variances of gi1 ,...,is (Xi1 , . . . , Xis ) are defined as:
Z
Di1 ,...,is =
gi1 ,...,is (Xi1 , . . . , Xis )2 dxi1 . . . dxis .
57
(3.4)
Squaring then integrating both sides of (3.1), we obtain:
D=
X
Di +
i
X
Dij + . . . + D12...m ,
(3.5)
i<j
which states that the variance of the original function equals the sum of the variances
of the sub-functions, as shown in Figure 3-1. We will return to this notion of uncertainty apportionment in Section 3.4.1, where the ANOVA-HDMR is used in deriving
the entropy power decomposition.
In the Sobol’ method, global sensitivity indices are defined as:
Si1 ,...,is =
Di1 ,...,is
,
D
s = 1, . . . , m.
(3.6)
By definition, the sum of all sensitivity indices for a given function is one. Two
particularly important types of sensitivity indices are the main effect sensitivity index
(MSI) and the total effect sensitivity index (TSI). For a system modeled by (1.3), the
MSI of the ith factor, denoted by Si , represents the expected relative reduction in the
variance of Y if the true value of Xi is learned (i.e., the variance of Xi is reduced to
zero):
Si =
var (Y ) − E [var (Y |Xi )]
.
var (Y )
(3.7)
Alternatively, Si can be expressed as the ratio of the variance of gi (Xi ) to the variance
of g(X):
Si =
Di
.
D
(3.8)
Whereas the MSI signifies the contribution to output variance due to Xi alone,
the total effect sensitivity index ST i denotes the contribution to output variance due
to Xi and its interactions with all other factors. Using the notation X∼i to denote
all factors except Xi , ST i represents the percent of output variance that remains
unexplained after the true values of all factors except Xi have been learned:
ST i =
var (Y ) − E [var (Y |X∼i )]
.
var (Y )
58
(3.9)
Equivalently, ST i equals the sum of all global sensitivity indices of the form specified
by (3.6) that contain the subscript i.
For a particular model,
Pm
i=1
Si ≤ 1, whereas
Pm
i=1
ST i ≥ 1; deviation from unity
reflects the magnitude of the interaction effects. The various factors of a model
can be ranked according to their MSI or TSI in a factor prioritization or factor
fixing setting [122]. The goal of factor prioritization is to determine the factors that,
once their true values are known, would result in the greatest expected reduction in
output variability; these correspond to the factors with the largest values of Si . In
factor fixing, the objective is to identify non-influential factors that can be fixed at a
given value without substantially affecting the output variance; in this case, the least
significant factors are those with the smallest values of ST i .
3.1.2
Vary-All-But-One Analysis
A related concept to Sobol’-based GSA is vary-all-but-one analysis (VABO), in which
two Monte Carlo simulations are conducted to determine a factor’s contribution to
output variability [123, 124]. In the first MC simulation, all factors are sampled
from their nominal distributions, whereas in the second, one factor is fixed at a
particular point on its domain while the other factors are allowed to vary. The
observed difference in output variance between the two simulations is then attributed
to the contribution of the fixed factor to output variability.
While the VABO method is simple to implement, its main limitation is that there is
no clear guidance as to where each factor ought to be fixed on its domain; depending
on the chosen location, the resulting apportionment of variance among the factors
can change [122]. By contrast, GSA circumvents this problem by accounting for all
possible values of each factor within its domain, such that the resulting sensitivity
index represents the expected reduction in variance, where the expectation is taken
over the distribution of values at which the factor can be fixed.
59
3.1.3
Distributional Sensitivity Analysis
Building on global sensitivity analysis methods, distributional sensitivity analysis
(DSA) can also be applied to parameters whose uncertainty is epistemic in nature.
The DSA procedure was developed by Allaire and Willcox in order to address one of
the inherent limitations of GSA [8, 125]: the assumption that all epistemic uncertainty
associated with a particular factor may be reduced to zero through further research
and improved knowledge. This generalization is optimistic, and can in fact lead
to inappropriate allocations of resources. Distributional sensitivity analysis, on the
other hand, treats the portion of a factor’s variance that can be reduced as a random
variable, and therefore may be more appropriate than GSA for the prioritization of
efforts aimed at uncertainty reduction, as it could convey for which input(s) directed
research will yield the greatest return.
Letting Xio be the random variable corresponding to the original distribution for
factor i, and Xi0 be the random variable whose distribution represents the uncertainty
in that factor after some design modification, the ratio of remaining (irreducible)
variance to initial variance for factor i is defined by the quantity δ:
δ=
var (Xi0 )
.
var (Xio )
(3.10)
When we do not know the true value of δ, we treat it as a uniform random
variable on the interval [0, 1]. Whereas GSA computes for each factor Xi the main
effect sensitivity index Si , the analogous quantity in DSA is adjSi (δ), the adjusted
main effect sensitivity index of Xi given that it is known that only 100(1 − δ)% of its
variance can be reduced.1 Additionally, the average adjusted main effect sensitivity
index, or AASi , is the expected value of adjSi (δ) over all δ ∈ [0, 1].
Another advantage of DSA is that it does not require additional MC simulations;
rather, it can be implemented by reusing samples from a previous Sobol’-based GSA.
Finally, the current DSA methodology assumes that a design activity that reduces
the variance of Xi by 100(1−δ)% does not change the distribution family to which Xi
1
The main effect sensitivity indices computed from GSA correspond to the case where δ = 0.
60
2
belongs. That is to say, if Xi is initially a Gaussian random variable with variance σX
,
i
its distribution remains Gaussian after the specified activity, with an updated variance
2
of δσX
. In Section 3.2, we extend the existing DSA methodology to permit changes
i
to a factor’s distribution type. This contribution will complement the discussion in
Section 3.4, which demonstrates that a distribution’s shape can play a significant role
in uncertainty estimation.
3.1.4
Relative Entropy-Based Sensitivity Analysis
In addition to variance-based methods, there has also been work done to develop
methods for global and regional sensitivity analysis using information entropy as a
measure of uncertainty. One method uses the Kullback-Leibler divergence, or relative
entropy, to quantify the distance between two probability distributions [126].
Suppose that Y o represents the original estimate of a system’s quantity of interest,
and Y 0 the modified result once some factor Xi has been fixed at a particular value
(e.g., its mean), as in a VABO approach. In relative entropy-based sensitivity analysis,
(2.22) is used to compute the K-L divergence between the two QOI estimates, which
serves to quantify the impact of the factor that has been fixed. That is to say, the
larger the value of DKL (Y o ||Y 0 ), the more substantial the contribution of factor i to
uncertainty in the QOI.
This approach can be used to perform both global and regional sensitivity analysis
by simply adjusting the limits of integration, and is suitable for distributions where
variance is not a good measure of uncertainty (e.g., bimodal or skewed distributions).
However, as in the VABO approach, the results of relative entropy-based sensitivity
analysis vary depending on where a particular factor is fixed on its domain. Furthermore, the K-L divergence is not normalized and does not have physical meaning —
it can be used to rank factors, but does not have a proportion interpretation such as
shown in Figure 3-1 for variance-based sensitivity indices.
61
3.2
Extending Distributional Sensitivity Analysis
to Incorporate Changes in Distribution Family
In the existing distributional sensitivity analysis methodology, it is assumed that when
a design activity is performed that reduces the variance of factor X by 100(1 − δ)%,
only the parameters of X’s distribution are altered, not the type of distribution itself.
If the activity were to also change the distribution family to which X belongs (e.g.,
from Gaussian to uniform), any additional fluctuations expected in the variance of
the QOI due to the switch would not be reflected in the computed adjusted main
effect sensitivity indices. Of this, Allaire writes:
“[If] the distribution family of a given factor was expected to change
through further research (e.g. from an original uniform distribution to a
distribution in the triangular family), then reasonable distributions from
the new family, given that the original distribution was from another
family, could be defined” [8, pp. 60].
Here, we extend DSA to permit changes in distribution family by defining reasonable new factor distributions, which represent the result of further research or additional knowledge about a factor [8]. These new distributions reflect a 100(1 − δ)%
reduction in the factor’s variance; however, for each value of δ ∈ [0, 1], there is usually
no unique distribution, and thus we average over k possible reasonable distributions
to estimate the adjusted main effect sensitivity index.2
As a note on terminology, we will use X o to denote the original random variable for
some factor, and X 0 to represent the updated random variable such that var (X 0 ) =
δvar (X o ), as in (3.10). Allaire’s DSA methodology uses acceptance/rejection (AR)
sampling to generate samples from a desired distribution by sampling from a different
distribution. In the current work, we also employ AR sampling to obtain samples of
X 0 (the desired distribution) from samples of X o ; in order to do this, there must
2
To illustrate this point, Allaire provides the example of a factor X that is uniformly distributed
√
on the√interval [0,√1]. For δ√= 0.5, there are infinitely many new distributions (e.g., U[0, 2/2],
U[1 − 2/2, 1], U[ 2/4, 1 − 2/4], etc.) for which the variance is 50% of the original variance.
62
exist a constant κ with the property that κfX o (x) ≥ fX 0 (x) for all x [105]. The AR
sampling procedure is outlined below; the steps may be repeated until sufficiently
many samples of X 0 have been obtained.
Acceptance/rejection method
1. Draw a sample, xo , from X o .
2. Draw a sample, u, from a uniform random variable on [0, 1].
3. If
fX 0 (xo )
fX o (xo )
≥ κu, let x0 = xo .
4. Otherwise, discard xo and return to Step 1.
In Sections 3.2.1 through 3.2.3, we propose procedures to select reasonable new
distributions and perform AR sampling for cases where X o and X 0 are uniform,
triangular, or Gaussian. For these cases, the analytical expressions for δ are listed
in Table 3.1. Note that the proposed procedures represent but one possible set of
methods to generate reasonable distributions for X 0 , and are not intended to be
definitive. Accordingly, the distributions obtained using the procedures are but a
subset of possible manifestations of X 0 , and thus should not be taken to be exhaustive.
In Appendix A, we will examine the corresponding cases in which X 0 is a bimodal
uniform random variable.
63
Xo
X0
Uniform
U[ao , bo ]
Uniform
U[a0 , b0 ]
2
b0 − a0
b o − ao
(b0 − a0 )2
12σ o2
2 02
(a
3
Triangular
T (a0 , b0 , c0 )
+ b02 + c02 − a0 b0 − a0 c0 − b0 c0 )
(bo − ao )2
a02 + b02 + c02 − a0 b0 − a0 c0 − b0 c0
18σ o2
Gaussian
N (µ0 , σ 0 )
12σ 02
(bo − ao )2
σ 02
σ o2
3 0
(b − a0 )2
Triangular
a02 + b02 + c02 − a0 b0 − a0 c0 − b0 c0
18σ 02
2
T [ao , bo , co ] ao2 + bo 2 + co2 − ao bo − ao co − bo co ao2 + bo 2 + co2 − ao bo − ao co − bo co ao2 + bo 2 + co2 − ao bo − ao co − bo co
Gaussian
N [µo , σ o ]
0
var (X )
for changes in factor distribution between the uniform, triangular, and Gaussian families.
Table 3.1: Ratio of δ = var
(X o )
Rows represent the initial distribution for X o and columns correspond to the updated distribution for X 0 .
64
3.2.1
Cases with No Change in Distribution Family
First, we focus on cases where X o and X 0 belong to the same family of distributions
(i.e., the diagonal entries of Table 3.1). In these cases, the extent of X 0 is always
less than or equal to that of X o , and thus AR sampling works well for generating
samples of X 0 from samples of X o . The procedures provided below are adapted from
Allaire [8]. Examples of reasonable distributions generated using the procedures are
Probability density
provided in Figure 3-2 for select values of δ and k.
3.5
7
3
6
2.5
5
2
4
1.5
3
1
2
0.5
1
0
−0.2
0
0.2
0.4
0.6
0.8
1
0
0
1.2
0.2
0.4
0.6
0.8
1
x
(a) Uniform to uniform: X o ∼ U[0, 1], δ = 0.1,
k=3
(b) Triangular to triangular: X o ∼ T (0, 1, 0.7),
δ = 0.5, k = 3
0.8
Probability density
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
−3
−2
−1
0
x
1
2
3
(c) Gaussian to Gaussian: X o ∼ N (0, 1),
δ = 0.3, k = 1
Figure 3-2: Examples of reasonable uniform, triangular, and Gaussian distributions.
In each figure, the solid red line denotes fX o (x), the dashed blue lines represent various
possibilities for fX 0 (x), and the dotted black lines show κfX o (x), which is greater than
or equal to the corresponding density fX 0 (x) for all x.
65
In the uniform to uniform and triangular to triangular cases, a key distinction
between the current work and the procedures proposed by Allaire is that the k distributions for X 0 are made to be evenly spaced on the interval [ao , bo ]; this is to
ensure the inclusion of a wide spectrum of possible distributions when estimating the
adjusted main effect sensitivity index.
Uniform to uniform: X o ∼ U[ao , bo ] → X 0 ∼ U[a0 , b0 ]
Selecting a0 and b0 :
1. From δ =
b0 −a0 2
,
bo −ao
solve for extent L = b0 − a0 .
2. Evenly space k samples of b0 on the interval [ao + L, bo ].
3. For each sample of b0 , find a0 = b0 − L.
Specifying κ:
1. Set κ =
bo −ao
.
b0 −a0
For triangular distributions, the variance is not solely a function of extent, but
√
also of the peak location. For a factor X o , the extent can vary between 18δσ o2 and
√
24δσ o2 . The former case occurs when the peak coincides with either the lower or
upper bound (i.e., co = ao or co = bo ), resulting in a skewed distribution. The latter
case corresponds to a symmetric triangular distribution in which the peak coincides
with the mean value (i.e., co = (ao + bo )/2).
Triangular to triangular: X o ∼ T (ao , bo , co ) → X 0 ∼ T (a0 , b0 , c0 )
Selecting a0 , b0 , and c0 :
1. Compute the minimum and maximum possible extent for X 0 : Lmin =
√
√
18δσ o2 and Lmax = 24δσ o2 .
2. Evenly space k samples of L on the interval [Lmin , Lmax ].
3. For each sample of L, randomly sample midpoint M from the interval [ao , bo ].
• Find a0 = M −
L
2
and b0 = M + L2 .
66
• If a0 ≥ ao and b0 ≤ bo , accept a0 and b0 . Otherwise, repeat Step 3.
4. From δ =
a02 +b02 +c02 −a0 b0 −a0 c0 −b0 c0
,
ao2 +bo 2 +co2 −ao bo −ao co −bo co
solve for c0 . Select the root that is closer
to co (this is to circumvent the problem of sample impoverishment towards
the tails of the distribution).
Specifying κ:
1. Using (1.6), evaluate fX 0 (c0 ) and fX o (c0 ).
2. Set κ =
fX 0 (c0 )
.
fX o (c0 )
Gaussian to Gaussian: X o ∼ N (µo , σ o ) → X 0 ∼ N (µ0 , σ 0 )
Selecting µ0 and σ 0 :
1. Set µ0 = µo .
√
2. Set σ 0 = δσ o2 .
Specifying κ:
1. Using (1.7), evaluate fX 0 (c0 ) and fX o (c0 ).
2. Set κ =
3.2.2
fX 0 (µ0 )
.
fX o (µ0 )
Cases Involving Truncated Extent for X o
In this section, we consider three cases in which it is not always feasible to generate
samples of X 0 from the entire distribution for X o . These cases are triangular to
uniform, Gaussian to uniform, and Gaussian to triangular, which form the lower left
entries of Table 3.1. In the first case, the probability density of a triangular random
variable X o equals zero at ao and bo ; thus, there are very few samples of X o in the
tails of the distribution, making it nearly impossible to generate samples of X 0 from
those regions using the AR method. Similarly, in the second and third cases, X o is
Gaussian and has infinite extent; however, sample impoverishment in the tails again
limits the feasibility of the AR method to generate sufficiently many samples of X 0 .
67
To circumvent this problem, we instead use the AR method to generate samples
of X 0 from a truncated approximation for X o . We specify lower and upper bounds
(BL and BU , respectively) for the interval on which X 0 can be supported, such that
the vast majority of the probability mass of X o is contained in [BL , BU ].
The lower and upper bounds can be computed for a desired level using the cumulative distribution function FX o (x), which is equal to the area under the probability
density fX o (x) from −∞ to x:
FX o (x) = P (X o ≤ x)
Z x
=
fX o (ξ) dξ.
(3.11)
(3.12)
−∞
As an example, to specify the bounds [BL , BU ] that delimit the middle 95% of the
probability mass of X o , we can simply solve (3.12) for the values of x corresponding
to FX o (x) = 0.025 and FX o (x) = 0.975, respectively. The analytical expressions for
the cumulative distribution function of triangular and Gaussian random variables are
given in (3.13) and (3.14), respectively:
X o ∼ T (ao , bo , co ) :
o
o
o
X ∼ N (µ , σ ) :

(x − a)2




(b − a)(c − a)




(b − x)2
FX o (x) = 1 −

(b − a)(b − c)






0
if a ≤ x ≤ c,
if c < x ≤ b,
(3.13)
otherwise.
x − µo
1
FX o (x) =
1 + erf √
.
2
2σ o2
(3.14)
Empirical evidence suggests that reasonable bounds for the three cases in this
section (and the percentage of the area under fX o (x) that is captured), are:
• Triangular to uniform: BL = FX−1o (0.025), BU = FX−1o (0.975); middle 95%.
• Gaussian to uniform: [BL , BU ] = µo ∓ 2σ o ; middle 95.4%.
• Gaussian to triangular: [BL , BU ] = µo ∓ 3σ o ; middle 99.7%.
68
The procedures to estimate the parameters of X 0 and specify κ for AR sampling are
listed below. Examples of reasonable distributions for these three cases are shown in
12
3
10
2.5
Probability density
Probability density
Figure 3-3 for select values of δ and k.
8
6
4
2
0
0
2
1.5
1
0.5
0.2
0.4
0.6
0.8
0
−2
1
−1.5
−1
−0.5
x
(a) Triangular to uniform: X o ∼ T (0, 1, 0.3),
δ = 0.5, k = 2
0
x
0.5
1
1.5
2
(b) Gaussian to uniform: X o ∼ U[0, 1], δ = 0.7,
k=2
1.4
Probability density
1.2
1
0.8
0.6
0.4
0.2
0
−3
−2
−1
0
x
1
2
3
(c) Gaussian to triangular: X o ∼ N (0, 1),
δ = 0.6, k = 2
Figure 3-3: Examples of reasonable uniform, triangular, and Gaussian distributions
for cases involving a truncated extent for X o
69
Triangular to uniform: X o ∼ T (ao , bo , co ) → X 0 ∼ U[a0 , b0 ]
Selecting a0 and b0 :
1. Using (3.13), find bounds BL = FX−1o (0.025) and BU = FX−1o (0.975).
2. From δ =
3(b0 −a0 )2
,
2(ao2 +bo 2 +co2 −ao bo −ao co −bo co )
solve for extent L = b0 − a0 .
3. Evenly space k samples of b0 on the interval [BL + L, BU ].
4. For each sample of b0 , find a0 = b0 − L.
Specifying κ:
1. Using (1.6), evaluate fX o (a0 ) and fX o (b0 ).
2. If fX o (a0 ) ≤ fX o (b0 ), set κ =
1
.
fX o (a0 )(b0 −a0 )
Otherwise, set κ =
1
.
fX o (b0 )(b0 −a0 )
Gaussian to uniform: X o ∼ N (µo , σ o ) → X 0 ∼ U[a0 , b0 ]
Selecting a0 and b0 :
1. Set lower bound BL = µo − 2σ o and upper bound BU = µo + 2σ o .
2. From δ =
(b0 −a0 )2
,
12σ o2
solve for extent L = b0 − a0 .
3. Evenly space k samples of b0 on the interval [BL + L, BU ].
4. For each sample of b0 , find a0 = b0 − L.
Specifying κ:
1. Using (1.7), evaluate fX o (a0 ) and fX o (b0 ).
2. If fX o (a0 ) ≤ fX o (b0 ), set κ =
1
.
fX o (a0 )(b0 −a0 )
Otherwise, set κ =
1
.
fX o (b0 )(b0 −a0 )
Gaussian to triangular: X o ∼ N (µo , σ o ) → X 0 ∼ T (a0 , b0 , c0 )
Selecting a0 , b0 , and c0 :
1. Compute the minimum and maximum possible extent for X 0 : Lmin =
√
√
18δσ o2 and Lmax = 24δσ o2 .
70
2. Evenly space k samples of L on the interval [Lmin , Lmax ].
3. Set lower bound BL = µo − 3σ o and upper bound BU = µo + 3σ o .
4. For each sample of L, randomly sample midpoint M from the interval
[BL , BU ].
• Find a0 = M −
L
2
and b0 = M + L2 .
• If a0 ≥ BL and b0 ≤ BU , accept a0 and b0 . Otherwise, repeat Step 4.
5. From δ =
a02 +b02 +c02 −a0 b0 −a0 c0 −b0 c0
,
18σ o2
solve for c0 . Select the root that is closer to
µo .
Specifying κ:
1. If a0 6= c0 , find αL and a0 < xL < c0 such that αL fX o (x) and fX 0 (x) are
tangent at xL . Otherwise, set xL = a0 .
2. If b0 6= c0 , find αR and c0 < xR < b0 such that αR fX o (x) and fX 0 (x) are
tangent at xR . Otherwise, set xR = b0 .
3. Using (1.6) and (1.7), find κL =
fX 0 (xL )
,
fX o (xL )
κR =
fX 0 (xR )
,
fX o (xR )
and κC =
fX 0 (c0 )
.
fX o (c0 )
4. Set κ = max{κL , κR , κC }.
3.2.3
Cases Involving Resampling
Finally, we consider the uniform to triangular, uniform to Gaussian, and triangular
to Gaussian cases, which comprise the upper right corner of Table 3.1. In these cases,
AR sampling may not be a valid method to generate samples of X 0 from the existing
samples of X o . This is because for larger values of δ, although var (X 0 ) < var (X o ),
the extent of X 0 can actually exceed that of X o . For example, in Figure 3-4(a), X o is a
uniform random variable with variance σ o2 ; two possibilities for X 0 are shown, where
X 0 is a triangular random variable and δ = 0.9. In both cases the new extent b0 − a0
exceeds original extent bo − ao , making it impossible to generate samples in the tail
regions of X 0 through AR sampling. Therefore, in order to compute the adjusted main
71
effect sensitivity index for δ = 0.9, we must perform a new Monte Carlo simulation to
generate samples of X 0 and propagate them through the appropriate analysis models
to obtain estimates of Y .
Recall that the minimum extent for a triangular distribution is Lmin =
√
18δσ o2 ;
for Lmin to be wholly encompassed within the bounds of a uniform distribution on
[ao , bo ] requires that δ ≤ 2/3. Thus, in the uniform to triangular case, AR sampling
is feasible for δ ≤ 2/3; otherwise, resampling is necessary.
For the uniform to Gaussian and triangular to Gaussian cases, the extent necessarily increases between X o and X 0 due the Gaussian distribution having infinite
support. Here, we employ the same strategy as in Section 3.2.2 by using a truncated
distribution to approximate X 0 and ensuring that it lies entirely within the interval
[ao , bo ]. By requiring that ao ≤ µ0 − 3σ 0 and bo ≥ µ0 + 3σ 0 , this approximation is
sufficient to capture more than 99.7% of the probability mass of X 0 . This condition
is satisifed for δ ≤ 1/3 in the uniform to Gaussian case; in the triangular to Gaussian
case, the corresponding threshold varies between 1/2 and 2/3, depending on the skewness of the triangular distribution for X o . If δ exceeds the specified value, resampling
must be used in lieu of the AR method to generate samples of X 0 .
For the three cases in this section, the procedures for specifying the parameters
of X 0 and generating samples from the distribution are provided below. Examples of
reasonable distributions for these cases are shown in Figure 3-4 for select values of δ
and k.
72
3
2.5
2.5
Probability density
Probability density
3
2
1.5
1
1.5
1
0.5
0.5
0
−0.2
2
0
0.2
0.4
0.6
0.8
1
0
−0.2
1.2
0
0.2
0.4
x
0.6
0.8
1
1.2
x
3
3
2.5
2.5
Probability density
Probability density
(a) Uniform to triangular: X o ∼ U[0, 1], δ = 0.9, (b) Uniform to triangular: X o ∼ U[0, 1], δ = 0.4,
k=2
k=3
2
1.5
1
1.5
1
0.5
0.5
0
−0.2
2
0
0.2
0.4
0.6
0.8
1
0
−0.2
1.2
0
0.2
0.4
x
0.6
0.8
1
1.2
x
6
6
5
5
Probability density
Probability density
(c) Uniform to Gaussian: X o ∼ U[0, 1], δ = 0.7, (d) Uniform to Gaussian: X o ∼ U[0, 1], δ = 0.3,
k=1
k=1
4
3
2
3
2
1
1
0
0
4
0.2
0.4
0.6
0.8
0
0
1
x
(e) Triangular to Gaussian: X o ∼ T (0, 1, 0.8),
δ = 0.5, k = 1
0.2
0.4
0.6
0.8
1
x
(f) Triangular to Gaussian: X o ∼ T (0, 1, 0.8),
δ = 0.2, k = 1
Figure 3-4: Examples of reasonable uniform, triangular, and Gaussian distributions
for cases involving resampling. The left column shows instances where δ is too large,
and resampling is required. The right column shows instances where δ is sufficiently
small, and AR sampling can be used to generate samples of X 0 from samples of X o .
73
Uniform to triangular: X o ∼ U[ao , bo ] → X 0 ∼ T (a0 , b0 , c0 )
Selecting a0 , b0 , and c0 :
1. Compute Lmin =
√
√
18δσ o2 and Lmax = 24δσ o2 .
2. Sample L and set lower and upper bounds for sampling M .
• If δ ≤ 2/3:
– Evenly space k samples of L on the interval [Lmin , min{Lmax , bo −
ao }].
– Set lower bound BL = ao and upper bound BU = bo .
• If δ > 2/3:
– Evenly space k samples of L on the interval [Lmin , Lmax ].
– Set lower bound BL = bo − Lmax and upper bound BU = ao + Lmax .
3. For each sample of L, randomly sample midpoint M from the interval
[BL , BU ].
• If L = BU − BL , set M =
• Find a0 = M −
L
2
BL +BU
.
2
and b0 = M + L2 .
• If a0 ≥ BL and b0 ≤ BU , accept a0 and b0 . Otherwise, repeat Step 3.
4. From δ =
2(a02 +b02 +c02 −a0 b0 −a0 c0 −b0 c0 )
,
3(bo −ao )2
solve for c0 . Select either root at random.
Specifying κ:
• If δ ≤ 2/3, set κ = (bo −ao )fX 0 (c0 ) and use AR sampling to generate samples
of X 0 ∼ T (a0 , b0 , c0 ) from samples of X o ∼ U[ao , bo ].
• If δ > 2/3, use MC sampling to generate samples of X 0 ∼ T (a0 , b0 , c0 ).
74
Uniform to Gaussian: X o ∼ U[ao , bo ] → X 0 ∼ N (µ0 , σ 0 )
Selecting µ0 and σ 0 :
1. Set µ0 =
0
2. Set σ =
ao +bo
.
2
q
δ(bo −ao )2
.
12
Specifying κ:
• If δ ≤ 1/3, set κ = (bo −ao )fX 0 (µ0 ) and use AR sampling to generate samples
of X 0 ∼ N (µ0 , σ 0 ) from samples of X o ∼ U[ao , bo ].
• If δ > 1/3, use MC sampling to generate samples of X 0 ∼ N (µ0 , σ 0 ).
Triangular to Gaussian: X o ∼ T [ao , bo , co ] → X 0 ∼ N (µ0 , σ 0 )
Selecting µ0 and σ 0 :
1. Set µ0 =
2. Set σ 0 =
ao +bo +co
.
3
q
δ(ao2 +bo 2 +co2 −ao bo −ao co −bo co )
.
18
Specifying κ:
• If ao ≤ µ0 − 3σ 0 and bo ≥ µ0 + 3σ 0 :
1. Using (1.6) and (1.7), find x∗ and α such that αfX o (x) and fX 0 (x) are
tangent at x∗ .
2. Set κ =
fX 0 (x∗ )
fX o (x∗ )
and use AR sampling to generate samples of X 0 ∼
N (µ0 , σ 0 ) from samples of X o ∼ T [ao , bo , co ].
• Otherwise, use MC sampling to generate samples of X 0 ∼ N (µ0 , σ 0 ).
3.3
Local Sensitivity Analysis
In this section, we explore how complexity and risk estimates are affected by small
changes in the mean or standard deviation of a system’s design parameters or QOI.
Revisiting Figures 1-4(a) and 1-4(b), we consider the separate cases of a constant
75
scaling of fY (y) that shrinks the standard deviation versus a constant shift in fY (y)
that alters the mean value. These design activities have direct implications for system
robustness and reliability. The results derived herein represent local sensitivities
obtained by linearizing the system about the current estimate.
3.3.1
Local Sensitivity to QOI Mean and Standard Deviation
To evaluate the local sensitivity of complexity and risk with respect to design activities
that perturb the mean (µY ) or standard deviation (σY ) of the QOI Y , we compute
the partial derivative of C(Y ) and P (Y < r) with respect to µY and σY . The partial
derivatives can then be linearized about the current design to predict the change
in complexity (∆C) or risk (∆P ) associated with a perturbation in mean (∆µY ) or
standard deviation (∆σY ). These expressions are derived in Appendix B; the key
results are summarized in Table 3.2.
Sensitivity of C(Y ) with respect to µY and σY
Partial derivatives
∂C(Y )
=0
∂µY
∂C(Y )
C(Y )
=
∂σY
σY
(3.15)
(3.16)
Local approximations
∆C(Y ) = 0
(3.17)
∆C ≈
∆σY
C(Y )
σY
(3.18)
Sensitivity of P (Y < r) with respect to µY and σY
Partial derivatives
∂P (Y < r)
= −fY (r)
∂µY
∂P (Y < r)
(µY − r)
=
fY (r)
∂σY
σY
(3.19)
(3.20)
Local approximations
∆P ≈ −∆µY fY (r)
(3.21)
∆P ≈
∆σY
(µY − r)fY (r)
σY
(3.22)
Table 3.2: Local sensitivity analysis results for complexity and risk
From (3.15), we see that perturbations in µY do not affect complexity. This is
76
unsurprising, as we know that C(Y ) contains no information about the expected value
of Y . From (3.16), we see that the sensitivity of complexity to σY is a constant, and
simply equals the ratio of C(Y ) to σY . Since both C(Y ) and σY are non-negative,
this implies that as a local approximation (see (3.18)), reducing standard deviation
also reduces complexity.
More interestingly, (3.19) and (3.20) show that the sensitivity of risk to both µY
and σY is proportional to the probability density of Y evaluated at the requirement
r. Since fY (y) ≥ 0 for all values of y, this implies that the sign of ∆µY or µY − r
determines whether risk is increased or decreased (see (3.21) and (3.22)). Figure 3-5
helps to illustrate this point for (3.22), where a reduction in standard deviation can
alter risk in either direction, depending on the relative
locations of r and µY .
Refinement: Change Std Dev
Refinement: Change Std Dev
0.8
𝑓𝑌 𝑦
0.6
f X(x)
f X(x)
0.6
0.4
𝑃 𝑌<𝑟
0.2
0
-0.2
-4
𝑓𝑌 𝑦
0.8
-2
𝑃 𝑌<𝑟
0.2
𝑦0
𝑟 𝜇𝑌
-3
0.4
(a)
µY −
r > 10
-1
0
2
3
𝑦
𝜇𝑌 𝑟
-0.2
-4
4
x
-3
-2
(b) µY − r < 0
-1
0
1
2
3
4
x
Figure 3-5: The relative locations of r and µY greatly impact the change in risk
associated with a decrease in σY . Moving from the red probability density to the
blue, P (Y < r) decreases if µY − r > 0, and increases if µY − r < 0.
3.3.2
Relationship to Variance-Based Sensitivity Indices
In this section, we seek to relate the local approximations in (3.18) and (3.22) to
sensitivity indices computed from variance-based GSA and DSA. We let Y o and Y 0
represent initial and new estimates of the QOI (the red and blue densities in Figure
3-5, respectively) corresponding to a design activity that reduces variance (and thus
standard deviation). Similarly, we let var (Y o ) and var (Y 0 ) represent the variance of
77
the QOI before and after the update, respectively. If the activity is one that results
in learning the true value of factor i, then we know from (3.7) that:
var (Y 0 ) = var (Y o ) − Si var (Y o ).
(3.23)
The ratio of var (Y 0 ) to var (Y o ) is given by:
var (Y 0 )
= 1 − Si .
var (Y o )
(3.24)
We can relate the above ratio to the quantity ∆σY /σY from (3.18) and (3.22):
p
p
var (Y 0 ) − var (Y o )
p
var (Y o )
s
var (Y 0 )
−1
=
var (Y o )
p
= 1 − Si − 1.
∆σY
=
σY
(3.25)
This allows us to rewrite (3.18) and (3.22) as:
p
∆C ≈ ( 1 − Si − 1)C(Y ),
p
∆P ≈ ( 1 − Si − 1)(µY − r)fY (r),
(3.26)
(3.27)
noting that C(Y ), µY , and fY (r) in the above expressions refer to the complexity,
mean, and probability density (evaluated at r) of the QOI for the initial design.
Finally, if the design activity is instead one that reduces the variance of factor i by
100(1 − δ)%, then (3.23) can be modified to:
var (Y 0 ) = var (Y o ) − adjSi (δ)var (Y o ),
(3.28)
and we can similarly substitute adjSi (δ) for Si in (3.26) and (3.27).
Since both Si and adjSi (δ) can only assume values in the interval [0, 1], the above
local approximations imply that a decrease in variance is concurrent with a reduction
78
in complexity, as well as in risk if µY − r > 0. While an attractive result, these
relations are simply local approximations and should not be generalized. As we will
demonstrate in Chapter 4, variance, complexity, and risk each describes a different
aspect of system uncertainty, and can therefore trend in opposite directions.
3.3.3
Local Sensitivity to Design Parameter Mean and Standard Deviation
In order to use sensitivity information about the QOI to update the design, it is
necessary to translate this information into tangible actions for modifying the design
parameters. For this, we extend (3.19) and (3.20) to compute the sensitivity of risk
to perturbations in the mean or standard deviation of the design parameters.
We define the vectors µX = [µX1 , µX2 , . . . , µXm ]T and σX = [σX1 , σX2 , . . . , σXm ]T ,
which consist of the mean and standard deviation estimates of the entries of X. Using
the chain rule, the partial derivative of P (Y < r) with respect to µXi is given by:
∂P (Y < r) ∂µY
∂P (Y < r)
=
.
∂µXi
∂µY
∂µXi
(3.29)
To estimate ∂µY /∂µXi , we use (1.4) and make the following approximation for µY :
µY = E [g(X)] ≈ g(E [X]) = g(µX ).
(3.30)
This approximation is exact if g(X) is linear. Therefore, we have:
∂µY
∂g(µX )
≈
.
∂µXi
∂Xi
(3.31)
Combining (3.19), (3.29), and (3.31), we obtain:
∂P (Y < r)
∂g(µX )
≈ −fY (r)
,
∂µXi
∂Xi
∂g(µX )
∆P ≈ −∆µXi fY (r)
.
∂Xi
79
(3.32)
(3.33)
Similarly, the partial derivative of P (Y < r) with respect to σXi is given by:
∂P (Y < r)
∂P (Y < r) ∂σY
=
.
∂σXi
∂σY
∂σXi
(3.34)
To approximate ∂σY /∂σXi , we can use results from DSA to estimate ∆σY /∆σXi . The
adjusted main effect sensitivity index adjSi (δ) relates the expected variance remaining
in Y (given by (3.28)) to the variance remaining in factor i (given by (3.10)) as δ
ranges between 0 and 1, which allows us to compute ∆σY /∆σXi as follows:
p
p
var (Y o ) − adjSi (δ)var (Y o ) − var (Y o )
p
p
δvar (Xio ) − var (Xio )
p
σY ( 1 − adjSi (δ) − 1)
√
=
.
σXi ( δ − 1)
∆σY
=
∆σXi
(3.35)
Thus, we can combine (3.20), (3.34) and (3.35) to obtain:3
p
fY (r)(µY − r)( 1 − adjSi (δ) − 1)
∂P (Y < r)
√
≈
,
∂σXi
σXi ( δ − 1)
p
∆σXi fY (r)(µY − r)( 1 − adjSi (δ) − 1)
√
∆P ≈
.
σXi ( δ − 1)
(3.36)
(3.37)
The key result of this section is that we can relate system risk to the specific
changes in the design parameters. As an example, let ∆P denote the desired change
in risk. Rearranging (3.33) and (3.37), we can estimate the requisite change in the
mean and standard deviation of each design parameter in order to achieve that goal.
For small values of ∆P , these changes are approximated by:
∆µXi
∆σXi
−1
∆P ∂g(µX )
≈−
,
fY (r)
∂Xi
√
∆P σXi ( δ − 1)
p
≈
.
fY (r)(µY − r)( 1 − adjSi (δ) − 1)
3
(3.38)
(3.39)
While (3.35) is a valid statement of ∆σY /∆σXi for any value of δ ∈ [0, 1), the choice of δ can
greatly affect the accuracy of the local approximation ∂σY /∂σXi ≈ ∆σY /∆σXi , especially for small
values of δ (corresponding to large reductions in factor variance), or if the relationship between
var (Y 0 ) and var (Xi0 ) is highly nonlinear.
80
The above expressions allow the designer to obtain a first-order estimate of the parameter adjustments needed to achieve a desired decrease in risk. They can also highlight different trends, tradeoffs, and design tensions present in the system. However,
we note again that these relations are merely local approximations whose predictive
accuracy cannot be guaranteed. Furthermore, they do not account for interactions
among the design parameters, nor do they imply that the suggested changes in mean
or standard deviation are necessarily feasible, as limitations due to physical or budgetary constraints are not accounted for. It is the responsibility of the designer to
use these tools as a guideline for cost-benefit analysis of various design activities, and
ultimately select the most appropriate action for risk mitigation.
3.4
Entropy Power-Based Sensitivity Analysis
In this section, we extend the ideas of variance-based GSA to an analogous methodology for complexity-based GSA. For this, we use entropy power as the basis for our
sensitivity analysis approach, recalling from (2.27) that entropy power is proportional
to the square of exponential entropy.
3.4.1
Entropy Power Decomposition
We rewrite the ANOVA-HDMR presented in (3.1) using the random variables Y and
Zi , Zij , . . . , Z12...m to denote the outputs of the original function and various subfunctions, respectively. That is, we let Y = g(X) and Zi1 ,...,is = gi1 ,...,is (Xi1 , . . . , Xis ),
where the indices i1 , . . . , is represent all sets of s integers that satisfy 1 ≤ i1 < . . . <
is ≤ m, as described in Section 3.1.1. The ANOVA-HDMR can thus be expressed as:
Y = g0 +
X
i
Zi +
X
Zij + . . . + Z12...m .
(3.40)
i<j
Since all sub-functions in (3.1) are orthogonal and zero-mean, Zi , Zij , . . . , Z12...m
(henceforth termed “auxiliary random variables”) are all uncorrelated. According
to (3.5), the variance of Y is the sum of the partial variances of Z1 through Zm and
81
their higher-order interactions:
var (Y ) =
X
var (Zi ) +
X
i
var (Zij ) + . . . + var (Z12...m ).
(3.41)
i<j
Normalizing by var (Y ), the proportional contribution to var (Y ) from each of the
auxiliary random variables is given by:
1=
X var (Zi )
i
var (Y )
+
X var (Zij )
i<j
var (Y )
+ ... +
var (Z12...m )
.
var (Y )
(3.42)
We desire to derive a similar relationship for N (Y ), the entropy power of the
QOI. As discussed in Section 2.3.3, although there is no general relationship between
entropy power and variance, for a given random variable, entropy power is always
less than or equal to variance, with equality in the case where the random variable is
Gaussian (see Table 2.1). Therefore, we seek to relate the disparity between variance
and entropy power to the random variable’s degree of non-Gaussianity, quantified
using the K-L divergence. For this, we define an “equivalent Gaussian distribution,”
which is a Gaussian random variable with the same mean and variance as the original
random variable. For an arbitrary random variable W , we denote its equivalent
Gaussian distribution using the superscript G, and define it as W G ∼ N (µW , σW ).
Using (2.22), we compute the K-L divergence between each random variable in
(3.40) and its equivalent Gaussian distribution to obtain DKL (Y ||Y G ), DKL (Z1 ||Z1G ),
DKL (Z2 ||Z2G ), etc. In Appendix C, we derive the following sum, which relates the
entropy power and non-Gaussianity of the QOI to the corresponding quantities for
the auxiliary random variables:
N (Y )exp[2DKL (Y ||Y G )] =
X
N (Zi ) exp[2DKL (Zi ||ZiG )]
i
+
X
N (Zij ) exp[2DKL (Zij ||ZijG )] + . . .
i<j
G
+ N (Z12...m ) exp[2DKL (Z12...m ||Z12...m
)].
82
(3.43)
The basis for (3.43) is that for any arbitrary random variable W , var (W ) =
N (W ) exp[2DKL (W ||W G )].4 This implies that the variance of each of Y and Zi , Zij ,
. . . , Z12...m is equal to the product of the variable’s entropy power and the exponential
term, where the argument of the exponential is two times the K-L divergence of the
variable with respect to its equivalent Gaussian distribution. Thus, (3.41) and (3.43)
are equivalent.
Note that in the derivation we have not made any assumptions about the underlying distributions of Y or the auxiliary random variables. The only requirement is
that all the auxiliary random variables must be uncorrelated, which is automatically
satisfied by the ANOVA-HDMR. We emphasize the distinction between uncorrelated
random variables Zi and Zj (i 6= j), for which the variance of the sum is equal to
the sum of the variances (i.e., var (Zi + Zj ) = var (Zi ) + var (Zj )), and independent
random variables Xi and Xj , for which the joint probability density fXi ,Xj (xi , xj ) is
equal to the product of the marginal densities fXi (xi ) and fXj (xj ). While independent
implies uncorrelated, the converse is not true. Recall that in Section 1.2 we made the
assumption that all design parameters X1 , . . . , Xm are independent. On the other
hand, the auxiliary random variables Z1 , Z2 , . . . , Z12...m are uncorrelated. While the
auxiliary random variables representing main effects (e.g., Z1 , Z2 , . . . , Zm ) are also
mutually independent, those that correspond to interaction effects are functions of
two or more design parameters, and thus are necessarily dependent. For example,
while Zi and Zj , i 6= j, are independent, Zij is dependent on both Zi and Zj , but all
three variables are uncorrelated.
Having obtained analogous expressions for the decomposition of output variance
and entropy power into comprising terms, we can normalize (3.43) to establish a
similar interpretation in terms of proportional contributions from the input factors
4
This result is consistent with the principle of maximum entropy, which states that for a given
mean and variance, differential entropy (and therefore entropy power) is maximized with a Gaussian
distribution. Since the K-L divergence is a non-negative quantity, exp[2DKL (W ||W G )] ≥ 1, which
implies that var (W ) ≥ N (W ). In the limiting case that W is itself a Gaussian random variable,
DKL (W ||W G ) = 0, exp[2DKL (W ||W G )] = 1, and var (W ) = N (W ), as required.
83
and their interactions:
1=
X N (Zi ) exp[2DKL (Zi ||Z G )]
i
N (Y ) exp[2DKL (Y ||Y
+ ... +
i
G )]
+
X N (Zij ) exp[2DKL (Zij ||ZijG )]
i<j
N (Y ) exp[2DKL (Y ||Y G )]
G
)]
N (Z12...m ) exp[2DKL (Z12...m ||Z12...m
.
G
N (Y )
exp[2DKL (Y ||Y )]
(3.44)
In (3.44), the proportion of output uncertainty directly due to Xi consists of two
parts: one that is the ratio of the entropy power of Zi to that of Y , and one that is the
ratio of the exponential of twice the K-L divergence from Zi to ZiG to the analogous
quantity for Y . Recalling that exponential entropy — proportional to the square root
of entropy power — measures the intrinsic extent of a random variable, we conclude
that the first ratio is directly influenced by the intrinsic extent of Xi (and thus Zi ).
On the other hand, the second ratio is directly related to the non-Gaussianity of Xi
(and thus Zi ). To illustrate this, we introduce the entropy power sensitivity indices
ηi and ζi , defined as:
N (Zi )
,
N (Y )
exp[2DKL (Zi ||ZiG )]
ζi =
.
exp[2DKL (Y ||Y G )]
ηi =
(3.45)
(3.46)
The above expressions correspond to the main effect indices for factor i; analogous
expressions can also be derived for the higher-order interactions. Due to the equivalence of (3.42) and (3.44), ηi and ζi are related to the variance-based main effect
sensitivity index Si as follows:
Si = ηi ζi .
(3.47)
Substituting ηi and ζi into (3.44) gives:
1 =η1 ζ1 + η2 ζ2 + . . . + η12...m ζ12...m .
(3.48)
The proportion of output uncertainty due to each factor directly can be broken down
84
into an intrinsic extent effect characterized by the sensitivity index ηi , which is related
to complexity, and a non-Gaussianity effect characterized by the sensitivity index ζi ,
which is related to distribution shape. The product of the two effects equals the MSI
of the factor. This allows us to associate (3.44) with the uncertainty apportionment
notion depicted in Figure 3-1.
Note that in the previous paragraph we used the phrase “due to each factor
directly” instead of “due to each factor alone.” This choice of wording reflects the
fact that a change in any auxiliary random variable Zj can affect both the intrinsic
extent and the non-Gaussianity of Y . Thus, the quantities N (Y ) and DKL (Y ||Y G ) are
impacted, which indirectly affects the indices ηi and ζi of all other auxiliary variables
Zi where i 6= j. Because it is usually difficult to determine a priori how changing
Zj would modify N (Y ) and DKL (Y ||Y G ), it is typically impractical to decouple the
direct and indirect effects that alterations to Zj would impose on the entropy power
decomposition. Despite this limitation, the entropy power sensitivity indices still
reveal useful information about how the spread and distribution shape of each input
factor contribute to uncertainty in the output quantity of interest.
3.4.2
The Effect of Distribution Shape
The K-L divergence between Zi and its equivalent Gaussian distribution is both shiftand scaling-invariant. The former is easy to comprehend, since by definition Zi and
ZiG have the same mean. To see the latter, we examine the case where Zi0 = αZi . It
is easy to show that h(Zi0 ) = h(Zi ) + log α (see Table 2.2) [88]. Making use of the
relations fZi0 (z) = α1 fZi ( αz ) and ξ = αz , we can compute the cross entropy h(Zi0 , Zi0G )
85
as follows:
h(Zi0 , Zi0G )
∞
1
z
1
z
fZi ( ) log
f G ( ) dz
=−
α
α Zi α
−∞ α
Z ∞
1
1
=−
fZi (ξ) log
f G (ξ) α dξ
α Zi
−∞ α
Z ∞
Z
=−
fZi (ξ) log fZiG (ξ) dξ + log α
Z
−∞
∞
fZi (ξ) dξ
−∞
= h(Zi , ZiG ) + log α.
(3.49)
Combining (2.24) and (3.49), we obtain:
DKL (Zi0 ||Zi0G ) = DKL (Zi ||ZiG ).
(3.50)
The above result implies that the non-Gaussianity of a random variable remains
constant if its underlying distribution maintains the same shape to within a multiplicative constant. For example, if Zi and Zi0 are both uniform random variables, but
Zi0 has a narrower distribution, the fundamental shape of probability density is not
affected, and therefore the two variables have the same degree of non-Gaussianity.
However, if Zi and Zi0 are both triangular random variables, but one has a symmetric
distribution and the other a skewed distribution, then Zi and Zi0 have differing levels
of non-Gaussianity. This is because the underlying distribution shape has changed,
even though both variables are of the triangular family. Finally, as a note of caution,
we reiterate that even if a design activity has no effect on DKL (Zi ||ZiG ), ζi can still
be impacted indirectly through changes to DKL (Y ||Y G ) imparted by other factors.
Since it is a measure of distance, the K-L divergence is always non-negative. As a
measure of non-Gaussianity, DKL (Zi ||ZiG ) = 0 if and only if Zi itself is Gaussian, and
is positive otherwise. This implies that exp[2DKL (Zi ||ZiG )] ≥ 1. However, ζi can be
greater than, less than, or equal to one, depending on the relative magnitude of the
numerator and denominator in (3.46). That is to say, ζi indicates whether an auxiliary
random variable Zi is less Gaussian (ζi > 1), more Gaussian (ζi < 1), or equally as
Gaussian (ζi = 1) as the QOI Y . Below, we consider the three cases separately for a
86
system modeled by Y = X1 + X2 , where X1 and X2 are either uniform or Gaussian
random variables. Note that this simple system does not contain interaction effects,
and thus Z1 = X1 , Z2 = X2 , and Z12 = 0.
𝑿𝟏 ~𝑼 −𝟎. 𝟓, 𝟎. 𝟓
-0.5
0
𝑿𝟐 ~𝑼 −𝟎. 𝟓, 𝟎. 𝟓
0.5
-0.5
0
𝒀~𝑻 −𝟏, 𝟏, 𝟎
0.5
-1 -0.5
0
0.5
1
0.5
1
(a) Case 1: Example with ζ1 > 1 and ζ2 > 1
2.5
𝑿𝟏 ~𝑼 −𝟎. 𝟓, 𝟎. 𝟓
2
2.5
𝑿𝟐 ~𝑵 𝟎, 𝟎. 𝟐𝟓
1.5
1
1
0.5
0.5
0
0
-0.5
0
0.5
-0.8
𝒀
2
1.5
-0.5
-0.6
-0.4
-0.2
0
0
0.2
0.5
0.4
0.6
0.8
-1.5
-1 -0.5
-1
-0.5
0
0
0.5
1
1.5
(b) Case 2: Example with ζ1 > 1 and ζ2 < 1
2.5
2
2.5
𝑿𝟏 ~𝑵 𝟎, 𝟎. 𝟐𝟓
2
1.5
𝑿𝟐 ~𝑵 𝟎, 𝟎. 𝟐𝟓
2
1.5
1.5
1
1
1
0.5
0.5
0.5
0
-0.8
0
-0.5
-0.6
-0.4
-0.2
0
0
0.2
0.5
0.4
0.6
0.8
-0.8
𝒀~𝑵 𝟎, 𝟎. 𝟑𝟓
2.5
0
-0.5
-0.6
-0.4
-0.2
0
0
0.2
0.5
0.4
0.6
0.8
-1.5
-1 -0.5
-1
-0.5
0
0
0.5
0.5
1
1
1.5
(c) Case 3: Example with ζ1 = ζ2 = 1
Figure 3-6: Examples of Y = X1 + X2 with increase, decrease, and no change in
Gaussianity between the design parameters and QOI
For the first case, we let X1 and X2 be i.i.d. uniform random variables on the
interval [−0.5, 0.5]. The sum Y = X1 + X2 has a symmetric triangular distribution
on the interval [−1, 1]. In this case, both ζ1 and ζ2 exceed one (Figure 3-7(a)). Figure
3-6(a) illustrates that in moving from the design parameters to the QOI, the system
increases in Gaussianity. Figure 3-6(b) shows an example where X1 is uniform and
X2 is Gaussian, resulting in a distribution for Y that is more Gaussian than X1 but
less Gaussian than X2 . In this case, the corresponding indices for non-Gaussianity
are ζ1 > 1 and ζ2 < 1 (Figure 3-7(b)). Finally, if both X1 and X2 are Gaussian, then
X1 + X2 is also Gaussian (Figure 3-6(c)), and ζ1 = ζ2 = 1 (Figure 3-7(c)).
87
1.6
1.4
1.2
1.6
Si
1.37
1.37
ηi
ζi
1
0.8
0.8
0.4
0.6
0.50
0.50
0.99
0.57
0.43 0.43
0.4
0.2
0
ζi
0.40
0.37
0.37
1.42
ηi
1.2
1
0.6
Si
1.4
0.2
Factor 1
0
Factor 2
(a) Case 1: X1 , X2 ∼ U[−0.5, 0.5]
Factor 1
Factor 2
(b) Case 2: X1 ∼ U[−0.5, 0.5],
X2 ∼ N (0, 0.25)
1.6
1.4
1.2
Si
ηi
ζi
1.00
1.00
1
0.8
0.6
0.50 0.50
0.50 0.50
0.4
0.2
0
Factor 1
Factor 2
(c) Case 3: X1 , X2 ∼ N (0, 0.25)
Figure 3-7: Sensitivity indices Si , ηi , and ζi for three examples of Y = X1 + X2 . For
each factor, Si equals the product of ηi and ζi .
3.5
Chapter Summary
In the design of complex systems, sensitivity analysis is crucially important for understanding how various sources of uncertainty influence estimates of the quantities of
interest. In this chapter, we built upon the existing methodology for variance-based
distributional sensitivity analysis to consider changes in distribution family and thus
explicitly study the effect of distribution shape. We also computed the local sensitivity of complexity and risk to perturbations in mean and standard deviation, and
interpreted the results in the context of design for robustness and reliability. Furthermore, we connected these sensitivities to indices from GSA and DSA to generate
local approximations for the change in complexity or risk associated with a reduction
in variance.
88
Another contribution of this chapter is the derivation of the entropy power decomposition, which categorizes uncertainty contributions into intrinsic extent and
non-Gaussianity effects. The entropy power decomposition is analogous to the notion of variance apportionment in GSA; however, it provides additional information
about how different aspects of a factor’s uncertainty are compounded in the system,
which can be valuable to the designer when selecting specific strategies to reduce
uncertainty.
In Chapter 4, we will connect the sensitivity analysis methods from this chapter
with design budgets to trade off various options for uncertainty mitigation and guide
decisions for the allocation of resources.
89
Chapter 4
Application to Engineering System
Design
In this chapter, we apply the framework shown in Figure 1-2 to a simple engineering
system in order to showcase its utility for designing for robustness and reliability.
Throughout this discussion, we will use as an example a R-C high-pass filter circuit,
which is introduced in the Section 4.1. Section 4.2 demonstrates how the sensitivity
analysis techiques developed in this thesis can be used to better our understanding
of the effects of uncertainty in the system, as well as analyze the influence of various
modeling assumptions. In Section 4.3, we introduce design budgets for cost and uncertainty, and connect them with sensitivity analysis results in order to identify specific
options for uncertainty mitigation, visualize tradeoffs between various alternatives,
and inform resource allocation decisions.
4.1
R-C Circuit Example
As an example system, consider the circuit shown in Figure 4-1, which consists of a
resistor and a capacitor in series. The design parameters of the system are resistance
(R) and capacitance (C), which have nominal values 100 Ω and 4.7 µF, respectively.
The corresponding component tolerances are Rtol = ±10% and Ctol = ±20%.
91
𝐶
𝑉in
𝑅
𝑉out
Figure 4-1: R-C high-pass filter circuit
As configured, the circuit acts as a passive high-pass filter, whose magnitude and
phase Bode diagrams are shown in Figure 4-2. We select the quantity of interest to be
the cutoff frequency fc , which has a nominal value of 339 Hz. To achieve the desired
circuit performance, fc must exceed 300 Hz.
(dB)
Magnitude
Magnitude (dB)
0
-10
-20
-30
(deg)
Phase
Phase (deg)
-40
90
45
0
1
10
10
2
10
3
10
4
Frequency (Hz)
(Hz)
Frequency
Figure 4-2: Bode diagrams for magnitude (top) and phase (bottom). The dashed
black line indicates the required cutoff frequency of 300 Hz.
We use the random variables X1 , X2 , and Y to represent uncertainty in R, C, and
fc , respectively. Due to stochasticity in the system, the functional requirement on
the QOI will be treated as a probabilistic design target (i.e., the probability that the
cutoff frequency falls below 300 Hz must not exceed a specified limit). The relationship
between the design parameters and the QOI is given by:
Y =
1
.
2πX1 X2
92
(4.1)
We model X1 and X2 as uniform random variables on the intervals specified
by their component tolerances, such that X1 ∼ U[a1 , b1 ] = U[90, 110] Ω and X2 ∼
U[a2 , b2 ] = U[3.76, 5.64] µF. For simplicity, we shall refer to this initial scenario as
Case 0. For this system, we can evaluate the ANOVA-HDMR of (4.1) in closed form
to obtain the following expressions for the expected value of Y , as well as the auxiliary
random variables Z1 , Z2 , and Z12 :
log(b1 /a1 ) log(b2 /a2 )
,
2π(b1 − a1 )(b2 − a2 )
(4.2)
Z1 =
log(b2 /a2 )
− E [Y ],
2π(b2 − a2 )X1
(4.3)
Z2 =
log(b1 /a1 )
− E [Y ],
2π(b1 − a1 )X2
(4.4)
E [Y ] =
Z12 =
1
log(b2 /a2 )
−
2πX1 X2 2π(b2 − a2 )X1
log(b1 /a1 )
log(b1 /a1 ) log(b2 /a2 )
−
+
.
2π(b1 − a1 )X2
2π(b1 − a1 )(b2 − a2 )
(4.5)
1400
var(Y) = 2015 Hz2
Number of samples
1200
C(Y) = 181 Hz
1000
P(Y<300Hz) = 18.6%
800
600
400
200
0
250
300
350
400
450
500
Cutoff frequency (Hz)
Figure 4-3: Histogram of cutoff frequency for Case 0 generated using 10,000 MC
samples. The dashed black line indicates the required cutoff frequency of 300 Hz.
In this example, we use Monte Carlo simulation with 10,000 samples to characterize the propagation of uncertainty from the design parameters to the quantity of interest. Figure 4-3 shows the histogram of Y for Case 0. The corresponding uncertainty
93
estimates are: var (Y ) = 2015 Hz2 , C(Y ) = 181 Hz, and P (Y < 300 Hz) = 18.6%.
4.2
4.2.1
Sensitivity Analysis of the R-C Circuit
Identifying the Drivers of Uncertainty
Having obtained uncertainty estimates for Case 0, we next apply existing variancebased sensitivity analysis methods to identify the key drivers of output variance.
Global sensitivity analysis of Case 0 reveals that approximately 80% of the variability
in Y can be attributed to X2 (capacitance), and the remaining 20% to X1 (resistance),
with interactions playing a negligible role (Figure 4-4(a)). Distributional sensitivity
analysis (assuming fixed distribution shape) indicates that the relationship between a
decrease in factor variance and the expected decrease in output variance is linear for
both X1 and X2 (Figure 4-4(b)). These results point to prioritizing X2 for uncertainty
reduction, as it dominates X1 in terms of contribution to output variance.
< 1%
1
0.8
20%
Factor 1
Factor 2
Interactions
Factor 1
Factor 2
adjSi(δ)
0.6
80%
0.4
0.2
0
−0.2
0
0.2
0.4
0.6
0.8
1
1−δ
(a) Main effect sensitivity indices
(b) Adjusted main effect sensitivity indices
Figure 4-4: Global and distributional sensitivity analysis results for Case 0
4.2.2
Understanding the Impacts of Modeling Assumptions
In this section, we extend our sensitivity analysis of the R-C circuit to study the effects
of modeling assumptions, in particular assumptions regarding the input distributions.
94
Focusing on X2 , we apply several methods developed in this thesis, including entropy
power-based sensitivity analysis and DSA that permits changes to distribution family.
Two specific cases are considered: in Case 1, we allow the distribution of X2 to vary
between several types while holding its variance constant; in Case 2, the interval on
which X2 is supported is fixed while the distribution type is varied across several
families. In both cases, we are interested in studying how assumptions regarding the
input factor distribution impact uncertainty in the QOI.
Case 1: Fixing the Variance of X2
As described in Case 0, we initially assumed that X1 and X2 are uniformly distributed
on the intervals [90, 110] Ω and [3.76, 5.64] µF, respectively, which correspond to to
variances of var (X1 ) = 33.3 Ω2 and var (X2 ) = 0.30 µF2 . Here, we use the procedures
developed in Section 3.2 to extend DSA from Figure 4-4(b) to permit changes between
distribution families.
1
0.8
1
Uni −> Uni
Uni −> Tri
Uni −> Gaussian
Uni −> Bmd Uni
0.8
0.6
adjSi(δ)
adjSi(δ)
0.6
0.4
0.4
0.2
0.2
0
0
−0.2
0
Uni −> Uni
Uni −> Tri
Uni −> Gaussian
Uni −> Bmd Uni
0.2
0.4
0.6
0.8
−0.2
0
1
1−δ
0.2
0.4
0.6
0.8
1
1−δ
(a) Factor 1: Resistance
(b) Factor 2: Capacitance
Figure 4-5: Distributional sensitivity analysis results with changes to distribution
family
Figure 4-5(a) shows that for X1 , there is virtually no difference in the adjusted
main effect sensitivity indices between the four scenarios considered: the initial case
where var (X1 ) is reduced while maintaining X1 as a uniform random variable, and
the alternative options where a decrease in var (X1 ) is concomitant with a switch to
a triangular, Gaussian, or bimodal uniform distribution. This result suggests that
95
the shape of the underlying distribution for X1 does not contribute significantly to
uncertainty in the QOI; that is, the relationship between a decrease in var (X1 ) and
the expected reduction in var (Y ) is not affected by the distribution family (of the
four considered) to which X1 belongs.
Figure 4-5(b) shows a similar trend for X2 ; however, the four curves diverge slightly
for small values of 1−δ. To study this region, we concentrate on the leftmost points in
Figure 4-5(b) where 1−δ = 0, which correspond to no reduction in var (X2 ) from Case
0. Figure 4-6 shows four possible distributions for X2 , each of which has a variance
of 0.30 µF2 . Note that the triangular and Gaussian distributions have supports that
extend beyond the original interval.1
Probability density
1.5
U
T
G
BU
1
0.5
0
3
3.5
4
4.5
5
5.5
6
6.5
Capacitance (µF)
Figure 4-6: Examples of uniform (U), triangular (T), Gaussian (G), and bimodal
uniform (BU) distributions for X2 , each with a variance of 0.30 µF2
Case
var (X2 )
C(X2 )
var (Y )
C(Y )
P (Y < 300 Hz)
1-T
0%
+16.6%
+2.1%
+1.2%
−1.5%
1-G
0%
+19.3%
+1.3%
−1.9%
−2.3%
1-BU
0%
−50.0%
−2.9%
−4.5%
+2.6%
Table 4.1: Change in uncertainty estimates for Cases 1-T, 1-G, and 1-BU, as compared
to Case 0
1
Although a Gaussian random variable is infinitely supported, in this example we have truncated the Gaussian distribution at ±3 standard deviations (which captures more than 99.7% of the
probability), as it is not realistic for capacitance to approach ±∞.
96
Table 4.1 shows how uncertainty estimates for the system are altered as a result
of changing the distribution of X2 from uniform (Case 0) to triangular (Case 1-T),
truncated Gaussian (Case 1-G), or bimodal uniform (Case 1-BU). It is easy to see
that estimates of variance, complexity, and risk can trend in opposite directions. For
example, in both Cases 1-T and 1-G there is a significant increase in C(X2 ), which results in a slight increase in output variance but a slight decrease in risk. Furthermore,
Case 1-G, which represents the maximum entropy distribution for X2 given a variance
of 0.30 µF2 , does not necessarily correspond to the most conservative estimates for
output uncertainty. In Case 1-BU, the complexity of X2 is reduced by 50%, which in
turn decreases both output variance and complexity but increases risk. These results
demonstrate the presence of competing objectives for uncertainty mitigation, such
that efforts to reduce one measure of uncertainty can result in an increase in the others. In general, the extent to which the various measures of uncertainty trend jointly
or divergently is problem-dependent.
Case 2: Fixing the Extent of X2
In Case 2, we investigate the effect of our initial assumption of a uniform distribution for X2 to capture the uncertainty associated with a ±20% capacitor tolerance.
Figure 4-7 shows several alternative distributions on the interval [3.76, 5.64] µF: triangular and truncated Gaussian distributions centered at 4.7 µF (Cases 2-T and 2G, respectively),2 and a symmetric bimodal uniform distribution parameterized by
aL = 3.76 µF, bL = 4.23 µF, aR = 5.17 µF, and bR = 5.64 µF (Case 2-BU). Since these
distributions lie wholly within the original interval, we can sample from them using
the acceptance/rejection procedures provided in Section 3.2; thus, no additional MC
simulations are required for the following analysis.
The distributions shown in Figure 4-7 differ greatly in variance as well as complexity (see Table 4.2). Cases 2-T and 2-G can represent reasonable choices for X2 if there
is strong confidence in the nominal capacitance. A possible scenario for Case 2-BU is
2
As before, we approximate the Gaussian distribution with a truncated version in which ±3
standard deviations of the distribution fall within the specified interval.
97
Probability density
1.5
U
T
G
BU
1
0.5
0
3
3.5
4
4.5
5
5.5
6
6.5
Capacitance (µF)
Figure 4-7: Examples of uniform (U), triangular (T), Gaussian (G), and bimodal
uniform (BU) distributions for X2 supported on the interval [3.76, 5.64] µF
Case
var (X2 )
C(X2 )
2-T
−50.0%
2-G
2-BU
var (Y )
C(Y )
P (Y < 300 Hz)
−17.6% −41.2%
−21.6%
−7.4%
−66.7%
−31.1% −54.6%
−30.8%
−10.5%
+75.0%
−50.0%
+11.9%
+11.2%
+60.7%
Table 4.2: Change in uncertainty estimates for Cases 2-T, 2-G, and 2-BU, as compared
to Case 0
one in which the capacitors with ±10% tolerance are removed during the manufacturing process and sold for a higher price, leaving the worst-performing components
in the ±20% tolerance category.
Figure 4-8 shows the histograms of Y generated using 10,000 MC samples for the
various cases.3 It is easy to see that the choice of X2 greatly affects the shape of the
resulting distribution for Y , as well as the corresponding uncertainty estimates. Table
4.2 shows that in Cases 2-T and 2-G, there is a significant decrease in both var (X2 )
and C(X2 ) over Case 0, which also results in large reductions in the variance, complexity, and risk in the QOI. These outcomes are consistent with the adjusted main effect
sensitivity indices shown in Figure 4-5(b), which predict the expected reduction in
var (Y ) to be approximately 40% and 55% for Cases 2-T and 2-G, respectively, corresponding to reductions in var (X2 ) of 50% and 66.7%. Case 2-BU, on the other hand,
3
Note that Figure 4-3 is reproduced as Figure 4-8(a) as a point of comparison in Case 2.
98
1400
1400
2
var(Y) = 2015 Hz
C(Y) = 181 Hz
1000
P(Y<300Hz) = 18.6%
800
600
400
200
0
250
var(Y) = 1182 Hz2
1200
Number of samples
Number of samples
1200
C(Y) = 142 Hz
1000
P(Y<300Hz) = 10.9%
800
600
400
200
300
350
400
450
0
250
500
300
Cutoff frequency (Hz)
350
(a) Case 0
500
1400
2
var(Y) = 898 Hz
1200
C(Y) = 124 Hz
1000
var(Y) = 3270 Hz2
1200
Number of samples
Number of samples
450
(b) Case 2-T
1400
P(Y<300Hz) = 7.7%
800
600
400
200
0
250
400
Cutoff frequency (Hz)
C(Y) = 203 Hz
1000
P(Y<300Hz) = 29.7%
800
600
400
200
300
350
400
450
0
250
500
Cutoff frequency (Hz)
300
350
400
450
500
Cutoff frequency (Hz)
(c) Case 2-G
(d) Case 2-BU
Figure 4-8: Histogram of Y resulting from various distributions for X2 considered in
Case 2
represents a 50% reduction in C(X2 ) and a 75% increase in var (X2 ), and produces
sizable increases the uncertainty estimates for the QOI.4 This illustrates that, as in
Case 1, the maximum entropy distribution for X2 given the available information (in
this case, the uniform distribution since we are focusing on the interval [3.76, 5.64] µF)
does not necessarily lead to the most conservative estimates for output uncertainty.
Figure 4-9 shows the entropy power-based sensitivity indices ηi and ζi alongside
the main effect indices Si computed from GSA. We see that for Cases 0, 2-T, and
2-G, Si is similar to ηi for both factors, and that ζi is equal to or slightly greater than
4
Case 2-BU does not fall along the DSA curves shown in Figure 4-5(b). Since it entails a
75% increase in var (X2 ), one would need to extrapolate the uniform → bimodal uniform curve to
1 − δ = −0.75; doing so confirms that a 60.7% increase in var (Y ) can reasonably be expected.
99
2.5
2.5
S
S
i
i
ηi
2
ζi
1.5
ζi
1.5
1.17
1.14
1.11
0.98
1
0.80
0.72
0.67
0.5
0.20 0.18
0.00 0.00
Factor 1
Factor 2
0
Interactions
1.13
1.02
1
0.5
0
ηi
2
0.33
0.59
0.27
0.00 0.00
Factor 1
Factor 2
(a) Case 0
Interactions
(b) Case 2-T
2.5
2.5
Si
Si
ηi
2
2
ζi
1.5
ηi
2.06
ζi
1.5
1.19
1.18
1.00
1
1
0.85
0.87
0.75
0.57 0.57
0.5
0.43
0.5
0.35
0.38
0.13 0.13
0
0.00 0.00
Factor 1
Factor 2
0
Interactions
(c) Case 2-G
0.00 0.00
Factor 1
Factor 2
Interactions
(d) Case 2-BU
Figure 4-9: Main effect and entropy power-based sensitivity indices for Case 2
one. This indicates that uncertainty apportionment using variance and entropy power
produce similar results. Furthermore, it reveals that Gaussianity increases in moving
from the input factors to the QOI. For Case 2-BU, however, ζ1 is less than one, ζ2
is exceedingly large, and η2 is significantly smaller than S2 . This implies that the
proportion of output variance due to Factor 2 alone (87%) is dominated by the nonGaussianity of the bimodal uniform distribution, with only a small contribution due
to the intrinsic extent of X2 . This is consistent with the result that the complexity of
Y increases by only 11.9% over Case 0, whereas variance rises by a substantial 60.7%
(see Table 4.2).
For Case 2-BU, our analysis has highlighted that the propagation of uncertainty
from the inputs to the QOI is dominated by the bimodality of X2 , which introduces
100
significant non-Gaussianity into the system. Thus, in using entropy power decomposition, we learned more about the underlying features of the system that contribute to
output uncertainty than can be revealed by variance-based sensitivity analysis alone.
We conclude from Case 2 that initial assumptions regarding distribution shape can
greatly impact uncertainty in the QOI, especially for systems involving multimodality
or other highly non-Gaussian features. Therefore, one must be careful when selecting
the appropriate distribution to characterize uncertainty in the system inputs, as misattributions can lead to gross over- or underestimates of uncertainty in the quantities
of interest.
4.3
Design Budgets and Resource Allocation
Once the sources of uncertainty in a system are identified through sensitivity analysis, the task of uncertainty mitigation centers on making decisions regarding design
modifications in the subsequent iteration [5, 6, 127]. Such decisions often relate to
the allocation of resources in order to improve one or more aspects of the design.
Resources can be allocated to a variety of different activities — for example, to direct
future research, to conduct experiments, to improve physical and simulation-based
modeling capabilities, or to invest in superior hardware or additional personnel. These
steps complete the resource allocation feedback loop shown in Figure 1-2.
Typically, the appropriate decisions for minimizing system uncertainty at each step
of the design process are not known a priori. Instead, sensitivity analysis results are
often used to infer the potential of various options for mitigating uncertainty. As discussed in Chapter 3, variance-based GSA and DSA, as well as relative entropy-based
sensitivity analysis all enable the prioritization of factors according to contribution to
output variability. However, these methods do not address how to identify or select
between specific strategies.
In the following sections, we use the R-C circuit example to demonstrate how the
sensitivity analysis methods developed in this thesis can be combined with budgets
for cost or uncertainty and used to uncover design alternatives for reducing uncer-
101
tainty, visualize tradeoffs among the available options, and ultimately guide decisions
regarding the allocation of resources to improve system robustness and reliability.
4.3.1
Visualizing Tradeoffs
First, we are interested in studying how uncertainty estimates for the QOI are affected
by reductions in the variance of one or both design parameters. For simplicity, we will
assume that uncertainty in both resistance and capacitance can be well-characterized
using uniform random variables. This assumption is justified by the DSA results in
Figure 4-5, which showed that the variance of Y is not significantly affected by the
shape of the underlying distributions for X1 and X2 . Therefore, reductions in var (X1 )
and var (X2 ) serve to shrink the extent of the corresponding uniform distributions.5
We note, however, that the techniques employed in the following analysis can be
adjusted to also accommodate changes in the distribution family of X1 or X2 .
Figure 4-10 shows how different levels of reduction in var (X1 ) and var (X2 ) trade
against one another with respect to variance, complexity, and risk remaining in the
QOI. The bottom left corner of each figure represents the initial design (Case 0).
The colored contours were generated using 10,000 MC simulations to evaluate the
QOI at each design parameter combination. Hence, they represent the most accurate
uncertainty estimates available for the system.
For many realistic engineering systems, however, evaluating the QOI is computationally expensive, and it may not be feasible to perform numerous MC simulations
in order to visualize tradeoffs in the design space. In those situations, it is still
possible to approximate variance, complexity, and risk using sensitivity information.
For variance, we can use the adjusted main effect sensitivity indices from DSA (i.e.,
Figure 4-5) to relate a 100(1 − δ)% decrease in var (X1 ) or var (X2 ) to the expected
reduction in var (Y ). Similarly, we can combine the DSA results with local sensitivity
estimates from (3.26) and (3.27) to predict the expected change in complexity and
risk. The contours generated using these approximations are indicated by the dashed
5
We assume that reductions in variance reflect improvements in component tolerance, and thus
decrease distribution extent symmetrically about R = 100 Ω or C = 4.7 µF.
102
1
200
500
200
Percent reduction in var(X2)
Percent reduction in var(X2)
1
500
0.8
800
500
800
0.6
1100
800
1100
0.4
1400
1100
0.2 170
0
02000
0
0.2
1400
1400
0.6
0.8
0.6
90
105
90
135
120
105
135
120
150
135
0.4 150
0.2 165
150
165
0180
0
1
75
120
Percent reduction in var(X1)
0.2
0.4
0.6
0.8
1
Percent reduction in var(X1)
(a) Lines of constant variance
(b) Lines of constant complexity
1 0
Percent reduction in var(X2)
75
1
0.8 05
1700
0.4
75
90
0
0.025
0.8 0.05
0.075
0.025
0.6 0.1
0.07
0.0
25
0.0
5
0.075
0.1
5
0.1
0.1
0.4 25
0.125
0.15
0.2
0.125
0.15
0.175
0
0
0
0.05
0.2
0.175
0.4
0.6
0.15
0.8
0.175
1
Percent reduction in var(X1)
(c) Lines of constant risk
Figure 4-10: Contours for variance, complexity, and risk corresponding to reductions
in factor variance, generated from 10,000 MC simulations (solid colored lines) or
approximated using distributional or local sensitivity analysis results (dashed black
lines)
black lines in Figure 4-10. In Figure 4-10(a), there is a close match between the variance contours predicted from DSA results and those generated using MC simulations.
For complexity and risk (Figures 4-10(b) and 4-10(c)), however, the correspondence
deteriorates for large deviations from Case 0, especially as var (X1 ) is decreased.
Nevertheless, the trends in variance, complexity, and risk resulting from reductions in
var (X1 ) and var (X2 ) are correctly captured, which render the approximated contours
useful for visualizing tradeoffs between uncertainty mitigation options in the input
factors, while circumventing the need to perform additional MC simulations.
Figure 4-10 confirms that decreasing var (X2 ) yields the most dramatic reductions
103
in output uncertainty; in fact, similar trends are observed for variance, complexity,
and risk. This is consistent with the prioritization of X2 over X1 in Section 4.2.1.
Next, we will connect the potential for uncertainty reduction with the cost of implementing the requisite design change, which leads to a discussion of budgets for cost
and uncertainty.
4.3.2
Cost and Uncertainty Budgets
In typical system design processes, uncertainty reduction efforts are constrained by
the availability of resources, such as funds to acquire superior tools or components,
manpower to conduct additional research or experiments, or time to allot to computational models. In this thesis, we assume that there is a cost associated with each
activity that mitigiates uncertainty in the QOI. As an illustrative example, let us consider the notional curves shown in Figure 4-11(a) for the R-C circuit, which depicts
the cost associated with a 100(1 − δ)% reduction in the variance of X1 or X2 .6 In
Figure 4-11(b), the same information is visualized as contour lines of equal cost with
respect to different proportions of variance reduction in the two factors. We observe
that although reducing var (X2 ) is more effective in abating output uncertainty than
decreasing var (X1 ), it is also more expensive to implement. Furthermore, the reduction in var (X2 ) is capped at 75% (that is, 25% of the variability in X2 is irreducible),
whereas var (X1 ) can be reduced by as much as 99%.
In order to make design decisions that adequately mitigate uncertainty yet are
feasible to implement, we must take into account the available budgets, both in terms
of cost and uncertainty. A cost budget specifies the amount of resources that can be
expended to improve the design. An uncertainty budget, on the other hand, refers
to the total level of uncertainty that is deemed tolerable for the system. In both
cases, we seek to determine how much of the prescribed amount ought to be allocated
to each design parameter. For the R-C circuit, we impose the following budgetary
6
In general, it may be difficult to determine the relationship between cost and reductions in
input uncertainty. We assume in this thesis, however, that such relationships exist and have already
been established using, for example, historical data, cost estimation methods, or expert opinion
elicitation.
104
1
25
20
Cost
15
10
5
0
0
0.2
0.4
0.6
0.8
0.8
22.5
20
17.5
0.6
15
12.5
10
7.5
0.2
5
0.4
0
0
1
0.2
12
.5
10
7.5
5
0.4
0.6
3027.525 2.5 20 7.5 15
2
1
Percent reduction in var(X2)
Factor 1
Factor 2
0.8
1
Percent reduction in var(X1)
1−δ
(a) Cost of reducing variance of X1 or X2
(b) Lines of constant cost
Figure 4-11: Notional cost associated with a 100(1 − δ)% reduction in the variance
of X1 or X2
constraints:
1. Complexity: The complexity with respect to the QOI shall not exceed 150 Hz.
2. Risk: The probability of violating the cutoff frequency requirement shall not
exceed 10%.
3. Cost: The cost of uncertainty reduction shall not exceed 20 units of cost.
Next, we examine each constraint individually in turn to study the resulting optimal allocation. Figure 4-12 overlays contours of variance, complexity, and risk in the
QOI (solid colored lines) with those corresponding to cost of implementation. These
contours allow us to consider the cost and uncertainty budgets in conjunction. We
see that for a given cost, the maximum possible uncertainty reduction is achieved
by decreasing the variance of both X1 and X2 simultaneously, rather than focusing on either factor alone. Table 4.3 lists the lowest possible uncertainty and cost
estimates for the system when each of the budgetary constraints (C(Y ) ≤ 150 Hz,
P (Y < 300 Hz) ≤ 10%, Cost ≤ 20) is in turn made active. Of the three budgets, the
complexity constraint is the cheapest to satisfy; however, the resulting allocation of
factor variance reduction does not satisfy the risk constraint. To ensure that risk does
not exceed 10%, a minimum cost of 14.8 is required; this corresponds to a complexity
of 132 Hz. Finally, a cost budget of 20 is more than adequate to guarantee that the
105
1
200
500
500
0.8
800
20
10
1400
0.4
0.6
10 1400
0.8
120
0.6
135
15
0.4 150
0180
0
1
75
105
90
120 20
105
120
135 15
10
150
0.2 165
5
Percent reduction in var(X1)
135
10 1
50
5
165
0.2
0.4
0.6
0.8
1
Percent reduction in var(X1)
(a) Lines of constant variance and cost
1 0
0
0.025
0.8 0.05
0.075
20
15
0.0
0.07 20
5
0.1 1
5
25
0.0
5
0.075
0.1
0.2
0.175
0.4
0.6
0.125
10 0.15
15
0.15
5
5
20
0.125
10
0.175
0
0
0.05
25
0.15
0.2
0
0.025
0.6 0.1
0.1
0.4 25
(b) Lines of constant complexity and cost
30
Percent reduction in var(X2)
20
15
5
1700
90
20
02000
0
0.2
1100
15
0.2 170
50
110015
20
0.4
1400
800
75
1
0.8 05
25
15
800 20
25
0.6
1100
500
75
90
30
Percent reduction in var(X2)
200
30
Percent reduction in var(X2)
1
0.8
0.175
1
Percent reduction in var(X1)
(c) Lines of constant risk and cost
Figure 4-12: Contours for variance, complexity, and risk corresponding to reductions
in factor variance (solid colored lines) overlaid with contours for cost of implementation (dashed green lines)
uncertainty budgets are also satisfied. As we observe in Figures 4-12(b) and 4-12(c),
the expenditure of 20 units of cost is sufficient to decrease var (X2 ) by up to 67%;
even without a simultaneous reduction in var (X1 ), it is enough to reduce complexity
and risk to acceptable levels of 125 Hz and 8%, respectively.
The results in Table 4.3 suggest that the cost and uncertainty budgets for the
R-C circuit design are mutually compatible and achievable. There are a host of
solutions that satisfy all three budgetary constraints, which lie within the shaded
region in Figure 4-13. The main tradeoff in these solutions is between cost and risk:
when both of those constraints are satisfied, the complexity budget is automatically
satisfied as well. In the following section, however, we will consider the cost and
106
Active constraint
var (Y ) C(Y )
P (Y < 300 Hz)
Cost
Complexity
1378
150
13.4%
8.8
Risk
1055
132
10%
14.8
Cost
780
114
6.4%
20
Table 4.3: Best achievable uncertainty mitigation results given individual budgets for
cost and uncertainty. Entries in red denote budget violations.
Figure 4-13: Contours for risk (solid blue lines) overlaid with contours for cost (dashed
green lines). Shaded region denotes area in which all budget constraints are satisfied.
uncertainty budgets alongside available design options, which can lead to different
conclusions regarding their feasbility.
4.3.3
Identifying and Evaluating Design Alternatives
In this section, we study how the nominal value and tolerance of the components
in the R-C circuit trade against one another in terms of contribution to uncertainty
in the QOI. The present analysis builds upon the design budget discussion from
Section 4.3.2, and enables designers to visualize strategies for uncertainty mitigation
in terms of actionable items in the design space.
For each of X1 and X2 , there are two parameters that can be modified: the nominal
value of the circuit component (R or C) and the tolerance (Rtol or Ctol ). We vary
R and C from their nominal values by up to ±20%, and decrease Rtol and Ctol from
the initial tolerances of ±10% and ±20%, respectively, down to ±1%. Figures 4-14
107
6
18
0.5
0.7
0.9
0
5.5 .1
0
5 210
Capacitance (µF)
Capacitance (µF)
3
0
5.5
150
180
4.5
240
18
21
0
0
4
270
240
3.5
3
80
0.
6
15
300
330
210 0
Case
Option A
Option B
240
110
120
270
90
100
5
0.0
1
0.7
0.5
0.3
0.1
4.5
0.0
1
0.3
0.1
4
3.5
3
80
0.01
Case 0
Option A
Option B
90
100
Resistance (Ω)
110
120
Resistance (Ω)
(a) Lines of constant complexity
(b) Lines of constant risk
0.5
0.3
0
20
22220
0
21201
0
140
140
130
0.02
Case 0
Option A
Option C
110
0
80
120
Resistance (Ω)
0.1
0.04
0.2
150
160
15
0
170
16
0
100
Resistor tolerance
16
0
18
17
0
0
0
19
180
0
17
190
0
18
200
200
0.6
190
0.6
Resistor tolerance
0.5
0.4
90
0.06
0.01
0
80
0.4
0
14
0.02
0.2
0.08
0
15
0.06
0.04
0.1
0.08
0.1
0.01
0.1
0.3
Figure 4-14: Uncertainty contours for variations in resistance and capacitance, generated from 10,000 MC simulations (solid colored lines) or approximated using entropy
power-based sensitivity analysis results (dashed black lines)
Case 0
Option A
Option C
90
100
110
120
Resistance (Ω)
(a) Lines of constant complexity
(b) Lines of constant risk
Figure 4-15: Uncertainty contours for variations in resistance and resistor tolerance
through 4-17 show contour lines for complexity (left) and risk (right) for different
combinations of the four quantities; a black X in each figure denotes the location of
Case 0 in the design space.
These contours reveal several interesting trends. First, we see that raising either
R or C lowers complexity in the QOI but also increases risk, and vice-versa (Figure
4-14); this indicates the presence of competing objectives. For a fixed value of R,
decreasing Rtol appears to have little effect on complexity, and almost no effect on
risk (Figure 4-15). Figure 4-16(a), on the other hand, shows that both C and Ctol
108
1
0
100
80
0
0.05
14
80
12
0
10
3.5
4
4.5
5
5.5
4.5
5
Case 0
Option B
Option D
0
3.5
6
0.5
0.6
0.05
Case 0
Option B
Option D
0
0
3
0.1
0.4
0.3
0.2
0.1
12
0.7
0
16
0.1
0.15
0.8
120
40
0.5
0.6
0
18
4
Capacitance (µF)
0.7
0.8
0.9
0
20
Capacitor tolerance
0.15
0.01
Capacitor tolerance
140
0.4
0
16
0.3
0.2
0
18
0.1
0
20
0
22
0.2
0
26 240
1
0.0
0.2
5.5
6
Capacitance (µF)
(a) Lines of constant complexity
(b) Lines of constant risk
Figure 4-16: Uncertainty contours for variations in capacitance and capacitor tolerance
0.2
0.2
160
160
140
120
0.15
120
100
0.1
Capacitor tolerance
Capacitor tolerance
140
100
80
80
60
0.05
40
60
20
0
0
0.02
0.04
0.06
0.15
0.1
0.16
0.13
0.1
0.07
0.04
0.04
0.0
1
0.
04
0.01
0.0
0.05
Case 08
0
Option C
Option D
0.08
0.16
0.16
0.13
0.13
0.1
0.1
0.07
0.04
0.010.07
1
Case 0
Option C
Option D
0
0
0.1
0.02
Resistor tolerance
0.04
0.06
0.08
0.1
Resistor tolerance
(a) Lines of constant complexity
(b) Lines of constant risk
Figure 4-17: Uncertainty contours for variations in resistor and capacitor tolerance
are influential with respect to complexity; however, Figure 4-16(b) indicates that
depending on the value of C, decreasing Ctol can either increase or decrease risk.7
Finally, Figure 4-17 shows that when trading Rtol against Ctol , decreasing the latter
is more effective for reducing both complexity and risk.
As was the case in Section 4.3.1, the colored contours in Figures 4-14 through 4-17
were generated using 10,000 MC simulations to evaluate Y at each design parameter
combination. The dashed black lines correspond to local approximations obtained
7
This is because changing C also shifts the distribution of Y relative to the requirement r = 300 Hz
(see Figure 3-5).
109
using sensitivity information, and can be produced without performing additional MC
simulations. The approximated risk contours in Figures 4-14(b), 4-15(b) 4-16(b), and
4-17(b) were computed using local sensitivity results from (3.33) and (3.37). Although
these approximations vary in accuracy, the overall trends in risk due to changing R,
C, Rtol , and Ctol are captured, especially for design parameter combinations close
to Case 0. The approximated complexity contours (Figures 4-14(a), 4-15(a) 4-16(a),
and 4-17(a)), on the other hand, rely on the results of entropy power-based sensitivity
analysis, and were estimated using the procedure below.
First, we introduce the notation X1o and X2o , corresponding to the random variables
for resistance and capacitance in Case 0, respectively, as well as X10 and X20 , denoting
the updated estimates for those factors after some design modificaton. Similarly,
letting i = 1, 2, we write the following relations:
Zio = gi (Xio )
Zi0 = gi (Xi0 )
Y o = g(X1o , X2o )
Y 0 = g(X10 , X20 )
ζio =
exp[2DKL (Zio ||ZioG )]
exp[2DKL (Y o ||Y oG )]
ζi0 =
exp[2DKL (Zi0 ||Zi0G )]
exp[2DKL (Y 0 ||Y 0G )]
Based on the entropy power decomposition given in (3.44), we can express the
entropy power of Y o as the sum of the following terms:
o
o
N (Y o ) = N (Z1o )ζ1o + N (Z2o )ζ2o + N (Z12
)ζ12
.
(4.6)
Since the interaction effects are negligible for this system, we can drop the last term
in (4.6). Then, we employ the following approximation to predict N (Y 0 ), which uses
the non-Gaussianity indices ζ1o and ζ2o evaluated in Case 0:
N (Y 0 ) ≈ N (Z10 )ζ1o + N (Z20 )ζ2o .
(4.7)
The relationship between each input factor Xi and the corresponding auxiliary random variable Zi is available from the ANOVA-HDMR (computed either analytically
110
or numerically); for the R-C circuit, these relationships are given in (4.2)–(4.5). Thus,
we are able to calculate N (Z10 ) and N (Z20 ) based on X10 and X20 , which then allows
us to estimate N (Y 0 ) without having to explicitly evaluate Y 0 using MC simulations.
Finally, recognizing that entropy power is proportional to the square of complexity,
we can rewrite (4.7) in terms of C(Y 0 ):
C(Y 0 ) ≈
p
2πe[N (Z10 )ζ1o + N (Z20 )ζ2o ].
(4.8)
We observe that the approximations obtained using (4.8) closely match the colored
contours generated from brute force MC simulation. The largest discrepancies occur
in Figure 4-15(a) for small resistor tolerances, which is consistent with the trend
observed in Figures 4-10(b) and 4-10(c) for small values of var (X1 ). Even for this case,
however, the overall trends in complexity due to varying Rtol and R are preserved.
Thus, the approximation given in (4.8) provides an easy way to visualize trends in
complexity and identify regions in the design space that warrant additional analysis
— toward which additional MC simulations can be directed to generate more rigorous
estimates of uncertainty.
Figures 4-14 through 4-17 depict tradeoffs among R, C, Rtol , and Ctol as continuous within the design space, implying infinitely many design combinations. For many
engineering systems, however, the design space instead contains a discrete number of
feasible options dictated by component availability. Thus, we next investigate how
local sensitivity analysis results can be used to identify distinct design alternatives.
Using (3.38) and (3.39), we can compute the requisite change in the mean and
standard deviation of X1 and X2 needed to achieve a desired decrease in risk. In this
case, we set ∆P = −0.086 so as to meet the constraint that P (Y < 300 Hz) ≤ 10%.
Table 4.4 lists the change in mean and standard deviation of each factor required
to achieve this risk reduction as predicted using local sensitivities, as well as the
corresponding component nominal value and tolerance. Consistent with the risk
contours in Figure 4-14(b), we see that reducing risk requires decreasing R or C.
This suggests that we use a resistor with a nominal value of no more than 96.6 Ω, or a
111
Component
Change mean only
Change SD only
∆µXi
Nominal value
∆σXi
Tolerance
Resistor
−3.4 Ω
R = 96.6 Ω
−8.4 Ω
N/A
Capacitor
−0.16 µF
C = 4.54 µF
−0.17 µF
Ctol = ±14.4%
Table 4.4: Local sensitivity predictions for the change in mean and standard deviation
(SD) of each factor required to reduce risk to 10%
capacitor with a nominal value of no more than 4.54 µF. Alternatively, we can reduce
risk by using components with tighter tolerances. For the resistor, the change in the
standard deviation of X1 needed to satisfy the risk constraint is −8.4 Ω; however, in
Case 0, σX1 = 5.8 Ω, which implies that it is not possible to achieve the requisite risk
reduction by decreasing Rtol alone. This is consistent with the risk contours in Figures
4-15(b) and 4-17(b). For the capacitor, we observe that the risk constraint can be
met by decreasing the standard deviation of X2 by −0.17 µF, which corresponds to a
component tolerance of ±14.4%.
The local sensitivity analysis predictions point the designer to values of R, C,
Rtol , and Ctol that would satisfy the uncertainty budget for risk. Consulting standard
off-the-shelf resistor and capacitor values [128], we identify four design alteratives
— Options A–D — which correspond most closely to the desired component specifications shown in Table 4.4. Options A–D are listed in Table 4.5 along with their
associated estimates of variance, complexity, risk, and cost of implementation. Each
option perturbs one of R, C, Rtol , and Ctol from its initial value in Case 0. We assume in this analysis that it is comparatively cheaper to purchase components with
different nominal values (Options A and B) than it is to acquire components with
significantly tighter tolerances (Options C and D).8 The locations of the four options
within the design space are indicated in Figures 4-14 through 4-17.
Figure 4-18 illustrates the tradeoffs between complexity and risk for the various
8
Assuming that X1 and X2 are both uniform random variables, a decrease in Rtol from ±10%
in Case 0 to ±1% in Option C corresponds to a 99% reduction in var (X1 ), and a decrease in
Ctol from ±20% in Case 0 to ±10% in Option D corresponds to a 75% reduction in var (X2 ). The
associated costs of implementation correspond to the rightmost red and blue points in Figure 4-11(a),
respectively.
112
Option
R (Ω)
Rtol
C (µF)
Ctol
var (Y )
C(Y ) P (Y < 300 Hz)
Cost
Case 0
100
±10%
4.7
±20%
2015
181
18.6%
N/A
A
82
±10%
4.7
±20%
2997
221
0%
5
B
100
±10%
3.9
±20%
2927
218
0%
5
C
100
±1%
4.7
±20%
1627
154
18.1%
13.2
D
100
±10%
4.7
±10%
788
115
7.1%
22.5
Table 4.5: Uncertainty and cost estimates associated with various design options.
Entries in red denote budget violations.
options. Immediately, we see that none of the options satisfies both the cost and
uncertainty budgets. Options A and B fall well within the cost budget and reduce
risk to zero; however, they both increase complexity to unacceptable levels, and thus
can be eliminated from consideration.9 In Option C, the resistor tolerance is reduced
to just ±1% at a cost of 13.2; yet, there is virtually no change in risk compared to
Case 0. Option D appears to be the most promising choice, as it is the only one to
satisfy both the complexity and risk constraints. However, it exceeds the cost budget
of 20.
0.25
0.2
Case 0
Option A
Option B
Option C
Option D
Risk
0.15
0.1
0.05
0
0
50
100
150
200
250
Complexity (Hz)
Figure 4-18: Tradeoffs in complexity and risk for various design options. Dashed
green lines bound region where both complexity and risk constraints are satisfied.
9
Had we instead identified options that raised the nominal values of R and C, the result would
have decreased complexity and increased risk, which is similarly undesirable.
113
These results suggest that in order to meet both the cost and uncertainty budgets, the designer must either seek alternative options, or relax one or more of the
constraints. In Figure 4-18, we see that Option D falls well within the region of acceptable uncertainty. Focusing on the top right corner of that region, we find that when
Ctol = ±12.5%, the system achieves C(Y ) = 130 Hz and P (Y < 300 Hz) = 10.0%
at a cost of 18.3, which meets all design budgets. However, while this design is
mathematically possible, it is not realistically feasible given limitations in component
availability, as standard capacitor tolerances include ±20%, ±10%, ±5%, ±2%, and
±1%. From this analysis, we conclude that the most promising course of action is
to allocate resources toward improving capacitor tolerance; furthermore, we recommend increasing the cost budget so that all uncertainty targets can be satisfied using
standard components.
4.4
Chapter Summary
In this chapter, we demonstrated the uncertainty quantification and sensitivity analysis methods introduced in Chapters 2–3 on a simple R-C high-pass filter circuit.
This example highlighted the applicability of our methods to assess the implications
of initial assumptions about the system for uncertainty in the quantities of interest.
We also showed how various sensitivity results can be used to formulate local approximations of output uncertainty, which enable the visualization of tradeoffs in the
design space without expending additional computational resources to perform model
evaluations. Finally, we connected our sensitivity analysis techniques to budgets for
cost and uncertainty, which facilitates the identification and evaluation of various
design alternatives and helps inform decisions for allocating resources to mitigate
uncertainty.
In Chapter 5, we will apply the methods developed in this thesis to perform costbenefit analysis of a real-world aviation environmental policy, further demonstrating
their utility for improving system robustness and reliability.
114
Chapter 5
Application to a Real-World
Aviation Policy Analysis
This chapter presents a case study in which the methods developed in this thesis are
applied to analyze a set of real-world aviation policies pertaining to increases in aircraft noise certification stringency. Section 5.1 provides some background on aviation
noise impacts and frames the objectives of the case study. Section 5.2 introduces
the APMT-Impacts Noise Module and explains how it is used in the present context
to perform a cost-benefit analysis of the proposed policy options, identify sources of
uncertainty, and analyze assumptions and tradeoffs. Section 5.3 discusses the case
study results and recommends specific courses of action to support the design of
cost-beneficial policy options.
5.1
5.1.1
Background and Problem Overview
Aviation Noise Impacts
The demand for commercial aviation has risen steadily over the last several decades,
and is expected to continue growing at a rate of 5% per year over at least the next two
decades [129–131]. This wave of growth brings with it increasing concerns regarding
the environmental impacts of aviation, which include aircraft noise, air quality degra-
115
dation, and climate change. Of these issues, aircraft noise has the most immediate
and perceivable community impact [132–134], and was the first to be regulated by
the International Civil Aviation Organization (ICAO) in 1971 with the publication of
Annex 16: Environmental Protection, Volume I: International Noise Standards [135].
There are many effects associated with exposure to aircraft noise, which can be
broadly classified as either physical or monetary. The physical effects of aviation noise
span a range of severities, and include annoyance, sleep disturbance, interference with
school learning and work performance, and health risks such as hypertension and heart
disease [136, 137]. The monetary effects, on the other hand, include housing value
depreciation, rental loss, and the cost associated with lost work or school performance.
The assessment of these monetary impacts is of particular interest to policymakers,
researchers, and aircraft manufacturers, as they provide quantitative measures with
which to perform cost-benefit analysis of various policy options.1
Motivated by a 2004 report to the US Congress on aviation and the environment
[138], the Federal Aviation Administration (FAA) is developing a comprehensive suite
of software tools that can characterize and quantify a wide spectrum of environmental
effects and tradeoffs, including interdependencies among aviation-related noise and
emissions, impacts on health and welfare, and industry and consumer costs under
various scenarios [139]. This effort, known as the Aviation Environmental Tools Suite,
takes as inputs proposed aviation policies or scenarios of interest, which can pertain to
regulations (e.g., noise and emissions stringencies, changes to aircraft operations and
procedures), finances (e.g., fees or taxes), or anticipated technological improvements.
These inputs are processed through various modules of the Tools Suite — in particular
the Aviation environmental Portfolio Management Tool for Impacts (APMT-Impacts)
— to produce cost-benefit analyses that explicitly detail the monetized noise, climate,
and air quality impacts of the proposed policy or scenario with respect to a welldefined baseline [140].
1
While the physical and monetary effects of aviation noise are usually presented separately, the
two categories are not necessarily independent – that is, monetary impacts (typically assessed from
real estate-related damages) often serve as surrogate measures for the wider range of interrelated
effects that are difficult to quantify individually. However, it is recognized that such estimates may
undervalue the full impacts of noise.
116
The Noise Module within APMT-Impacts was developed by He et al. and relates
city-level per capita income to an individual’s willingness to pay for one decibel (dB) of
noise reduction [141, 142]. It is based on a meta-analysis of 63 hedonic pricing aviation
noise studies, which used observed differences in housing markets between noisy and
quiet areas to determine the implicit value of quietness (or conversely, the cost of
noise).2 Because income data are typically readily available for most airport regions
around the world, the income-based model facilitates the assessment of aviation noise
impacts on a global scale. For example, a previous study used the APMT-Impacts
Noise Module to estimate that capitalized property value depreciation due to aviation
noise in 2005 totaled $23.8 billion around 181 airports worldwide, including $9.8
billion around 95 airports in the US [142]. In the following sections, we will discuss
the APMT-Impacts Noise Module in more detail, as well as introduce the aviation
policy case study to which it is applied.
5.1.2
CAEP/9 Noise Stringency Options
Since ICAO’s initial publication of Annex 16 in 1971, international aircraft noise
certification standards have been updated on three different occasions to reflect advances in technology. The most recent stringency increase was adopted in 2001 as
part of the ICAO Committee for Aviation Environmental Protection’s (ICAO-CAEP)
CAEP/5 cycle. In 2007, a formal review of existing noise certification standards was
initiated as part of CAEP/7. Subsequently, five options for noise stringency increase
were proposed during CAEP/9 in 2013, which correspond to changes in noise certification stringency of −3 (Stringency 1), −5 (Stringency 2), −7 (Stringency 3), −9
(Stringency 4), and −11 (Stringency 5) cumulative EPNdB relative to current limits
2
Hedonic pricing (HP) studies typically derive a Noise Depreciation Index (NDI) for one airport,
which represents the percentage decrease in property value corresponding to a one decibel increase
in the noise level in the region. Previous meta-studies have shown that NDI values tend to be
similar across countries and stable over time [143–145]. An alternative to HP is contingent valuation
(CV), which uses survey methods to explicitly determine individuals’ willingness to pay for noise
abatement, or alternatively, willingness to accept compensation for noise increases. While HP and
CV methods both have respective advantages, some key drawbacks which limit their applicability
to cost-benefit analysis of aviation policies include insufficient availability of real-estate data (HP)
[146], inconsistency of results between different study sites (CV) [147, 148], and prohibitive cost of
performing new valuation studies (both).
117
[149].3
The CAEP/9 noise stringency goals are achieved through a mixture of changes
to fleet composition and aircraft operation projected by the ICAO Forecasting and
Economic Support Group (FESG). For the purpose of policy cost-benefit analysis,
the noise impacts of the proposed stringency options are modeled for 99 airports in
the US using the FAA’s Aviation Environmental Design Tool (AEDT) [152]. These
99 airports are listed in Appendix D and belong to the FAA’s database of 185 Shell-1
airports worldwide that account for an estimated 91% of total global noise exposure
[153]. The noise contours for each airport are expressed in day-night average sound
level (DNL) in increments of 5 dB: 55–60 dB, 60–65 dB, 65–70 dB, 70–75 dB, and 75 dB
and above.4 The noise contours corresponding to each stringency option, along with
a baseline scenario with no stringency change, are forecasted from 2006 (the reference
year of aviation activity) to 2036, with a policy implementation year of 2020.
Along with forecasts of future noise, each CAEP/9 noise stringency option also
has an associated cost of implementation, which is the sum of the recurring and nonrecurring costs estimated by the FESG.5 Recurring costs are incurred annually and
consist of fuel costs, capital costs, and other direct operating costs, and were calculated using the Aviation environmental Portfolio Management Tool for Economics
(APMT-Economics) under the assumptions documented in Ref. [149]. Non-recurring
costs include loss in fleet value and cost to manufacturers, and are provided as low
and high estimates for each stringency option.
A policy is said to be cost-beneficial (resulting in a “net benefit”) if its associated
environmental benefits (typically negative) exceed the implementation costs (typically
3
The effective perceived noise in decibels (EPNdB) is the standard metric used in aircraft noise
certification limits, and represents the weighted sum of the sound pressure level from nine contiguous
frequency ranges, with added penalties for discrete pure tones and duration of noise [150, 151].
4
The day-night average sound level (DNL) is the 24-hour A-weighted equivalent noise level with a
10 dB penalty applied for nighttime hours. A similar measure, the day-evening-night average sound
level (DENL), is commonly used in Europe; DENL is very similar to DNL, except that it also applies
a 5 dB penalty to noise events during evening hours.
5
Recurring and non-recurring costs were estimated on a global scale. To obtain US-only policy
implementation costs, it was assumed that US operations comprise approximately 27% of the global
sum, based on previous analyses conducted using APMT-Economics [154].
118
positive), producing a negative value for net monetary impacts.6 Conversely, a policy
is not cost-beneficial (resulting in a “net dis-benefit”) if the net monetary impact is
a positive value.
The objectives of this case study are to explicitly model uncertainties in the
CAEP/9 noise stringency options and the APMT-Impacts Noise Module, understand
their implications for policy net cost-benefit, and identify resource allocation strategies for improving robustness and reliability. The outcomes of this study can help
inform decisions to support the design of cost-beneficial policy implementation responses. Prior work by Wolfe and Waitz used various modules within APMT-Impacts
to estimate the net present value of each stringency option with respect to noise, climate, and air quality impacts [155]. The cost of policy implementation was treated
as a deterministic value equal to the low estimate for each stringency and discount
rate. It was found that Stringencies 4 and 5 resulted in a net dis-benefit across all
discount rates considered, even though the lowest (least conservative) cost estimates
were used. In this case study we will use random variables to characterize uncertainty
in the implementation cost, and focus only on CAEP/9 Stringencies 1–3. Furthermore, we will limit the scope of environmental effects to consider only noise impacts.
Under these assumptions, it is expected that the present analysis will provide a more
conservative estimate of policy net cost-benefit for Stringencies 1–3.
5.2
APMT-Impacts Noise Module
The APMT-Impacts Noise Module is a suite of scripts and functions implemented in
R
the MATLAB
(R2013a, The MathWorks, Natick, MA) numerical computing envi-
ronment. The next sections describe the inputs, assumptions, and impacts estimation
procedure used in the module.
6
We note, however, that depending on the policy, environmental impacts can be either positive
or negative; the same is true of implementation costs.
119
5.2.1
Model Inputs
There are a number of inputs to the APMT-Impacts Noise Module, which can be
classified into three broad categories: scenario, scientific, and valuation. Below, we
describe each category and its respective inputs in greater detail.
Scenario Inputs
The scenario inputs correspond to the policy under consideration, and include demographic data for population and personal income, alternative forecasts of future
aviation activities or situations, and the cost of implementing the proposed policy.
As discussed in Section 5.1.2, noise contours from AEDT correspond to the DNL
of aircraft noise at a particular location, and are computed as yearly averages around
each airport. To evaluate the monetary impacts of a particular aviation policy, two
sets of noise contours are needed: baseline and policy. The baseline contours for the
reference year are constructed using actual aircraft movement data from a representative day of operations. The baseline forecast for future years represents an estimate
of the most likely future noise scenario while maintaining the status quo for technology, fleet mix, and aviation demand. The policy forecast reflects the expected future
noise levels after the implementation of the proposed policy. The monetary impacts
due to policy implementation correspond to the difference between the policy and the
baseline scenarios (henceforth termed the “policy-minus-baseline” scenario).
As will be further explained in Section 5.2.2, the APMT-Impacts Noise Module
assesses monetary noise impacts based on a per person willingness to pay (WTP)
for one decibel of noise reduction, which is computed as a function of the average
city-level personal income. Thus, locality-specific population and income data are
required. Income data may be obtained from a variety of sources, such as national
or local statistical agencies, or the US Bureau of Economic Analysis, which publishes
the annual personal income for each Metropolitan Statistical Area (MSA) [156]. The
income level for each of the 99 airports in the CAEP/9 analysis is listed in Appendix
D. Similarly, population data can be obtained from several sources, including the
decadal US Census, the European Environmental Agency’s population maps, the
120
Gridded Rural-Urban Mapping Project, and national or local statistical agencies.
For the CAEP/9 noise stringency analysis, noise and population data are provided
jointly in the form of the number of persons residing within each 5 dB contour band
surrounding an airport in the years 2006, 2026, and 2036. The population data were
derived from the 2000 US Census, and thus population changes since that time are
not accounted for. Finally, under the current capabilities of the APMT-Impacts Noise
Module, uncertainty associated with income and population data are not modeled.
The cost of policy implementation (IC) is another example of a scenario input. In
the CAEP/9 analysis, the estimated total cost for a particular stringency option can
be obtained by summing the recurring and non-recurring costs for a selected discount
rate. There are two cost totals for each stringency — low and high — corresponding
to the best- and worst-case estimates, respectively. To characterize uncertainty in the
implementation cost, we model this input using a uniform random variable on the
interval bounded by the low and high estimates.
Scientific Inputs
The scientific inputs enable the estimation of monetary impacts for a scenario of
interest, and consist of the background noise level (BNL), noise contour uncertainty
(CU), and the regression parameters (RP) that relate personal income to the WTP for
noise abatement. The uncertainties associated with these inputs arise from limitations
in scientific knowledge or modeling capabilities and are represented probabilistically
using random variables.
Background noise level refers to the level of ambient noise in an airport region,
which is important since monetary impacts are assessed only when the level of aviation
noise exceeds this threshold. Although BNL can vary from region to region, typical
values range between 50–60 dB DNL [143, 148]. To capture uncertainty in this input,
we model BNL as a symmetric triangular distribution between 50 dB and 55 dB, with
a mode of 52.5 dB.
Under the current capabilities of AEDT, the computed noise contours are fixed
values. In order to account for uncertainty in the noise contour level, we assume
121
a triangular distribution for CU with minimum, maximum, and mode at −2 dB,
2 dB, and 0 dB, respectively. This distribution represents an engineering estimate and
captures only uncertainty in the noise contour level, not in the area of the contour,
which can also affect the estimated monetary noise impacts [157].
500
0
−500
Intercept ($/dB/person)
1000
Mean = 39.11
SD = 159.65
−1000
2000
1500
1000
500
0
0
Income coefficient (1/dB/person)
−0.02
0
0.02
0.04
0.06
Number of samples
Number of samples
500
1000
Mean = 0.0114
SD = 0.0061
1500
2000
Figure 5-1: Joint and marginal distributions for the regression parameters obtained
using 10,000 bootstrapped samples
Finally, the regression parameters consist of two quantities, income coefficient and
intercept, which specify a linear relationship between income and WTP for 1 dB of
reduction in aviation noise. For US airports, this relationship is given by:
WTP = Income coefficient × Income + Intercept.
122
(5.1)
Since income coefficient and intercept are not independent, they are modeled jointly
as a random variable in R2 . The joint distribution and corresponding marginal distributions are shown in Figure 5-1.7 These distributions are approximately Gaussian
in shape, and were obtained by bootstrapping the 63 meta-analysis observations of
WTP versus per capita income and performing weighted least-squares regression using
a robust bisquare estimator [142].
Valuation Inputs
Finally, the valuation inputs pertain to subjective judgments regarding the value of
money over time or the significance of environmental impacts, rather than quantities
rooted in scientific knowledge. These factors are modeled as deterministic values instead of random variables. The valuation inputs of the APMT-Impacts Noise Module
consist of the discount rate, income growth rate, and significance level.
The discount rate (DR) captures the depreciation in the value of money over
time, and is expressed as an annual percentage. It is closely related to the time
span of the analyzed policy, which is based on the typical economic life of a building
and the duration of future noise impacts that is considered by the house buyers. In
this analysis, the policy time span is 30 years (2006–2036). Consistent with previous
analyses of the CAEP/9 noise stringency options, the nominal discount rate is selected
to be 3% [155]. However, other discount rates (5%, 7%, 9%) will also be considered
to investigate the effect of this parameter on net cost-benefit. We assume that the
same discount rate is used to compute the net present value of both monetary noise
impacts and recurring implementation costs.
The income growth rate (IGR) represents the annual rate of change in the citylevel average personal income. It is universally applied to the income levels of all
7
In the case of non-US airports, a third regression parameter, the interaction term, is added to
(5.1), such that WTP is given by:
WTP = (Income coefficient + Interaction term) × Income + Intercept.
(5.2)
The interaction term has an approximately Gaussian marginal distribution with mean and standard
deviation equal to 39.11 and 159.65, respectively [142]. For this US-based case study, however, we
will not further discuss the interaction term.
123
airports in the analysis when calculating the WTP for noise abatement. While this
parameter may be user-selected to be any reasonable value (even negative growth
rates), in this analysis it is set to zero so as to consider noise impacts solely due to
the growth of aviation, rather than due to changes in economic activity.
The significance level (SL) is the threshold DNL above which aircraft noise is
considered to have significant impact on the surrounding community. It does not affect
the value of the computed monetary noise impacts, but rather designates impacts as
significant or insignificant, and thereby includes or excludes them from the reported
results. In this analysis, the significance level is set to 52.5 dB, as consistent with
the midrange lens definition in the APMT-Impacts Noise Module.8 Noting that the
lowest noise contour level in the CAEP/9 analysis is 55 dB, this choice implies that
any aviation noise above the background noise level is perceived as having a significant
impact on the community.
Symbol
Factor (Units)
Definition
BNL
Background noise level (dB)
T (50, 55, 52.5)
CU
Contour uncertainty (dB)
T (−2, 2, 0)
Income coefficient (1/dB/person)
Mean = 0.0114, SD = 0.0061
Intercept ($/dB/person)
Mean = 39.11, SD = 159.65
SL
Significance level (dB)
52.5 dB
IGR
Income growth rate (%)
0%
IC
Implementation cost ($)
U [low estimate, high estimate]
DR
Discount rate (%)
3%, 5%, 7%, 9%
RP
Table 5.1: Inputs to CAEP/9 noise stringency case study. The definitions for BNL,
CU, RP, SL, and IGR correspond to the midrange lens of the APMT-Impacts Noise
Module, whereas IC and DR are used in the post-processing of impacts estimates to
compute net cost-benefit.
Table 5.1 summarizes the main model inputs presented in this section. We note
that with the exception of implementation cost, which is added directly to the net
present value of monetary noise impacts (to be discussed in Section 5.2.2), all inputs
8
A lens is a ready-made set of input values that can be used to evaluate policy options given a
particular perspective or outlook. The midrange lens represents a most likely scenario, where all
model inputs are set to their nominal value or distribution.
124
listed in the table are used in impacts estimation for both the baseline and policy
scenarios. Thus, while uncertainties in these inputs affect the baseline or policy
scenario results to the first-order, when considering a policy-minus-baseline scenario,
their effects become second-order. Finally, for each of the probabilistic inputs (BNL,
CU, RP, and IC), 10,000 MC samples are used to characterize the distribution.9
5.2.2
Estimating Monetary Impacts
In this case study, we use the APMT-Impacts Noise Module to assess the net costbenefit of each CAEP/9 noise stringency option, which we estimate as the net present
value of the sum of the monetary noise impacts and implementation cost. A schematic
of the module is shown in Figure 5-2.
For the k th airport in the analysis, the capitalized monetary impacts of aviation
in year t in the dth 5 dB noise contour band is given by V (k, d, t):
V (k, d, t) = WTP(k, t) × ∆dB(k, d, t) × Number of persons(k, d, t),
(5.3)
where WTP(k, t) is obtained using (5.1), and ∆dB(k, d, t), the aviation noise level
above the BNL, is computed as:
∆dB(k, d, t) = Noise contour level(k, d, t) + CU − BNL.
(5.4)
To estimate the total capitalized monetary impacts in year t, we sum V (k, d, t) over
all noise contour bands and across all airports:
V (t) =
XX
k
V (k, d, t).
(5.5)
d
It is important to note that V (t) is a capitalized value that encompasses not only
the property value depreciation due to aviation noise in year t, but also the future
9
In order to check for convergence of the MC samples, the running mean and variance of each
input were plotted versus the sample number. After 10,000 samples, fluctuations in the running
mean and variance were on the order of 0.1% for all inputs.
125
Model Inputs
Scenario
Scientific &
Valuation
Quantities of
Interest
Contour
Uncertainty
Exposure
Area
(dB)
(m2)
(dB)
AEDT
Background
Noise Level
Exposed
Population
Exposed
Population
(dB)
(Persons)
(Persons)
(Persons/m2)
US Census,
EEA, GRUMP
Significance
Level
ΔdB
Personal
Income
Income
Growth Rate
Willingness
to Pay
Capitalized
Impacts
($)
USBEA,
Statistical
Agencies
(%)
($/Person/dB)
($)
Regression
Parameters
Capital
Recovery
Factor
Annual
Impacts
Implementation Cost
($)
Net Present
Value
Physical Impacts
Noise
Contours
Intermediate
Results
Population
Data
(dB)
(dB)
($)
APMTEconomics
Discount
Rate
(%)
($)
Monetary Impacts
Implementation Cost
($)
Figure 5-2: Schematic of APMT-Impacts Noise Module [142, adapted from Figure 4]
noise damages anticipated by the house buyers. This quantity can be transformed
into a net present value (NPV) by applying a selected discount rate:
NPV = V (0) +
T
X
V (t) − V (t − 1)
t=1
(1 + DR)t
,
(5.6)
where V (0) denotes the capitalized impacts in the reference year (2006), and T = 30
is the policy time span. Because baseline and policy scenarios are provided for only a
subset of the years in the policy timespan (2006, 2026, 2036), the year-by-year results
are obtained by linearly interpolating the monetary impacts between fixed contour
years.
We select the quantity of interest for this case study to be the net cost-benefit
of each noise stringency option, given as the net present value of the policy-minus-
126
baseline scenario plus the corresponding cost of implementation. This quantity can
be written as:
Y = NPVpolicy − NPVbaseline + IC.
(5.7)
All monetary values are expressed year 2006 USD. The complexity of Y represents
the intrinsic extent of the distribution for the net monetary impacts of the proposed
stringency option. The risk associated with Y is defined as P (Y > 0), and can be
interpreted as the probability that the stringency option is not cost-beneficial.
5.3
5.3.1
Results and Discussion
Comparison of Stringency Options and Discount Rates
NPV (Billion 2006 USD)
4
3
2
1
0
-1
-2
Str 1
DR = 3%
Str 2
DR = 5%
DR = 7%
Str 3
DR = 9%
Figure 5-3: Comparison of mean estimates for Y in Stringencies 1–3 and discount
rates between 3%–9%. Error bars denote 10th , 90th percentile values.
Figure 5-3 shows a comparison of the mean net monetary impacts of Stringencies
1–3 at discount rates of 3%, 5%, 7%, and 9%. Table 5.2 provides the corresponding
uncertainty estimates. We see that across all discount rates considered, the mean of
Y is negative in Stringency 2, positive in Stringency 3, and mixed in Stringency 1.
The complexity of Y remains relatively constant within each stringency, but increases
significantly between the stringencies. This suggests that an increase in noise stringency is concurrent with a growth in complexity (and thus, a decrease in robustness).
127
Str.
Stringency 1
DR
3%
Mean
-0.12
0
10th
-0.31
-0.19
th
0.07
0.19
SD
0.14
Comp.
Risk
90
5%
7%
Stringency 2
Stringency 3
9%
3%
5%
7%
9%
3%
5%
7%
9%
0.08
0.13
-1.19
-0.72
-0.41
-0.19
2.91
2.29
1.86
1.56
-0.11
-0.06
-1.42
-0.94
-0.63
-0.42
2.42
1.80
1.37
1.07
0.26
0.32
-0.96
-0.50
-0.19
0.03
3.41
2.78
2.35
2.05
0.14
0.14
0.14
0.17
0.16
0.16
0.16
0.37
0.36
0.36
0.35
0.56
0.54
0.53
0.53
0.70
0.66
0.64
0.63
1.49
1.43
1.40
1.38
24%
49%
66%
77%
0%
0%
0%
15%
100% 100% 100% 100%
Table 5.2: Comparison of mean and uncertainty estimates of Y for Stringencies 1–
3 and discount rates between 3%–9%. All monetary values (mean, 10th and 90th
percentiles, standard deviation, and complexity) are listed in 2006 USD (billions).
The risk associated with Y increases with discount rate within Stringency 1, equals
zero in Stringency 2 (except when DR= 9%), and remains at 100% in Stringency 3.
From these results, we conclude that under the current assumptions, Stringency 2
is virtually always cost-beneficial and Stringency 3 is never cost-beneficial. The most
interesting option is Stringency 1, which straddles the border, with risk estimates
ranging between 24–77%. Because of this, we will focus our attention on Stringency
1, and in the following section investigate the factors that drive its uncertainty.
5.3.2
Global and Distributional Sensitivity Analysis
Figure 5-4(a) shows global sensitivity analysis results for Stringency 1 when a discount
rate of 3% is assumed. With a MSI of 0.93, implementation cost is by far the largest
contributor to output variability. On the other hand, the scientific inputs (BNL,
CU, and RP), along with factor interactions (Int.), each contributes only a minuscule
proportion. This apportionment result does not change when the discount rate is
increased (see Table 5.3, left side). Thus, in the remainder of this section, we will
focus on the case where DR = 3%.
The distributional sensitivity analysis results for this case are shown in Figure
5-5(a). We see that the adjusted main effect sensitivity index for each factor varies
linearly with δ. If we permit BNL, CU, RP, and IC to deviate from their nominal
distribution families specified in Table 5.1, the resulting effect on adjSi (δ) is negli-
128
2%
3%1% 1%
4%
32%
43%
93%
BNL
CU
RP
IC
Int.
BNL
CU
RP
Int.
20%
(a) IC ∼ U[low est., high est.]
(b) IC = mean estimate
Figure 5-4: Main effect sensitivity indices for Stringency 1, DR = 3%, with IC modeled probabilistically in (a) and as a deterministic value in (b)
IC ∼ U[low est., high est.]
IC = mean estimate
DR
3%
5%
7%
9%
3%
5%
7%
9%
BNL
0.02
0.01
0
0
0.32
0.32
0.32
0.32
CU
0.01
0
0
0
0.20
0.20
0.20
0.20
RP
0.03
0.01
0
0
0.43
0.43
0.43
0.43
IC
0.93
0.97
0.99
0.99
N/A
N/A
N/A
N/A
Int.
0.01
0.01
0.01
0.01
0.04
0.04
0.04
0.04
Table 5.3: Main effect sensitivity indices for Stringency 1 at various discount rates
gible.10 Thus, we can conclude that irrespective of distribution shape, the relative
factor contribution to output variance does not change with δ.
Next, we investigate the influence of our initial assumption that IC is a uniform
random variable on the interval delimited by the low and high cost estimates. Figures
5-6(a) through 5-6(c) show the distribution of Y for 10,000 MC samples when IC is
assumed to be a uniform, symmetric triangular, or Gaussian random variable on this
interval.11 The variance of IC is decreased steadily (by 50% between the uniform and
10
In DSA, the regression parameters are modeled as a multivariate Gaussian random variable
(MVN) in R2 parameterized by the vector of means M and the covariance matrix Σ. This is an
approximation since the RP distribution is not truly Gaussian (as can be seen from Figure 5-1).
Furthermore, we have not permitted changes to the RP distribution family, as its shape is specified
according to empirical evidence rather than modeling assumptions.
11
Consistent with the approach in Chapter 4, we approximate the Gaussian distribution with
129
0.8
0.8
0.6
0.6
adjSi()
1
i
adjS ()
1
0.4
0.2
0
0
BNL: Tri->Uni
BNL: Tri->Tri
BNL: Tri->Gaussian
BNL: Tri->Bmd Uni
CU: Tri->Uni
CU: Tri->Tri
CU: Tri->Gaussian
CU: Tri->Bmd Uni
IC: Uni->Uni
IC: Uni->Tri
IC: Uni->Gaussian
IC: Uni->Bmd Uni
RP: MVN->MVN
0.4
0.2
0.2
0.4
0.6
0.8
0
0
1
0.2
1-
0.4
1-
0.6
0.8
1
(a) IC ∼ U[low est., high est.]
1
adjSi(δ)
0.8
0.6
0.4
0.2
0
0
0.2
0.4
0.6
0.8
1
1−δ
(b) IC = mean estimate
Figure 5-5: Adjusted main effect sensitivity indices for Stringency 1, DR = 3%, with
IC modeled probabilistically in (a) and as a deterministic value in (b)
triangular cases, and by 66.7% between the uniform and Gaussian cases), which in
turn reduces the standard deviation and complexity of Y . Since µY < 0, a reduction in var (Y ) also decreases P (Y > 0), hence increasing the probability that the
policy will result in a net benefit (i.e., improving reliability).12 Despite these significant reductions in uncertainty, however, IC still remains the largest driver of output
variability, as its main effect sensitivity index (SIC ) equals 0.86 for the symmetric
triangular case and 0.82 for the Gaussian case.
In the limiting case, Figure 5-6(d) shows the distribution of Y when IC is fixed
at a deterministic value equal to the mean cost estimate for Stringency 1. Compared
to the probabilistic cases, there are significant reductions in the standard deviation
a truncated version in which ±3 standard deviations of the distribution fall within the specified
interval.
12
We note that this outcome is not necessarily true of other stringencies and discount rates in the
CAEP/9 analysis. For example, when DR = 7% in Stringency 1, µY > 0 (see Table 5.2), and thus
a decrease in var (Y ) actually increases risk.
130
2500
Mean = $−0.12B
SD = $0.14B
Comp = $0.56B
Risk = 23.9%
2000
Number of samples
Number of samples
2500
1500
SIC = 0.93
1000
500
2000
1500
SIC = 0.86
1000
500
0
−0.6
−0.4
−0.2
0
0.2
0
−0.6
0.4
Net present value (Billion 2006 USD)
−0.4
−0.2
0
0.2
0.4
Net present value (Billion 2006 USD)
(a) IC ∼ U[low est., high est.]
(b) IC ∼ T [low, high, mean]
2500
2500
Mean = $−0.12B
SD = $0.09B
Comp = $0.36B
Risk = 7.3%
2000
Number of samples
Number of samples
Mean = $−0.12B
SD = $0.10B
Comp = $0.43B
Risk = 12.7%
1500
SIC = 0.82
1000
500
2000
Mean = $−0.11B
SD = $0.04B
Comp = $0.15B
Risk = 0%
1500
1000
500
0
−0.6
−0.4
−0.2
0
0.2
0
−0.6
0.4
Net present value (Billion 2006 USD)
−0.4
−0.2
0
0.2
0.4
Net present value (Billion 2006 USD)
(c) IC ∼ N [mean, (high − low)/6]
(d) IC = mean estimate
Figure 5-6: Histogram of Y under various IC assumptions for Stringency 1, DR =
3%
and complexity of Y . Furthermore, the risk of a net policy dis-benefit is reduced
to zero. Figure 5-4(b) shows the GSA results for the deterministic IC case. Of the
scientific inputs, the regression parameters contribute the largest proportion to output
variance at 43%, followed by BNL at 32% and CU at 20%, with interactions again
playing a minor role. As in the probabilistic case, the variance apportionment for
the deterministic IC case remains constant across all discount rates considered (see
Table 5.3, right side). Likewise, the DSA results in Figure 5-5(b) reveal that the
adjusted main effect sensitivity indices tend to vary linearly with δ, and incur only
small deviations when changes between distribution families are considered.
131
5.3.3
Visualizing Tradeoffs
Building off of the variance-based sensitivity analyses from the previous section, we
now investigate how changes in the mean or variance of the model inputs affect complexity and risk in the QOI. For this analysis, we focus on BNL, CU, and IC, whose
distributions are specified according to engineering assumptions. The regression parameters are excluded, as additional research is needed to extend local sensitivity
analysis and visualization techniques to multivariate random variables.
0.2
0.7
0.2
0.8
0.6
IC mean ($B)
0.1
0
−0.1
50
0.56
0.57
−0.2
0.55
IC mean ($B)
0.1
51
52
53
0
0.7
0.5
0.6
0.4
0.5
0.3
0.4
0.2
0.3
0.1
0.2
−0.1
0.1
−0.2
54
55
50
51
52
BNL mean (dB)
53
54
55
BNL mean (dB)
(a) Lines of constant complexity due to changes (b) Lines of constant risk due to changes in
in mean
mean
1
0.25
0.8
Percent reduction in var(IC)
Percent reduction in var(IC)
1
0.25
0.3
0.3
0.35
0.35
0.4
0.4
0.6
0.4
0.2
0
0
0.45
0.45
0.5
0.5
0.55
0.2
0.4
0.55
0.6
0.8
0.6
0.04
0.08
0.08
0.12
0.16
0.12
0.16
0.16
0.4
0.2
0.08
0.04
0.12
0.2
0.2
0
0
1
Percent reduction in var(BNL)
(c) Lines of constant complexity due to
reductions in variance
0.08
0.8
0.12
0.16
0.2
0.2
0.2
0.4
0.6
0.8
1
Percent reduction in var(BNL)
(d) Lines of constant risk due to reductions in
variance
Figure 5-7: Uncertainty contours associated with changes in the mean and variance
of BNL and IC, generated from 10,000 MC simulations (solid colored lines) or approximated using sensitivity analysis results (dashed black lines)
In Figures 5-7 through 5-9, we let the mean of BNL, CU, and IC vary between the
132
0.2 0.8
0.7
0.2
IC mean ($B)
0.1
0
−0.1
−2
−1
0.57
0.55
−0.2
0.56
IC mean ($B)
0.1
0.7
0
0
0.6
0.6
0.5
0.5
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
−0.1
−0.2
1
2
−2
−1
CU mean (dB)
0
1
2
CU mean (dB)
(a) Lines of constant complexity due to changes (b) Lines of constant risk due to changes in
in mean
mean
1
0.25
0.8
0.6
0.4
0.2
0.25
0.3
0.3
0.35
0.35
0.4
0.4
0.45
0.45
0.5
0.5
0.55
0
0
0.2
0.4
0.55
0.6
0.8
Percent reduction in var(IC)
Percent reduction in var(IC)
1
0.8
0.6
0.08
0.12
0.16
0.12
0.16
0.16
0.4
0.2
0.08
0.04
0.12
0.08
0.04
0.2
0.2
0
0
1
Percent reduction in var(CU)
(c) Lines of constant complexity due to
reductions in variance
0.08
0.12
0.16
0.2
0.2
0.2
0.4
0.6
0.8
1
Percent reduction in var(CU)
(d) Lines of constant risk due to reductions in
variance
Figure 5-8: Uncertainty contours associated with changes in the mean and variance
of CU and IC
lower and upper bounds of their respective distributions, as well as allow their variance
to be reduced by 0–100% (assuming that such reductions decrease the extent of the
underlying distribution symmetrically about the mean). No changes in distribution
family are considered. The solid colored lines denote contours generated using 10,000
MC simulations at each design parameter combination, whereas the dashed black
lines indicate approximations based on distributional, local, or entropy power-based
sensitivity analysis results. A black X in each figure designates the location of the
initial case as defined in Table 5.1.
Figures 5-7 and 5-8 show how changes in IC trade off against BNL and CU,
respectively. Figures 5-7(a) and 5-8(a) reveal that varying the mean implementation
133
2
2
9
56
.5
1.5 0
0.
1
56
0.
54
0
25
15
0.
.
−0.5 0
−1
0.
0.57
55
−1
0.5
−2
50
51
52
−1.5
53
54
0.
0.2
−2
50
55
35
0
−0.5
3
0.
2
0.
0.
0.5
5
5
0.
CU mean (dB)
0.
.
1 0
7
5
0.
58
0.
CU mean (dB)
1
−1.5
25
0.
15
0.
1.5
51
52
BNL mean (dB)
3
53
54
55
BNL mean (dB)
0.4
0.6
0.8
0.238
0.2
37
0.2
0.237
7
23
0.2
36
38
0.2
0.237
0.4
0
0
1
Percent reduction in var(BNL)
(c) Lines of constant complexity due to
reductions in variance
0.6
0.
55
0.2
38
37
0.
0
0
56
0.2
0.2
0.2
0.4
0.8
0.2
Percent reduction in var(CU)
0.6
0.
Percent reduction in var(CU)
5
5
0.5
0.5
0.8
0.23
1
37
1
7
(a) Lines of constant complexity due to changes (b) Lines of constant risk due to changes in
in mean
mean
0.2
0.4
0.6
0.8
1
Percent reduction in var(BNL)
(d) Lines of constant risk due to reductions in
variance
Figure 5-9: Uncertainty contours associated with changes in the mean and variance
of BNL and CU
cost estimate has no effect on output complexity. This is unsurprising since IC is
simply added to the NPV of the policy-minus-baseline noise impacts according to
(5.7), and thus contributes only a constant shift to the estimates of Y . However,
when the mean value of IC is decreased, the risk of incurring a net policy dis-benefit
is greatly reduced (Figures 5-7(b) and 5-8(b)). A reduction in the mean BNL value
slightly increases complexity and decreases risk, whereas a reduction in the mean
CU value slightly decreases complexity and increases risk. This incongruity can be
explained by (5.4), which states that a decrease in BNL results in a larger value of
∆dB and thus a higher level of estimated noise impacts, whereas a decrease in CU
has the opposite effect. Figures 5-7(c), 5-7(d), 5-8(c), and 5-8(d) show that reducing
134
the variance of IC greatly decreases both complexity and risk in the QOI. This effect
is so dominant that reductions in the variance of BNL or CU have negligible impact
by comparison. These results corroborate the findings of variance-based sensitivity
analysis, and confirm that the most effectual strategy for reducing uncertainty in
policy net cost-benefit is to invest additional resources toward improving estimates
of implementation cost.
Finally, Figure 5-9 illustrates the tradeoffs in output complexity and risk associated with changes in the mean or variance of BNL and CU. Consistent with the trends
in Figures 5-7(a), 5-7(b), 5-8(a), and 5-8(b), we see that complexity decreases when
the mean of BNL is increased, or when the mean of CU is decreased, whereas risk
decreases when the mean of BNL is decreased, or when the mean of CU is increased.
Figures 5-9(c) and 5-9(d) show that reductions in the variance of BNL and/or CU
have little effect on complexity, and virtually no effect on risk. These results validate
our conclusion that from an uncertainty mitigation point of view, it is not worthwhile
to expend resources to learn more about background noise level or noise contour uncertainty; instead, efforts should be directed toward reducing uncertainty in policy
implementation cost.
5.4
Chapter Summary
In this chapter, we applied the uncertainty and sensitivity analysis methods developed
in this thesis to study the net cost-benefit of the CAEP/9 noise stringency options
as estimated using the APMT-Impacts Noise Module. Based on our analysis, we
conclude that among all sources of uncertainty considered, scenario uncertainty in
the form of forecasted policy implementation cost outpaces scientific uncertainty in
the noise monetization model by an order of magnitude. This result holds true across
all stringencies and discount rates considered, as well as for several reasonable specifications of input distribution shape. Additional sensitivity analyses revealed that
as defined, the scientific inputs of the APMT-Impacts Noise Module comprise only a
small contribution to the complexity and risk associated with the net cost-benefit of
135
the proposed policy options. For Stringency 1, we recommend that the best course of
action for ensuring net policy benefit is to improve economic forecasting capabilities
and obtain more precise estimates of implementation cost.
In this case study, we have also highlighted the disparity in uncertainty estimates
for policy cost-benefit when implementation cost is modeled probabilistically versus as
a deterministic value, which represents an extension to previous cost-benefit analyses
of the CAEP/9 noise stringency options. We note, however, that a key difference
between this work and the prior analysis by Wolfe and Waitz lies in having modeled
only the monetized noise impacts of the proposed stringency options, such that the
net present value estimates do not capture the full suite of environmental benefits
associated with policy implementation. A natural extension of the current work is to
incorporate the climate and air quality co-benefits of the CAEP/9 noise stringency
options, as well as characterize the effects of uncertainty in those estimates.
136
Chapter 6
Conclusions and Future Work
The overarching goal of this thesis was to develop methods that help quantify, understand, and mitigate the effects of uncertainty in the design of engineering systems. In
this chapter, we summarize the progress made toward the thesis objectives, highlight
key intellectual contributions, and suggest directions for future research.
6.1
Thesis Summary
For many engineering systems, current design methodologies do not adequately quantify and manage uncertainty as it arises during the design process, which can lead
to unacceptable risks, increases in programmatic cost, and schedule overruns. In this
thesis, we developed new methods for uncertainty quantification and sensitivity analysis, which can be used to better understand and mitigate the effects of uncertainty.
In Chapter 2, we defined complexity as the potential of a system to exhibit unexpected behavior in the quantities of interest, and proposed the use of exponential
entropy to quantify complexity in the design context. This complexity metric can be
used alongside other measures of uncertainty, such as variance and risk, to afford a
richer description of system robustness and reliability.
In Chapter 3, we extended existing variance-based global sensitivity analysis to
derive an analogous interpretation of uncertainty apportionment based on entropy
power, which is proportional to complexity squared. This derivation also revealed
137
that a factor’s contribution to output entropy power consists of two effects, which
are related to its spread (as characterized by intrinsic extent) versus distribution
shape (as characterized by non-Gaussianity). To further examine the influence of
input distribution shape on estimates of output variance, we also broadened the
existing distributional sensitivity analysis methodology to accommodate changes to
distribution family.
Another key aspect the work in this thesis is the development of local sensitivity
analysis techniques to predict changes in the complexity and risk of the quantities of
interest resulting from various design modifications. These local approximations are
particularly useful for systems whose simulation models are computationally expensive, as they can be used to identify tradeoffs and infer trends in the design space
without performing additional model evaluations. Furthermore, they can be connected with budgets for uncertainty and cost in order to identify options for improving robustness and reliability and inform decisions regarding resource allocation. In
Chapters 4 and 5, we demonstrated the uncertainty and sensitivity analysis methods
developed in this thesis on two engineering examples.
6.2
Thesis Contributions
The main contributions of this thesis are:
• The proposal of an information entropy-based complexity metric, which can be
used in the design context to supplement existing measures of uncertainty with
respect to specified quantities of interest;
• The development of local sensitivity analysis techniques that can approximate
changes in complexity and risk resulting from modifications to design parameter
mean or variance, as well as enable the visualization of design tradeoffs;
• The identification of intrinsic extent and non-Gaussianity as distinct features
of a probability distribution that contribute to complexity, as well as the development of techniques to analyze the effect of distribution shape on uncertainty
138
estimates;
• A demonstration of how uncertainty estimates and sensitivity information can
be combined with design budgets to guide decisions regarding the allocation of
resources to improve system robustness and reliability.
6.3
Future Work
There are a number of possible extensions to the research presented in this thesis.
The first is to broaden the application of the uncertainty and sensitivity analysis
methods to problems of higher dimensionality. Although the theory underlying many
of the techniques (e.g., estimation of complexity and entropy power, characterizing
non-Gaussianity using K-L divergence) can easily be extended to higher dimensions,
practical issues often hinder their applicability. For example, density estimation —
upon which entropy estimation is prefaced — is typically straightforward in 1-D; for
higher dimensions, however, it becomes much more challenging, or even downright
intractable. High-dimensional density estimation is an active area of research; it is
expected that advancements in this field will allow the methods developed in this
thesis to be applied to larger classes of engineering problems.
A related issue in computational tractability is the expense of pseudo-random
Monte Carlo simulation. One avenue of future work could lead to the incorporation
of efficient sampling methods, such as Latin hypercube or quasi-random Monte Carlo,
to reduce computational cost. In particular, it would be interesting to study how
such techniques can be combined with acceptance/rejection sampling or importance
sampling to enable sample reuse in distributional sensitivity analysis.
In this thesis, we focused primarily on characterizing the downstream effects of parameter uncertainty, which is due to not knowing the true values of the model inputs.
Another important area for future research is to extend the proposed system design
framework to consider other sources of uncertainty, such as model inadequacy and
observation error. This issue is particularly relevant for systems such as the CAEP/9
noise stringency case study presented in Chapter 5, in which we discovered that by
139
modeling implementation cost probabilistically, we identified a major contributor to
uncertainty in policy net cost-benefit. Within the context of the APMT-Impacts
Noise Module, other potential sources of uncertainty include the spatial and temporal resolution of the AEDT noise contours (which are currently provided in increments
of 5 dB for only three of the years in the 30-year policy duration), as well as changes
in population and income level during the analysis timespan. By constructing a more
accurate picture of the various sources of uncertainty present in a system, we can
better understand their effects on the quantities of interest, as well as take steps to
mitigate potentially detrimental impacts in order to ensure system robustness and
reliability.
140
Appendix A
Bimodal Distribution with
Uniform Peaks
In Section 1.2, we introduced a bimodal distribution consisting of two uniform peaks
(i.e., a “bimodal uniform” distribution), parameterized by six quantities: aL , bL , fL ,
aR , bR , and fR . Below, we reproduce Figure 1-3(d) as well as the expressions for the
𝑓𝑋 𝑥
mean and variance of such a distribution from Table 2.1.
𝑓𝐿
𝑎𝐿
𝑓𝑅
𝑏𝐿
𝑥
𝑎𝑅
𝑏𝑅
Figure A-1: Bimodal uniform distribution with peak bounds aL , bL , aR , and bR , and
peak heights fL and fR
fL 2
(b − a2L ) +
2 L
fL
σ 2 = (b3L − a3L ) +
3
µ=
fR 2
(b − a2R )
2 R
fR 3
(b − a3R ) − (E [X])2
3 R
141
(A.1)
(A.2)
A.1
Specifying Distribution Parameters
Because there are five degrees of freedom in the parameter specification, mean and
variance alone are not sufficient to uniquely define a bimodal uniform random variable.
Thus, we must make additional assumptions about the shape of the distribution. In
this thesis, we assume that the two peaks are symmetric about the mean and that
each peak has width equal to 1/4 of the extent of a uniform random variable with
the same variance. This allows us to compute aL , bL , fL , aR , bR , and fR using the
following procedure.
Algorithm 1 Specify a bimodal uniform random variable X given µ and σ
1. Find the lower and upper bound of a uniform distribution with mean µ and
standard deviation σ:
√
⇒
a = µ − (b − a)/2, b = µ + (b − a)/2
b − a = σ 12
2. Assume that the peaks of X are symmetric, each with a width equal to (b−a)/4.
Thus, the peak heights are fL = fR = 2/(b − a).
3. Let b∗R = b and a∗R = 43 (b − a) + a.
4. Find a∗L and b∗L such that the bimodal uniform random variable X ∗ ∼
BU{a∗L , b∗L , fL , a∗R , b∗R , fR } has a variance of σ 2 . This requires solving the following system of equations:
(b − a)/4 = b∗L − a∗L
2
fL ∗3
fR ∗3
fL ∗2
fR ∗2
∗3
∗3
∗2
∗2
σ = (bL − aL ) + (bR − aR ) −
(b − aL ) + (bR − aR )
3
3
2 L
2
2
5. The mean of X ∗ is µ∗ = (b∗L + a∗R )/2. Compute the difference ∆µ = µ∗ − µ.
Then, shift each of a∗L , b∗L , a∗R , and b∗R by ∆µ to obtain aL , bL , aR , and bR :
aR = a∗R − ∆µ
bR = b∗R − ∆µ
aL = a∗L − ∆µ
bL = b∗L − ∆µ
6. Define X ∼ BU(aL , bL , fL , aR , bR , fR ).
142
A.2
Options for Variance Reduction
Algorithm 1 can be used to define a symmetric bimodal uniform random variable
o
X BU ∼ BU(aoL , boL , fLo , aoR , boR , fRo ) with mean µo and variance σ o2 . For a given value
of δ, we desire to specify parameters for a new bimodal uniform random variable X 0 ∼
BU(a0L , b0L , fL0 , a0R , b0R , fR0 ) such that σ 02 = δσ o2 . Some possible options for variance
reduction include:
1. Decreasing the width of one or both peaks.
2. Decreasing the separation between the peaks.
3. Increasing the height of one peak versus the other.
In this section, we will focus on Options 2 and 3, as Option 1 is typically not
sufficient to reduce variance to the requisite level for most values of δ. Algorithms 2
and 3 provide the procedures to define X 0 using Options 2 and 3, respectively. Note
that neither option is valid for the full range of possible values δ ∈ [0, 1]; instead, each
option is applicable for δ ∈ (δ ∗ , 1].
Using Option 2, the minimum achievable variance occurs when there is no separation between the peaks (i.e., the symmetric bimodal uniform distribution becomes
a uniform distribution), such that b0L = a0R . In this case, σ 02 = (boL − aoL )2 /3 =
(boR − aoR )2 /3. Similarly, using Option 3, variance is minimized when the height of
one peak is reduced to zero while that of the other is increased, such that σ 02 =
(boL − aoL )2 /12 = (boR − aoR )2 /12. The values of δ ∗ for these options are listed below. Finally, we note that Algorithm 1 can also be used to specify a new symmetric bimodal
o
uniform random variable X 0 based on X BU for all values of δ ∈ [0, 1].
Algorithm 1:
Algorithm 2:
Algorithm 3:
δ∗ = 0
(boL − aoL )2
3σ o2
o
(bL − aoL )2
∗
δ =
12σ o2
δ∗ =
143
Algorithm 2 Decrease peak separation to achieve σ 02 = δσ o2
1. Set peak heights fL0 = fLo and fR0 = fRo .
2. Select at random which peak to move. If the left peak is chosen, go to Step 3.
Else, go to Step 4.
3. Set a0R = aoR and b0R = boR . Find a0L and b0L by solving the following system of
equations:
b0L − a0L =
δσ
o2
fRo o
(b − aoR )
fLo R
o
2
fLo 03
fRo o3
fL 02
fRo o2
03
o3
02
o2
= (bL − aL ) + (bR − aR ) −
(b − aL ) + (bR − aR )
3
3
2 L
2
4. Set a0L = aoL and b0L = boL . Find a0R and b0R by solving the following system of
equations:
b0R − a0R =
δσ
o2
fLo o
(bL − aoL )
o
fR
o
2
fL o2
fRo 03
fRo 02
fLo o3
o3
03
o2
02
(b − aL ) + (bR − aR )
= (bL − aL ) + (bR − aR ) −
3
3
2 L
2
5. Define X 0 ∼ BU(a0L , b0L , fL0 , a0R , b0R , fR0 ).
Algorithm 3 Vary peak heights achieve σ 02 = δσ o2
1. Set peak bounds a0L = aoL , b0L = boL , a0R = aoR , and b0R = boR ,
2. Find fL0 and fR0 by solving the following system of equations:
1 = fL0 (boL − aoL ) + fR0 (boR − aoR )
δσ
o2
0
2
fL0 o3
fR0 o3
fL o2
fR0 o2
o3
o3
o2
o2
= (bL − aL ) + (bR − aR ) −
(b − aL ) + (bR − aR )
3
3
2 L
2
3. Select at random which peak height to decrease. If the left peak is chosen, set
fL0 equal to the smaller of {fL0 , fR0 } obtained in Step 2, and fR0 equal to the
larger. Otherwise, set fR0 to equal the smaller of {fL0 , fR0 }, and fL0 equal to the
larger.
4. Define X 0 ∼ BU(a0L , b0L , fL0 , a0R , b0R , fR0 ).
144
o
Figure A-2 shows a bimodal uniform distribution X BU obtained using Algorithm
1 for µo = 0.5 and σ o = 0.29 as well as several examples of bimodal uniform distributions X 0 obtained using Algorithms 1–3 for δ = 0.5. Each realization of X 0 has
σ 0 = 0.20.
5
o
Alg. 1 (XBU )
Alg. 1 (X’)
Alg. 2 (X’)
Alg. 3 (X’)
Probability density
4
3
2
1
0
0
0.2
0.4
0.6
0.8
1
x
Figure A-2: Examples of distributions for X 0 obtained using Algorithms 1–3 (solid
o
lines), based on an initial distribution X BU (dashed green line) and δ = 0.5
A.3
Switching to a Bimodal Uniform Distribution
In this section, we outline the procedure for switching from a uniform, triangular,
or Gaussian random variable X o to a bimodal uniform random variable X 0 . As in
Appendix 3.2, the key steps are to estimate the parameters of X 0 and to specify κ
so that the acceptance/rejection method can be used to generate samples of X 0 from
samples of X o . Examples of reasonable bimodal uniform distributions generated using
the procedure below are shown in Figure A-3.
Selecting a0L , b0L , fL0 , a0R , b0R , and fR0 :
o
1. Use Algorithm 1 to define X BU ∼ BU(aoL , boL , fLo , aoR , boR , fRo ) based on µo ,
σo.
2. Determine which of Algorithms 1–3 can be used to achieve σ 02 = δσ o2 .
145
3. Of the possible algorithms, select one at random and use it to specify X 0 ∼
BU(a0L , b0L , fL0 , a0R , b0R , fR0 ).
Specifying κ:
• Evaluate fX o (x) at each of a0L , b0L , a0R , and b0R .
o
n 0
f0
f0
f0
f
• Set κ = max fX o L(a0 ) , fX o L(b0 ) , fX o R(a0 ) , fX o R(b0 ) .
L
R
R
3.5
7
3
6
2.5
5
Probability density
Probability density
L
2
1.5
1
0.5
0
−0.2
4
3
2
1
0
0.2
0.4
0.6
0.8
1
0
0
1.2
0.2
0.4
x
0.6
0.8
1
x
(a) Uniform to bimodal uniform: X o ∼ U[0, 1],
δ = 0.7, k = 1
(b) Gaussian to uniform: X o ∼ T (0, 1, 0.6),
δ = 0.3, k = 1
2.5
Probability density
2
1.5
1
0.5
0
−3
−2
−1
0
x
1
2
3
(c) Gaussian to triangular: X o ∼ N (0, 1),
δ = 0.9, k = 1
Figure A-3: Examples of switching from a uniform, triangular, or Gaussian distribution to a bimodal uniform distribution. In each figure, the solid red line denotes
fX o (x), the dashed blue line represents fX 0 (x), and the dotted black line shows
κfX o (x), which is greater than or equal to fX 0 (x) for all x.
146
Appendix B
Derivation of Local Sensitivity
Analysis Results
In Section 3.3, we introduced expressions for the sensitivity of complexity and risk to
local changes in the mean or standard deviation of the quantity of interest. Here, we
provide the derivations for (3.15), (3.16), (3.19), and (3.20). Letting Y and Y 0 represent initial and new estimates of the QOI, respectively, the corresponding changes in
mean, standard deviation, complexity, and risk, are given below:
∆µ = µY 0 − µY ,
∆σ = σY 0 − σY ,
∆C = C(Y 0 ) − C(Y ),
∆P = P (Y 0 < r) − P (Y < r).
B.1
Perturbations in Mean
First, we consider the case where Y and Y 0 are related by a constant shift ∆µ, while
∆σ = 0:
Y 0 = Y + ∆µ.
147
(B.1)
B.1.1
Effect on Complexity
A constant shift has no effect on a random variable’s differential entropy, as h(Y +
∆µ) = h(Y ) [88]. Therefore, the exponential entropy of Y and Y 0 are also equivalent:
C(Y 0 ) = exp[h(Y + ∆µ)] = C(Y ).
(B.2)
Since ∆C = 0 for all values of ∆µ, we have that the partial derivative of C(Y ) to µY
is zero:
∂C(Y )
= 0.
∂µY
B.1.2
(B.3)
Effect on Risk
Here, we define a new variable z = y − ∆µ, and note that the probability densities of
Y and Y 0 are related by:
fY 0 (y) = fY (y − ∆µ) = fY (z).
(B.4)
Therefore, the failure probability associated with Y 0 is:
Z
0
r
Z
P (Y < r) =
r
fY 0 (y) dy =
−∞
fY (z) dy.
(B.5)
−∞
Taking the partial derivative of P (Y 0 < r) with respect to µY 0 , and noting that
∂z/∂µY 0 = ∂[y − (µY 0 − µY )]/∂µY 0 = −1, we obtain:
∂P (Y 0 < r)
=
∂µY 0
Z
r
−∞
∂fY (z) ∂z
dy = −
∂z ∂µY 0
Z
r
−∞
∂fY (z)
dy.
∂z
(B.6)
In the limit as ∆µ → 0, z → y and µY 0 → µY . Then, the above expression becomes:
∂P (Y < r)
=−
∂µY
Z
r
−∞
r
∂fY (y)
dy = −fY (y) .
∂y
−∞
148
(B.7)
Commonly, fY (y) → 0 as y → −∞, so we have:
∂P (Y < r)
= −fY (r).
∂µY
B.2
(B.8)
Perturbations in Standard Deviation
Next, we shrink the standard deviation of Y while leaving the mean unchanged. For
this derivation, we scale Y by a positive constant γ, and define new random variable
Y 0 such that:1
Y0 =
Y − µY
,
γ
(B.9)
fY 0 (y) = γfY (γy + µY ).
(B.10)
Note that we have included in Y 0 a constant shift of ∆µ = −µY , such that µY 0 = 0.
This shift serves to recenter Y 0 at the origin so as to simplify the derivation; in
general we do not restrict Y 0 to be zero mean. To account for the shift, we adjust
the requirement on Y 0 accordingly by defining a new constant r0 = r − µY . We also
define a new variable z = γy + µY . The variance of Y 0 is then given by:
σY2 0
Z
∞
1
=γ
y fY (γy + µY ) dy = 2
γ
−∞
2
Z
∞
(z − µY )2 fY (z) dz =
−∞
σY2
,
γ2
(B.11)
which corresponds to the following change in standard deviation:
σY
∆σ =
− σY = σY
γ
1
−1 .
γ
(B.12)
Thus, for γ > 1, standard deviation decreases between Y and Y 0 . The differential
entropy estimates h(Y ) and h(Y 0 ) are related by:
0
Z
∞
h(Y ) = −
fY (z) log[kfY (z)] dz = h(Y ) − log γ.
(B.13)
−∞
Note that in Sections 2.3 and 3.4.2, we described a multiplicative scaling of Y as Y 0 = αY or
Y = α(Y − µY ); in this section we use γ for convenience, where γ = 1/α.
1
0
149
B.2.1
Effect on Complexity
The change in complexity is given by:
exp[h(Y )]
− exp{h(Y )} = C(Y )
∆C =
exp[log γ]
1
−1 .
γ
(B.14)
Combining (B.12) and (B.14), we obtain:
∆C
C(Y )
=
.
∆σ
σY
(B.15)
∂C(Y )
C(Y )
=
.
∂σY
σY
(B.16)
In the limit as ∆σ → 0, we have:
B.2.2
Effect on Risk
The probability of failure associated with Y 0 is:
0
0
Z
r0
P (Y < r ) = γ
fY (γy + µY ) dy = P (Y < γr0 + µY ).
(B.17)
−∞
Taking the partial derivative with respect to γ, we obtain:
∂P (Y < γr0 + µY )
∂
=
∂γ
∂γ
Z
γr0 +µY
fY (z) dz = r0 fY (γr0 + µY ).
(B.18)
−∞
Then, if we let γ → 1 and make the substitution r0 = r − µY , we have:
∂P (Y < r)
= (r − µY )fY (r).
∂γ
(B.19)
We seek to compute the partial derivative of P (Y < r) with respect to σY . Using
the chain rule, we obtain:
∂P (Y < r)
∂P (Y < r) ∂γ
=
∂σY
∂γ
∂σY
150
(B.20)
Recall that from (B.11), γσY 0 = σY . Differentiating both sides with respect to σY 0
allows us to estimate ∂γ/∂σY :
γ+
∂γ
σY 0 = 0,
∂σY 0
∂γ
−γ
=
.
∂σY 0
σY 0
(B.21)
(B.22)
As γ → 1, σY 0 → σY , and the above expression becomes:
∂γ
−1
=
.
∂σY
σY
(B.23)
Combining (B.19), (B.20), and (B.23), we have:
(µY − r)
∂P (Y < r)
=
fY (r).
∂σY
σY
(B.24)
Finally, we note that in the case where risk is expressed as P (Y > r) = 1 − P (Y <
r), the corresponding sensitivities ∂P (Y > r)/∂µY and ∂P (Y > r)/∂σY are related
to (B.8) and (B.24), respectively, through only a change in sign.
151
Appendix C
Derivation of the Entropy Power
Decomposition
In Section 3.4.1, we introduced the following entropy power decomposition, which
apportions N (Y ), the entropy power of the QOI, into contributions due to the entropy
power and non-Gaussianity of the auxiliary random variables Z1 , . . . , Zm and their
interactions:
N (Y )exp[2DKL (Y ||Y G )] =
X
N (Zi ) exp[2DKL (Zi ||ZiG )]
i
+
X
N (Zij ) exp[2DKL (Zij ||ZijG )] + . . .
i<j
G
+ N (Z12...m ) exp[2DKL (Z12...m ||Z12...m
)].
(C.1)
The superscript G in the above expression denotes equivalent Gaussian random variables that have the same mean and variance as the original distribution.
In this section, we derive the entropy power decomposition starting from the
ANOVA-HDMR given in (3.1) and (3.40). We limit our consideration to a system
consisting of only two input factors (m = 2), although the results can be generalized
153
to higher dimensions. For such a system, (3.40) and (3.41) reduce to:
Y = g0 + Z1 + Z2 + Z12 ,
(C.2)
var (Y ) = var (Z1 ) + var (Z2 ) + var (Z12 ).
(C.3)
The auxiliary random variables Z1 , Z2 , and Z12 are uncorrelated, though not necessarily independent.1 For this two-factor system, the distributions of the equivalent
Gaussian random variables and the corresponding probability densities are:
Z1G
∼ N (µZ1 , σZ1 )
fZ1G (z) =
Z2G ∼ N (µZ2 , σZ2 )
fZ2G (z) =
G
Z12
∼ N (µZ12 , σZ12 )
G (z) =
fZ12
Y G ∼ N (µY , σY )
fY G (z) =
−(z − µZ1 )2
q
exp
2σZ2 1
2πσZ2 1
−(z − µZ2 )2
1
q
exp
2σZ2 2
2πσZ2 2
−(z − µZ12 )2
1
q
exp
2σZ2 12
2πσZ2 12
−(z − µY )2
1
p
exp
2σY2
2πσY2
1
To verify the entropy power decomposition, we need to show that:
G
exp[2h(Y, Y G )] = exp[2h(Z1 , Z1G )] + exp[2h(Z2 , Z2G )] + exp[2h(Z12 , Z12
)].
(C.4)
where h(Z1 , Z1G ) represents the cross entropy between Z1 and its equivalent Gaussian
random variable Z1G (likewise Y , Z2 , and Z12 ). By definition, h(Z1 , Z1G ) is computed
as:
1
Since Z12 is a function of both X1 and X2 , it is necessarily dependent on both Z1 and Z2 , which
are respectively functions of X1 and X2 alone. However, Z1 and Z2 are independent, due to our
assumption in Section 1.2 that all design parameters are independent.
154
h(Z1 , Z1G )
Z
∞
=−
fZ1 (z) log fZ1G (z) dz


Z ∞

2 
1
−(z − µZ1 )
fZ1 (z) log q
=−
dz
exp
 2πσ 2

2σZ2 1
−∞
Z1


Z ∞
1
 dz
fZ1 (z) log  q
=−
2
−∞
2πσZ1
Z ∞
−(z − µZ1 )2
fZ1 (z) log exp
−
dz
2σZ2 1
−∞
Z ∞
Z ∞
q
1
2
fZ1 (z) dz + 2
(z − µZ1 )2 fZ1 (z) dz.
= log 2πσZ1
2σ
−∞
Z1 −∞
(C.5)
−∞
Z
∞
Noting that
Z
−∞
h(Z1 , Z1G )
(C.7)
(C.8)
∞
fZ1 (z) dz = 1 and
−∞
(C.6)
(z − µZ1 )2 fZ1 (z) dz = σZ2 1 , (C.8) simplifies to:
q
1
= log 2πσZ2 1 + .
2
(C.9)
Multiplying both sides by two and taking the exponential, we get:
exp[2h(Z1 , Z1G )]
q
1
2
= exp 2 log 2πσZ1 + 2
2
= exp log(2πσZ2 1 ) + 1
= 2πeσZ2 1 .
(C.10)
(C.11)
(C.12)
Similarly, the other terms of (C.4) are given by:
exp[2h(Z2 , Z2G )] = 2πeσZ2 2 ,
(C.13)
G
exp[2h(Z12 , Z12
)] = 2πeσZ2 12 ,
(C.14)
exp[2h(Y, Y G )] = 2πeσY2 .
(C.15)
155
Substituting (C.12)–(C.15) into (C.4), we obtain:
2πeσY2 = 2πeσZ2 1 + 2πeσZ2 2 + 2πeσZ2 12 ,
σY2 = σZ2 1 + σZ2 2 + σZ2 12 .
(C.16)
(C.17)
We have already established that (C.17) is true, as it is equivalent to (C.3). We
conclude therefore that (C.4) also holds true.
Using (2.24) to relate cross entropy to differential entropy and K-L divergence, we
can rewrite (C.4) as follows:
exp[2h(Y ) + 2DKL (Y ||Y G )] = exp[2h(Z1 ) + 2DKL (Z1 ||Z1G )]
+ exp[2h(Z2 ) + 2DKL (Z2 ||Z2G )]
G
+ exp[2h(Z12 ) + 2DKL (Z12 ||Z12
)].
(C.18)
Using (2.26), we can rewrite the above expression in terms of N (Y ):
(2πe)N (Y ) exp[2DKL (Y ||Y G )] = (2πe)N (Z1 ) exp[2DKL (Z1 ||Z1G )]
+ (2πe)N (Z2 ) exp[2DKL (Z2 ||Z2G )]
G
+ (2πe)N (Z12 ) exp[2DKL (Z12 ||Z12
)].
(C.19)
Dividing both sides by (2πe), the result becomes:
N (Y ) exp[2DKL (Y ||Y G )] = N (Z1 ) exp[2DKL (Z1 ||Z1G )]
+ N (Z2 ) exp[2DKL (Z2 ||Z2G )]
+ N (Z12 )
G
exp[2DKL (Z12 ||Z12
)]
,
G
exp[2DKL (Y ||Y )]
(C.20)
which is exactly (C.1) for m = 2. Thus, we have verified the entropy power decomposition from Section 3.4.1.
156
Appendix D
CAEP/9 Analysis Airports
Table D.1 lists the 99 US airports used in the CAEP/9 noise stringency case study in
Chapter 5, along with the average personal income in the corresponding Metropolitan
Statistical Area obtained from the US Bureau of Economic Analysis [156]. Each
airport is identified using its four-letter ICAO designation. All income figures are in
2006 USD.
No.
Airport
City
State
Income
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
KABE
KABQ
KACK
KALB
KATL
KAUS
KBDL
KBFI
KBHM
KBNA
KBOI
KBOS
KBUF
KBWI
KCAE
KCLE
KCLT
KCMH
Allentown
Albuquerque
Nantucket
Albany
Atlanta
Austin
Hartford
Seattle (Boeing Field)
Birmingham
Nashville
Boise
Boston
Buffalo
Baltimore
Columbia
Cleveland
Charlotte
Columbus
PA
NM
MA
NY
GA
TX
CT
WA
AL
TN
ID
MA
NY
MD
SC
OH
NC
OH
35,717
32,935
71,837
37,828
39,160
37,022
46,456
47,006
38,087
38,471
35,591
52,018
33,402
44,704
33,330
38,156
39,542
36,695
157
No.
Airport
City
State
Income
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
KCOS
KCVG
KDAL
KDAY
KDCA
KDFW
KDSM
KDTW
KELP
KEWR
KFAT
KFLL
KGEG
KGON
KGRR
KGSO
KHOU
KHUF
KIAD
KIAH
KILN
KIND
KJAX
KJFK
KLAS
KLAX
KLGA
KLIT
KMCI
KMCO
KMDW
KMEM
KMHT
KMIA
KMKE
KMMH
KMSP
KMSY
Colorado Springs
Cincinnati
Dallas (Love Field)
Dayton
Washington (Reagan)
Dallas/Fort Worth
Des Moines
Detroit
El Paso
Newark
Fresno
Fort Lauderdale
Spokane
Groton/New London
Grand Rapids
Greensboro
Houston (Hobby)
Terre Haute
Washington (Dulles)
Houston (Bush)
Wilmington
Indianapolis
Jacksonville
New York (Kennedy)
Las Vegas
Los Angeles
New York (LaGuardia)
Little Rock
Kansas City
Orlando (International)
Chicago (Midway)
Memphis
Manchester
Miami
Milwaukee
Mammoth Lakes
Minneapolis/St. Paul
New Orleans
CO
OH
TX
OH
DC
TX
IA
MI
TX
NJ
CA
FL
WA
CT
MI
NC
TX
IN
DC
TX
OH
IN
FL
NY
NV
CA
NY
AR
MO
FL
IL
TN
NH
FL
WI
CA
MN
LA
35,649
38,096
40,614
33,649
53,384
40,614
40,432
37,928
24,718
50,843
29,111
42,148
31,626
41,792
32,734
33,978
42,984
26,936
53,384
42,984
31,423
38,338
39,197
50,843
38,183
42,329
50,843
35,840
38,914
34,625
43,294
36,417
43,160
42,148
41,162
39,404
45,002
42,416
158
No.
Airport
City
State
Income
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
KOAK
KOKC
KOMA
KONT
KORD
KORF
KORL
KPDX
KPHL
KPHX
KPIT
KPVD
KRDU
KRIC
KRNO
KROC
KRSW
KSAN
KSAT
KSDF
KSEA
KSFO
KSHV
KSJC
KSLC
KSMF
KSRQ
KSTL
KSWF
KSYR
KTEB
KTOL
KTPA
KTUL
KTUS
KTYS
KVNY
PANC
Oakland
Oklahoma City
Omaha
Ontario
Chicago (O’Hare)
Norfolk
Orlando (Executive)
Portland
Philadelphia
Phoenix
Pittsburgh
Providence
Raleigh/Durham
Richmond
Reno
Rochester
Fort Myers
San Diego
San Antonio
Louisville
Seattle/Tacoma
San Francisco
Shreveport
San Jose
Salt Lake City
Sacramento
Sarasota
St. Louis
Newburgh
Syracuse
Teterboro
Toledo
Tampa
Tulsa
Tucscon
Knoxville
Van Nuys
Anchorage
CA
OK
NE
CA
IL
VA
FL
OR
PA
AZ
PA
RI
NC
VA
NV
NY
FL
CA
TX
KY
WA
CA
LA
CA
UT
CA
FL
MO
NY
NY
NJ
OH
FL
OK
AZ
TN
CA
AK
59,640
36,357
40,476
29,330
43,294
36,567
34,625
38,416
43,548
37,173
38,808
37,777
39,215
39,730
43,803
35,385
41,089
43,967
32,553
35,954
47,006
59,640
33,418
56,124
36,906
38,852
47,497
39,126
36,563
32,574
50,843
32,812
36,470
38,470
33,263
33,150
42,329
42,256
159
No.
Airport
City
State
Income
95
96
97
98
99
PHKO
PHLI
PHNL
PHOG
PHTO
Kailua-Kona
Lihue
Honolulu
Kahului
Hilo
HI
HI
HI
HI
HI
29,185
32,442
39,938
34,160
29,185
Table D.1: CAEP/9 US airports and income levels
160
Bibliography
[1] P. Eremenko. Formal Model-Based Design & Manufacture: A Template for
Managing Complexity in Large-Scale Cyber-Physical Systems. In Conference
on Systems Engineering Research, Atlanta, GA, 2013.
[2] L. Allen, R. Angel, J. D. Mangus, G. A. Rodney, R. R. Shannon, and C. P.
Spoelhof. The Hubble Space Telescope Optical Systems Failure Report. Technical Report NASA-TM-103443, National Aeronautics and Space Administration,
1990.
[3] C. Bolkcom. V-22 Osprey Tilt-Rotor Aircraft: Congressional Research Service
Report for Congress. Technical report, Congressional Research Service, Jan. 7,
2005.
[4] M. Hiltzik. 787 Dreamliner Teaches Boeing Costly Lesson on Outsourcing. Los
Angeles Times, Feb. 15, 2011.
[5] National Aeronautics and Space Administration, Washington, DC. NASA Systems Engineering Handbook, NASA/SP-2007-6105 Rev1 edition, 2007.
[6] N. P. Suh. The Principles of Design. Oxford University Press, New York, NY,
1990.
[7] J. J. Deyst. The Application of Estimation Theory to Managing Risk in Product
Developments. In Proceedings of the 21st Digital Avionics Systems Conference,
volume 1, pages 4A3–1–12, Irvine, CA, 2002.
[8] D. L. Allaire. Uncertainty Assessment of Complex Models with Application
to Aviation Environmental Systems. PhD thesis, Massachusetts Institute of
Technology, Cambridge, MA, 2009.
[9] A. Haldar and S. Mahadevan. Probability, Reliability and Statistical Methods
in Engineering Design. John Wiley & Sons, Inc., New York, NY, 2000.
[10] C. J. Roy and W. L. Oberkampf. A Comprehensive Framework for Verification,
Validation, and Uncertainty Quantification in Scientific Computing. Computer
Methods in Applied Mechanics and Engineering, 200:2131–2144, 2011.
161
[11] R. C. Smith. Uncertainty Quantification: Theory, Implementation, and Applications. SIAM Computational Science & Engineering, Philadelphia, PA, 2014.
[12] W. L. Oberkampf, S. M. DeLand, B. M. Rutherford, K. V. Diegert, and K. F.
Alvin. Error and Uncertainty in Modeling and Simulation. Reliability Engineering and System Safety, 75:333–357, 2002.
[13] A. H-S. Ang and W. H. Tang . Probability Concepts in Engineering: Emphasis
on Applications to Civil and Environmental Engineering. John Wiley & Sons,
Inc., Hoboken, NJ, 2nd edition, 2007.
[14] M. C. Kennedy and A. O’Hagan. Bayesian Calibration of Computer Models.
Journal of the Royal Statistical Society, Series B, 63(3):425–464, 2001.
[15] B. M. Ayyub. Elicitation of Expert Opinions for Uncertainty and Risks. CRC
Press, Boca Raton, FL, 2001.
[16] A. Babuscia. Statistical Risk Estimation for Communication System Design.
PhD thesis, Massachusetts Institute of Technology, Cambridge, MA, 2012.
[17] A. Babuscia an K. Cheung. A Quantitative Approach to Perform Expert Elicitation for Engineering Risk Analysis: Methodology and Experimental Results.
Royal Statistical Society (Series A), 2013 (to appear).
[18] H. Agarwal, J. E. Renaud, E. L. Preston, and D. Padmanabhan. Uncertainty
Quantification Using Evidence Theory in Multidisciplinary Design Optimization. Reliability Engineering and System Safety, 85:281–294, 2004.
[19] J. C. Helton, J. D. Johnson, C. J. Sallaberry, and C. B. Storlie. Survey of
Sampling-Based Methods for Uncertainty and Sensitivity Analysis. Reliability
Engineering and System Safety, 91:1175–1209, 2006.
[20] I. M. Sobol’. On the Distribution of Points in a Cube and the Approximate
Evaluation of Integrals. Zh. Vychisl. Mat. Mat. Fiz., 7(4):784–802, 1967.
[21] H. Niederreiter. Random Number Generation and Quasi-Monte Carlo Methods.
Society for Industrial and Applied Mathematics, Philadelphia, PA, 1992.
[22] T. Bui-Thanh, K. Willcox, and O. Ghattas. Model Reduction for Large-Scale
Systems with High-Dimensional Parametric Input Space. SIAM Journal on
Scientific Computing, 30(6):3270–3288, 2008.
[23] C. Lieberman, K. Willcox, and O. Ghattas. Parameter and State Model Reduction for Large-Scale Statistical Inverse Problems. SIAM Journal on Scientific
Computing, 32(5):2523–2542, 2010.
162
[24] O. Bashir, K. Willcox, O. Ghattas, B. van Bloemen Waanders, and J. Hill.
Hessian-Based Model Reduction for Large-Scale Systems with Initial Condition
Inputs. International Journal for Numerical Methods in Engineering, 73(6):844–
868, 2008.
[25] D. Allaire and K. Willcox. Surrogate Modeling for Uncertainty Assessment
with Application to Aviation Environmental System Models. AIAA Journal,
48(8):1791–1803, 2010.
[26] M. S. Eldred, A. A. Giunta, and S. S. Collis. Second-Order Corrections for
Surrogate-Based Optimization with Model Hierarchies. In 10th AIAA/ISSMO
Multidisciplinary Analysis and Optimization Conference, AIAA Paper 20044457, 2004.
[27] T. W. Simpson, J. D. Peplinski, P. N. Koch, and J. K. Allen. Metamodels
for Computer-Based Engineering Design: Survey and Recommendations. Engineering with Computers, 17(2):129–150, 2001.
[28] N. M. Alexandrov, R. M. Lewis, C. R. Gumbert, L. L. Green, and P. A. Newman. Approximation and Model Management in Aerodynamic Optimization
with Variable-Fidelity Models. Journal of Aircraft, 38(6):1093–1101, 2001.
[29] T. D. Robinson, M. S. Eldred, K. E. Willcox, and R. Haimes. Surrogate-Based
Optimization Using Multidelity Models with Variable Parameterization and
Corrected Space Mapping. AIAA Journal, 46(11):2814–2822, 2008.
[30] R. Spence and R. S. Soin. Tolerance Design of Electronic Circuits. Imperial
College Press, London, UK, 1997.
[31] K. Premphet and W. Paecharoenchai. Polypropylene/Metallocene EthyleneOctene Copolymer Blends with a Bimodal Particle Size Distribution: Mechanical Properties and Their Controlling Factors. Journal of Applied Polymer
Science, 85(11):2412–2418, 2002.
[32] M. E. Fowler, H. Keskkula, and D. R. Paul. Synergistic Toughening in Rubber
Modified Blends. Polymer, 28(10):1703–1711, 1987.
[33] T. Usami, Y. Gotoh, and S. Takayama. Generation Mechanism of Short-Chain
Branching Distribution in Linear Low-Density Polyethylenes. Macromolecules,
19(11):2722–2726, 1986.
[34] E. A. Murphy. One Cause? Many Causes?: The Argument from the Bimodal
Distribution. Journal of Chronic Diseases, 17(4):301–324, 1964.
[35] P. Seeman, C. Ulpian, C. Bergeron, P. Riederer, K. Jellinger, E. Gabriel, G. P.
Reynolds, and W. W. Tourtellotte. Bimodal Distribution of Dopamine Receptor
Densities in Brains of Schizophrenics. Science, 225(4663):728–731, 1984.
163
[36] Y. Wu, C. Zhou, J. Xiao, J. Kurths, and H. J. Schellnhuber. Evidence for a
Bimodal Distribution in Human Communication. Proceedings of the National
Academy of Sciences, 107(44):18803–18808, 2010.
[37] D. G. Xenikos. Modeling Human Dialogue – The Case of Group Communications in Trunked Mobile Telephony. Physica A, 388(23):4910–4918, 2009.
[38] D. Conway, E. Allison, R. Felstead, and M. Goulden. Rainfall Variability in
East Africa: Implications for Natural Resources Management and Livelihoods.
Philosophical Transactions of the Royal Society A, 363(1826):49–54, 2005.
[39] M. E. Angeles, J. E. González, N. D. Ramı́rez-Beltrán, C. A. Tepley, and D. E.
Comarazamy. Origins of the Caribbean Rainfall Bimodal Behavior. Journal of
Geophysical Research: Atmospheres, 115(11):D11106, 2010.
[40] S. Touati, M. Naylor, and I. G. Main. Origin and Nonuniversality of the Earthquake Interevent Time Distribution. Physical Review Letters, 102(16):168501,
2009.
[41] T. Parsons. Global Omori Law Decay of Triggered Earthquakes: Large Aftershocks Outside the Classical Aftershock Zone. Journal of Geophysical Research:
Solid Earth, 107(B9):ESE 9–1–20, 2002.
[42] K. v. Sichart and R.-P. Vollertsen. Bimodal Lifetime Distributions of Dielectrics for Integrated Circuits. Quality and Reliability Engineering International, 7(4):299–305, 1991.
[43] K. L. Wong. The Physical Basis for the Roller-Coaster Hazard Rate Curve for
Electronics. Quality and Reliability Engineering International, 7(6):489–495,
1991.
[44] F. Knoll and T. Vogel. Design for Robustness. IABSE (International Association
for Bridge and Structural Engineering), Zurich, Switzerland, 2009.
[45] A. Karl. Robust Design and Stochastics: Key Methods, Case Studies and
Challenges. In MDO Consortium Workshop, Stanford, CA, 2013.
[46] C. M. Creveling, J. L. Slutsky, and D. Antis, Jr. Design for Six Sigma in
Technology and Product Development. Prentice Hall, Upper Saddle River, NJ,
2003.
[47] K. C. Kapur and L. R. Lamberson. Reliability in Engineering Design. John
Wiley & Sons, Inc., New York, NY, 1977.
[48] L. A. Ealey. Quality By Design: Taguchi Methods and U.S. Industry. ASI Press,
Burr Ridge, IL, 2nd edition, 2009.
164
[49] G. Taguchi. The System of Experimental Design: Engineering Methods to Optimize Quality and Minimize Costs (2 vols.). UNIPUB/Kraus International
Publications and American Supplier Institute, White Plains, NY and Dearborn, MI, 1987.
[50] M. S. Phadke. Quality Engineering Using Robust Design. Prentice Hall, Englewood Cliffs, NJ, 1989.
[51] R. A. Fisher. The Design of Experiments. Oliver & Boyd, Edinburgh, UK,
1935.
[52] G. E. P. Box, W. G. Hunter, and J. S. Hunter. Statistics for Experimenters:
An Introduction to Design, Data Analysis, and Model Building. John Wiley &
Sons, Inc., New York, NY, 1978.
[53] W. J. Welch, T.-K. Yu, S. M. Kang, and J. Sacks. Computer Experiments for
Quality Control by Parameter Design. Journal of Quality Technology, 22(1):15–
22, 1990.
[54] A. C. Shoemaker, K.-L. Tsui, and C. F. J. Wu. Economical Experimentation
Methods for Robust Design. Technometrics, 33(4):415–427, 1991.
[55] R. H. Myers, A. I. Khuri, and G. Vining. Response Surface Alternatives to
the Taguchi Robust Parameter Design Approach. The American Statistician,
46(2):131–139, 1992.
[56] D. D. Frey, F. Engenhardt, and E. M. Greitzer. A Role for One Factor a a
Time Experimentation in Parameter Design. Research in Engineering Design,
14(2):65–74, 2003.
[57] D. D. Frey and H. Wang. Adaptive One-Factor-at-a-Time Experimentation and
Expected Value of Improvement. Technometrics, 48(3):418–431, 2006.
[58] K.-L. Tsui. An Overview of Taguchi Method and Newly Developed Statistical
Methods for Robust Design. IIE Transactions, 24(5):44–57, 1992.
[59] D. D. Frey and R. Jugulum. The Mechanisms by Which Adaptive One-Factorat-a-Time Experimentation Leads to Improvement. Journal of Mechanical Design, 128(5):1050–1060, 2006.
[60] T. B. Savoie and D. D. Frey. Detecting Mistakes in Engineering Models: The
Effects of Experimental Eesign. Research in Engineering Design, 23(2):155–175,
2012.
[61] G. Eckes. The Six Sigma Revolution. John Wiley & Sons, Inc., New York, NY,
2001.
165
[62] G. Tennant. Six Sigma: SPC and TQM in Manufacturing and Services. Gower
Publishing Company, Burlington, VT, 2001.
[63] A. M. Hasofer and N. C. Lind. Exact and Invariant Second-Moment Code
Format. Journal of the Engineering Mechanics Division, 100(1):111–121, 1974.
[64] R. Rackwitz and B. Fiessler. Structural Reliability Under Combined Random
Load Sequences. Computers and Structures, 9(5):489–494, 1978.
[65] B. Fiessler, H. J. Neumann, and R. Rackwitz. Quadratic Limit States in Structural Reliability. Journal of Engineering Mechanics, 1095(4):661–676, 1979.
[66] H. O. Madsen, S. Krenk, and N. C. Lind. Methods of Structural Safety. Prentice
Hall, Englewood Cliffs, NJ, 1986.
[67] W. Weaver. Science and Complexity. American Scientist, 36:536–544, 1948.
[68] B. A. Huberman and T. Hogg. Phase Transitions in Artificial Intelligence Systems. Artificial Intelligence, 33:155–171, 1987.
[69] G. Johnson. Researchers on Complexity Ponder What It’s All About. The New
York Times: Technology, May 6, 1997.
[70] S. Balestrini-Robinson. A Modeling Process to Understand Complex System
Architectures. PhD thesis, Georgia Institute of Technology, Atlanta, GA, 2009.
[71] P. A. Corning. Complexity is Just a Word! Technological Forecasting and Social
Change, 59:197–200, 1998.
[72] S. C. Bankes. Tools and Techniques for Developing Policies for Complex and
Uncertain Systems. In Proceedings of the National Academy of Sciences of the
United States of America, volume 99, pages 7263–7266, 2002.
[73] J. D. Summers and J. J. Shah. Mechanical Engineering Design Complexity Metrics: Size, Coupling, and Solvability. Journal of Mechanical Design,
132(2):021004–1–11, 2010.
[74] D. Braha and O. Maimon. The Measurement of a Design Structural and Functional Complexity. IEEE Transactions on Systems, Man and Cybernetics, Part
A: Systems and Humans, 28(4):527–535, 1998.
[75] A. J. Albrecht and Jr. J. E. Gaffney. Software Function, Source Lines of Code,
and Development Effort Prediction: A Software Science Validation. IEEE
Transactions on Software Engineering, SE-9(6):639–648, 1983.
[76] P. A. Murmann. Expected Development Time Reductions in the German Mechanical Engineering Industry. Journal of Product Innovation Management,
11(3):236–252, 1994.
166
[77] A. Griffin. The Effect of Project and Process Characteristics on Product Development Cycle Time. Journal of Marketing Research, 34(1):24–35, 1997.
[78] M. H. Meyer and J. M. Utterback. Product Development Cycle Time and Commercial Success. IEEE Transactions on Engineering Management, 42(4):297–
304, 1995.
[79] T. J. McCabe. A Complexity Measure. IEEE Transactions on Software Engineering, 2(4):308–320, 1976.
[80] M. V. Tatikonda and S. R. Rosenthal. Technology Novelty, Project Complexity, and Product Development Project Execution Success: A Deeper Look at
Task Uncertainty in Product Innovation. IEEE Transactions on Engineering
Management, 47(1):74–87, 2000.
[81] S. Novak and S. D. Eppinger. Sourcing by Design: Product Complexity and
the Supply Chain. Management Science, 47(1):189–204, 2004.
[82] J. Kim and D. Wilemon. An Empirical Investigation of Complexity and Its
Management in New Product Development. Technology Analysis & Strategic
Management, 21(4):547–564, 2009.
[83] K. M. M. Hölttä and K. N. Otto. Incorporating Design Effort Complexity
Measures in Product Architectural Design and Assessment. Design Studies,
26(5):463–485, 2005.
[84] J. Hartmanis and R. E. Stearns. On the Computational Complexity of Algorithms. Transactions of the American Mathematical Society, 117:285–306,
1965.
[85] S. Arora and B. Barak. Computational Complexity. Cambridge University
Press, New York, NY, 2009.
[86] A. N. Kolmogorov. Three Approaches to the Quantitative Definition of Information. Problems of Information Transmission, 1(1):1–7, 1965.
[87] G. J. Chaitin. On the Simplicity and Speed of Programs for Computing Infinite
Sets of Natural Numbers. Journal of the ACM, 16(3):407–422, 1969.
[88] T. M. Cover and J. A. Thomas. Elements of Information Theory. John Wiley
& Sons, Inc., New York, NY, 1991.
[89] C. A. Rodriguez-Toro, S. J. Tate, G. E. M. Jared, and K. G. Swift. Complexity
Metrics for Design (Simplicity + Simplicity = Complexity). In Proceedings of
the Institution of Mechanical Engineers, Part B: Journal of Engineering Manufacture, volume 217, pages 721–725, 2003.
167
[90] M. Tribus and E. C. McIrvine. Energy and Information (Thermodynamics and
Information Theory). Scientific American, 225(3):179–188, 1971.
[91] C. E. Shannon. A Mathematical Theory of Communication. The Bell System
Technical Journal, 27:379–423 and 623–656, 1948.
[92] J. Rifkin and T. Howard. Entropy: A New World View. The Viking Press, New
York, NY, 1980.
[93] J. S. Coleman. An Introduction to Mathematical Sociology. The Free Press,
New York, NY, 1964.
[94] K. D. Bailey. Sociological Entropy Theory: Toward a Statistical and Verbal
Congruence. Quality and Quantity, 18(1):113–133, 1983.
[95] K. D. Bailey. Social Entropy Theory. State University of New York Press,
Albany, NY, 1990.
[96] S. Lloyd and H. Pagels. Complexity as Thermodynamic Depth. Annals of
Physics, 188:186–213, 1988.
[97] D. Allaire, Q. He, J. Deyst, and K. Willcox. An Information-Theoretic Metric of
System Complexity With Application to Engineering System Design. Journal
of Mechanical Design, 134(10):100906–1–10, 2012.
[98] N. P. Suh. Complexity: Theory and Applications. Oxford University Press, New
York, NY, 2005.
[99] L. L. Campbell. Exponential Entropy as a Measure of Extent of a Distribution.
Probability Theory and Related Fields, 5(3):217–225, 1966.
[100] G. H. Hardy, J. E. Littlewood, and G. Pólya. Inequalities. Cambridge University
Press, Cambridge, UK, 1934.
[101] H. V. Dannon. Power Means Calculus, Product Calculus, Harmonic Mean
Calculus, and Quadratic Mean Calculus. Gauge Institute Journal, 4(4):1–88,
2008.
[102] E. T. Jaynes. Information Theory and Statistical Mechanics. The Physical
Review, 106(4):620–630, 1957.
[103] E. T. Jaynes. Information Theory and Statistical Mechanics II. The Physical
Review, 108(2):171–190, 1957.
[104] B. M. Ayyub and G. J. Klir. Uncertainty Modeling and Analysis in Engineering
and the Sciences. CRC Press, Boca Raton, FL, 2010.
[105] M. H. DeGroot and M. J. Schervish. Probability and Statistics. Addison-Wesley,
Boston, MA, 3rd edition, 2002.
168
[106] D. P. Bertsekas and J. N. Tsitsiklis. Introduction to Probability. Athena Scientific, Nashua, NH, 2nd edition, 2008.
[107] D. Mukherjee and M. V. Ratnaparkhi. On the Functional Relationship Between
Entropy and Variance with Related Applications. Communications in Statistics
– Theory and Methods, 15(1):291–311, 1986.
[108] N. Ebrahimi, E. Maasoumi, and E. S. Soofi. Ordering Univariate Distributions
by Entropy and Variance. Journal of Econometrics, 90(2):317–336, 1999.
[109] S. Kullback and R. A. Leibler. On Information and Sufficiency. The Annals of
Mathematical Statistics, 22(1):79–86, 1951.
[110] S. Verdú and D. Guo. A Simple Proof of the Entropy-Power Inequality. IEEE
Transactions on Information Theory, 52(5):2165–2166, 2006.
[111] S. Kaplan and B. J. Garrick. On the Quantitative Definition of Risk. Risk
Analysis, 1(1):11–27, 1981.
[112] L. A. Cox, Jr. What’s Wrong with Risk Matrices? Risk Analysis, 28(2):497–512,
2008.
[113] J. Li and D. Xiu. Evaluation of Failure Probability Via Surrogate Models.
Journal of Computational Physics, 229(23):8966–8980, 2010.
[114] A. Saltelli, K. Chan, and E. M. Scott. Sensitivity Analysis. John Wiley & Sons,
Inc., New York, NY, 2000.
[115] R. I. Cukier, C. M. Fortuin, K. E. Shuler, A. G. Petschek, and J. H. Schaibly.
Study of the Sensitivity of Coupled Reaction Systems to Uncertainties in Rate
Coefficients I: Theory. Journal of Chemical Physics, 59:3873–3878, 1973.
[116] J. H. Schaibly and K. E. Shuler. Study of the Sensitivity of Coupled Reaction Systems to Uncertainties in Rate Coefficients II: Applications. Journal of
Chemical Physics, 59:3879–3888, 1973.
[117] A. Saltelli and R. Bolado. An Alternative Way to Compute Fourier Amplitude
Sensitivity Test. Computation Statistics & Data Analysis, 26:445–460, 1998.
[118] T. Homma and A. Saltelli. Importance Measures in Global Sensitivity Analysis
of Nonlinear Models. Reliability Engineering & System Safety, 52(1):1–17, 1996.
[119] I. M. Sobol’. Sensitivity Estimates for Nonlinear Mathematical Models. Mathematical Modeling and Computational Experiment, 1(4):407–414, 1993.
[120] I. M. Sobol’. Global Sensitivity Indices for Nonlinear Mathematical Models
and Their Monte Carlo Estimates. Mathematics and Computers in Simulation,
55:271–280, 2001.
169
[121] I. M. Sobol’. Theorems and Examples on High Dimensional Model Representation. Reliability Engineering & System Safety, 79(2):187–193, 2003.
[122] A. Saltelli, S. Tarantola, F. Campolongo, and M. Ratto. Sensitivity Analysis
in Practice: A Guide to Assessing Scientific Models. John Wiley & Sons, Inc.,
Hoboken, NJ, 2004.
[123] J. J. Lee, I. A. Waitz, B. Y. Kim, G. G. Fleming, L. Maurice, and C. A.
Holsclaw. System for Assessing Aviation’s Global Emissions (SAGE), Part
2: Uncertainty Assessment. Transportation Research Part D: Transport and
Environment, 12(6):381–395, 2007.
[124] K. Marais, S. P. Lukachko, M. Jun, A. Mahashabde, and I. A. Waitz. Assessing
the Impact of Aviation on Climate. Meteorologische Zeitschrift, 17(2):157–172,
2008.
[125] D. Allaire and K. Willcox. A Variance-Based Sensitivity Index Function for
Factor Prioritization. Reliability Engineering and System Safety, 107:107–114,
2012.
[126] H. Liu, W. Chen, and A. Sudjianto. Relative Entropy Based Method for Probabilistic Sensitivity Analysis in Engineering Design. Journal of Mechanical Design, 128(2):326–336, 2006.
[127] H. Takeda, P. Veerkamp, T. Tomiyama, and H. Yoshikawa. Modeling Design
Processes. AI Magazine, 11(4):37–48, 1990.
[128] J. D. Irwin and R. M. Nelms. Basic Engineering Circuit Analysis. John Wiley
& Sons, Inc., Hoboken, NJ, 8th edition, 2005.
[129] FAA Aerospace Forecast, Fiscal Years 2009-2025. Technical report, United
States Federal Aviation Administration, 2009.
[130] B. Metz, O. R. Davidson, P. R. Bosch, R. Dave, and L. A. Meyer, editors.
Contribution of Working Group III to the Fourth Assessment Report of the
Intergovernmental Panel on Climate Change. Cambridge University Press, New
York, NY, 2007.
[131] A. Schäfer and I. Waitz. Air Transportation and the Environment. Transport
Policy, 2014. http://dx.doi.org/10.1016/j.tranpol.2014.02.012.
[132] Results From a Survey of the Nation’s 50 Busiest Commercial Service Airports.
Technical Report GAO/RCED-00-222, United States General Accounting Office, 2000.
[133] Y. Schipper. Environmental Costs in European Aviation. Transport Policy,
11(2):141–154, 2004.
170
[134] P. J. Wolfe, S. H. L. Yim, G. Lee, A. Ashok, S. R. H. Barrett, and I. A. Waitz.
Near-airport Distribution of the Environmental Costs of Aviation. Transport
Policy, 2014. http://dx.doi.org/10.1016/j.tranpol.2014.02.023.
[135] Annex 16: Environmental Protection Volume I: Aircraft Noise. Technical report, International Civil Aviation Organization, 2005.
[136] S. McGuire. Sleep Disturbance Due to Aircraft Noise. Technical report, Partnership for AiR Transportation Noise and Emissions Reduction (PARTNER),
Project 25, 2009.
[137] H. Swift. A Review of the Literature Related to Potential Health Effects of
Aircraft Noise. Technical report, Partnership for AiR Transportation Noise
and Emissions Reduction (PARTNER), Project 19, 2009.
[138] I. Waitz, J. Townsend, J. Cutcher-Gershenfeld, E. Greitzer, and J. Kerrebrock.
Aviation and the Environment: A National Vision Statement, Framework for
Goals and Recommended Actions. Technical report, Report to the United States
Congress, 2004.
[139] A. Mahashabde, P. Wolfe, A. Ashok, C. Dorbian, Q. He, A. Fan, S. Lukachko,
A. Mozdzanowska, C. Wollersheim, S. R. H. Barrett, M. Locke, and I. A. Waitz.
Assessing the Environmental Impacts of Aircraft Noise and Emissions. Progress
in Aerospace Sciences, 47(1):15–52, 2011.
[140] I. Waitz, S. Lukachko, Y. Go, P. Hollingsworth, K. Harback, and F. Morser.
Requirements Document for the Aviation Environmental Portfolio Management
Tool. Technical Report PARTNER-COE-2006-001, Partnership for AiR Transportation Noise and Emissions Reduction (PARTNER), Project 19, 2006.
[141] Q. He. Development of an Income-Based Hedonic Monetization Model for the
Assessment of Aviation-Related Noise Impacts. Master’s thesis, Massachusetts
Institute of Technology, Cambridge, MA, 2010.
[142] Q. He, C. Wollersheim, M. Locke, and I. Waitz. Estimation of the Global
Impacts of Aviation-Related Noise Using an Income-Based Approach. Transport
Policy, 2014. http://dx.doi.org/10.1016/j.tranpol.2014.02.020.
[143] J. P. Nelson. Meta-Analysis of Airport Noise and Hedonic Property Values:
Problems and Prospects. Journal of Transport Economics and Policy, 38(1):1–
28, 2004.
[144] J. P. Nelson and R. B. Palmquist. Review of APMT Benefit Valuation Methods and Suggestions for Modifications. Technical report, Prepared for the Massachusetts Institute of Technology, 2008.
171
[145] Z. Wadud. A Systematic Review of Literature on the Valuation of Local Environmental Externalities of Aviation. Technical report, Omega: Aviation in a
Sustainable World, Cambridge, UK.
[146] C. Kish. An Estimate of the Global Impact of Commercial Aviation Noise.
Master’s thesis, Massachusetts Institute of Technology, Cambridge, MA, 2008.
[147] P. A. Diamond and J. A. Hausman. Contingent Valuation: Is Some Number
Better Than No Number? Journal of Economic Perspectives, 8(4):45–64, 1994.
[148] S. Navrud. The State-of-the-Art on Economic Valuation of Noise. Technical
report, Final Report to the European Commission DG Environment, 2002.
[149] Economic Assessment of the Noise Stringency Scenarios. Technical Report
CAEP/9-IP/12, ICAO Committee for Aviation Environmental Protection,
2013.
[150] K. D. Kryter. The Meaning and Measurement of Perceived Noise Level. Noise
Control, 6(5):12–17, 1960.
[151] Federal Aviation Regulations, Noise Standards: Aircraft Type and Airworthiness Certification. Technical Report 14 CFR Part 36, United States Federal
Aviation Administration, 2009.
[152] Report on the Environmental Portion of the Noise Stringency Assessment. Technical Report CAEP/9-IP/20, ICAO Committee for Aviation Environmental
Protection, 2013.
[153] Environmental Tool Suite Frequently Asked Questions.
http:
//www.faa.gov/about/office_org/headquarters_offices/apl/research/
models/toolsfaq/. United States Federal Aviation Administration [Online
Resource]; Accessed March 11, 2014.
[154] Cost-Benefit Analysis of CAEP/9 Noise Stringency Options. Technical Report CAEP/9-IP/25, ICAO Committee for Aviation Environmental Protection,
2013.
[155] P. J. Wolfe and I. A. Waitz. Aircraft Noise Certification Stringency: Benefits,
Costs, and Environmental Co-benefits. Transportation Research Part D, 2014
(in review).
[156] Local Area Personal Income. http://www.bea.gov/regional/. United States
Department of Commerce Bureau of Economic Analysis [Online Resource]; Accessed June 15, 2010.
[157] R. Tam, P. Belobaba, K. R. Polenske, and I. Waitz. Assessment of Silent
Aircraft-Enabled Regional Development and Airline Economics in the UK. In
Proceedings of the 45th AIAA Aerospace Sciences Meeting and Exhibit, Reno,
NV, 2007.
172
Download