FTC 2005 Abstracts

advertisement
FTC 2005 Abstracts
Thursday, 9:15 to 10: 00 am
Logarithmic SPC Session
An SPC Control Chart Procedure Based on Censored Lognormal Observations
Uwe Koehn
Koehn Statistical Consulting LLC
Purpose: To present an SPC control chart procedure based on censored lognormal
observations
Abstract: A bicycle manufacturer was interested in testing whether or not the lifetimes of
purchased bicycle frames were in statistical control. Based on prior data, it was
determined that the lifetimes had a lognormal distribution. Censoring the observations
was desirable since time till failure can be very long, especially when the process is in
control or better. A procedure using censoring was developed and its properties
explored. Graphical techniques, a corresponding method for spread, and sequential
methods will be discussed. The procedure is seen to compare very favorably to using
non-censored data. From a practical standpoint, the procedure using censoring is better
not only from a time standpoint, but usually because it better models the data.
Robust Parameter Design Session
Process Optimization through Robust Parameter Design in the Presence of
Categorical Noise Variables
Timothy J. Robinson,
University of Wyoming
William A. Brenneman
William R. Myers,
The Procter & Gamble Company
Purpose: To demonstrate the use of response surface methods for robust design
categorical noise factors are present and when some control factors have levels which
are not equally desirable due to cost and/or time issues.
Abstract: When categorical noise variables are present in the Robust Parameter Design
(RPD) context, it is possible to reduce process variance by not only manipulating the
levels of the control factors but also by adjusting the proportions associated with the
levels of the categorical noise factor(s). When no adjustment factors exist or when the
adjustment factors are unable to bring the process mean close to target, a popular
approach to determining optimal operating conditions is to find the levels of the control
factors which minimize the estimated mean squared error of the response. Although this
approach is effective, engineers may have a difficult time translating mean squared error
into quality. We propose the use of a parts per million (PPM) defective objective function
in the case of categorical noise variables in the RPD setting. Furthermore, we point out
that in many situations the levels of the control factors are not equally desirable due to
cost and/or time issues. We have termed these types factors non-uniform control
factors. We propose the use of desirability functions to determine optimal operating
conditions in the RPD with categorical noise factor setting when non-uniform control
factors are present and illustrate this methodology with an example from industry.
Business Process Modeling Session
Business Process Characterization Using Categorical Data Models
Cathy Lawsom
General Dynamics
Douglas Montgomery
Arizona State University
Purpose: The purpose of this paper is to demonstrate how logistic regression is
used to model a complex business process.
Abstract: Variation exists in all processes especially business processes where critical
input and process variables may be controlled by human intervention. Significant returns
may be realized by identifying and removing sources of variation from business
processes. Because business processes tend to be heavily dependent on human
interaction, they can be difficult to characterize and model. This research develops a
methodology for synthesizing the qualitative information about the performance of a
business process and transforming it into specifically defined categorical data that can
be used for statistical modeling and optimization. The process under investigation is the
identification and pursuit of new business opportunities for a Department of Defense
(DoD) prime contractor. This process is heavily dependent on the people who obtain
information about potential opportunities and make decisions about whether to pursue
an identified opportunity. This research explores methods for taking the demographic,
anecdotal and qualitative data associated with particular business opportunities and
creating categorical data sets that can be statistically modeled. This research illustrates
how binary logistic regression was used to analyze these data and establish significant
relationships between these key process attributes and the process outcome which is
either the win or loss of the opportunity.
Thursday, 10:30 to 12:00 am
Mulitvariate SPC Session
Using Nonparametric Methods to Lower False Alarm Rates in Multivariate
Statistical Process Control
Luis A Beltran
Linda Malone
University of Central Florida
Purpose: To devise a systematic distribution-free approach by extending current
developments and focusing on the dimensionality reduction using Principal Component
Analysis (PCA) without restricting the technique or techniques to normality requirements.
Abstract: Although there has been progress in the area of multivariate SPC, there are
numerous limitations as well as unanswered questions with the current techniques.
MSPC charts plotting Hotelling’s T 2 require the normality assumption for the joint
distribution among the process variables, which is not feasible in many industrial
settings. The motivation to investigate nonparametric techniques for multivariate data in
quality control is that fewer restrictive assumptions of the data are imposed; as such the
assumption of normality is not a requirement. The nonparametric approach is also less
sensitive to outliers, hence more robust and prone to fewer false alarms. In this
research, the goal will be to create a systematic distribution-free approach by extending
current developments and focusing on the dimensionality reduction using PCA without
restricting the technique or techniques to normality requirements. The proposed
technique is different from current approaches in that it creates a unified distribution-free
approach to non-normal multivariate quality control. The proposed technique will be
better based on its ease of use and robustness to outliers in MSPC. By making the
approach simple to use in an industrial setting, recommendations on the process could
be obtained efficiently, resulting in a cost savings and consequently improving quality.
Statistical Monitoring of Dose-Response Quality Profiles from High-Throughput
Screening
James D. Williams
General Electric
Jeffrey B. Birch,
William H. Woodall,
Virginia Tech
Purpose: To present a statistical monitoring procedure for researchers in
pharmaceutical and chemical companies to evaluate the quality of their bioassay test
procedures over time.
Abstract: In pharmaceutical drug discovery and agricultural crop product discovery, in
vitro bioassay experiments are used to identify promising compounds for further
research. The reproducibility and accuracy of the bioassay is crucial to be able to
correctly distinguish between active and inactive compounds. In the case of agricultural
product discovery, a replicated dose response of commercial crop protection products is
assayed and used to monitor test quality. The activity of these compounds on the test
organisms, the weeds, insects, or fungi, is characterized by a dose-response curve
measured from the bioassay. These curves are used to monitor the quality of the
bioassays. If undesirable conditions in the bioassay arise, such as equipment failure or
problems with the test organisms, then a bioassay monitoring procedure is needed to
quickly detect such issues. In this paper we illustrate a proposed nonlinear profile
monitoring method to monitor the within variability of multiple assays, the adequacy of
the dose-response model chosen, and the estimated dose-response curves for aberrant
cases. We illustrate these methods with in vitro bioassay data collected over one year
from DuPont Crop Protection.
Topics In DOE Session
Bayesian Analysis of Data from Split-Plot Designs
Steven G. Gilmour
University of London
Peter Goos
Universiteit Antwerpen
Purpose: In this talk, a reliable method to analyze data from split-plot experiments will
be presented.
Abstract: Often, industrial experiments involve one or more hard-to-change variables
which are not reset for every run of the experiment. The resulting experimental designs
are of the split-plot type and fall in the category of multi-stratum designs (see, for
example, Trinca and Gilmour (2001)). A proper classical statistical analysis requires the
use of generalized least squares estimation and inference procedures and, hence, the
estimation of the variance components in the statistical model under investigation.
In most split-plot or multi-stratum designs utilized in practice, the hard-to-change
variables are reset a small number of times, such that estimation of the variance
components corresponding to the whole plots stratum of the experiment is either
impossible or inefficient. As a consequence, these variance components are estimated
by most packages to be zero (see, for example, Goos, Langhans and Vandebroek
(2004)) and the generalized least squares inferences collapse to ordinary least squares
ones. The resulting statistical analysis may lead to erroneous decisions regarding the
significance of certain effects in the model.
In the presentation, it will be shown that this problem can be avoided by incorporating
the researcher’s prior beliefs regarding the magnitude of the variance components in a
Bayesian analysis of the data. This approach guarantees a more correct analysis of the
data as it will only produce ordinary instead of generalized least squares results if the
data contain enough information to contradict the researcher’s prior. A split-plot
experiment conducted at the University of Reading to identify the factors influencing the
aroma of freeze-dried coffee will be used as an illustration.
How will the status quo be changed?
There is still a lot of research on the topic of analyzing industrial split-plot and other
multi-stratum experiments. The research described here shows that it is essential to
incorporate prior knowledge regarding the variance components in the analysis.
Otherwise, the statistical analysis of the data may be flawed.
Adapting Second Order Designs for Specific Needs: a Case Study
James R Simpson.
FAMU-FSU
Drew Landman
Old Dominion University
Rupert Giroux
FAMU-FSU
Purpose: To develop an experiment and analysis method for efficiently characterizing
and calibrating large load strain gauge balances for detecting aerodynamic forces and
moments on aircraft models in wind tunnel testing
Abstract: Strain gauge balances can be used to capture aerodynamic forces and
moments on aircraft and other vehicles tested in wind tunnels. These balances must be
calibrated periodically using static load testing. The calibration testing procedure at
NASA was originally developed during the 1940’s and is based on a modified one-factorat-a-time method that is time intensive and does not provide a statistically rigorous
estimate of model quality. An approach using experimental design is proposed to
characterize the relationships between applied load (force) and response voltages for
each of the six aerodynamic forces and moments. A second order experimental design
based on the structure of the Box Behnken design was constructed to suit the unique
requirements of the calibration operation. Monte Carlo simulation was used to compare
the proposed design’s performance potential relative to existing designs prior to
conducting a set of actual experiments. Lessons were learned in constructing
nonstandard designs for second order models and in leveraging simulation effectively
prior to live testing. The new calibration process will require significantly fewer tests to
achieve the same or improved precision in characterization.
Six Sigma Session
Six Sigma beyond the Factory Floor
Ron Snee
Tunnell Consulting
Abstract: Six Sigma is a process-focused, statistically-based approach to business
improvement that companies as diverse as Motorola, Honeywell, General Electric,
DuPont, 3M, American Express, Bank of America and Commonwealth Health
Corporation have used to produce millions of dollars in bottom-line improvements.
Initially Six Sigma initiatives were focused on improving the performance of
manufacturing processes. Today there is a growing awareness that additional gains in
efficiency and operational performance can be achieved by widening the scope of Six
Sigma beyond the factory floor to include the improvement of non-manufacturing,
administrative and service functions. The economic advantage of such efforts are
potentially very significant because (a) the administrative component of modern
manufacturing is large and has a great influence on overall economic performance, and
(b) the service industry is in a modern society well over two-thirds of the entire economy.
This presentation will discuss how Six Sigma can be used to improve processes beyond
the factory floor including dealing with the “We’re different” barrier, appropriate
roadmaps and common technical challenges for the methods and tools used. Several
illustrative examples will be presented
Some Trends in Six Sigma Education
Douglas Montgomery
Arizona State University
Thursday, 2:00 to 3:30 pm
TECHNOMETRICS Session
Control Charts and the Efficient Allocation of Sampling Resources
Marion R Reynolds, Jr.
Virginia Tech
Zachary G. Stoumbos
Rutgers University
Abstract: Control charts for monitoring the process mean  and process standard
deviation  are often based on samples of n > 1 observations, but in many applications,
individual observations are used (n = 1). In this paper, we investigate the question of
whether it is better, from the perspective of statistical performance, to use n = 1 or n > 1.
We assume that the sampling rate in terms of the number of observations per unit time
is fixed, so using n = 1 means that samples can be taken more frequently than when
n > 1. The best choice for n depends on the type of control chart being used, so we
consider Shewhart, exponentially weighted moving average (EWMA), and cumulative
sum (CUSUM) charts. For each type of control chart, a combination of two charts is
investigated; one chart designed to monitor , and the other designed to monitor .
Most control chart comparisons in the literature assume that a special cause produces a
sustained shift in a process parameter that lasts until the shift is detected. We also
consider transient shifts in process parameters, which are of a short duration, and drifts
in which a parameter moves away from its in-control value at a constant rate. Control
chart combinations are evaluated using the expected detection time for the various types
of process changes, and a quadratic loss function. When a signal is generated, it is
important to know which parameters have changed, so the ability of control chart
combinations to correctly indicate the type of parameter change is also evaluated. Our
overall conclusion is that it is best to take samples of n = 1 observations and use an
EWMA or CUSUM chart combination. The Shewhart chart combination with the best
overall performance is based on n > 1, but this combination is inferior to the EWMA and
CUSUM chart combinations on almost all performance characteristics (the exception
being simplicity). This conclusion seems to contradict the conventional wisdom about
some of the advantages and disadvantages of EWMA and CUSUM charts, relative to
Shewhart charts.
The Inertial Properties of Quality Control Charts
William H. Woodall
Virginia Tech
Mahmoud A.Mahmoud
Cairo University
Abstract: Many types of control charts have an ability to detect process changes that
can weaken over time depending on the past data observed. This is often referred to as
the “inertia problem.” We propose a new measure of inertia, the signal resistance, to be
the largest standardized deviation from target not leading to an immediate out-of-control
signal. We calculate the signal resistance values for several types of univariate and
multivariate charts. Our conclusions support the recommendation that Shewhart limits
should be used with EWMA charts, especially when the smoothing parameter is small.
DOE for Computer Simulation Session
Application of Design of Experiments in Computer Simulation Studies
Shu Yamada
Hiroe Tsubaki
University of Tsukuba
Purpose: This paper presents an approach of design of experiments in computer
simulation with some case studies in automobile industry.
Abstract: In recent days, computer simulation has been applied in many fields, such as
Computer Aided Engineering in manufacturing industry and so forth. In order to apply
computer simulation effectively, we need to consider the following two points: (1)
Exploring a model for computer simulation, (2) Effective application of simulation based
on the explored model. As regard (1), once a tentative model is derived based on
knowledge in the field, it is necessary to examine validity of the model. At this
examination, design of experiments plays an important role. After exploring a computer
model, the next stage is (2), such as optimization of the response by utilizing computer
simulation. This paper presents an approach of design of experiments in computer
simulation in terms of (1) and (2) with some case studies in automobile industry. For
example, in order to optimize a response by many factors, the first step may be
screening active factors from many candidate factors. Design of experiments, such as
supersaturated design, etc help at this screening problem. After fining some active
factors, the next step may be approximation of the response by an appropriate function.
Composite design, Uniform design is helpful to fit second order model as an
approximation.
Computer Experimental Designs to Achieve Multiple Objectives
Leslie M. Moore
Los Alamos National Laboratory
Purpose: Issues and strategies for designing computer experiments are reviewed for
conducting sensitivity analysis and construction of an emulator.
Abstract: Simulator codes are a basis for inference in many complex problems including
weapons performance, materials aging, infrastructure modeling, nuclear reactor
production, and manufacturing process improvement. Goals of computer experiments
include sensitivity analysis to gain understanding of the input space and construction of
an emulator that may form a basis for uncertainty analysis or prediction. Orthogonal
arrays, or highly fractionated factorial designs, and near-orthogonal arrays are used for
computer experiments for sensitivity analyses. Latin hypercube samples, possibly
selected by space-filling criterion, are in common use when Gaussian spatial processes
are the modeling paradigm or uncertainty analysis is the objective. Orthogonal-array
based Latin hypercube designs are used to achieve both objectives. Improvement in
terms of obtaining a space-filling design will be demonstrated for orthogonal-array based
Latin hypercube design. The impact of competing experiment objectives will be
discussed in terms of loss of efficiency in sensitivity analysis conducted with data from a
Latin hypercube design.
Common Mistakes in Statisical Applications Session
Common Mistakes When
Using SPC (and What to do About Them)
Douglas Fair
InfinityQS International, Inc.
Abstract:
Common Mistake Number 1
Incorrectly Using Related Data in a Subgroup. This discussion will center upon the
common mistake of using a subgroup of data points that have been collected from a
single part. For example, measuring a single can's height in 3 places and calling it a
subgroup. This violates one of the fundamental rules of subgrouping, that of data
independence. However, this error can be rectified by using the proper innovative control
chart, a "3D Control Chart."
Common Mistake Number 2
Combining multiple process streams in the same subgroup. This discussion involves
data collected from multiple machines, production lines or filling heads to be used on the
same control chart. In this case, if a single product weight value is collected from say, 5
production lines, and all 5 weight values are used within a subgroup, then the resulting
Xbar and Range control chart will be less than useful. In fact, alarms will be masked and
process control will be impossible to assess. However, this error can be rectified by
using the proper innovative control chart, a "Group Chart."
Each of these two topics will include discussions concerning
1. Real-life situations that have been encountered
2. Software that can be used to automatically enter data and display it on the innovative
charts discussed above.
Common Practitioner Mistakes in Data Analysis
Scott M. Kowalski
Minitab, Inc.
Purpose: To share some experiences with statistical misconceptions about data
analysis.
Abstract: In companies around the globe, the majority of data analysis is done by people
with very little training in statistics. For example, most green/black belts are given 3-5
weeks of Six Sigma training and only part of that training focuses directly on statistics.
This leaves the typical data analyst armed with a statistical software program, but
without sufficient savvy to know right from wrong. Unfortunately, many analysts make
mistakes when using everything from basic statistical tools to quality tools to design of
experiments. In this talk, I present some of the statistical misconceptions that I have
encountered over the last several years of traveling to companies that are using
statistics to analyze data.
Friday, 8:00 to 9:30 am
Journal of Quality Technology Session
A Dual-Response Approach to Robust Parameter Design for a Generalized Linear
Model
William R. Myers
William A. Brenneman
The Procter & Gamble Company
Raymond H. Myers
Virginia Tech
Absract: Robust Parameter Design (RPD) has been used extensively in industrial
experiments since its introduction by Genichi Taguchi. RPD has been studied and
applied, in most cases, assuming a linear model under standard assumptions. More
recently, RPD has been considered in a generalized linear model (GLM) setting. In this
paper, we apply a general dual-response approach when using RPD in the case of a
GLM. We motivate the need for exploring both the process mean and process variance
by discussing situations when a compromise between the two is necessary. Several
examples are provided in order to further motivate the need for applying a dual-response
approach when applying RPD in the case of a GLM.
Analysis of Performance Measures in Experimental Designs Using the Jackknife
Asokan Mulayath Variyath
Bovas Abraham
Jiahua Chen
University of Waterloo
Abstracts: Experimental designs with performance measures as responses are common
in industrial applications. The existing analysis methods often regard performance
measures as sole response variables without replicates. Consequently, no degrees of
freedom are left for error variance estimation in these methods. In reality, performance
measures are obtained from replicated primary-response variables. Precious information
is hence lost. In this paper, we suggest a jackknife-based approach on the replicated
primary responses to provide an estimate of error variance of performance measures.
The resulting tests for factor effects become easy to construct and more reliable. We
compare the proposed method with some existing methods using two real examples and
investigate the consistency of the jackknife variance estimate based on simulation
studies.
Response Surface Methodology Session
Comparison of Global Characterization Techniques in Response Surfaces
Francisco Ortiz Jr.
James R. Simpson
FAMU-FSU
Drew Landman
Old Dominion University
Purpose: A comparative study of different modeling techniques for the characterization
of systems with a
complex underlying response model.
Abstract: The optimization of multiple responses in experimental design requires
satisfying different, sometimes conflicting objectives. Individual objectives typically
correspond to optimality regions in different locations in the experimental design space.
The practitioner must then locate a region of the design space where the trade-off
between objectives is acceptable. In order to find these acceptable tradeoff regions it
may be useful to characterize the larger region of operability rather than the region
around the current operating setting. This expansion of the region of interest may involve
fitting complex, highly nonlinear surfaces. The purpose of this research is to perform
comparative study of relevant modeling techniques for the characterization of systems
with a complex underlying model. A case study from wind tunnel tests of a model X-31
aircraft is used to demonstrate the advantages and disadvantages of these techniques.
Results from this research could also prove useful for the single response situation
involving relatively large design spaces or optimization cases where it would be useful to
perform a global search in lieu of the traditional steepest ascent local search approach.
Response Surface Design Evaluation Using Mean Square Error Criteria
Christine Anderson-Cook
Los Alamos National Lab
Connie M. Borror
University of Illinois
Abstract: Evaluation of designs for response surface models has, for the most part,
been based on assessment of the scaled prediction variance (SPV). Bias and mean
square error received some attention as bases for choosing satisfactory designs in the
early 1990s, but have largely been ignored as design assessment criteria. A drawback
to the use of bias or mean square error as the evaluation criteria has been the difficulty
to visually assess these quantities over various design regions. This difficulty is a result
of the fact that the user must not only assume a particular form of the model of interest,
but must also state the form of the model to be protected against. As a result, numerous
graphical displays (such as variance dispersion graphs) would be needed to adequately
assess competing designs.
A recently developed graphical display, the fraction of design space (FDS) plot, will be
used to evaluate competing response surface designs using bias-related criteria. The
FDS plot was originally developed to display SPV over the operating region of interest
plotted against the fraction of design space. A single line summarizes the distribution of
the prediction variance by plotting SPV values versus the fraction of the design space at
or below that value. The FDS plot shows promise as an excellent way of showing
graphically the potential bias for various forms of assumed models and various
experimental designs. We will present the results of these comparisons and provide
recommendations.
Graphical Methods Session
Using a Pareto Chart to Select Effects for a Two-Level Factorial DOE
Pat Whitcomb
Stat-Ease, Inc.
Purpose: To present a new method of using a Pareto chart of t-values for properly
displaying relative effect sizes, along with t-limits that aid in the selection of effects.
Abstract : In the past, we have exclusively recommended using the half-normal plot of
effects for identification of potentially significant effects. However, if constructed
properly, the Pareto chart is a useful tool for showing the relative size of effects,
especially to non-statisticians. Problems with either technique arise if the design is not
an orthogonal balanced 2k-p design. In these situations the effects have differing
standard errors, thus the size of the effect may not indicate its statistical significance.
This problem has been addressed for the half-normal (and full normal) plot of effects by
standardizing the effects1. This presentation shows that, for Pareto charts, plotting the tvalues of the effects addresses the problem of non-orthogonal and/or imbalanced
designs. Furthermore, limits based on the significant t-value and the Bonferroni
corrected t-value are added to the Pareto chart and aid in the selection of effects.
Extreme Makeover: Data Edition
Julia C. O’Neill
Lori B. Pfahler
Merck & Co., Inc.
Purpose: To present practical guidelines for more effective presentation of data in tables
and graphics.
Abstract: Even the best educated scientists earn graduate degrees without a single
course focused on presenting data clearly. Small changes in the formatting of tables and
graphics can have a tremendous impact on the clarity of the information displayed.
During our redesign of the statistics training curriculum of a large chemical company, a
comprehensive needs analysis of over 1000 technologists uncovered this previously
unvoiced expectation: the technologists needed advice on effectively presenting data.
This launched our research into the body of knowledge on data presentation and
perception.
Although some excellent works by authorities such as Cleveland and Tufte have been
published, these are not easily accessible by most technologists. We synthesized the
principles and guidelines we found in the literature, and reduced them to a practical set
of rules for data presenters. We use many interesting and relevant examples to illustrate
the impact of each rule.
This presentation has been shared with enthusiastic audiences of scientists, quality
professionals, and teachers and parents of high-school science fair participants. Those
who apply the guidelines presented will see a dramatic improvement in the clarity of their
tables and graphics.
Friday, 10:00 to 11:30 am
Measurement Systems Session
Two-Dimensional Guidelines for Measurement System Indices
T. Kevin White
Voridian
Abstract:
Motivation or Background
When evaluating the adequacy of a measurement process, it is common to perform a
Gauge Repeatability & Reproducibility (R&R) study and look at the % Gauge R&R and
the Precision/Tolerance (P/T) Ratio. The acceptable guidelines for these measurement
system indices were originated in the parts industry and are often not realistic for many
measurement systems in the chemical and process industries. The result is that many
process improvement efforts, including Six Sigma projects, get stuck or detoured into
investigating the measurement process when there is still plenty to be gained by working
on the process. Six Sigma is partially to blame, as these guidelines from the parts
industry have been carried over into the chemical and process industries in the last
several years due to more and more companies starting Six Sigma efforts.
Description (Describe the work done)
The traditional guidelines will be challenged. It will be shown that even when these
guidelines are exceeded, only very small improvements in overall process capability can
be achieved by improving the measurement system. Work will also be shown that
explains the relationship between % Gauge R&R, P/T Ratio, and Cp. From this work,
two-dimensional guidelines for % Gauge R&R and P/T Ratio have been developed that
are more realistic for the chemical and process industries. Understanding how these
indices are inter-related is the key to making a decision regarding the adequacy of the
measurement system.
Significance (Are there improvements, applications, new abilities, new points of
view, etc.? How will the status quo be changed?)
These two-dimensional guidelines will be especially helpful to Black Belts and Green
Belts (and other quality improvement professionals) when performing a measurement
systems analysis. These new guidelines will challenge the traditional guidelines and be
a simple, yet improved way of making a determination regarding measurement system
adequacy. Ultimately, there will be fewer improvement efforts getting side-tracked by
the measurement system when there is still plenty to be gained by working on the
process itself.
On the Comparison of Two Measurement Devices
Joseph G. Voelkel
Rochester Institute of Technology
Bruce E. Siskowski
Reichert Inc
Purpose: To review the problem and difficulties associated with comparing measurement
devices and illustrate how to do it in a methodologically correct way.
Abstract: Measurement devices in the medical industry often have no reference
standards with which they may be compared. In such a case, devices are usually
compared to each other.
Frequently-used methods of comparison include regression, correlation, or the so-called
Bland-Altman plotting method. We review each of these, including any shortcomings
they may have—for example, regression is inappropriate because both the X’s and Y’s
include measurement error. We also contrast our problem to the standard Gage R&R
study that is commonly performed in industry.
A reasonable approach to take is based on so-called structural equation modeling. We
show how such modeling is appropriate for comparing two measuring devices. We show
how two devices may be compared by a sequence of likelihood-ratio tests. An example
based on two devices used to measure intra-ocular pressure of the human eye is used
to illustrate the technique, as well as possible extensions to it.
These methods are not new, but they seem to be unknown to many statisticians. We
hope this talk will help the audience learn about this important topic.
Generalizing Gage R&R summaries beyond two-way crossed models
Annie Dudley Zangi
Nicole Hill Jones
SAS Institute, Inc.
Purpose: To encourage more extended use of Gage R&R in the workplace.
Abstract: Traditional Gage R&R methods and summaries are based on a two-way
crossed model, with operator and part in the model. Often your environment doesn’t fit
into a two variable crossed model, but you still want to identify the contribution of part
versus every other factor you have available (i.e. repeatability and reproducibility). This
paper extends the definitions to include three and four way crossed models, main effects
models, and nested models, then illustrates how to analyze these using JMP® software.
Reliability Session
The Analysis and Comparison of Start-up Demonstration Tests
Michelle L. Depoy Smith
William S. Griffith
University of Kentucky
Purpose: The purpose of this paper is to review briefly the start-up demonstration test
literature and to present new results for existing tests and to propose new alternative
tests which are easier to implement and which, in a number of situations, outperform the
existing tests.
Abstract: Start-up demonstration tests were introduced by Hahn and Gage in the
Journal of Quality Technology in 1983. They can be used to determine the acceptability
of equipment. For example, a vendor may be required to demonstrate the reliability of
the equipment before a customer accepts it. In a start-up demonstration test, there are
successive start-up attempts and some procedure for deciding whether to accept or
reject the equipment. In the Hahn and Gage paper, a specified number of consecutive
successes was required to accept the equipment. In 2000, Balakrishnan and Chan
introduced a modification of the Hahn and Gage test where termination of the test and
rejection of the unit would occur if a total of f failures occurred prior to achieving k
consecutive successful start-ups (CSTF test). Smith and Griffith (2003 and to appear)
and Martin (2004) in separate papers have further investigated the CSTF test. In the
present paper, some alternative test schemes which are easier for the practitioner to
implement and which, in a number of situations, outperform this CSTF test are
introduced and studied. Markov chain techniques are used to do the probabilistic
analysis of the tests and we provide some practical advice on choosing a good start-up
test based on various criteria.
An Early Detection Test for the Compatibility of Two Software Environments
Daniel R. Jeske
Qi Zhang
University of California, Riverside
Purpose: Develop a test to determine if software failure data from a field environment
can be pooled with the failure data from its test environment for the purpose of improving
the precision of failure rate estimates that are made early in the early stages of the field
interval.
Abstract: In order to achieve higher precision in software failure rate predictions, it is
often the case that failure data from the test interval is pooled with failure data from the
field interval. The validity of pooling these two sources of data depends on the two
environments being compatible with respect to the manner in which the software is
used. In this paper, we formulate the hypothesis of compatible environments in terms of
a statistical hypothesis and develop an appropriate test procedure. We assume the
underlying failure process of the software follows a nonhomogeneous Poisson process
and develop a methodology to test whether or not the same Poisson process that is fit to
the test data continues to hold in the field environment. Our procedure includes an early
detection test for the case where there is insufficient field data for asymptotic tests to be
suitable, such as would often be the case in the early stages of field deployment. The
early detection test is conservative, but we outline a bootstrap algorithm that can be
used to approximately calibrate its size. We illustrate our test procedure by applying it to
two real software projects, and also report on a simulation study that confirms its size
properties.
Meeting Challenges in New Product Development Phases Using Accelerated Life
Testing
Sarath Jayatilleka
Maytag Appliances
O. Geoffrey Okogbaa
University of South Florida
Abstract: Some concepts of the uses of accelerated life tests (ALT) at different stages of
the new product development process are discussed here. Specific accelerated life tests
at component and subsystem levels were designed and data were analyzed to identify
early failures and achieve the improvements of product reliability. Goals were defined in
reliability design specifications for the product performance under the specified
operational conditions over the expected life. Several examples from product
development processes are provided in order to explain the above concepts. In the first
phase of a design, time-compressed accelerated life tests (ALT) and Highly Accelerated
Life Tests (HALT) were used. Based on such test results, further design iterations were
tested with HALTs and benchmarked with the previous design iterations. Test fixtures
were designed in order to simulate the normal operating levels and also to accelerate
specific failure modes. Failures in the eyes of the customer such as noise levels were
captured during the normal level operations. Weibull probability plots were successfully
used in reliability growth monitoring processes. In summary, above tests and analyses
were effectively and efficiently used to increase the degree of reliability improvements
between design iterations and shorten the design cycle time.
Process Analytical Technology
Engineering a Proactive Decision System for Pharmaceutical Quality
Ajaz S. Hussain
Food and Drug Administration
Purpose: To engineering a proactive decision system for pharmaceutical quality,
articulate key challenges, and seek solutions to these challenges in both short and long
term.
Abstract: This presentation briefly highlights important opportunities for engineering a
more proactive decision system, discusses its desired characteristics and attempts to
identify operational, tactical and strategic elements to be considered. Uncertainty in the
current decision environment needs to be reduced to facilitate effective and efficient
[risk-based] decisions. This may be achieved through enhanced quality by design (QbD)
principles which may include development of knowledge centric relationships between
material properties, controls, product quality attributes and/or intended clinical
performance. Although in the current system it is well recognized that quality can not be
tested into products - it has to be by design, yet testing to document quality is the most
prevalent approach. Utility of knowledge centric relationships for regulatory decisions
pose many challenges; for example, what are acceptable formulations of knowledge
centric relationships and how should these be evaluated? Two relatively recent
proposals, namely; "Science of Design" and "Engineering Systems", will be examined
and interpreted in a pharmaceutical context to identify and articulate key challenges and
to seek short and long term solutions to these challenges.
Multivariate Calibration for Analysis of Content and Coating Uniformity in
Pharmaceutical Tablets
John F. Kauffman
Food and Drug Administration
John A. Spencer
Food and Drug Administration
Purpose: To present a multivariate calibration for near infrared (NIR) spectra that
simultaneously quantifies tablet coating thickness and tablet compression force.
Abstract: Near infrared spectrometry (NIR) is widely used in process analysis and
control. NIR reflectance of solid pharmaceutical tablets can be used to quantify the
amount of active pharmaceutical ingredient in the tablet, but it is also well known that
NIR spectra are influenced by the compression force used to prepare the tablet.
Typically, spectra exhibit a decrease in the high energy portion of the spectrum as the
tableting force increases. Addition of a polymeric coating to the tablet surface has a
similar effect. This presents a problem when NIR spectrometry is used to monitor both
compression uniformity and coating thickness uniformity in pharmaceutical tablets. We
will present a multivariate calibration method for making simultaneous determinations of
coating thickness and tablet compression force in pharmaceutical tablets. The work is
based on analysis of NIR spectra from a set of coated tablets whose coating thickness is
known, and a separate set of tablets prepared with varying compression forces. We will
also discuss the applicability of similar calibration methods to spectroscopic image
analysis for the purpose of monitoring content and coating uniformity in coated
pharmaceutical tablets.
Application of PAT for Development of a Pharmaceutical Unit Operation
Steven M. Short
Carl A. Anderson
James K. Drennen III
Robert P. Cogdill
Zhenqi Shj
Duquesne University
Purpose: To illustrate the utility of the application of PAT to a unit operation for
secondary pharmaceutical manufacturing.
Abstract: One of the important ideas of process analytical technology is the optimization
of pharmaceutical manufacturing processes. In this presentation, the ideas espoused in
the FDA's PAT guidance are used for unit operation development. Specific tools used for
the development of a unit operation were risk analysis, design of experiments,
application of sensors, and multivariate analysis. The reliability of the unit operation
under study is thereby optimized as an input to subsequent unit operations. An important
objective of this exercise is to demonstrate how application of PAT increases the level of
process understanding thereby allowing less restrictive operating conditions.
Friday, 1:30 to 3:00 pm
Screening DOE Session
Using Fractional Factorial Split-Plots: Minimum Aberration or Optimum Blocking
James M. Lucas
J. M. Lucas and Associates
Frank Anbari
George Washington University
Purpose: To compare the Minimum Aberration and the Optimum Blocking approaches to
designing Fractional Factorial Split-Plot experiments and to suggest a combined
approach
Abstract: Industrial experiments often involve factors that are hard-to-change and
factors that are easy-to-change. A split-plot experiment is often appropriate in this
situation. When a full factorial is too expensive or time consuming to conduct, a
fractional factorial split-plot can be used. This is the situation we address.
A Minimum Aberration (MA) Split-Plot chooses the defining contrast to minimize the
number of “words” of shortest length. Before the publication of Bingham, Sitter and
Shone (2004) all MA papers considered visiting each level of the whole-plot factors only
once. Therefore, the MA literature considered only a small subset of the potential splitplot experiments. BSS removed this restriction and considered blocking patterns that
allowed multiple visits.
Optimum Blocking (OB) chooses the fractional factorial to minimize the maximum
variance of prediction; that is to maximize the global (G-) efficiency. This is achieved by
minimizing the number of model terms that are confounded with blocks. The MA
approach does not sufficiently penalize poor blocking, the criterion can allow too many
model terms to be confounded with blocks. Therefore the two approaches can differ.
We compare MA with OB and show when the two approaches differ.
The 26-1 case will be described in detail. An industrial example will be discussed. Future
work combining MA and OB will be discussed. The combined approach will be preferred
for future applications.
Sequential Supersaturated Designs for Efficient Screening
Angela M. Jugan
David Drain
University of Missouri – Rolla
Purpose: Generating supersaturated designs sequentially for efficient screening.
Abstract: Screening experiments are used to determine which factors have a
significant effect on a response with the assumption of effect sparsity. This can be
crucial in testing potential drugs or improving an industrial process. Supersaturated
designs are one answer to this problem, but most existing designs assume the entire
experiment will be done at the same time. In many situations, the experiment can be
done sequentially, with experimental runs later in the experiment chosen on the basis of
results from earlier runs. We evaluate a population of candidate supersaturated designs
with eighteen factors and eight initial runs, then choose eight additional runs
sequentially. The designs are analyzed with stepwise regression with the use of the
SWEEP operator to find three or less significant variables, and then one new run is
added to the design using the information found. After eight new runs are added, a
fitness value is assigned to the design based on how accurately it identified the true
significant parameters. The genetic algorithm takes candidates with the highest fitness
values and creates a new population of designs to evaluate. The process continues until
set criteria are met. This eighteen factor design is more efficient than any published
design and controls alpha and beta risks at least as well.
Multivariate Regression Session
Evolutionary Algorithms in Multicollinearity Situations: A Case study with
Stabilizing Transformations
Flor A Castillo
Carlos M. Villa,
The Dow Chemical Company
Purpose: In this paper the potential of using computing evolutionary algorithms (GPgenerated symbolic regression) for alleviating multicollinearity problems is presented
with a case study in an industrial setting
Abstract: In many industrial applications, observational data is collected and stored to
later become the focus of a modeling exercise usually for the purpose of process control.
The most desirable models, from the point of view of plant personnel, are often those
developed by multiple linear regression because of the relative simplicity of
interpretation. However, modeling with this type of data provides many challenges. Of
special importance is data multicollinearity or strong relationships between inputs.
Severe multicollinearity can badly affect the precision of the estimated regression
coefficients making them sensitive to the data collected and producing models with poor
precision. In these cases, techniques such as Principal Component Regression, and
Partial Least Squares are often used. However variable interpretation of the resulting
principal components is often difficult, especially for plant personnel. In order to improve
multiple linear regression modeling in the presence of severe multicollineaity, an
approach based on genetic programming (GP) symbolic regression was developed in
the Dow Chemical Company.
GP-generated symbolic regression is a computer-based algorithm that can generate
several functional mathematical expressions (equations) that fit a given set of data. The
algorithm starts by assembling a specified number of equations (a population) by
combining input variables and linking them by mathematical operators. Assembled
equations are evaluated in terms of a fitness function such sum square errors (SSE) or
correlation coefficient. Equations with low fitness are removed and the fittest solutions
are allowed to pass as members of a new generation of solutions (self-reproduction).
The algorithm also creates new solutions by combining bits and pieces of good ones
(crossover) and occasionally making a random change and introducing a complete new
solution (mutation). The new generation of equations then goes to the same procedure.
This process is repeated until a predefined number of generations of equations is
reached. This procedure produces several possible models of the response as a
function of the input variables. This set of equations or models offer a unique opportunity
since it shows possible relationships among variables that, when applied to multiple
regression modeling, have the potential to alleviate severe multicollinearity. Thus
providing a stable regression model in terms of the original variables which is easier to
implement and understand . In this paper the potential of GP-generated symbolic
regression (GP) for alleviating multicollinearity problems in multiple regression is
presented with a case study in an industrial setting. The GP-generated algorithm was
implemented as a tool box in MATLAB
Historical plant data was obtained and a first order multiple linear regression model was
fit to the data. Severe multicollinearity was confirmed by large variance inflation factors
(VIF). The GP algorithm was used and a set of resulting models or equations was
obtained. Equations were ranked in terms of R2 and the one with the largest R2 was
considered first. From this equation, the relationships between the variables were
observed and the original variables were transformed according to the functional form of
this relationship. Later, a first order polynomial was fitted to the transformed variables
and the resulting model was assessed in terms of R2. A residuals analysis was also
performed to check the IIDN(0,σ2) assumption. If the error assumptions were correct,
the VIF were evaluated to ensure that no severe multicollinearity was observed. If the
error structure was still incorrect, then other GP equation from the selected set was
chosen and the process was repeated until a model with no severe Multicollinearity and
with appropriate error structure was obtained. This approach was successfully applied in
a chemical process.. The key advantage of the approach is that produces a model that is
in terms of the original variables with no severe Multicollinearity. This simple, stable
polynomial model is easier for plant personnel to understand and implement.
Hierarchical Monitoring of Defect Rates Using Process Data
David R. Forrest
Virginia Institute of Marine Science
Christina M. Mastrangelo
University of Washington
Purpose: To present a hierarchical GLM and to demonstrate its application in a
manufacturing environment where it is ‘oriented’ toward a specific product goal.
Abstract: In semiconductor manufacturing, discovering the processes that are
attributable to defect rates is a lengthy and expensive procedure. This talk presents a
viable approach for understanding the impact of process variables on defect rates. By
using a process-based hierarchical model, we can relate sub-process manufacturing
data to layer-specific defect rates. This talk will demonstrate a hierarchical modeling
method using process data drawn from the Gate Contact layer, Metal 1 layer, and
Electrical Test data to produce estimates of defect rates. A binary logistic model is used
for the sub-process and meta-models because of the dichotomous nature of the data. A
semiconductor chip is either conformant or non-conformant, and logistic regression
provides a method of estimating the probability of non-conformance. Logistic modeling
is a form of generalized linear models, maximizing the likelihood of a linear model to
estimate the logit link function of the response variable. A benefit of the hierarchical
approach is that the parameters of the high-level model may be interpreted as the
relative contributions of the sub-models to the overall yield. In addition, the talk will
demonstrate how control charts may be used to monitor sub-process outputs where
these outputs have been oriented towards a product performance characteristic and
compares these to multivariate control charts on the sub-processes.
Measurement System Analysis Session
Bayesian Models for the Characterization of Reference Materials
Will Guthrie
National Institute of Standards and Technology
Purpose: To use Bayesian methods to characterize the physical properties of Standard
Reference Materials (SRM's) for use in calibration and measurement assurance.
Abstract: As part of the nation's metrology system, the National Institute of Standards
and Technology (NIST) characterizes the physical properties of a wide range of
materials for use as calibration and measurement assurance standards. Straight-forward
application of linear models often can be used successfully for the summarization of a
Standard Reference Material's properties. However, the measurement processes used
in the certification of some materials require more complex models. This talk will
describe the use of Bayesian models, fit using Markov-Chain Monte Carlo methods, as a
convenient and flexible way to extend the family of models typically used for SRM
certification. Examples will include the use of a three-parameter lognormal model for the
certification of sulfur in diesel fuels and a hierarchical errors-in-variables regression
model for the certification of the specific surface area of cement.
Open Source Excel Tools for Statistical Analysis of Complex Measurements
Hung-kung Liu
Will Guthrie
John Lu
Juan Soto
National Institute of Standards and Technology
Purpose: To develop statistical metrology tools for scientists and engineers that facilitate
statistical analysis by combining a familiar interface with powerful, high-quality, flexible
statistical software.
Abstract: Excel is the tool of choice for data collection for many scientists. However,
it has limited statistical capabilities. The S language is often the vehicle for research in
statistical methodology, and R provides an open source route to participation in that
activity. We utilize existing Excel plugins to harness the power of R scripts to enable
Excel users to use NIST recommended methodologies for the statistical analysis of their
data. In this talk, we will demonstrate two such examples: an automatic uncertainty
calculator, and a sensitivity coefficients generator for mass calibration.
Bayesian 3D Reconstruction of Chemical Composition from 2D Electron
Microscopy
Donald Malec
National Institute of Standards and Technology
Purpose: To develop and study new models for identifying and classifying the chemical
composition of unknown compounds.
Abstract: A 3D adaptation of 2D Markov Random Field models to electron microscopy
will be presented, including the incorporation of energy spectrum recordings to
identify chemical elements. Inherent differences, such as electron scattering, between
electron microscopy and other areas where tomography is used, such as X-ray
tomography, will be discussed.
Download