For my Father, whom we have to assume would have wanted me to get a degree at MIT; for my sister Frances, who never doubted; and for my Mother, who saw us all through. 1 Objective Comparison of Design of Experiments Strategies in Design and Observations in Practice by Ion Chalmers Freeman Submitted to the Engineering Systems Division on May 7, 2004, in partial fulfillment of the requirements for the degree of Master of Science in Engineering and Management Abstract Design of Experiments (DoE) strategies in robust engineering determine which prototypes and how many of each are created and tested. A better strategy is one that delivers a closer-to-optimal performance at a lower experimental cost. Prototype testers who may use statistical DoE, design-build-test, or one-at-a-time methods in a wide variety of industries were sought out and interviewed to examine the strategies used in practice and how they fit into a proposed five-layer process support model. From these interviews, we see that DoE are competently and widely practiced. Some improvements to the state of the practice may include • contracts to specify and reward quality engineering among suppliers to complex product systems • wider use in light of new computing power of system level mathematical models for experimentation on complex systems This thesis also examines the relative value of strategies in a particular response surface using a software-based comparator. The data is modified to simulate data environments with other levels of repeatability and interactions, and the way that these variables effect the performance of strategies is examined. The concept of an optimal design of experiments strategy is developed by abstracting the characteristics of a generic strategy and letting it develop in a genetic algorithm in that comparator. The framework for the evaluation of DoE strategies is one significant output to come out of this work that may be of use in future research. Further, the particular abstraction chosen for DoE strategies is offered to other researchers as an exemplar of a particular perspective, to help engender dialogue about methods for optimizing prototype testing policy. Thesis Supervisor: Christopher L. Magee, Ph. D. Title: Professor of the Practice of Engineering Systems and Mechanical Engineering Acknowledgments I would like to thank Professor Chris Magee, my supervisor, for his vision and constant guidance during this research. Other MIT and Sloan faculty contributed greatly to this work. Professor Dan Frey supplied me with preprints of his papers, as well as the source data used in much of the work. Professor Dan Whitney shared with me his enthusiasm and perspective. Professor Emeritus Don Clausing shared with me Antje Peters’ thesis, and his time, contributing several points about the practice of engineering experimentation to this thesis. Professor Diane Burton was generous with her advice, and at Harvard, Professor Stefan Thomke was a great help in navigating his own work in this area. Several of my System Design and Management classmates, notably Wei Shen and Frank Lanni, and alumni, notably Scott Ahlman, contributed substantially to this material. My graduate studies were supported by the Center for Innovation in Product Development at MIT. I have been lucky enough to enjoy the feedback and encouragement of my officemates Jagmeet Singh, Ryan Whitaker, Tom Speller, Mo-Han Hsieh (who provided the ANOVA analysis in Appendix A,) especially Mike Gray, and other CIPD students including Victor Tang and Katja Höllta. The CIPD staff, Nils Nordal, Kathleen Sullivan, Susan MacPhee and Michael Mack, provided valuable assistance and information at various times, and encouraged me to create a seminar on this research. The SDM staff, notably Dennis Mahoney, Ted Hoppe, Jeff Shao and Bill Foley have rendered me great assistance throughout my stay at MIT. Finally, I wish to thank the following: my roommate Tim Baldwin, for always pulling me back into the world; Tabitha Baker, for giving me hope; and Marisa Joelson, for letting me work on my thesis when I should have been at the gym. Cambridge, Massachusetts May 7, 2004 I. Freeman Contents 1 Problem Statement 12 1.1 Product Development . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 1.2 Performance and Variability . . . . . . . . . . . . . . . . . . . . . . . 13 2 Background 15 2.1 Taguchi Methods & Robust Engineering . . . . . . . . . . . . . . . . 15 2.2 Parameter Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 2.3 Design of Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . 17 2.3.1 The Resistance of Statisticians . . . . . . . . . . . . . . . . . . 19 2.4 Strategies in Design of Experiments . . . . . . . . . . . . . . . . . . . 19 2.5 Adoption . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 2.5.1 Morrow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 2.5.2 Peters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 2.5.3 Thomke . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 3 GABackground 23 3.1 History . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 3.2 Appropriateness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 3.3 Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 3.4 Design Decisions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 3.4.1 Population Size . . . . . . . . . . . . . . . . . . . . . . . . . . 26 3.4.2 Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 3.4.3 Crossover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 5 3.4.4 Mutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Practice 4.1 4.2 29 Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 4.1.1 Conversations . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 Profiles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 4.2.1 Vehicle Handling at a High Performance Car Development Interest . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.2 30 A Master Black Belt at an Automatic Transmission at an Automotive Manufacturing Company . . . . . . . . . . . . . . . 33 4.2.3 Engine Design at an Automotive Design Concern . . . . . . . 34 4.2.4 Consulting at a Corporate Engineering Center . . . . . . . . . 34 4.2.5 Black Belt at an auto parts manufacturer . . . . . . . . . . . . 35 4.2.6 Software project manager . . . . . . . . . . . . . . . . . . . . 37 4.2.7 Decision Support at a Large Commercial and Retail Loan Generator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 Naval Marine Propulsion . . . . . . . . . . . . . . . . . . . . . 39 Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 4.3.1 43 4.2.8 4.3 28 Improvements . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Comparator 45 5.1 Concept . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 5.2 Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 5.2.1 Platform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 5.2.2 Modules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 5.2.3 Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48 The Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49 5.3.1 Paper Airplane Game . . . . . . . . . . . . . . . . . . . . . . . 49 5.3.2 Normal Data Games . . . . . . . . . . . . . . . . . . . . . . . 52 5.3.3 Manipulating the nonlinearity . . . . . . . . . . . . . . . . . . 54 5.3.4 Manipulating the noise . . . . . . . . . . . . . . . . . . . . . . 55 5.3 6 5.4 The Central Command . . . . . . . . . . . . . . . . . . . . . . . . . . 56 5.5 The Strategies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 5.5.1 Validating the experimental designs . . . . . . . . . . . . . . . 62 5.5.2 Guess . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 5.5.3 Full Factorial . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 5.5.4 Adaptive One At A Time . . . . . . . . . . . . . . . . . . . . 64 5.5.5 Orthogonal Array . . . . . . . . . . . . . . . . . . . . . . . . . 66 5.6 Payoff . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 5.7 Evaluating the experimental designs . . . . . . . . . . . . . . . . . . . 67 5.8 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68 5.9 Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 6 Abstraction 73 6.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73 6.2 Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74 6.3 Available optimization methods . . . . . . . . . . . . . . . . . . . . . 75 6.4 Genetic Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 6.4.1 The Codings . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 6.4.2 The Fitness Function . . . . . . . . . . . . . . . . . . . . . . . 78 6.4.3 Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79 6.4.4 Crossover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86 6.4.5 Mutation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 6.5 Optimization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 6.6 Results and Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 6.6.1 Strategy Competence . . . . . . . . . . . . . . . . . . . . . . . 90 6.6.2 Two model runs . . . . . . . . . . . . . . . . . . . . . . . . . . 92 Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97 6.7 7 Conclusion 7.1 98 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 7.1.1 98 Problem Statement . . . . . . . . . . . . . . . . . . . . . . . . 7 7.2 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7.3 Further Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 A Grewen Data 99 103 A.1 The Contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107 B Interview Questions in Design of Experiments 112 C Tables 115 C.1 Selected Configurations for the Strategies . . . . . . . . . . . . . . . . 116 C.1.1 Full Factorial . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 C.1.2 Adaptive One At A Time . . . . . . . . . . . . . . . . . . . . 123 C.1.3 Orthogonal Array . . . . . . . . . . . . . . . . . . . . . . . . . 125 C.2 Performance of the Strategies . . . . . . . . . . . . . . . . . . . . . . 127 C.2.1 Interpreting this Table . . . . . . . . . . . . . . . . . . . . . . 127 D Five Layer Process Support Model 141 D.1 Layers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141 D.2 Effecting Process Change . . . . . . . . . . . . . . . . . . . . . . . . . 143 D.3 Describing Current Processes . . . . . . . . . . . . . . . . . . . . . . 143 D.4 Accountability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144 D.5 Fads . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144 D.6 An Example: Balancing the Registers . . . . . . . . . . . . . . . . . . 144 D.6.1 A process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145 D.6.2 Conflict with other corporate needs . . . . . . . . . . . . . . . 145 8 List of Figures 5-1 Class Structure for Comparator . . . . . . . . . . . . . . . . . . . . . 46 5-2 Comparator Functions . . . . . . . . . . . . . . . . . . . . . . . . . . 56 5-3 Stacked Histogram of Mean Payoff . . . . . . . . . . . . . . . . . . . 69 5-4 Stacked Histogram of Mean Wealth . . . . . . . . . . . . . . . . . . . 69 5-5 Strategy values at different nonlinearity values . . . . . . . . . . . . . 70 5-6 Strategy values at different signal-to-noise ratios . . . . . . . . . . . . 71 6-1 Evolution of Payoff over ten generations(Run 1) . . . . . . . . . . . . 90 6-2 Evolution of Payoff over ten generations(Run 2) . . . . . . . . . . . . 90 6-3 Evolution of Diligence(Run 1) . . . . . . . . . . . . . . . . . . . . . . 92 6-4 Evolution of Diligence(Run 2) . . . . . . . . . . . . . . . . . . . . . . 93 6-5 Evolution of Profligacy(Run 1) . . . . . . . . . . . . . . . . . . . . . . 93 6-6 Evolution of Profligacy(Run 2) . . . . . . . . . . . . . . . . . . . . . . 94 6-7 Evolution of Planning(Run 1) . . . . . . . . . . . . . . . . . . . . . . 94 6-8 Evolution of Planning(Run 2) . . . . . . . . . . . . . . . . . . . . . . 95 6-9 Evolution of Care(Run 1) . . . . . . . . . . . . . . . . . . . . . . . . 95 6-10 Evolution of Care(Run 2) . . . . . . . . . . . . . . . . . . . . . . . . 96 6-11 Evolution of Fairness(Run 2) . . . . . . . . . . . . . . . . . . . . . . . 96 6-12 Evolution of Fairness(Run 1) . . . . . . . . . . . . . . . . . . . . . . . 97 9 List of Tables 2.1 An orthogonal array experimental design of order two for a four-parameter, three-level experiment . . . . . . . . . . . . . . . . . . . . . . . . . . 18 5.1 Variance Components in the Grewen Data . . . . . . . . . . . . . . . 50 6.1 Canonical strategies as phenotypes . . . . . . . . . . . . . . . . . . . 78 6.2 80-Goodness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 6.3 Selections, Run 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 6.4 Selections, Run 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 A.1 Grewen Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104 A.2 Grewen ANOVA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108 C.1 Strategy Value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128 10 Listings 5.1 Player.run() method . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 5.2 The Strategy Interface . . . . . . . . . . . . . . . . . . . . . . . . . . 59 5.3 Listing the Configurations . . . . . . . . . . . . . . . . . . . . . . . . 62 5.4 AOAT Parabolic update() . . . . . . . . . . . . . . . . . . . . . . . . 64 5.5 SQL Query to retrieve comparative value . . . . . . . . . . . . . . . . 67 11 Chapter 1 Problem Statement This thesis examines the application of prototype testing approaches in a number of enterprises, and considers the drivers and perceived utility of these design of experiments strategies. It then creates a framework for comparing strategies, and presents an method for developing a testing approach within that framework. 1.1 Product Development In the course of a product life cycle, decisions are made and commitments created. Each of these decisions has a cost. The earlier in the product development life cycle a good decision is made, the greater the benefits reaped[1]. The spread of time from the first articulation of need that becomes a product requirement, to the final collection of the last retired product is the product life time. However, as there are multiple products in a typical family, it is useful to speak of a product life cycle. A product life cycle comes about through iteration. One product will feed information and infrastructure the next, making subsequent versions better suited to customer needs, as well as cheaper and easier to design and produce. The life cycle, generically, has four phases: Envision, Design, Develop and Deploy.1 It is during the design phase that decisions regarding the implemented form 1 This breakdown, and the alternate Conceive, Design, Implement, Operate, come from [2, Slide 12 and included features are made. 1.2 Performance and Variability When the system design is complete, a component can be expected to deliver a particular function. The measure of the degree to which this function is provided is the component’s performance. However, the measured performance will vary from measurement to measurement, and from prototype to prototype. These are both types of variability that come out of prototype testing. It is during the validated design phase that prototype testing takes place. A component’s performance will indicate how well it suits your need, and its variability will tell you how much you can trust it. All of this activity lies within a single one of the four Ps of marketing, Product2 . The only way to affect profitability in the detailed design phase is through product changes. We can change the product in two ways – we can change its cost, or we can change its quality. Reducing the cost of goods sold (COGS) increases profit to first order. The definition of quality is somewhat flexible; it should be clear here that we are speaking of ’manufacturing-based’ quality[3], which can be considered as performance minus variability, and can be measured in terms of the total loss to society due to functional variation and harmful side effects[4]. An experimental design can reveal to the experimenter the effects on quality of design changes including cost reduction, and uncover ways that quality can be improved. To manipulate the quality of a designed product, and to see how proposed cost reductions might change the quality, a product designer engages in parameter design by selecting modifiable characteristics of the system as control factors and creating prototypes that have particular configurations, which is to say some specific combination of chosen levels for the control factors. This process is referred to as pro8] 2 The other three being Place, Promotion and Price 13 totype testing, and the specific configurations to build and the number of experiments to run on those configurations are together known as Design of Experiments(DoE). Design of Experiments strategies determine how configurations will be selected for experimentation, and how many times an experiment will be repeated. Design of Experiments strategies are the central topic of this thesis. We examine the question of how DoE is practiced in a company, why product designers engage in it, what causes the practice to vary and its perceived benefits. In the first part, we investigate the adoption and practice of DoE in product development concerns. While Some work has investigated how widely adopted DoE is, and how rational it is for a company to pursue it, as discussed in section 2.5.3. While the drivers and inhibitors of Taguchi method adoption have been well explicated[5], the activity of adopting the practice in a corporation has not been researched to the same extent. We felt the practice and understanding of DoE would benefit from some standard way to compare these four types of strategies. In the second part, we therefore created a comparative framework within which different strategies and response surfaces can be evaluated. Given a generic comparative framework for DoE strategies, we naturally asked in the third part how one would use it to find the optimal strategy in a particular data environment under a particular set of rules. So, whereas with the framework we could determine the best of any set of strategies, our new goal was to find the best DoE strategy for a given experimental environment. 14 Chapter 2 Background Taguchi methods have become progessively better adopted since their initial introduction in the 1940s, and have focused attention on robustness and effects of noise in product development. They have come to dominate statistical performance enhancing methods at many American corporations, and continue to be a well-regarded method which engineering firms can adopt to improve the quality of their goods[7]. The selection of control factors and the levels which they can take frames the subsequent work on DoE. This series of tasks occupies the parameter design phase. Taguchi Methods, Robust Engineering and Design of Experiments can contribute meaningfully to a company’s profitability. DoE is most effective if introduced as part of a meaningful quality initiative[5]. However, the performance of one system may have nothing to do with the performance of another. The proper choice of a set of experiments to do, and how often to repeat each member of that set, depends quite strongly on what preconceptions the designer has about the system’s response to changes in parameter levels. 2.1 Taguchi Methods & Robust Engineering Genichi Taguchi, an electrical engineer charged by Nippon Telephone and Telegraph to develop a methodology to produce high-quality products and continue to improve quality in a scarcity environment(Japan) in the 1950s and 1960s, devised an accessi15 ble way to involve reduction of variability in the design phase1 . A narrow interest in reducing variability in response for launching munitions[4, p.2] spread through various industries, and eventually Dr. Taguchi was invited to America by Bell Labs in 1980. Ford Motor Company created what is now the American Supplier Institute to elaborate on and promulgate his ideas. Simply put, Taguchi methods seek to maximize the signal-to-noise ratio, the performance’s mean over its standard deviation. They do this by creating a Quality Loss Function. The Quality Loss Function seeks to capture the total cost to society of variation from an optimum, and is generally modeled as a parabola. The width of the parabola is difficult to gauge, and is set so that components in a system that have less need of precision lose quality more slowly than others, and so that the value does not drop below zero inside the acceptable bounds. 2.2 Parameter Design Parameter design in the second step in Taguchi methods, which consider all that has come before as system design. In the validated design subphase, the shape and specific characteristics of the product are well described, and the designer is seeking information on fine tuning to improve performance and reduce variability. She does this by selecting parameters or control factors from several aspects of the product design. Parameters are selected with criteria including expected influence on the objective signal-to-noise function and ease of manipulation2 , and are kept few enough that the number of tested configurations does not grow too high. Hirschi and Frey[6] note that in a coupled system, more than a few parameters become very difficult for people to consider. These parameters each have a number of values that they can take, or levels, 1 Most of the history here comes from [5, Appendix A], [8], [4, p.2] and the websites for Leeds Metropolitan University and the American Supplier Institute. A more complete history and explanation is given in [7] 2 Influence on signal to noise and ease of manipulation can be difficult to reconcile. Taguchi specifically recommends transformations in the control factors to maximize their impact; transformed control factors are rather difficult to manipulate as they are composed of real control factors 16 defined. A collection of parameters with levels set to particular values is referred to herein as a configuration. The number of tested configurations depends on both the experimental design and the total number of configurations, and parameter design reduces the latter. The total number of configurations T in the parameter space is T = Y np (2.1) 1≤p≤P where P is the number of parameters and np is the number of levels for the pth parameter. So, the number of configurations to test increases geometrically with the number of levels of each parameter. Chapters 5 and 6.5, constrain themselves to rectangular parameter spaces, which is to say systems where each parameter has the same number of settings as any other. This simplifies the equation somewhat to T = NP (2.2) where N is the number of levels for any parameter. To keep the total number of configurations low, parameters are frequently chosen to be ’two-level’, especially for screening DoE (as discussed in section 4.2.4), or ’threelevel.’ Three-level parameters are chosen with the nominal value that comes out of system design and reasonable minimum and maximum values. In a full experimental regime for optimizing performance and reducing variability, a designer might choose many more than three levels, or run tests repeatedly on the same system making finer distinctions. It is useful, therefore, to be able to test many fewer than the total number of configurations. 2.3 Design of Experiments The creation and testing of a prototype with a particular configuration is an experiment. The prospect of testing every possible configuration can be daunting as discussed in section 2.2 and in any case one must decide how many prototypes she 17 Table 2.1: An orthogonal array experimental design of order two for a four-parameter, three-level experiment Parameter Configuration Configuration Configuration Configuration Configuration Configuration Configuration Configuration Configuration 1 2 3 4 5 6 7 8 9 A 1 1 1 2 2 2 3 3 3 B 1 2 3 1 2 3 1 2 3 C D 1 1 2 2 3 3 2 3 3 1 1 2 3 2 1 3 2 1 will build, or how many times – called repetitions – she will repeat an experiment, with a given configuration. These decisions are referred to as Design of Experiments (DoE). The origin of Design of Experiments has nothing to do with product development. Ronald Fisher developed a statistical method at Rothamsted Agricultural Station to investigate the main effects of varying influences on crop yield. The central idea is that a parameterized statistical model links expected response with the values of the control factors. The DoE strategy tries the least number of configurations necessary to establish the parameters in the statistical equation. The Orthogonal Array is a particular type of DoE useful for pulling out main effects and first order interactions. We can speak of DoE strategies in terms of their experimental design, which configurations the strategy visits and in which order. We shall now illustrate the orthogonal array strategy with an example. The paper airplane game is more fully described in section 5.3.1; briefly, it is a 3-parameter, 4-level rectangular parameter space with an easily measurable response. We build an experimental design using an orthogonal array strategy. The order of an orthogonal array design is the number of columns for which – and for less than which – there is no repetition (or, strictly, no more repetition for any one grouping of a that number of columns than for any other.)[9] In Table 2.1, 18 no two columns have any repetition in them, so this is an orthogonal array of order two. The statistical model that corresponds to an orthogonal array of order two is a simple main effects model, which is to say, one that assumes that the response is the sum of some constant and a value for each parameter according to its level. 2.3.1 The Resistance of Statisticians Jeffery Morrow’s 1988 thesis[5] discusses a tension in its first appendix. Dr. Taguchi employed the orthogonal array in teaching his methods, and Robust EngineeringTM is tied to the use of the orthogonal array in many people’s minds. Statisticians can therefore be antagonistic toward the adoption of Taguchi methods, as they perceive it to underplay the complexity of the response surface. The signal-to-noise objective function and the quality loss function are therefore ignored as part of a statistically naı̈ve endeavor. The quality loss function is rejected in part on its own merits with the complaint that the coefficient is difficult to establish. 2.4 Strategies in Design of Experiments The most natural and intuitive way to do product testing is to design the product, make it, and see if it works, adjusting the design if the product fails. This approach is herein called design-build-test and is still popular, as we shall see in section 4.2. Over the course of the Twentieth Century, however, the balanced orthogonal array developed by Fisher has gained wide currency among enlightened product developers, as later described in section 4.3. Box, Hunter and Hunter were great advocates of statistical approaches in experimentation. As we shall see in section 4.3, statistical DoE approaches are often driven by a corporate Design for Six Sigma (DFSS) policy framework. DFSS is essentially a recognition ”that upstream decisions made during the design phase profoundly affect the quality and cost of all subsequent activities[10],” and enacts a Define, Measure, Analyze, Improve and Control model[11]. Policies created within the DFSS 19 framework in a company which has created an environment amenable to devoting significant resources to quality improvements therefore specify instructions for actions that reduce variability in later stages of the product development process explained in section 1.2. Policies in DFSS are set by Green Belts, managers trained in process improvement; Black Belts, trained statistical process improvement specialists; and Master Black Belts, teachers and leaders in service to the rest of the organization. While they can be hired in, their training is frequently provided as part of a corporate push for quality improvement in the company’s products and processes. The presence of belted DFSS staff at a company is a clear signal that some attempt at a framework for statistical quality improvement has been adopted, although not necessarily sign of a conducive environment – which can be developed through Senior Executive Six Sigma training – or a harbinger of successful polity. Neither Fisher nor Taguchi invented the idea of changing parameter settings to find the best effect. A more traditional way to determine the best configurations is to choose control factors to be parameters after the system design has been done, and to vary each one in turn from its nominal position. In a system with no interactions, this nominal one at a time strategy would give the correct influences and point to the best answer. Recent work[12] has reexamined the one-at-a-time strategy and found that it can beat statistical methods under particular circumstances in systems with high repeatability and strong interactions between parameter settings. The more intuitive adaptive one at a time (AOAT), in which a parameter has to be left at its best setting, not its nominal setting, can work better than Orthogonal Array[13]. Vary everything, then one at a time (VEOT)[14], is a similar strategy where in a rectangular parameter space the setting for every parameter is changed with each new configuration until every setting has been tried, then the best is selected and the strategy proceeds as in adaptive one at a time. 20 2.5 Adoption Several MIT theses have considered the adoption of Taguchi methods at American companies. This section reviews some of them. 2.5.1 Morrow In 1988, Jeffrey Morrow surveyed several five large American companies – Ford Motor Company, Davidson Rubber, AT&T, and ITT – for his Master of Technology degree[5]. From this he developed an ’impressionistic and anecdotal’ idea of to what extent Taguchi methods had penetrated these enterprises, and what factors had determined the success of that process implementation. He found some factors common to process implementation in general, that they require strong upper management support, are encouraged by market pressure to change business methods, and that they are more likely to be successful in stably structured companies; some factors common to process improvement methods, that they are hindered by poor training materials and less effective when applied piecemeal; and some special factors, which were excessive ’segmentalism,’ which Morrow defines as ’tightly drawn lines of responsibility between functionally organized departments,’ and the resistance of statisticians. 2.5.2 Peters In 1992 and 1993, Antje Peters worked with Don Clausing[7] to inquire of several US corporations how involved they were in the use of Taguchi methods. She used the determinants laid out by Morrow[5] and surveyed engineers regarding the adoption of Taguchi methods at the three organizations that first brought them to America, Xerox, AT & T and Ford Motor Company, as well as ten others3 . She found that management buy-in, internal experts and reinforced training using the case method were the promoters most often cited by her respondents. 3 comprising Polaroid, General Motors Corporation, Hewlett Packard, Hughes Aircraft Company, Packard-Hughes Interconnect, LSI Logic, Eastman Kodak Company, Lexmark International, Analog Devices Corporation and the Boeing Company 21 2.5.3 Thomke Thomke[15] examined the decisions that stakeholders have to make in the design process around expending resources on experimentation. Thomke references Middendorf’s[16] division of experimentation into three types: device evolution, repeated analysis and synthesis. Device evolution is parameter manipulation in the absence of a statistical model, an adaptive optimization like the adaptive-one-at-a-time strategy. Repeated analysis is experimentation in well developed statistical models, such as orthogonal array or analytic modeling in automatic transmissions. Synthesis recalls the final step in the Taguchi DoE strategy, the confirmation trial, as it is a simple affirmation of an existing mathematical model. Thomke examines mode switching, the tendency of a product design entity to stop doing practical prototype testing and move to some other experimental mode, which is to say computer simulation, rapid prototyping or mass screening. Mass Screening is the creation of a particular test for fitness, which is then applied to many possible design candidates. This experimental mode is largely applied in pharmaceutical drug discovery. Rapid Prototyping is popular in easily configurable or simple to make products. Thomke uses the example of a programmable application-specific integrated circuit (ASIC) chip, and further explores the concept in his case study on IDEO[17]. Computer Simulation, as we will see in section 4.2.1, is a useful substitute for practical prototype building when the parameters’ settings effects are well-known and subject to modeling. While simulation can be a fast and cheap way to get results, it does not supply variability information. In a follow up paper, Thomke, et al.[18] ”propose that strategies and modes of experimentation can be an important factor in the effectiveness of a firms innovation processes and its relative competitive position.” 22 Chapter 3 GABackground Genetic algorithms belong to empirical mathematics, and are a powerful way to solve a problem that is unlikely to yield to analytic methods. The central idea is to allow ensembles of agents to perfect mutable strategies to solve a problem over the course of generations. Genetic algorithms will be applied in Chapter 6.5 to optimize an abstracted design of experiments. This chapter introduces the notion of genetic algorithms and provides some background. 3.1 History The notion of genetic algorithms – solutions that are perfected by iterated selection – may be as old as digital computers themselves, and go back at least as far as Baricelli[19] and Box[20] (separately) in 1957. University of Michigan researchers, notably John Holland, initially developed the Genetic Algorithm approach leaning on Rechenberg’s evolutionary strategy and Fogel, Owens and Walsh’s evolutionary programming 1 . The missing element is these was the notion of crossover or structured recombination discussed below in section 3.4.3. With the 1975 release of both Holland’s Adaptation in Natural and Artificial Sys1 the history of genetic algorithms is more completely explored in Marczyk[21] and in chapter 4 of Goldberg[22] 23 tems and Kenneth De Jong’s An Analysis of the Behavior of a Class of Genetic Adaptive Systems[23], genetic algorithms were first defined as an area of study. De Jong found that genetic algorithms could be used to study nonconvex, stochastic processes. He defined the on-line performance xe (s) of strategy s in environment e over the course of T generations as[24, section 4.6] T X n 1 X xe (s) = fe (dti ) T · n t=1 i=1 (3.1) where n is the population size, f (d) is the fitness of agent d and dti is the ith agent in generation t. The on-line performance is thus the running average of the objective function value. De Jong also defined an off-line performance x∗e (s) = 1X ∗ fe (t) T (3.2) where fe∗ (t) = max (fe (1), fe (2), fe (3), ..., fe (t)) (3.3) and is the running average of the highest fitness attained by an agent in each generation. The off-line performance retains the best achieved value in a generation, the online performance averages across all agents with the strategy in the environment. While smaller populations have better on-line performance, larger populations are able to find a better answer, and have ultimately better off-line performance[22, p. 111]. Genetic Algorithms were first implemented to try to reproduce nature, and slowly gained ground in illustrating the shapes of solutions to nonconvex problems, problems in noisy environments, and discontinuous problems. Some recent applications of genetic algorithms have been in machine learning and design optimization, as well as in generic nonlinear optimization techniques. Genetic algorithms are a rather broad field, and are today readily encapsulated in platforms such as MatLab or Swarm. The genetic algorithm optimization scheme discussed in Chapter 6.5 was written ’from the ground up’ in a fourth generation language. 24 3.2 Appropriateness Genetic Algorithm models are a technique for the optimization of non-linear systems. They do not require that the fitness function, which for this study will be the payoff an agent gets less its experimental cost, or its wealth, be continuous between adjacent points. Nor do they require that a particular parameter choice get the same answer every time, particularly important here because not only might two players with the same DoE strategy behave differently, but due to variability in the response surface, two players with the same experimental design may select different final configurations, and two players with the same final configurations may get different payoffs! This variability represents the intrinsic variability in Product Development, where payoff and response surfaces are inconsistent and unpredictable, and where making the right decisions in a rational framework does not guarantee a good result. 3.3 Application It is useful when thinking of genetic algorithms to understand certain terms[22]. A gene is a bit in a genome or chromosome, which is a series of true or false values called a bit string. The gene’s locus is simply its position, which we will describe with numerical indices. The value of a particular bit string is referred to as an allele. Patterns found in the alleles of different objects are referred to as a schema (pl. schemata.) The genotype is the collection of all alleles associated with an object, and the phenotype is how the genotype affects the fitness function, which is the measure of the object’s success. Agents, or players in this work, are run simultaneously with many other players in a generation. They are then selected and bred to produce new players for the next generation. The schema theorem states that the patterns in successful agents receive exponentially more representation as the agents iterate through more generations[25]. This, along with implicit parallelism, the idea that n3 schemata are being tested with any n 25 agents[22, p. 41], is what gives the genetic algorithm approach its processing power. 3.4 Design Decisions In the creation of a genetic algorithm scheme, certain design decisions have to be made to distinguish one model from another. This section outlines what the decisions are, and attempts to point toward implications. The decisions as made are discussed in section 6.4. 3.4.1 Population Size Population size determines how subject the population is to premature convergence through gene starvation. Too large a population size will slow the emergence of a best answer. 3.4.2 Selection The fitness function, or objective function, comes into play when the selection mechanism is enacted. Fitness is determined through some scalar measure of the strategy’s performance, and is generally scaled in some way. Brindle[22, p. 121] suggests six schemes for selecting the strategies to be propagated to the next generation. A generic agent was to create a strategy by 1. Deterministic Sampling: probability of selection is fi / P f , the agent’s fitness over all the sum fitness for all agents. 2. Remainder stochastic sampling with replacement: fi /µf , the ratio between one agent’s fitness and the population’s mean fitness, is calculated. The strategy is awarded the whole part, the remainder is used as a probability for additional selection. 3. Remainder Stochastic Sampling without replacement: As in the previous scheme, but with the probability, which is less than one, zeroed out once the strategy is 26 selected. 4. Stochastic sampling with replacement: The ”roulette wheel.” All of the fitness evaluations fi are placed end to end, and a uniformly distributed random number U [0, P fi ] determines which strategy is added to the selection pool. If for example, the number 0.0132 was selected, the first agent had a fitness of 0.0072, the second 0.0031, and the third 0.0049, the third agent would be selected. This is referred to as a ’with replacement’ strategy because the same agent would be selected again if another random number u s.t. 0.0103 <= u < 0 is picked. 5. Stochastic sampling without replacement: Like the roulette wheel, but with the probability of selection dropping by 0.5 (or to 0)2 every time a strategy is selected as a parent 6. Stochastic tournament: The stochastic tournament selects a pair of agents using a roulette wheel scheme, then compares them to one another, discarding the one with the lower value. Goldberg[22] reports that Brindle found little difference in performance between genetic algorithms using these different methods, except for deterministic selection, which performed poorly. 3.4.3 Crossover In order to generate new strings, agents are mated by splitting the genome and pasting together different alleles from different agents. So, therefore, at which point the parental genomes are split and how they are constructed are important design considerations for the genetic modeler. A model could use a single or multiple genomes to represent a player, could take advantage of diploidy to allow alleles that are currently unhelpful to persist in a dynamic environment. 2 De Jong developed this scheme, and selected the parents separately, which is why this number is 0.5 and not 1. 27 Crossover has the greatest impact on how well the phenotypes of the children represent those that were successful in their parents. 3.4.4 Mutation However, mutation can also have a great impact. Mutation is the easiest concept in genetic algorithms to express or conceive, as it is essentially the chance that any given bit, once determined by generation, selection and crossover, will flip. A mutation of 0.1 reduces a genetic algorithm to little better than a random search, and at a rate of 0.5 they are literally the same[22, p.111][23]. 28 Chapter 4 Practice 4.1 Method To assess the actual usefulness and status of Design of Experiments in various enterprises, a survey was constructed and delivered by phone to four or five practitioners in several industries. We found these DoE practitioners by use of three techniques: 1. Through personal connections with professional engineers 2. Through academic ties, using fellow students, alumni or faculty 3. Through professional associations We were able to recruit seven of these practitioners into confidential conversations regarding their practice of DoE. The interview notes were then edited for clarity and sent to the subjects for review and approval. These conversations were meant neither to provide a holistic look at industrial use of DoEs nor to follow up in detail on any previous study. Our intent is merely to ask practitioners close to hand how they use design of experiments, what are their approaches to it and issues with it, and how well they see it as working for them. 29 4.1.1 Conversations The interview guide is included in Appendix B. Interviews took place at a time of the subject’s choosing over the phone, and took approximately an hour. The interview guide investigates such concerns as mode switching, parameter design, policy implementation, response to management time pressure, testing budget and DoE philosophy. Interview subjects were forthcoming about each of these topics, except where their experience did not help them answer the questions. Conversations did not follow the guide exactly. Often, the questions that were in the guide had been answered or rendered irrelevant in the answer to an earlier section, and interesting comments the interviewee made were followed up on, rendering the conversations at times discursive, and at times digressive. 4.2 Profiles The following profiles correspond roughly one to one with the interviewees. In one case, an interviewee was willing to speak to DoE strategies in two different industries, and has been split into two profiles. 4.2.1 Vehicle Handling at a High Performance Car Development Interest Respondent #1 is engaged in design and development of the chassis his company produces. While his corporation has no official policy or process for design of experiments in quality engineering, his division has a semi-formal policy set around the use of DoE. He uses D-Optimal Designs/citeneubauer:dopt to get the coefficients of his quadratic model for parameter effects, which he has been using for the last seven of the ten years he has been doing design of experiments. Respondent #1’s experiments are analytical primarily, with no variance in response. He uses DoE to unwind the strength of a control factor in a particular mix. The DoE experiments, of which he typically does between 100 and 120 for his 1030 factor, 3-level system can take up to an hour of computing time each. Once he has run them all, he can very quickly see what the effects of changing particular parameters would be. Typically, his statistical model has an R-squared adjusted fit of 0.98 or better with the response of his finite element analytic model. Once the model is in place, Respondent #1 uses genetic algorithms to find what he calls the ”best balance” configuration for his objective function. He does not call it ”optimum” because it uses weighting factors based on engineering judgment. Respondent #1 chooses his three levels for the parameter design pragmatically. Knowing that his decisions will go into the final detailed design for an actual product, he selects the experiment run order in physical DoE’s that varies the factor which is hardest to change the least amount. The range for his factor levels are based on experience with the system and what is judged feasible within design and development and many constraints including the laws of physics, package, design for manufacture and cost. Vehicle dynamics, the performance category Respondent #1 works in, heavily relies on subjective ratings of vehicle attributes. DoEs are used at times in a physical system on the road and track to match this subjective feeling to measurable performance objectives like various vehicle accelerations, gains and displacements. Various parametric and detailed vehicle dynamic models are also utilized heavily for this task. In addition, Respondent #1 and his team completes these DoEs in labs on hydraulic ride simulators, otherwise known as a four-post rig because the vehicles four tires sits on four hydraulic rams that move vertically to simulate the road. The fourpost rig is a more controlled environment than the road and track as it does not suffer ambient weather changes, resultant track changes, driver line changes and ultimately tire behavior change. Therefore, it reduces the amount of noise to deal with in the system experiment. Initially, he starts with a two-level screening DoE to find out which of around 15 factors are important, then returns with a three-level DoE on a few selected factors. It has been about three years since he has felt a need to do a practical DoE, because 31 of the high r-squared fit of the model; he estimates that he has only done twenty in his ten-year career. As noted above, the outer array of the test track, with rain and lain rubber, can be hard to control, and so he prefers to do his testing in the lab. He can do lab experiments on components, which have fewer control factors and typically fewer interactions, using a quarter-car or a whole-car test rig. He will do five or six repetitions on the track, and only three to five in the lab. Respondent #1 is trying to improve twenty or more vehicle responses. Most of them are improved through minimization, although others are best maximized, like Aerodynamic downforce and average vertical tire loading, and some are ”optimum” at set values. Respondent #1 has also worked in designing automotive race cars, for which the vehicle dynamic models were not as well developed. That experience, he feels, is more typical of the industry. ”80 Respondent #1 feels that most try to build ”mental models” and engineering knowledge through design-build test and one-factor-at-a-time strategies. He believes this is insufficient in complex dynamic systems, like a automobile, which invariably exhibit significant interactions, non-linearities and far too many factors and responses to keep track in a person’s head. Respondent #1 believes in a balance of strong theoretical laws of physics knowledge, analytical modeling, physical design-build-test iterates and DoE application. Otherwise, one tends to end up at local ”optima” at best and have spent far more resources and time to get there. Ultimately, he has no doubt in the value of design of experiments. As one example, he points to how he completed four major kinematics and compliance DoEs recently on a project that he estimates would have taken six to eight months longer with lower performance using design-build-test. He defines his test budget in terms of scheduling; he’s constrained by prototype availability, parts availability and project timing. He uses a mostly intuitive process for determining the cost-benefit, driven by his confidence in the model he is using. ”All models have limitations. Models are just abstractions of reality. You have to understand any model’s strengths and weaknesses to gauge which information you 32 can use when and where.” 4.2.2 A Master Black Belt at an Automatic Transmission at an Automotive Manufacturing Company Respondent #2 performs DoEs on automotive transmissions, seeking ways to decrease costs in the automatic transmission in future production runs of the same vehicle. While other engineers in her group use analytic DoEs, her DoEs are purely practical. The automatic transmission is a complex system in and of itself. In order to create realistic models for the transmission, Respondent #2’s team had to consider both the theory of gear sets for the hardware component, and of hydraulic stability, which came out of analytic work on the origins of shift quality. As her main concern is reducing the vibrations due to shifting, all of her optimizations are of the ”smalleris-better” variety. She works with her six-person group to decide what the objectives should be, but typically measures gear noise, ’XYZ’ noise, torsional access measure and airborne noise. In order to provide as pleasant a ride as possible, the vehicle handling department at the product developer visited in section 4.2.1 needs to be able to predict the vibrations of components like the automatic transmission. It is therefore important to Respondent #2 that she be able to shift the mean of the transmission’s vibration as little as possible when the gear changes, as well as minimize its variation. To do this she uses design of experiments. The first step in her experimental design is to do some probe testing, choosing well spread configurations as in uniform design, and repeating them until the standard deviation bounds. A desire for a 90 % confidence level gives her the notion of variability she uses on her DoEs. While Respondent #2 assumes constant variability across her design space in the design of her experimental strategy, she acknowledges that this is sometimes clearly not the case, ”Sometimes in a solenoid coupled to a hydraulic package, the variability is quite large in one part of the parameter space compared to another.” In these cases, 33 where variability grows extremely large at particular configurations, she and her team have to do many repetitions in order to establish the variability. It is only in such systems that she sees DoEs granting no measurable result. Respondent #2 typically limits her parameter selection to three or four control factors, or ’knobs,’ and prefers to select knobs that do not have a large impact on cost, recalling ” I ran a gear noise DoE a few years ago, and burned through a halfmillion dollars pretty quickly.” The expense of her DoEs are driven by time; which is consumed in setting up and breaking down the experiments as well as by performing and analyzing them. While her experiments generally use two-level parameters, many of the physical systems do have curvature to them. She sees one great advantage in working on the continuous improvement of a released product, instead of its development. As the product has a warrantee history and a product scrap history. This lets her do realistic cost-benefit analyses when designing an experimental design. However, she suggests greater focus on variability reduction in product design. 4.2.3 Engine Design at an Automotive Design Concern Respondent #3 does practical and analytic DoEs on engines in new product development. The last product she worked on was designed from the inside out using Robust Engineering principles. The experimental design of the engines is driven by the deep engineering understanding of the system built up over decades, and by reverse engineer and X-ray analysis of competitors products. Analytic models using finite element analysis represent the engine, and the experiments are run against these models. 4.2.4 Consulting at a Corporate Engineering Center Respondent #4 has been working at a center for engineering excellence (CEE) at a large corporation with diverse products since graduating college in 1997. He took a 34 DFSS course in 1998, and perceives DoE as being part of his employer’s six sigma push. As an engineer at a CEE, Respondent #4’s responsibility is to help other business units fill gaps in their expertise to bring new ideas to market. He spends much of his time in concept selection using voice-of-the-customer methods like Kj to capture needs and translating those needs into requirements using tools like Triz, QFD and the Pew selection matrix. However, he is occasionally called upon to do DoEs. To do practical DoEs, Respondent #4 uses minitab to generate a list of experiments once he has input a presumed statistical model. He last used DoEs in the year 2000 to optimize the performance of a guidance system for a ground vehicle. Magnetic beacons were lain in the vehicle’s path, and Respondent #4 varied around a half-dozen parameters, all of which, like the type of sensor, had discrete values except for the speed of the vehicle, which could vary continuously. He used DoEs to determine how far apart the beacons could be placed and still define a path. Other engineers in his group run DoEs against finite element models. In Respondent #4’s team, they use their embedded engineering knowledge to decide when to run a DoE and which control factors to use. 4.2.5 Black Belt at an auto parts manufacturer Respondent #5 was trained as a black belt by her employer, which has its own training program. She has started to move into a Master Black Belt role, training other staff on statistical methods. She works with design of experiments frequently in three regimes: 1. Manufacturing process improvement 2. Manufacturing trouble shooting 3. New Product Development Respondent #5 recently worked on a muffler assembly. Her company formed an interdisciplinary team to develop a new muffler assembly. She was the six sigma 35 technical expert and was charged with creating the experimental design. After a few meetings, she proposed doing multiple repetitions on the experimental design in order to capture some of the variability, and was ”was laughed at in the meeting.” Respondent # 5 ruefully concludes ”One [experiment] is better than nothing.” Respondent #5 is sanguine about the lack of repetitions. While her statistical training recommends them, she feels the benefit would not be worth the cost, and that the testing environment and manufacturing processes are so well-controlled that any variability would be unimportant. The prototypes took hundred of dollars to build; the product was eventually canceled because of its cost. Respondent #5 and her team had six months for the project. A larger driver than prototype cost in the non-repetition of their experiments was perhaps the minimum standards they had to pursue. Generally in new product development, Respondent #5’s company is given specifications by a systems integrator and stop testing once the meet the standard. They have no incentive for reducing variability or delivering performance better than specified. Respondent #5’s team started off identifying important control factors an interactions using their physical intuition, physical analysis, knowledge and experience. They created a list of about 30 control factors this way. After considering them analytically, they drop to about 20, thence to around 12 when the uncontrollable or insignificant ones are dropped. Then the team ran an L12 Plackett-Burman design, which is a kind of orthogonal array[9, Section 7.4] for 11 factors at 2 levels. This screening DoE allowed Respondent #5 and her team to focus on the four most important factors, on which they ran a full factorial with a center point, for a total of 29 prototypes. The center point was to seek evidence of curvature, which was not forthcoming. If she had found curvature, she would have had to review the results from the other prototypes and see if any of them were already good enough. If they weren’t, she would have had to go back to her customer to negotiate for more money, perhaps enough to do a response surface design. 36 Respondent #5 is chary of aspects of Taguchi methods, calling the quality loss function ”just a teaching tool” and indicating that three-level parameter design gives up much of the advantage of DoEs by requiring so many more experiments. ”Taguchi provided a simple, practical method but sometimes other methods give better results. It all depends on the applications, situations, and how people use them.” 4.2.6 Software project manager Respondent #6 is a project manager at a transit consultancy. The consultancy provides a customizeable software platform to transit authorities, and performs safety inspections on transit system elements. In neither endeavor does the consultancy use DoE. Respondent #6 exclusively engages in design-build-test. ”We are not so sophisticated that we test the products against different platforms.” If his company sells additional components or reports with the system, they will test them. If the components do not work, they return and debug them. The testing budget is decided from experience with similar types of products for the same or similar customers. They stay on budget about 15 % of the time. ”Usually the smaller the task is, the easier it is for us to come in under budget.” 4.2.7 Decision Support at a Large Commercial and Retail Loan Generator While replacing terminology like ’performance’ and ’variability’ with ’reward’ and ’risk,’ Respondent #7 believes her work is fundamentally the same as that of a quality engineer working on an engineered product. She handles a diverse array of issues using DoEs, including capital allocation for the bank, process improvement, and individual loan decisions for retail and commercial customers. These regimes are all treated differently as far as the decision making process; they all use design of experiments, but somewhat differently. Belted DFSS specialists as discussed in section 2.4 work out statistical recipes to 37 use for evaluating applicants for different types of lending products. These recipes are adjusted by each process owner by his assigning weights to the different components. It is this customized recipe that is built into software that loan officers run on their desks. The lender does do some practical DoEs in their new product group, adjusting product parameters and testing the market response. However, in Respondent #7’s area, the testing is all on analytic models. Unlike in engineering, where parameters are considered independent to first order per the hierarchical ordering principle (c.f. Wu and Hamada[9, section 3.5]), the parameters in her models are strongly interdependent. However, the statistical model is not developed enough to create a fractional factorial to serve it, and Respondent #7 favors uniform design. ”Orthogonal array can require too many experiments, and not give the best coverage. I don’t need complete orthogonality; I need the best information. Uniform design techniques are far more popular in the practice.” In another distinction from physical engineering, whereas in something like Respondent #1’s work in vehicle handling quite complex finite element models can take hours to run, in financial modeling a complete experimental design with repetitions can be run in two hours. In capital allocation, Respondent #7 attempts to maximize value-at-risk. The stakes for the lender are much higher in a capital allocation model than in a lending scenario, and the choice to use design of experiments is made by the corporate leadership. First a model is built using the loss database, market information, historical context, and whatever other information is appropriate to a particular scenario. These models may include analytic models particular to a scenario – a valuation might use a Merton or Black-Scholes model. This is an analog to the physical analysis of an engineering system. Once the model is developed, it is used to predict the probability distribution of the output. Since the inputs are correlated and interact in a very complex way, normally distributed input parameters do not create normally distributed output 38 parameters, and the only way Respondent #7 is able to establish the distribution for her objective function is to perform a monte carlo simulation. In credit risk applications, Respondent #7 has some ten objective functions she attempts to optimize with hundreds of parameters. Typically, she will select a score or so of knobs to tweak. She uses this model and a monte carlo simulation to establish rules for making credit decisions. The model does not behave linearly – credit ratings of 41 or below will be treated the same, while above four the value of the loan will drop steeply. Respondent #7’s parameter level selection is driven by nonlinearity and competitive information. She needs to capture as much curvature as possible, and needs to see the implications of matching the other offers in her market. For example, price elasticity is a parameter that effects probability of default or mean time to default in strongly nonlinear fashion, and for which other lenders will have diverse values. To do a stochastic assessment, she makes her number of levels equal to her number of runs for uniform coverage. If she uses a latin hypercube, she can distinguish 27 parameter levels in a 27-run DoE. Respondent #7 has no formal methodology for balancing the number of repetitions she does with the number of configurations that she tests. However, she does need to do many repetitions to establish expectation values and variances in any of her models, and almost always has a large number of control factors. Finally, Respondent #7 uses DoEs to assess operational risk, which includes risk to the bank’s reputation or credit rating, risk of legal actions like law suits, and so forth. 4.2.8 Naval Marine Propulsion Respondent #8 develops shafts and rotors for sea-going vessels. As he is a defense contractor, the specifications that his group works with specify minimum performance, reliability and safety standards for his product. Surpisingly, given naval architecture’s proud engineering history, DoE is not generally practiced nor attended to. 1 Credit ratings vary from one to ten, getting worse as they get higher 39 In the same industrial sector, on the other hand, Power generation uses DoEs extensively and has a well-developed list of tradeoffs defined. According to Respondent #8, The power generation side of the business has become involved in DFSS and employes quality engineering tools. While Respondent #8 is aware of some DoE work that has gone on in his division, he maintains that it is ”rarely very formal, and tends to be full-factorial in nature.” 4.3 Analysis DoE is alive and well in American companies. One surprising result wwas how much each practitioner felt his or her experience represented all DoE practitioners in design. While the development of statistical models and the policy of experimentation vary widely, there are some broad similarities. DoE practitioners typically do a two-level balanced screening DoE – often one of the Plackett-Burman designs – to select control factors, then investigate curvature with three or more levels on those knobs, replacing real-system experimentation with numerical models where possible. Experimentation and theory eventually lead to reliable statistical models, at which point the experimental design is determined by a software tool. Based on these interviews, we can define a five layer model for intentional process support in an organization. At the broadest level, an environment is created that is amenable to the activity. This allows the creation of a framework in which policies can be articulated. Next come the policies themselves. Once the policies are in place, instructions can take place inside that policy, and on the most immediate level, actions are performed that enact the process. In business process change, they would enact the transformation, or, rather, the new way of doing things. These layers are all dependent, each on its neighbors. Actions without instructions are ill-coordinated, and very possibly harmful. A policy is moribund if not followed up with instructions, and suboptimally effective if not articulated within a framework. We can see from our investigations in 4.2 that the introduction of formal design 40 of experiments strategies into product development processes follows this model. • An environment is created when the senior management decides that reducing variability and improving quality are important goals to which company resources should be allocated. • A framework is put in place by working with consultants, academics, publications and peers to create an operation goal of Design for Six Sigma, Robust Design or some quality improvement program. This framework is supported by seminars, classes and incentive structures. • A policy of employing design of experiments to discover and manipulate control factors is articulated within that framework. • An instruction comes from the engineer, and engineering manager, or through software code to a practitioner to perform a DoE. • The action of performing the DoE is completed. Most of the interviewees were involved in stable DoE practices where these levels were fully realized and harder to see. However, we can look at Respondent #4’s experience with the ground vehicle as examples of instructions without a policy that did not persist. Respondent #2 and Respondent #1’s experiences differ because one is in an organization that has an explicit and defined policy in DoE, and the sets policy on the same level as instruction, whereas the company provides a framework in which these policies can be created. DoE was in many cases a ’Design for Six Sigma’ (DFSS) framework push. The main implementation mechanism were classes in experimental design, and the adoption of the DFSS Black Belt hierarchy. The DoE practice was introduced into organizations by a specific DFSS vendor, who taught classes to create Green Belts of design engineers. These Green Belts generally decide locally the best way to apply these statistical practices. From the ground-level view of this survey, this strategy works impressively well in creating statistical control in product design. 41 Tools such as minitab are useful to the statistical experimenter in creating the experimental design, and their availability and usability may explain why orthogonal array designs were not common in our survey group. GASolver can help determine the best configuration in a nonconvex response surface once the statistical model has been performed. Microsoft Excel is used by nearly all practitioners for some part of the process – tracking experiments, statistical analysis, or plotting the output. At what organizational level the policy was created depended on the stakes for the company – the lender specified DoE for capital allocation at a department level, but slipped it into loan approval in tools development – and on what level of the organization variability reduction had an advocate. These drivers are more explicitly lain out in Morrow[5]. As expected by Thomke[15], computer simulation has come to dominate design of experiments. The combination of a believable system level mathematical model, which allows experimentation to take place in a few hours, and a reliable statistical model, which obviates the need for running a full factorial with that system level mathematical model, creates a powerful ability to perfect well-understood product systems. However, in some cases, such as Respondent #7’s coworkers in new product marketing for financial instruments, the statistical model is not well-developed and the underlying mechanisms are not well-described. In these cases, DoE practitioners tend to stick with statistical models that handle main effects and, perhaps, first order interactions, depending on the time and money they have to spend. That time and money is afforded them gladly. In this study, we did not look at DoE evangelists who are working to overcome policy impediments in their organizations, and found that the engineers had enough space and money to complete DoEs that they found were worthwhile. To some extent the number of experiments was budget limited, but in no case did a practitioner express frustration with management pressure to cut schedule or cost by reducing the scope of their experimental design. The above is not to say they expressed no frustrations whatsoever. While the practioners all felt that reducing variability was worthwhile, and that their design of 42 experiments strategies were well chosen, most of them had a message that they wanted to include. These represented frustrations with their industries, tools or companies, and I have listed a sample here. • Concept selection is now driven by preconceived business needs whereas in the past our corporation has gotten a lot of benefit by creating new products and then finding needs for them. • The industry could be doing more optimization of subsystems • Variability reduction in product design could go a long way. We spend a lot of money on premium production processes when we could design sensitivity out. 4.3.1 Improvements We will follow Frey, et al.[12] in suggesting that the quest for a statistical model is not always worthwhile. In product development – especially in a case like Respondent #4’s, where the products are very diverse, or Respondent #2’s, where the performance is determined by complex interactions, an experimental design that implies a statistical model is not obviously preferable to one that follows some other type of optimization algorithm. The assumption of homoscedascity, which is to say, that variability will be constant in all parts of the parameter space, persists even in physical environments where the practitioner knows the assumption is false. This results in as many repetitions being performed on configurations where the variance is more accessible as on configurations where it is difficult to establish. One way to improve the cost and quality characteristics of an experimental strategy would be to distribute the number of repetitions according to the repeatability of the response. Indeed, as the variability of the response is one of the measure parameters, one can envision an iterative process where measured variability determines continued repetition. Many engineers at the system level were contacted and do not use design of experiments. While they generally insist that DoEs are necessary for component development, they generally feel that the state of the art in systems engineering is not 43 well developed enough for DoEs to be worth the considerable investment in time and money required to do repeated prototype testing. Further the benefit of doing DoEs is not expected to outweight the difficulty and expense of building the detailed system level mathematical analytic models that it would require. This stance seems to be belied somewhat by the fact that computers are now powerful enough that engineers like Respondent #3 can do DoEs against analytic models for automotive engines. While it would be facile to recommend that all systems engineers start performing designed experiments on analytic models, we would suggest that the decision to do so or not take into account the current quite powerful state of the art in computing and the demonstrable benefit some companies have obtained by developing these models. Respondent #5 points to a breakdown in the supplier relationship. To the extent that her specifier had a quality engineering framework in design, it clearly did not extend into her enterprise. While Respondent #5’s company had adopted quality engineering for its manufacturing process, its capability appears to lag in design. As they do not seem to be much engaged in the business of new product development, this may be the optimal place to be. However, her experience does suggest that tiered or relative reward structures be worked out to encourage vendors to overdeliver. Lastly, Respondent #8’s experience is somewhat worrisome. While some quality engineering recommendations are starting to be deployed in military – and notably United States Air Force[27] – specifications, the notion that defense contractors are complacently meeting minimum standards and eschewing opportunities for continuous quality improvements points to avoidable waste for which the tax payer is held to account. We would suggest that the other forces follow the Air Force’s lead and begin to incorporate quality engineering requirements in their contracts. 44 Chapter 5 Comparator The main output of this work is a software-based comparison framework for design of experiments strategies. The software assigns any codeable1 design of experiments strategy a particular value for a particular well-defined response surface. 5.1 Concept Before prototype testing can begin, the product designer must decide on his experimental strategy. Some analyses have been done to attempt to rationalize the decision making process by creating computer representations of experimental strategies and running them against different response surfaces[13]. We believed it would be useful to extend the concept to create a generic framework with pluggable interfaces for any arbitrary strategy run against a defined response surface, with variable parameters for repetitions, payoff structures, including targets, payoffs, balance and any creative response one wishes to test. As the enactment of a product design strategy will be quite variable in the payoff that it obtains, the model employs a monte carlo simulation and allows statistical moments of large ensembles of agents performing the coded strategy against the same data set. 1 as described in section 5.2.2 under ’Strategy’ 45 Figure 5-1: Class Structure for Comparator 5.2 Architecture The Comparator consists of an object oriented computer application, with a Player object that provides an interface for a DoE Strategy object. The central AirplaneWorld object also provides an interface for a Game object, which in turn would provide an interface to a Goal object. Its basic structure of the Comparator is shown in figure 5-1. 46 5.2.1 Platform This model runs on Windows 2000. It was developed in Java using Borland JBuilder 8 personal edition, 9 personal edition and X foundation edition. Data was reported by Microsoft Excel from the data store in Microsoft Access. The Comparator reached the data store using JDBC; the reporting tool used ActiveX Data Objects. 5.2.2 Modules There are five types of modules snappable into the framework: goals, games, data stores, parameter spaces and strategies. Each of these has a corresponding Java interface, and are implemented by the Goal, Game, DataStore, ParameterSpace and Strategy interfaces respectively. Goal This allows different sorts of goals to be set for the players. Whereas one game might have a parabolic quality loss function around a particular point, another might give full credit for any response below a particular level. The Goal interface lets the designer distinguish between less-is-better, more-is-better, and optimal-is-best environments, and to specify a shape for the payoff function. The data discussed in this chapter uses a ParabolicDistanceGoal, which delivers maximum value at a particular point then parabolically decaying value further from that point. Game The game provides the interface to the response surface. Any configuration passed to a Game.experiment() will deliver a result, which will be specified by data or a Game’s number generator. The game also serves to connect players to the goal object. The data discussed in this chapter uses a NormalDataGame, described below in section 5.3.2. DataStore This interface allows the development of data access modules for other types of data stores. The data store interface a player uses to interact with the Microsoft Access database is a ConcreteDataStore. 47 ParameterSpace This interface was intended to allow flexibility in the specification of parameters and their level settings. However, some of the software in the framework can only handle RectangularParameterSpaces, which have the same number of levels for each parameter, and the framework has not been used with anything other than a 4-parameter, 3-level parameter space. Strategy This is the main interface of interest, and defines what is mean by a codeable design of experiments strategy. It allows objects to report to the player which configuration to try next, and how many times to try it. It also provides a reporting mechanism for the player to record what response a particular configuration got. Finally, it takes this information and instructs the player which configuration to use to try for the final payoff. 5.2.3 Tables The Microsoft Access database stores game, flight, reward and player data in a number of tables. The Game table records the game identifier, the type, target and tolerance for the goal, the number of reward attempts (here always three), the signal to noise ratio and the interaction strength. The Player table lists the players, how many repetitions each does on a configuration, its strategy and the wealth the player has earned. The Flight table links to the game and player tables, and lists when a plane was thrown, the settings of parameters A, B, C, and D, and the distance the plane flew, as well as when the flight was recorded. The Reward table also lists the parameter settings and distances, as well as linking to the player and game tables. Instead of the time, however, it records how much payoff its player got for that reward experiment. The database also has a store for the Grewen data. To support the genetic algorithms, there is an alleles table, which lists the allele indices, their value, and to which player it belongs. A table named ZeroToFifteen links each allele index with an integer equivalent. 48 5.3 The Data Data can be selected from a predefined set of responses from any configuration, a pseudorandom generator that uses a configuration as a parameter, or some other number generator that satisfies the Comparator’s Game interface. For the results given in section 5.8, we used a parameterized NormalDataGame object. 5.3.1 Paper Airplane Game Eppinger[28] developed a paper airplane experiment to illustrate methods in prototype testing and robust engineering. The experiment uses a sheet of paper with a number of lines drawn on it. The lines correspond to design choices one can make by folding over a particular line or another according to whether the flaps should be up, down, or flat, how long the wings should be, how tightly they should angle, and where an ancillary paper clip should be placed for weighting. The simple format defines four design variables or control factors and 3 levels of each of these variables. Jennifer Grewen employed this airplane template in her undergraduate thesis[29]. She threw 10 airplanes in each of the 81 possible configurations as a full factorial design with 10 repetitions. This data set is hereafter referred to as the Grewen data. The Grewen data is reproduced in Appendix A. This set is the test bed for the model and all strategies that are encoded in the model. The original flight distance ranged from 5 to 27 feet; configuration means varied from 8.15 to 21.8 feet. In order to make the data easier to interpret, it was scaled such that the smallest mean (for configuration 32132 ) was zero and the largest mean (for configuration 2111) was 100. That scaling for the unadulterated data was x0 = 7.326(x − 8.15) (5.1) where x is the measured flight distance in feet, and x0 the same measure in scaled distance. 2 configurations here are referred to by their A-setting, their B-setting, their C-setting and their D-setting, each of which can be 1, 2 or 3 49 A1 C1 Parameter A: Weight Position : ter C h e m t Para e Leng s o N C2 A2 C3 A3 Expt. # 1 2 3 4 5 6 7 8 9 Weight A A1 A1 A1 A2 A2 A2 A3 A3 A3 D3 D1 D2 Stabiliz. B B1 B2 B3 B1 B2 B3 B1 B2 B3 MIT Design of Experiments Exercise v2.0 D3 D1 D2 D3 D1 D2 Wing Ang ___ _ _ _ t #_ _ _ _ _ _ n e _ erim _ _ _ _ _____ p x E ance ___ _ t s _ i D e __ m a Parameter D: N le Nose C C1 C2 C3 C2 C3 C1 C3 C1 C2 Wing D D1 D2 D3 D3 D1 D2 D2 D3 D1 B1 (up) B2 (flat) Parameter B: Stabilizer Flaps B3 (down) Table 5.1: Variance Components in the Grewen Data Input variables Constant term C D A*B*D C*D B*D A*C*D B A*B*C A*B A*D A A*C B*C*D B*C A*B*C*D Coefficient Std. Error 14.7127161 0.10078103 -1.01203704 0.12343105 -0.82537037 0.12343105 0.78458333 0.18514659 0.5138889 0.15117155 0.48277777 0.15117155 -0.45416668 0.18514659 -0.29759258 0.12343105 0.43958333 0.18514659 -0.35888889 0.15117155 0.28277779 0.15117155 0.10611111 0.12343105 0.11111111 0.15117155 -0.09166667 0.18514659 -0.05 0.15117155 -0.0625 0.22675732 p-value 0 0 0 0.00002694 0.00072983 0.00149454 0.01450977 0.01627129 0.01796453 0.01797388 0.0619926 0.39038122 0.46268675 0.62074649 0.74097294 0.78295088 SS 175335.8594 553.0782471 367.8675842 147.7370453 95.06944275 83.90677643 49.50416565 47.82313156 46.37604141 46.36844635 28.7867775 6.08016634 4.44444466 2.01666665 0.89999998 0.625 The individual data points thus vary from -23.1 to 138.1 in the scaled form; in the unmodified Grewen Data, the unscaled global mean of 14.7 feet corresponds to a scaled length of 48.1. As the reader will note in section 5.3.3, the data is modified to accommodate investigations of the effects of nonlinearity, and the scaling at that time is shifted to keep the means between 0 and 100. An analysis of the variance in the data set shows that two of the third order interactions are more significant than two of the main effects, as shown in table 5.3.13 . This table is sorted by p-value, the probability that a component’s effect differs from zero. The bottom three components, B*C*D, B*C, and A*B*C*D, have a better-than-even chance of not existing. This sorting roughly corresponds to magnitude, but differs because some components, like the main effect of B, have lower variability than their larger neighbors, and so a higher p-value. Sorting by size, B, the third largest main effect, would be the tenth-largest of the sixteen terms, not the seventh. This ordering completely 3 This table was generated by Mohan Hsieh using XLMiner from Cytel Statistical Software. 50 undercuts the hierarchical ordering principle introduced in section 4.2.7, one of the assumptions underlying the orthogonal array. As explained in section A.1, the tokens represent the control factor interactions; ”A*B*D” can be thought of as the effect that A has on how the level of B changes the contribution of the level of C to the response (the functions are commutative, so this could just as well read ”B*D*A”.) These sixteen components – the mean, the four main effects, the six cross terms, the four three-way terms and the singe four-way term – were selected for simplicity and clarity. There are a total of 81 terms, enough to explain all of the differences between the means of the configuration. The missing terms contain the effect of a parameter on itself. The first column of table A.2 is labled ’Dif,’ and shows the difference between the flight distance we would have expected based on our model, and the actual configuration mean. The Dif column is small, but significant. It is possible that interactions with low p-values (that are definitely different from zero) are among these higherorder terms and are dropped. These missing 55 terms, which is to say, the ’Dif’ column, together crest five at one point, but are more typically less than one; there is no reason to think the unexpressed terms will be larger than A ∗ C ∗ D. In any case, the point that the Grewen Data is violates the hierarchical ordering principle and will not easily yield to an orthogonal array would still hold true if it were. In addition, this data set is very noisy. The final column in the table A.1 is the signal-to-noise ratio for each configuration, which is calculated as ST Nt = abs( ȳt − ȳ ) σt (5.2) and gets as high as 21.88 for configuration 1233, but is for 51 of the 81 configurations less than 1, and for 26 less than 1 . 2 As we shall see in section 5.3.4, an average datum’s difference from the global mean is only about twice the noise. The outer array – effect external to the design decisions that affect the objective function – are quite important in the throwing of paper airplanes. While the thrower’s style 51 and height were constant in this case, we know nothing about the constancy of the attitude or thrust of the throws, nor of the folding of the airplane. Even the most disciplined paper airplane maker and thrower will have to deal with a large outer array – air is a turbulent medium and subject to all sorts of unrepeated drafts. The paper from which the airplanes are constructed will equilibrate with the atmospheric humidity over time, but only at its edges or when it is at the top of the stack. Add to these sorts of uncontrollable effects the poor discrimination in mensuration, and repeatable data becomes an unlikely affair. 5.3.2 Normal Data Games We created a class of number generators called DataGame objects, data pickers which selected data from predefined sets as discussed below in section 5.3.1. However, this presented two problems. 1. The EMV players described in section 5.5 presumed a normal data set, and could not adapt to the discrete distribution the actual 10 data given by the Grewen data for each configuration. 2. the Grewen data has only 0.5 foot (scaled to 3.664 ) precision, and results generally in the 12’ - 17’ range. This results in much repetition in the reported data, making statistical moments difficult to calculate. The agent can not in these cases get a sense of the variability associated with a configuration. We therefore created a NormalDataGame object, which calculated the means and populations standard deviations for the response associated with each configuration. It used these parameters to generate a gaussian pseudorandom response. The difficulty in striking a target can be varied in this comparator by moving the target. In the work thus far, two settings have been used, 60 and 80. 60 was very simple to hit, as the scaling tended to place the global mean very near there and thus many configurations had significant probabilities of going this distance. The 80 point, 4 One foot equals 7.326 scaled units 52 on the other hand, was very difficult to hit as only few configurations had significant probabilities with that distance. This is somewhat artificial, and it is unlikely a product developer would be required to find such an unlikely goal. This may unfairly impede AdaptiveOneAtATime’s performance, as OrthogonalArray will be better at precise results. The parabolic players were rewarded some amount every time they threw for an award. If the game had a payoff P = AR, where A = 3 is the number of attempts the players get at a payoff, then each throw could have a maximum payoff R. The quality loss function was constructed as a symmetric parabola so that its value was R at the target D = 80, and 0 at the minimum 05 , so that the player would only very rarely get a negative payoff. This differs somewhat from Frey, et al,frey:role) who preferred to consider larger-is-better problems. Once it had calculated a mean µ and population standard deviation σ from its experiments on a configuration t, a player would then expect a payoff from any throw of < P >t = R Z inf − inf (D − x)2 exp µt − x 2 dx σt2 (5.3) and the quality Q of the configuration t is (by Wu & Hamada [9, equation 3.2] Q = P [(D − µt )2 − σ 2 ] (5.4) This is the quantity the player uses to select the configuration to throw. While OrthogonalArrayParabolic players attempt to predict this for every configuration and take the best one, FullFactorialParabolic and AdaptiveOneAtATimeParabolic players just use the best one that they have experienced. Guess players, of course, just generate a random configuration, and provide a baseline for comparison. 5 the value went to 0 at the minimum as it was further from the target than the maximum 53 5.3.3 Manipulating the nonlinearity The nonlinearity of a system was defined as by Frey[12], as the sum of the square of the means for each configuration less the contributions due to global average and main effects, which are the linear components, over the total variability. The sum square of the factor effects SSF E is calculated for m configurations t each tested N times for a mean ȳt by SSF E = m NX (y¯t − ȳ)2 m t=1 (5.5) where ȳ is the global mean of the experimental strategy. Each of the configurations has an associated setting for each parameter, and we assume that the setting of every parameter can have some effect on the outcome of the response. We extract these effects with a main effects model. For a rectangular parameter space with s level settings in each of q parameters, the number of configurations t with a particular level set for parameter p will be m/s. The effect due to that level setting is M Ep(t) X s m/s (y¯t − ȳ) = m t=1 (5.6) We can then compose an expectation value for a configuration by summing the global mean yt and the main effects M Ep(t) calculated for each parameter at that configuration t’s level setting s. Having accomplished this, we can define the sum square variations due to main effects as SSM E = m NX (yM E (t) − ȳ)2 m t=1 where yM E (t) = ȳ + q X M Ep(t) (5.7) (5.8) p=1 which allows us to calculate the sum squared variation from interactions SSIN T = SS F E − SSM E 54 (5.9) and thence the Interaction Strength I I= SSIN T SSF E (5.10) For the Grewen data, this interaction strength was quite large IGrewen = 0.527 (5.11) meaning that more than half the variation was not explained by main effects. To examine the relative performances of strategies at different levels on nonlinearity, we took the Grewen data and made its nonlinearity tunable, by defining a scaling of the difference between the value in the main effects model and the statistical mean. The main effects of each parameter setting were stored in a database table, and a new mean generated for a desired Interaction Strength Id for each configuration by a simple scaling ȳt0 = yM E (t) + Id (ȳt − yM E (t)) I (5.12) This allows us to compare the same strategies at different levels of nonlinearity. 5.3.4 Manipulating the noise The ’signal’ in the Grewen data is less than twice the noise. It is calculated in the following way, again for m configurations t each tested N times for a mean ȳt and global mean ȳ, as in section 5.3.3. √ Pm (ȳt −ȳ)2 qP t=1 N n=1 ST N = (ytn −ȳt )2 m ST NGrewen = 1.93 (5.13) (5.14) As does the nonlinearity, this extreme value would press any strategy hard for success. While this hard problem may be of interest at times, to get variable solvability prob55 Figure 5-2: Comparator Functions lems for the agents to work on, we scale the noise in a similar way to the nonlinearity. Instead of the sigma calculated from the data, we generate normally distributed pseudorandom numbers using σt0 = ST N σt ST NGrewen (5.15) where STN is defined by the model. Now that we can manipulate the interactions and noise of a particular data set, we are ready to see how changes in those values affect the performance of the strategies. 5.4 The Central Command The operation of the framework is diagrammed in figure 5-2. The basic activity of enacting the DoE strategy and performing the experimental design is handled in the Player.run() method, reproduced below. The player asks the strategy for a configuration and number of repetitions, performs the repetitions, reports the results, and asks if it should continue. As this is the central piece of code in the framework, it may be useful to walk through it bit by bit. The first thing the agent does is 56 update the number of players currently in the body of the run method. This is to aid higher-level thread management. Listing 5.1: Player.run() method public void run ( ) { p l a y e r s r u n n i n g o n t h i s v m ++; The player starts performing its experiments by retrieving its strategy and acknowledging its experimental cost. The experimental cost in these experiments was always set to one, but flexibility was built into the model. The strategy here provides a list of configurations to the player, and the player begins to work through them. The hasNext() method returns true if there are more configurations the strategy wishes the player to try. S t r a t e g y how = t h i s . g e t S t r a t e g y ( ) ; int c o s t = t h i s . getGame ( ) . g e t E x p e r i m e n t a l C o s t ( ) ; try { // g i v e up i f t h i s t h r o w s an ArrayIndexOutOfBoundsException I t e r a t o r e x p e r i r a t o r = how . i t e r a t o r ( ) ; i f ( null != e x p e r i r a t o r ) { Game game = t h i s . getGame ( ) ; while ( e x p e r i r a t o r . hasNext ( ) ) { Request the configuration and number of repetitions for the experimental design. // g e t t h e n e x t bunch o f parameter s e t t i n g s HashMap s e t t i n g b y p a r a m = ( HashMap ) e x p e r i r a t o r . next ( ) ; int t r i a l c o u n t = how . g e t T r i a l C o u n t ( ) ; double [ ] r e s p o n s e s = new double [ t r i a l c o u n t ] ; // r e p e a t f o r a c e r t a i n number o f t r i a l s f or ( int t r i a l = 0 ; t r i a l < t r i a l c o u n t ; t r i a l ++) { We are now inside the repetitions loop. The configuration and number of repetitions have been decided. All that remains is to perform the experiment, store the results and account the cost. 57 // throw t h e a i r p l a n e double r e s p o n s e = game . e x p e r i m e n t ( t h i s . hashCode ( ) , settingbyparam ) ; if ( storeflights ) { // u p d a t e t h e d a t a s t o r e try { d a t a s t o r e . s t o r e d a t a ( t h i s . gameid , setti ngbyparam , response ) ;} catch ( OutOfMemoryError o ) { System . e r r . p r i n t l n ( "Player , line 106 , OutOfMemoryError " ) ; } } 10 responses [ t r i a l ] = response ; t h i s . w e a l t h −= c o s t ; } Update the strategy by passing back the configuration and all of the results it engendered. The strategy object will decide how to handle it – whether to get statistical moments of the responses, take an extremum, ignore them altogether, or some other approach. // l e t t h e s t r a t e g y know how i t d i d i f (0 < t r i a l c o u n t ) { how . update ( settin gbyparam , r e s p o n s e s ) ; } } } After all of the experiments in the experimental strategy have been performed with all of their repetitions, the strategy object has enough data to decide which configuration it is going to use to get the payoff. At this point, it starts performing the experiments, getting rewarded, and adding those rewards to its wealth. // Throw t h e p l a n e s Goal d e s i r e = t h i s . getGame ( ) . g e t G o al ( ) ; Map b e s t s e t t i n g s = how . g e t B e s t S e t t i n g s ( t h i s . g e t D a t a S t o r e ( ) ) ; 58 for ( int attempt = 0 ; attempt < d e s i r e . getAttemptCount ( ) ; attempt++) { double d i s t a n c e = game . e x p e r i m e n t ( t h i s . hashCode ( ) , bestsettings ) ; double p a y o f f = d e s i r e . g e t P a y o f f ( d i s t a n c e ) ; t h i s . w e a l t h += p a y o f f ; datastore . recordTrial ( bestsettings , distance , payoff ) ;} this . getDataStore ( ) . s t o r e w e a l t h ( this . wealth ) ; } 10 catch ( IndexOutOfBoundsException e ) { e . printStackTrace () ; System . e r r . p r i n t l n ( "run failed for player " + S t r i n g . v a l u e O f ( t h i s . hashCode ( ) ) ) ; } Finally, the player clears up some resources and announces its exit. this . getDataStore ( ) . clearCache ( ) ; S t r i n g n o t i c e = " finished Player .run for " + S t r i n g . v a l u e O f ( t h i s . hashCode ( ) ) + " at " + j a v a . t e x t . DateFormat . g e t T i m e I n s t a n c e ( j a v a . t e x t . DateFormat .SHORT) . format ( j a v a . u t i l . Calendar . g e t I n s t a n c e ( ) . getTime ( ) ) ; p l a y e r s r u n n i n g o n t h i s v m −−; // decrement t h e semaphor } 5.5 The Strategies Each of the strategies is an implementation of a generic Strategy interface reproduced here. Listing 5.2: The Strategy Interface /∗ ∗ Strategy . java ∗/ package edu . mit . c i p d . a i r p l a n e g a m e ; 59 import j a v a . u t i l . ∗ ; /∗ ∗ ∗/ public i n t e r f a c e S t r a t e g y { /∗ ∗ l e t t h e c l i e n t i t e r a t e t h r o u g h t h e d e s i g n s p a c e ∗ r e l e a s e a new parameter s e t t i n g e v e r y time 10 ∗/ public I t e r a t o r i t e r a t o r ( ) ; /∗ ∗ i n c r e a s e t h e a c c u r a c y o f t h e s t r a t e g y by u p d a t i n g i t −− t h i s is ∗ required for adaptive s t ra te gi e s ∗ @arg s e t t i n g s b y p a r a m which s e t t i n g s were used ∗ @arg d i s t a n c e t h e r e s u l t o f t h o s e s e t t i n g s ∗/ public void update ( HashMap s e t t i n g s b y p a r a m , double [ ] d i s t a n c e s ) ; /∗ ∗ use t h e d a t a s t o r e t o e v a l u a t e t h e e f f e c t s o f t h e parameters , and 20 ∗ e s t i m a t e t h e b e s t s e t o f v a l u e s f o r t h e d e s i r e d outcome ∗ @arg d a t a s t o r e t h e a c c e s s o b j e c t f o r t h e s t o r e d d a t a −− a d a p t i v e s t r a t e g i e s may not need i t ∗ @return t h e s e t t i n g s t o g i v e t h e b e s t r e s u l t hashed by t h e parameters ∗/ public Map g e t B e s t S e t t i n g s ( DataStore d a t a s t o r e ) ; /∗ ∗ A d a p t i v e s t r a t e g i e s need t o know t h e i r g o a l ∗ @arg g o a l t h e t a r g e t t o s h o o t f o r ∗/ public void s e t G o a l ( Goal g o a l ) ; /∗ ∗ 30 ∗ A l l o w s t h e c a l l e r t o s u g g e s t how many t i m e s an e x p e r i m e n t s h o u l d be r e p e a t e d 60 ∗ @param T r i a l C o u n t −− how many t i m e s t o r e p e a t an e x p e r i m e n t ∗/ public void s e t P r e f e r r e d T r i a l C o u n t ( int T r i a l C ou n t ) ; /∗ ∗ ∗ Allow t h e s t r a t e g y t o d i c t a t e how many t i m e s t h i s p a r t i c u l a r experiment i s repeated ∗ @return t h e number o f t i m e s t o r e p e a t t h e n e x t e x p e r i m e n t ∗/ public int g e t T r i a l C o u n t ( ) ; /∗ ∗ 40 ∗ r e t u r n t h e parameter s p a c e t h e s t r a t e g y i s w o r k i n g with , which s h o u l d have ∗ been s e t i n t h e c o n s t r u c t o r ∗/ public ParameterSpace g e t S p a c e ( ) ; } Each of the examples below is from the ’parabolic’ payoff structures. Strategies were each implemented three times. In the first version, they were merely meanseeking. In the second version, they used the error function to create an expectation value of their reward. In these first two versions, the players got full reward for landing in between two values, or goalposts. In the third version, the payoff structure was changed to be parabolic. There was one optimum point, and the payoff decayed parabolically from it. We will return to a goalposts structure in Chapter 6.5. For the remainder of this chapter, we will only consider the parabolic payoff agents. We distinguish four types of DoE strategies: 1. impulsive strategies, where parameter design is abandoned and a random configuration tried, 2. complete strategies (the full factorial strategy), where every possible combination of parameter levels are tested, 61 3. statistical strategies, which gather information to satisfy a statistical model and thereby predict a best setting, and 4. adaptive strategies or search routines, which move through a parameter space seeking an optimal result. For each of these types of strategies, we built a canonical example, respectively Guess, FullFactorial, OrthogonalArray and AdaptiveOneAtATime, described below. 5.5.1 Validating the experimental designs The function of the Strategy objects is to determine the configurations explored by the Player objects. So, the first step in validating the model is to ensure that the strategies are providing the correct configuration lists to the Players. Each time a player performs an experiment, it records the configuration and result in a database, as shown above in line 7 of listing 5.4. This data helps the strategy make a decision about which configuration the player should use in the payoff round. So, it is important that the strategy deliver exactly the configurations it purports to. Appendix C shows the results of the following query with @Strat set to the name of each strategy. Listing 5.3: Listing the Configurations Select Sample . PlayerID , A & B & C & D AS C o n f i g u r a t i o n , avg ( D i s t a n c e ) As Mu, count ( d i s t a n c e ) As Frequency , s t d e v p ( d i s t a n c e ) As Sigma FROM (FLIGHT INNER JOIN (SELECT TOP 10 mid ( P l a y e r . S t r a t e g y , l e n ( ’edu.mit.cipd. airplanegame .’ ) + 1 ) AS S t r a t e g y , PlayerID , T ri a l Co u n t FROM PLAYER Where P l a y e r . S t r a t e g y = ’edu .mit.cipd. airplanegame .’ + @Strat AND T ri a l C o u n t =7) as Sample ON FLIGHT . PlayerID = Sample . PlayerID ) GROUP BY Sample . PlayerID , A, B, C, D, Sample . S t r a t e g y , Sample . T r i al C o u n t ORDER BY Sample . PlayerID , MIN( F l i g h t .WHen) Each player is listed in the Player table with its playerid and strategy, which is prepended by the JavaTM package qualifier ’edu.mit.cipd.airplanes.’. 62 5.5.2 Guess The simplest strategy is Guess. Its hasNext() method6 always returns false, its getBestConfiguration() method returns a randomly selected configuration, and its update() method is not implemented. It is this last point that explains why there is no GuessParabolic class. The parabolic strategies are coded to employ their strategies against a parabolic reward. As guess is not strategic, it needed no extension. The flight data query for Guess comes up empty, as it should. Guess does no flights before entering the reward phase. We expect that guess will do equally well as far as payoff in all sorts of situations. Guess provides an implicit null hypothesis for considering the effectiveness of the other strategies. 5.5.3 Full Factorial Next clearest in the set of canonical strategies is full factorial. FullFactorial- Parabolic visits every possible configuration. It’s iterator’s next() function retrieves the next array of settings created by a function getAllSettingsCombinations(). public s t a t i c I n t e g e r [ ] [ ] g e t A l l S e t t i n g C o m b i n a t i o n s ( Parameter [ ] parameters , Map s e t t i n g s b y p a r a m ) { Iterator seterator = settingsbyparam . values () . i t e r a t o r () ; int s e t t i n g c o u n t = 1 ; while ( s e t e r a t o r . hasNext ( ) ) { s e t t i n g c o u n t ∗= ( ( C o l l e c t i o n ) s e t e r a t o r . next ( ) ) . s i z e ( ) ; } // a l l o c a t e an a r r a y t o h o l d t h e s e t t i n g s Integer [ ] [ ] r e t v a l = new I n t e g e r [ s e t t i n g c o u n t ] [ p a r a m e t e r s . length ] ; // i t e r a t e a c r o s s t h e p a r a m e t e r s 10 int p r e v c o u n t = 1 ; fo r ( int i n d e x =0; i n d e x < p a r a m e t e r s . l e n g t h ; i n d e x++){ 6 strictly, the hasNext() method of its associated iterator object 63 L i s t s e t t i n g s = ( L i s t ) settingsbyparam . get ( parameters [ index ] ) ; int mycount = s e t t i n g s . s i z e ( ) ; for ( int prev = 0 ; prev < p r e v c o u n t ; prev++){ fo r ( int s e t t i n g = 0 ; s e t t i n g < mycount ; s e t t i n g ++){ fo r ( int r e p e a t = 0 ; r e p e a t < s e t t i n g c o u n t / p r e v c o u n t / mycount ; r e p e a t++){ int pos = r e p e a t + prev ∗ s e t t i n g c o u n t / p r e v c o u n t + s e t t i n g ∗ s e t t i n g c o u n t / p r e v c o u n t /mycount ; r e t v a l [ pos ] [ i n d e x ] = ( I n t e g e r ) s e t t i n g s . g e t ( s e t t i n g ) ;}}} p r e v c o u n t ∗= mycount ; } 20 return r e t v a l ; } The update() function allows the strategy to calculate the Quality of each configuration, and getBestConfiguration() simply returns the highest-valued configuration. We expect full factorial to always find a reasonably good answer, assuming one exists, and to be as invariant as guess to noise and nonlinearity. 5.5.4 Adaptive One At A Time The AdaptiveOneAtATimeParabolic strategy goes through the parameters and tries each of the settings in turn, leaving the visited parameters at their best settings. The hasNext() method returns true until each parameter has had each setting visited. The update method calculates the quality loss function of the configuration given the test results. If the quality of the current configuration is highest, it preserves it and discards the previous best. Listing 5.4: AOAT Parabolic update() // use t h e q u a d r a t i c l o s s f u n c t i o n and t h e s t a t i s t i c a l moments to get the best expected 64 public void update ( HashMap s e t t i n g s b y p a r a m , double [ ] d i s t a n c e s ) { float l o s s = 0; double [ ] moments = A n a l y s i s . s t a t i s t i c a l m o m e n t s ( d i s t a n c e s ) ; //The f i r s t moment i s t h e mean . The second i s t h e s t a n d a r d deviation double mu = moments [ 0 ] ; double sigma = moments [ 1 ] ; /∗ c a l c u l a t e t h e q u a l i t y l o s s f u n c t i o n −− t h e c o e f f i c i e n t , t h e b r e a d t h o f t h e p a y o f f curve , doesn ’ t m a t t e r . we j u s t need t h e sigma−s q u a r e d + (mu − t a r g e t )−s q u a r e d ∗/ 10 l o s s = ( f l o a t ) ( Math . pow ( sigma , 2 ) + Math . pow ( (mu − t h i s . od ) , 2) ) ; // r e p l a c e t h e b e s t answer // n o t e t h a t b e s t answer h e r e i s t h e s m a l l e s t l o s s f u n c t i o n , not t h e b e s t d i s t a n c e i f ( l o s s < this . bestanswer ) { Parameter param = orderedparams [ l a s t p a r a m ] ; // g e t t h e c u r r e n t l y v a r y i n g parameter b e s t s e t t i n g s . put ( param , s e t t i n g s b y p a r a m . g e t ( param ) ) ; // s a v e its setting this . bestanswer = l o s s ; Subsequent calls to next() will use this new best setting to set the parameters that is not currently being varied. The call to getBestConfiguration() will simply return the bestsettings variable value. It is hard to develop an intuition about what to expect from adaptive one at a time, but we expect it to behave like a continuously improving search algorithm. It should be reasonably good in a situation we know nothing about, but somewhat subject to being trapped. 65 5.5.5 Orthogonal Array This is the statistical strategy, and as such is the least obvious in its operation. It is tied to a particular statistical model. As it supplies a symmetric second order orthogonal array as an experimental design, this strategy demands a main effects model. The OrthogonalArrayParabolic object does not calculate its own Orthogonal Array. It uses a particular orthogonal array taken from a lecture by Dan Frey, generalizing it by swapping the meanings of the parameters and levels in a randomized fashion. /∗ 10 A1 B1 C1 D1 A1 B2 C2 D2 A1 B3 C3 D3 A2 B1 C2 D3 A2 B2 C3 D1 A2 B3 C1 D2 A3 B1 C3 D2 A3 B2 C1 D3 A3 B3 C2 D1 ∗/ This strategy does some number of experiments at each visited configuration, then uses that information to estimate the main effects of parameter levels on flight distance and standard deviation. With that information, it then constructs a predicted quality as per equation 5.4 for all 81 configurations, then submits its best choice to the Game for the reward. We expect orthogonal array to not do terribly well against the Grewen data, as it depends on the hierarchical ordering assumption. Orthogonal array is fairly insensitive to noise, which is important here. Main effects are still fairly important, so we expect it to beat adaptive one at a time, but not to approach full factorial in 66 payoff. 5.6 Payoff After the agents had precessed through their experimental designs, they entered performance trials, were they would receive some payoff if they landed between two values (in the goalposts mode) or according to their distance from an optimum (in the parabolic mode.) In all cases, the agents had three tries for a payoff. Only the parabolic agents are discussed here. 5.7 Evaluating the experimental designs So, now the strategies can be compared with their success levels at different points, and for different noise and nonlinearity levels. The SQL Query Listing 5.5: SQL Query to retrieve comparative value SELECT Game . Payoff , Count ( ∗ ) AS Frequency , TrialCount , Noise , N o n l i n e a r i t y , mid ( P l a y e r . S t r a t e g y , l e n ( ’edu.mit.cipd. airplanegame .’ ) + 1 ) AS S t r a t e g y , AVG( Wealth ) AS [ $ \mu { Wealth }$ ] , STDEVP( Wealth ) AS [ $ \ s i g m a { Wealth } $ ] , [ $\mu { Wealth } $ ] / [ $\ s i g m a { Wealth }$ ] AS [ $STN { Wealth }$ ] , Game . AttemptCount ∗AVG( P a y o f f s . V i r t u e ) /Game . P a y o f f AS [ $ \mu { P a y o f f } $ ] , Game . AttemptCount ∗STDEVP( P a y o f f s . V i r t u e ) /Game . P a y o f f AS [ $\ sigma { P a y o f f } $ ] , [ $\mu { P a y o f f }$ ] / [ $\ s i g m a { P a y o f f } $ ] AS [ $STN { P a y o f f } $ ] FROM (GAME INNER JOIN [SELECT GameID , PlayerID , SUM( P a y o f f ) AS V i r t u e FROM Reward GROUP BY GameID , PLAYERID ] . AS P a y o f f s ON GAME.GAMEID = P a y o f f s .GAMEID) INNER JOIN PLAYER ON PLAYER.PLAYERID =P a y o f f s . PLAYERID WHERE GOAL=’edu.mit.cipd. airplanegame . PARABOLICDISTANCEGOAL ’ AND AttemptCOunt=3 AND TARGET=80 GROUP BY Game . AttemptCount , Game . Goal , Game . Target , 67 Game . T o l e r a n c e , S t r a t e g y , Game . Payoff , TrialCount , Noise , Nonlinearity ; gives us a table of results that allow us to compare the value of the DoE strategies at different settings, as shown in table C.1. The standard deviation and mean for the payoff is pulled from the preceding table to be shown in the following graph. 5.8 Results The query 5.5 creates the OLAP table C.1. This data forms the data source for both Figure 5-3, a stacked histogram for the mean value of payoff – the number of hits out of three the agent got – and Figure 5-4 for the mean value of wealth, which is the number of payoffs times the payoff amount (which is constant for every player in a particular game) less the number of experiments (which is the same for every player using a particular strategy and trial count). The charts therefore can be generated one from the other according to the relationship7 W = A RX p(a) − Cs A a=1 (5.16) Where W is the wealth, R is the reward for the goal, A is the number of attempts a player is given, a is the attempt index, p(a) is the payoff portion (up to 1) for a particular attempt and Cs is the experimental cost for a strategy s. On these following plots ’noise’ is the signal to noise ratio, and nonlinearity is the data scaled as explained in section 5.3.3. Figures 5-4 and 5-3 show success data for agents with each of the four canonical strategies and 2, 4 or 7 repeats on each experiment, operating in games with payoffs of 63, 567 or 3000, signal-to-noise ratios of 100, 10 or 1.94 (the level of the Grewen data,) and interactions of 0, 0.1, 0.527 (the level of the Grewen data) and 0.8. The model run to generate the high interaction data was incomplete, creating more unevenness in the height of the stacked columns than exists in the data. 7 see also equation 5-3. 68 45 Sum of $\mu_{Payoff}$ 40 Nonlinearity Noise 1.79769313486232E+308 - 0 0.8 - 100 0.8 - 10 0.8 - 1.93717750485 0.52705102742087 - 100 0.52705102742087 - 10 0.52705102742087 - 1.93717750485 0.1 - 100 0.1 - 10 0.1 - 1.93717750485 0 - 100 0 - 10 0 - 1.93717750485 35 30 25 20 15 10 5 0 2 4 7 2 4 7 4 2 4 7 2 4 7 2 4 7 2 4 7 2 4 7 2 4 7 4 2 4 7 2 4 7 2 4 7 4 2 4 7 63 567 972 3000 Guess 63 567 3000 63 AdaptiveOneAtATime 567 972 3000 63 FullFactorial 567 972 3000 OrthogonalArray Strategy Payoff TrialCount Figure 5-3: Stacked Histogram of Mean Payoff 35000 Sum of $\mu_{Wealth}$ 30000 Nonlinearity Noise 1.79769313486232E+308 - 0 0.8 - 100 0.8 - 10 0.8 - 1.93717750485 0.52705102742087 - 100 0.52705102742087 - 10 0.52705102742087 - 1.93717750485 0.1 - 100 0.1 - 10 0.1 - 1.93717750485 0 - 100 0 - 10 0 - 1.93717750485 25000 20000 15000 10000 5000 0 2 4 7 2 4 7 4 2 4 7 2 4 7 2 4 7 2 4 7 2 4 7 2 4 7 4 2 4 7 2 4 7 2 4 7 4 2 4 7 -5000 63 567 972 3000 Guess 63 567 3000 AdaptiveOneAtATime 63 567 972 3000 63 FullFactorial 567 972 3000 OrthogonalArray -10000 Strategy Payoff TrialCount Figure 5-4: Stacked Histogram of Mean Wealth 69 Noise 1.93717750485 Payoff 567 TrialCount 4 Nonlinearity vs. Strategy Performance 3 Sum of $\mu_{Payoff}$ 2.8 2.6 Payoff Strategy AdaptiveOneAtATimeParabolic FullFactorialParabolic Guess OrthogonalArrayParabolic 2.4 2.2 2 1.8 0 0.1 0.52705102742087 0.8 Interaction Strength Nonlinearity Figure 5-5: Strategy values at different nonlinearity values The stacked histogram is difficult to interpret, in part due to the missing 0.8 nonlinearity data for many of the 3000 payoff runs. We can use this data to look at strategy vs. signal-to-noise ratio or strategy vs. nonlinearity for any payoff value or number of repetitions. Figure 5-5 pulls out a little of this data. We see from this that adaptive one a time does outperform orthogonal array at the enhanced linearity level, 597 payout, four-trail case. The source table C.1 tells us that the standard deviations in payoff for AdaptiveOneAtATime and OrthogonalArray in this regime are 0.400 and 0.431 respectively, whereas the difference is only 0.1. The result does assure us that AdaptiveOneAtATime is not uniformly worse. Resetting the nonlinearity to the level of the Grewen data and varying the signal-to-noise ratio, we see that, as we expect, guess is insensitive to the noise level and full factorial essentially becomes perfect at reasonably low noise. From left to right, the payoff values for adaptive one at a time are (2.512, 2.646, 2.681) and for orthogonal array (2.640, 2.753, 2.771). Orthogonal array is somewhat less senstive to noise, improving its performance by only half as much as adaptive one at a time moving from an STN of 10 to an STN of 100. 70 Nonlinearity 0.52705102742087 Payoff 567 TrialCount 4 Signal To Noise Ratio vs. Strategy Performance 3 Sum of $\mu_{Payoff}$ 2.8 2.6 Payoff Strategy AdaptiveOneAtATimeParabolic FullFactorialParabolic Guess OrthogonalArrayParabolic 2.4 2.2 2 1.8 1.93717750485 10 100 SIgnal To Noise Ratio Noise Figure 5-6: Strategy values at different signal-to-noise ratios We see that, as we expect, the Orthogonal Array does well in noisy, linear environments. Full Factorial wins when the experimental cost becomes small compared with the payoff. Adaptive One At A Time loses to Orthogonal Array until the interactions become very strong. 5.9 Analysis Guess does well on wealth, but is beaten by the Full Factorial in the high payoff regime and by the Orthogonal Array and Adaptive One At A Time strategies in the lower payoff regimes. Full Factorial always wins on payoff, but loses in the areas of lower payoff. Adaptive One At A Time, here, never convincingly beats Orthogonal Array, only tying it in the enhanced-interaction Grewen data with the noise extracted, where we would expect it to do better. We believe this is due to the unrealistic constraint imposed on AOAT by the ’nominal is best’ target. We have shown that strategies are codeable, and perform relative to one another 71 in line with out expectations. We can therefore have some confidence in moving forward and applying this evaluative framework to other tasks. 72 Chapter 6 Abstraction With a framework for modeling an arbitrary design of experiments (DoE) strategy in place, it becomes possible to talk about what makes a DoE strategy. There is an essential value to being able to categorize a strategy along particular and familiar parameters. Formally, by an experimental design we mean a ordered list of configurations to test, and by an experimental strategy we mean an experimental design with repetitions assigned to each configuration. We use ’DoE strategy’ here to mean the means by which the experimental strategy is constructed. 6.1 Motivation In section 6.5, we discuss optimization of a DoE approach. It would be nice to be able to say of a particular response surface that you were using the best DoE. Abstraction is a necessary precursor. Tools like minitab, as discussed in section 4.3 will build the best experimental design for a known statistical model. When a response surface comports well with a particular statistical model, it would be hard to argue that a better experimental design existed. However, the presumption that a statistical model is known and available is an extreme one for DoE in general. It is therefore incumbent upon us to decide how to extract the essential nature 73 from a DoE strategy and parameterize it. It is in this way that we will be able to create novel or particularly appropriate DoE strategies. 6.2 Method As discussed in section 5.5, design of experiments strategies fall in four categories: arbitrary, complete, statistical and adaptive. Part of the challenge of abstracting and generalizing DoE is to create a measuring scheme under which any of the forms could be more or less expressed. Abstraction of DoE strategies is not straightforward, and there are several possible routes one could take. We could model learning or bias, as agents take successful configurations from a previous game or a preprogrammed list of configurations to try. We could grant each agent a statistical model and let those compete. We could allow agents to pick configurations at random for ’experimental designs’ of different lengths and just compare the effectiveness of a particular design length. We decided to establish certain tendencies in the agents that distinguish strategies that we had tested from one another. The four canonical strategies discussed in section 5.5 – Guess, Orthogonal Array, Adaptive One At A Time and Full Factorial – differ in the number of configurations that they test, the way they choose configurations to test and whether or not they’ve decided these questions before they start. We came up with these five tendencies • How far ahead do the players look? It takes a certain amount of discipline to not just assume that the best answer you’ve gotten so far will be quite close to the best answer possible. We call this the planning. • How much do the players want to spend in total? The plans are the atomic unit of experimental design, so the profligacy is the number of plans through which the player goes. • How seriously does this player regard variability in response? How many times 74 will it repeat an experiment? We call this the care. • The order 2 orthogonal array carefully balances the pairs of parameter settings. A1 B2 will show up as often as A2 B1 The concept of fairness captures this. Unlike the other concepts, this one does not vary between two values, but is just a set of decisions as to whether to balance groups of various sizes. A strategy might want to balance singletons, seeing A3 as often as D2 , pairs, triplets or quartets. • Finally, a tendency it made sense to include was that of Diligence. Diligence would be the earnestness with which a player approached its task, and its tendency to quit early when it felt it had a good answer. The investigation of diligence is left to further work, as the results with the first four alleles were not robust enough to include diligence. As diligence was intended to capture the practice of varying the number of trials between configurations, a Bayesian approach to noise inspired by the Bayesian approach to interaction in Chipman, et al.[30], it was deemed of a greater level of subtlety than the other inclinations, and its implementation was postponed until the rest of the model had settled. In short, all configurations tried by an agent were tried an equal number of times. As many of our investigations were done with ’goal posts,’ and not quality loss functions, the lazy players in this paradigm would gain an unfair and unrealistic advantage. 6.3 Available optimization methods What we proposed was a set of decision criteria for creating a DoE strategy. To create an objective function, the DoE strategy itself would have to be created and run against the data and payoff criteria. Therefore, we have a highly variable response, and there is no gradient available in the objective function. The first point argues for some sort of ensemble calculation, and the latter against numerical search or other traditional optimization schemes. Our options for optimization schemes are therefore somewhat limited. Genetic algorithms, however, require only that our parameters be codeable, the payoff function 75 be defined and our problem amenable to large populations of solvers. Our environment is so provided, as described in section 6.4 6.4 Genetic Algorithm Once we had decided to proceed with genetic algorithms, and therefore to code our five tendencies as genes, we had further design decisions to make. On top of the question of how a coding for care of ’1010,’ for example, would be interpreted, there are questions of initial population and number of generations to run, as well as design decisions – selection, crossover, and mutation – that separate genetic algorithm models one from another as described in section 3.3. A generic agent was to create a strategy by 1. Deciding on the length of a plan. A plan is a portion of an experimental design, which is to say, a list of configurations to try, 2. Creating the plan by populating the list with configurations selected according to some scheme, 3. Enacting the plan by visiting each configuration and testing it some determined number of times, 4. Reviewing the information obtained through the previous tests and iterating, creating a new plan, and 5. Continuing to iterate until some number of plans had been completed and merged into the experimental design. In a nod to the confirmation trial required in statistical models, the last configuration was determined by a simple main effects model, as explained in section 2.3. 76 6.4.1 The Codings A four-bit gene, either 0 to 16 or 4 yes/no decisions, was decided to give enough granularity so that effects could be considered, but not inappropriately complex. Each of the tendencies interprets the four-bit gene and lets it effect the phenotype, that is, the DoE strategy, in its own way. profligacy allele determined how many plans the agent goes through. The Adaptive One At A Time strategy goes through 4 plans (with one additional configuration,) and the Orthogonal Array goes through only one. The number of plans to test is allowed to vary from 0 to 15. The Guess strategy would have a profligacy of 0, and a zero profligacy renders the other alleles irrelevant. planning allele decided how many configurations would be in a particular plan. In the four-parameter, three-level parameter space of the paper airplane game, Adaptive One At A Time, for instance, would have two configurations in its plan, as it knows only the next two configurations it will test when it starts an allele. Orthogonal Array would have nine, as it knows which nine configurations it will test when it begins. The expression of the planning allele varied from 1 to 16. care allele told the agent how many times it should experiment with a given configuration. An agent with a care genotype of 0011, for instance, would throw 0 · 23 + 0 · 22 + 1 · 21 + 1 · 20 + 11 = 2 + 1 + 1 = 4 airplanes of each design it chose to test. fairness allele determined how those configurations were chosen. It created most of the complexity of this model. It is the seeking for balance of parameter settings in the experimental design. As a plan is constructed, an agent takes some care to balance the appearance of settings in a particular number of columns according to each yes or no gene. If the first bit was ’flipped,’ that is, true, 1 The choice is to add one to the binary equivalent or to make 0000 equal ’16.’ We chose the former in all cases. 77 Table 6.1: Canonical strategies as phenotypes Strategy Guess Orthogonal Array Adaptive One at a Time Full Factorial Profligacy 0000 0001 0101 1001 Planning any 1001 0010 1001 Care any gt 0000 gt 0000 any Fairness any 1100 0001 1111 then the groupings of size one were balanced. This means that for every A1, the player was inclined to include as many A2s and B1s. If the second bit was flipped, then the groupings of size two were balanced – A1B3 would tend to appear no more often than B2D1. Adaptive One At A Time balances four of a kind: only its last bit would be flipped. diligence allele was never implemented, and was used for sizing the populations and correcting the generation rule. If the diligence converged more quickly than another allele or converged to a particular value, this pointed out some problem in the model itself. 6.4.2 The Fitness Function At the end of its experimental strategy, a player would then try to get a payoff by running tests, as explained in section 5.6. The fitness function was simply the payoff less the experimental cost; alternately, we could have considered the payoff as the fitness function, and defined a penalty function to be the experimental cost; further work may find that distinction useful. It may be of interest to examine the signalto-noise ratio instead, as per the Taguchi method, but this would involve running multiple players with the same strategy. In the interest of simplicity, we ran only single players with single strategies. f = N (P )/T − Cost = N (P )/T − P r(P l)Ca (6.1) Where N is the number of successful trials, P is the total possible payoff, T is the 78 number of trials, and Pr, Pl and Ca are the profligacy, planning and care respectively. In all the trials we ran, the player had three chances to make the payoff, and the payoff was delivered in thirds. Therefore the T in the payoff equation was always ’3’, and N varied from ’0’ to ’3.’ 6.4.3 Selection The selection criteria are simple – we simply allow the strategies working for more successful players a greater chance of reproduction. This is the stochastic remainder selection mechanism described in section 3.4.2.We can look at some sample code from GAAirplaneWorld.java. /∗ ∗ ∗ // u p d a t e d 1/26/04 t o use t h e s t o c h a s t i c remainder s e l e c t i o n without ∗ r e p l a c e m e n t mechanism i n G o l d b e r g ’ s Genetic Algorithms , f i g u r e 4.24 ∗ ∗ Gives t h e p l a y e r s new s t r a t e g i e s , and z e r o e s t h e i r w e a l t h ∗ @param p l a y e r s −− t h e l i s t o f p l a y e r s ( s h o u l d be a p e r f e c t s q u a r e i n number ) ∗ @return −− t h e same l i s t o f p l a y e r s w i t h z e r o w e a l t h and new strategies ∗ @todo −− r e v i s i t t h e w e a l t h c a l c u l a t i o n . Maybe a v e r a g e mother and f a t h e r ? ∗/ 10 protected void g e n e r a t e ( P l a y e r [ ] p l a y e r s ) throws PlayerCreationException , j a v a . s q l . SQLException , IOException { // do a l i t t l e sigma t r u n c a t i o n and l i n e a r s c a l i n g // g e t t h e l i s t o f w e a l t h s int p o p s i z e = p l a y e r s . l e n g t h ; double [ ] w e a l t h s = new double [ p o p s i z e ] ; 79 f o r ( int pdex = 0 ; pdex < p o p s i z e ; pdex++) { w e a l t h s [ pdex ] = ( double ) p l a y e r s [ pdex ] . w e a l t h ; } // g e t t h e mean and l o w e r c u t o f f f o r v a l i d i t y double [ ] s t a t s = A n a l y s i s . s t a t i s t i c a l m o m e n t s ( w e a l t h s ) ; 20 double mean = s t a t s [ 0 ] ; double d e v i a n t = mean − 2 ∗ s t a t s [ 1 ] ; // i f t h e r e i s no d e v i a t i o n ( c o u l d happen ) a r b i t r a r i l y move t h e c u t o f f down i f ( 0 == s t a t s [ 1 ] ) { d e v i a n t = mean − 1 ; } double max = Double . MIN VALUE ; double min = Double .MAX VALUE; // change t h e w e a l t h t o z e r o i f i t ’ s more than two s t a n d a r d d e v i a t i o n s b e l o w t h e mean f or ( int pdex = 0 ; pdex < p o p s i z e ; pdex++) { w e a l t h s [ pdex ] = w e a l t h s [ pdex ] > d e v i a n t ? w e a l t h s [ pdex ] − deviant : 0; 30 // c a p t u r e t h e g r e a t e s t s c a l e d w e a l t h i f ( w e a l t h s [ pdex ] > max) { max = w e a l t h s [ pdex ] ; } i f ( w e a l t h s [ pdex ] < min ) { min = w e a l t h s [ pdex ] ; } } Now that we have zeroed the underacheivers, we calculate the new mean. Keeping that mean, we rescale the values of the objective function so that the greatest value is twice the new mean. We’ll limit that scaling a little to avoid negative numbers. Dividing by the mean gives us a scaled fitness function. // r e c a l c u l a t e t h e mean mean = A n a l y s i s . s t a t i s t i c a l m o m e n t s ( w e a l t h s ) [ 0 ] ; i f ( 0 == mean ) { mean = 1 ; 80 // so , I want max t o be s c a l e d t o t w i c e t h e mean w i t h o u t c h a n g i n g s a i d mean} double s l o p e = 1 . / (max / mean − 1 ) ; i f ( Double . i s I n f i n i t e ( s l o p e ) ) { s l o p e = 0 ; // a d j u s t f o r t h e non−v a r i a n t c a s e } 10 double i n t e r c e p t = ( 1 . − s l o p e ) ∗ mean ; // u n l e s s t h i s would p u t t h e l e a s t one l e s s than ze r o , i n which c a s e s c a l e i t t o z e r o i f ( 0 > s l o p e ∗ min + i n t e r c e p t ) { s l o p e = 1 . / ( 1 . − min / mean ) ; i n t e r c e p t = ( 1 . − s l o p e ) ∗ mean ; } f or ( int pdex = 0 ; pdex < p o p s i z e ; pdex++) { w e a l t h s [ pdex ] = w e a l t h s [ pdex ] ∗ s l o p e + i n t e r c e p t ; } At this point we have a reasonably well-scaled set of values for the objective function (the wealth, which is the sum payoff less the experimental cost,) with a dispersion approaching the magnitude of the largest value. This was necessary to more easily differentiate between fine performance differences[22]. Note that it preserves players with a negative objective function value, and may let the best of a bad lot dominate, preventing the growth of a more effective strain. Now, we preselect the players that have a scaled fitness function one or greater by moving them into a selection list. We attempted to scale the greatest value, recall, to twice the mean. If that was sucessful, it gets preselected twice here. //OK, so , I ’m a l l s c a l e d . P r e s e l e c t t h e p l a y e r s . boolean [ ] [ ] [ ] p r e s e l e c t e d = new boolean [ p o p s i z e ] [ 5 ] [ GAStrategy . ALLELE LENGTH ] ; double [ ] f r a c t i o n = new double [ p o p s i z e ] ; // t h i s i s t h e residual part int ndex = 0 ; 81 f o r ( int pdex = 0 ; pdex < p o p s i z e ; pdex++) { f r a c t i o n [ pdex ] = w e a l t h s [ pdex ] / mean ; /∗ f o r however many t i m e s t h e mean t h e y a r e ( which s h o u l d be bet we e n 0 and 10 2) , add one t o t h e p r e s e l e c t i o n . The remainder i s t h e chance t h e y g e t a n o t h e r e n t r y i n t h e preselection . ∗/ while ( f r a c t i o n [ pdex ] >= 1 ) { p r e s e l e c t e d [ ndex++] = ( ( GAStrategy ) p l a y e r s [ pdex ] . getStrategy () ) . getAlleles () ; f r a c t i o n [ pdex ]−−;} } Every player2 qualified for preselection subtracted one from its scaled fitness. This left a fraction for any non-integer multiples of the mean. We now interpret the fraction as the probability that the player will be moved into the selection list. Pi = Wi /µ0 − s (6.2) Where Pi is the probability the genotype will be selected, Wi is the player’s wealth, µ0 is the mean calculated after the removal of the negative values, and s is the number of times the genotype was already transferred to the select list. For example, if a player’s wealth was 372 and the recalculated mean after the negative values were removed was 248, the odds that it would end up on the selection list a second time after being preselected once (for being greater than the mean) is P372 = 372/248 − 1 = 0.5 = 50/ 2 or, more accurately, the set of alleles defining the player’s strategy 82 (6.3) If another player had come out of the same game with a wealth of 124, its probability of ending up on the selection list for the first time would be P124 = 124/248 = 0.5 = 50/ (6.4) try { // s t o p t h i s when t h e a l g o r i t h m t r i e s t o w r i t e p a s t t h e end o f t h e a r r a y //Now, we a s s i g n more a l l e l e s t o t h e p r e s e l e c t e d a r r a y a c c o r d i n g t o chance while ( ndex < p o p s i z e ) { // i t e r a t e t h r o u g h t h e p l a y e r s u n t i l we have enough alleles fo r ( int pdex = 0 ; pdex < p o p s i z e ; pdex++) { i f ( 0 < f r a c t i o n [ pdex ] ) { i f ( rand . nextDouble ( ) <= f r a c t i o n [ pdex ] ) { f r a c t i o n [ pdex ]−−; // t h a t ’ s a l l f o r you ! 10 p r e s e l e c t e d [ ndex++] = ( ( GAStrategy ) p l a y e r s [ pdex ] . getStrategy () ) . getAlleles () ;} } } } } catch ( ArrayIndexOutOfBoundsException e ) {} An element is chosen randomly from the select list, and the model • makes a new strategy object from it (line 17) • adds it, once to the fathers list and once to the mothers list (lines 18 and 19) • replaces the genotype with the last accessible item in the list (line 21) • decreases its range of selectable items by one (line 22.) // s e l e c t them two−by−two ParameterSpace s p a c e = p l a y e r s [ 0 ] . ge tP ar amet e rS pace ( ) ; 83 /∗ l e t me t a k e t h e t o p h a l f as p a r e n t s . So , I can l e t each winner be a mother and a f a t h e r ∗/ GAStrategy [ ] f a t h e r s = new GAStrategy [ p o p s i z e / 2 ] ; GAStrategy [ ] mothers = new GAStrategy [ p o p s i z e / 2 ] ; 10 int end = p o p s i z e / 2 ; f or ( int pdex = 0 ; pdex < p o p s i z e / 2 ; pdex++) { int c h o i c e = rand . n e x t I n t ( end ) ; GAStrategy how = new GAStrategy ( space , p r e s e l e c t e d [ c h o i c e ][0] , preselected [ choice ] [ 1 ] , preselected [ choice ] [ 2 ] , preselected [ choice ] [ 3 ] , preselected [ choice ] [ 4 ] ) ; mothers [ pdex ] = how ; f a t h e r s [ pdex ] = how ; 20 // copy t h e l a s t ( now i n a c c e s s i b l e ) c h o i c e o v e r t h e j u s t s e l e c t e d one p r e s e l e c t e d [ c h o i c e ] = p r e s e l e c t e d [ end ] ; end −−;} When I have chosen half the population to reproduce, I let them do so by selecting a mating partner for each strategy object randomly and letting them reproduce, first with one as the ’left parent,’ then with the other. The mechanics of mating are discussed more fully below in section 6.4.4, ’Crossover.’ j a v a . s q l . Connection con = t h i s . con ; // g e t t h e c o n n e c t i o n f o r p l a y e r −making 84 //OK, so , now I have t h e s e p a r e n t a l s t r a t e g i e s . Let ’ s p a i r them o f f // g e n e r a t e new s t r a t e g i e s and p l a c e them i n p l a y e r s f o r ( int pdex = 0 ; pdex < p o p s i z e / 2 ; pdex++) { // l e t each winner be a f a t h e r GAStrategy f a t h e r = f a t h e r s [ pdex ] ; int c h o i c e = rand . n e x t I n t ( p o p s i z e / 2 − pdex ) ; 10 GAStrategy mother = mothers [ c h o i c e ] ; GAStrategy son = new GAStrategy ( space , f a t h e r , mother ) ; p l a y e r s [ pdex ] = new P l a y e r (new C o n c r e t e D a t a S t o r e ( con , t h i s . getNextPlayerCode ( son . g e t C l a s s ( ) . getName ( ) , I n t e g e r . MIN VALUE) ) ) ; p l a y e r s [ pdex ] . s e t S t r a t e g y ( son ) ; GAStrategy d a u g h t e r = new GAStrategy ( space , mother , f a t h e r ) ; p l a y e r s [ p o p s i z e − pdex − 20 1 ] = new P l a y e r (new C o n c r e t e D a t a S t o r e ( con , this . getNextPlayerCode ( d a u gh t e r . g e t C l a s s ( ) . getName ( ) , I n t e g e r . MIN VALUE) ) ) ; p l a y e r s [ p o p s i z e − pdex − 1 ] . s e t S t r a t e g y ( d a u g h t e r ) ; i f ( pdex < p o p s i z e / 2 − 1 ) { // don ’ t do t h i s t h e l a s t time mothers [ c h o i c e ] = mothers [ p o p s i z e / 2 − pdex − 1 ] ; } } } I now have an output population equal in size to the input population, with a parentage represented by how well it had done in the objective function compared to the mean. Let us next look at how the child strategy objects differ from the parents. 85 6.4.4 Crossover Alleles are simply split in half, and the left half from the left parent is joined to the right half of the right parent. As with the care phenotype and many other things in the model, more complex and fruit-promising structures were employed, modified and discarded in the process of understanding the model. Variable crossover could be profitably reemployed here. /∗ ∗ ∗ G e n e r a t i o n a l C o n s t r u c t o r . Let a s t r a t e g y be t h e c h i l d o f two parents ∗ @param s p a c e t h e parameter s p a c e i n which t h e s t r a t e g y operates ∗ @param l e f t t h e p a r e n t t h a t c o n t r i b u t e s t h e l e f t p a r t o f t h e allele ∗ @param r i g h t t h e p a r e n t t h a t c o n t r i b u t e s t h e r i g h t p a r t o f the a l l e l e ∗ ∗/ public GAStrategy ( ParameterSpace space , GAStrategy l e f t , GAStrategy r i g h t ) { 10 this . space = space ; /∗ r i g h t , ok . Go t h r o u g h t h e a l l e l e s , d e t e r m i n e a s p l i t t i n g p o i n t , and j o i n ’em up ∗/ boolean [ ] [ ] mother = l e f t . g e t A l l e l e s ( ) ; boolean [ ] [ ] father = right . getAlleles () ; boolean [ ] [ ] a l l e l e s = new boolean [ mother . l e n g t h ] [ ] ; f or ( int genome = 0 ; genome < a l l e l e s . l e n g t h ; genome++) { // copy t h e mother ’ s a l l e l e boolean [ ] a l l e l e = new boolean [ mother [ genome ] . l e n g t h ] ; 86 20 // f i n d t h e s p l i t t i n g p o i n t int s p l i t = 2 ; // r e p l a c e t h e gene ’ s w i t h t h e f a t h e r ’ s a f t e r t h e s p l i t t i n g point fo r ( int gene = 0 ; gene < a l l e l e . l e n g t h ; gene++) { a l l e l e [ gene ] = gene < s p l i t ? mother [ genome ] [ gene ] : f a t h e r [ genome ] [ gene ] ; } a l l e l e s [ genome ] = a l l e l e ; } //OK, so now I have t h e same l i s t o f a l l e l e s as my p a r e n t s . Mutate them . a l l e l e s = mutate ( a l l e l e s ) ; setAlleles ( alleles ) ; 30 initializeStartingConfiguration () ;} 6.4.5 Mutation The mutate method call in listing 6.4.4 simply looks as each gene of an allele in turn, tests its probability of mutating a gene against a uniform random number between 0 and 1, and flips the bit on the gene if the random number is less than the mutation probability. If it has flipped any of the bits on an allele, it replaces it. This code follows. /∗ ∗ ∗ Unpredictably a l t e r s the strategy ’ s a l l e l e s ∗ @param a l l e l e s −− two d i m e n s i o n a l a r r a y o f b o o l e a n s ∗ @return −− argument a r r a y w i t h randomly s e l e c t e d b i t s f l i p p e d . The ∗ p r o b a b i l i t y o f s e l e c t i o n i s t h e parameter probabilityofmutation ∗/ public boolean [ ] [ ] mutate ( boolean [ ] [ ] 87 alleles ) { // go t h r o u g h each b i t . I f i t ’ s u n l uc k y , f l i p i t 10 f or ( int genome = 0 ; genome < a l l e l e s . l e n g t h ; genome++) { boolean [ ] a l l e l e = a l l e l e s [ genome ] ; boolean d i r t y b i t = f a l s e ; for ( int gene = 0 ; gene < a l l e l e . l e n g t h ; gene++) { i f ( rand . n e x t F l o a t ( ) < p r o b a b i l i t y o f m u t a t i o n ) { d i r t y b i t = true ; a l l e l e [ gene ] = ! a l l e l e [ gene ] ; } } // i f you changed a b i t , swap i t o u t if ( dirtybit ) { a l l e l e s [ genome ] = a l l e l e ; } } 20 return a l l e l e s ; } 88 6.5 Optimization To apply the genetic algorithm search method to the Comparator, we took the tendencies outlined in section 6.2, and expressed each as a four-bit allele3 as described in section 6.4.1. We then allowed populations to evolve along the lines described in section 6.4, passed the strategies to the players described in Chapter 5 and sought lessons in the result. The investigations with the genetic agents differed from the investigations of Chapter 5: 1. The rewards are not parabolic, and these agents operate without a quality loss function. They obtain one quantum payoffs for every payoff attempt for which the experimental result is within 5 % of the target, which is again 80. This corresponds to a range 18.524 to 19.616. Table A.1 is the data source yet has relatively few data in this range, making it difficult to hit with normally distributed pseudorandom numbers generated around each configuration’s mean according to its standard deviation. 2. Only the scaled Grewen data is used. The nonlinearity and noise, discussed in section 5.3, are not adjusted. 3. Neither adjusted in this study is the payoff. While the model is certainly capable to this sort of adjustment, and such model runs were made, the output was not instructive and is not presented here. As a nod to the hierarchical ordering principle, we instruct the strategy to construct an optimal configuration from a main effects model using the data it has gathered. As the configuration submitted to payoff testing is the best configuration the player has experienced, this creates some reward for seeking balance to the extent that main effects are important. 3 In Java, this was implemented as a four member boolean array 89 Average Payoff By Generation 800 0.35 600 0.3 0.25 200 0.2 0 1 2 3 4 5 6 7 8 9 10 0.15 Payoffs Value Units 400 -200 Cost Wealth Payoff 0.1 -400 0.05 -600 0 -800 Generation Figure 6-1: Evolution of Payoff over ten generations(Run 1) Average Payoff By Generation 800 0.35 600 0.3 0.25 200 0.2 0 1 2 3 4 5 6 7 8 9 10 0.15 Payoffs Value Units 400 -200 Cost Wealth Payoff 0.1 -400 0.05 -600 0 -800 Generation Figure 6-2: Evolution of Payoff over ten generations(Run 2) 6.6 Results and Analysis In figures 6-1 and 6-2, the players can be seen to suffer exponentially decaying cost with a more affine decay in payoff until the last few generations, wehre payoff suddently recovers after cost reduction has stopped increasing wealth. The agents have generations of sacrificing payoff for reduced experimental cost; they never get very good at finding the best configurations. 6.6.1 Strategy Competence Let us introduce a concept of strategy competence. In the NormalDataGame the genetic agents are playing each configuration t has a mean µt and a standard deviation 90 Table 6.2: Ten 80-best configurations Plane # 1311 3121 3112 3321 2112 1313 1321 3231 3313 2312 mean 83.15 84.25 86.45 73.63 87.91 71.43 83.88 69.23 90.84 80.95 stdevp 10.99 10.62 9.04 15.30 11.47 8.79 21.03 18.88 11.16 31.90 80 goodness 0.36 0.35 0.30 0.24 0.24 0.21 0.21 0.17 0.16 0.14 Table 6.3: Ten modal configurations, run 1, 200 players ten generations MODE 2 2111 3111 3 1223 4 1311 5 6 1313 1323 7 3211 8 1233 9 1211 10 1222 σt . The player will receive a full payoff of p(a) = 1 if experiment a yields a performance within tolerance of the target, which here means between 76 and 84. Since the performance (or flight distance) is a pseudorandom normal variable, p’s expectation value < p > is given by the error function. 2 2 Z 84 (µ−x) e σ2 dx < p >= √ π 76 (6.5) This expectation value for payoff varies from 0 to 1, and is called the ’80-goodness.’ The 80-goodness for the top ten configurations are listed in table 6.2. To measure the competence of the algorithm, let us run our model twice with populations of 200 agents for ten generations. The top ten selected configurations among all agents in these models are shown in table 6.3 and table 6.4. We get a sense of the competence of the populations by looking at how well the top ten selections congrue. The first, second and seventh most popular selections for the first run are respectively the second, first and ninth configurations for the second, Table 6.4: Ten modal configurations, run 2, 200 players ten generations MODE 2 3111 2111 3 1211 4 1213 5 6 1322 1212 91 7 1321 8 1221 9 3211 10 2211 Diligence Allele By Generation 40 35 Generation 1 30 Generation 2 25 Generation 3 20 Generation 4 Generation 5 Generation 6 Generation 7 Generation 8 Generation 9 15 10 5 0 0000 0011 Generation 10 0110 1001 1100 1111 Figure 6-3: Evolution of Diligence(Run 1) but none of them contain any of the top ten configurations for the space. These algorithm, then, is not terribly competent. 6.6.2 Two model runs Figures 6-5 and 6-6illustrate how profligacy evolves over the course of several generations. The agents tended to reduce costs for the first several generations, making their payoffs low. This is, we posit, because the noise and interaction levels were so high no viable strategy ever arose. That we allowed agents with negative values to propagate may have bred agents unable to operate in the data environment. We can now look at the evolution in two runs of the five alleles. Let us consider diligence first, as it has no operational role and will tell us when spurious convergence has taken place, either through some correlation with a useful allele – this is an artifact of our crossover rule – or through gene starvation. The figures show a clear, slowly growing convergence around 1001 in run 2 (figure 6-6) and a bimodal one around 0010 and 1111 in run 1 (figure 6-5). The model has been validated in that the convergences are around different numbers. A larger population size would converge more slowly. Since from the payoff chart we know what happens to profligacy, let us look at that next. Fully half the population is at 0000 Profligacy, which means the agent enacts 1 92 Diligence Allele By Generation 45 40 Generation 1 Generation 2 Generation 3 Generation 4 Generation 5 Generation 6 Generation 7 Generation 8 35 30 25 20 15 10 5 0 0000 Generation 9 0011 Generation 10 0110 1001 1100 1111 Figure 6-4: Evolution of Diligence(Run 2) Profligacy Allele By Generation 120 100 Generation 1 Generation 2 80 Generation 3 60 Generation 4 Generation 5 40 Generation 6 20 Generation 7 Generation 8 Generation 9 Generation 10 0 0000 0011 0110 1001 1100 1111 Figure 6-5: Evolution of Profligacy(Run 1) plan, by generation 6. While a full 30 % of run 2’s agents go to 0000 profligacy, there is another attractor at 0100 (which enacts 1 · 22 + 1 = 5 plans.) We are starting to see evidence of adaptive behavior, which requires multiple plans. Planning, though, is where adaptive and rigid behavior are distinguished. In run 1 (figure 6-7), 30 % of the population is at 0 – which is equivalent to Guess – at generation 6 and 40 % by generation 10, with as many again experimenting on only one or two configurations. As guess plays havoc with selected configurations, this explains the lack of competence of the populations. In run 2 (figure 6-8), we see a residual population slowly decaying at 1000 Run 2 (figure 6-10) has another mid-field 93 Profligacy Allele By Generation 90 80 Generation 1 Generation 2 Generation 3 Generation 4 Generation 5 Generation 6 Generation 7 Generation 8 70 60 50 40 30 20 10 0 0000 Generation 9 0011 Generation 10 0110 1001 1100 1111 Figure 6-6: Evolution of Profligacy(Run 2) Planning Allele By Generation 90 80 Generation 1 Generation 2 Generation 3 Generation 4 Generation 5 Generation 6 Generation 7 Generation 8 Generation 9 Generation 10 70 60 50 40 30 20 10 0 0000 0011 0110 1001 1100 1111 Figure 6-7: Evolution of Planning(Run 1) 94 Planning Allele By Generation 140 120 Generation 1 Generation 2 Generation 3 100 80 Generation 4 60 Generation 5 40 Generation 6 20 Generation 7 0 Generation 8 0000 Generation 9 0011 Generation 10 0110 1001 1100 1111 Figure 6-8: Evolution of Planning(Run 2) Care Allele By Generation 100 Generation 1 Generation 2 80 60 Generation 3 Generation 4 40 Generation 5 Generation 6 Generation 7 Generation 8 Generation 9 20 0 0000 0011 Generation 10 0110 1001 1100 1111 Figure 6-9: Evolution of Care(Run 1) population in Care, this one holding steady with about an eighth of the population. Otherwise, the drive to zero is very strong, capturing half the population of run 1 (figure 6-9) and 5/8 of run 2 by the tenth generation. Repetitions are not found to be adaptive in this environment. Fairness does not converge convincingly, in fact converging dramatically less than diligence in run 2 (figure 6-11, compare to (figure 6-12)). We would expect that fairness, which propels an agent though the parameter space, would only be a competitive advantage if the agents were doing long experimental designs. Since they tend to do so little experimentation, they can find no advantage in doing them evenly. 95 Care Allele By Generation 100 Generation 1 Generation 2 80 60 Generation 3 Generation 4 40 Generation 5 Generation 6 20 Generation 7 0 Generation 8 0000 Generation 9 0011 Generation 10 0110 1001 1100 1111 Figure 6-10: Evolution of Care(Run 2) Fairness Allele By Generation 35 30 Generation 1 Generation 2 Generation 3 25 20 Generation 4 15 Generation 5 10 Generation 6 Generation 7 Generation 8 Generation 9 Generation 10 5 0 0000 0011 0110 1001 1100 1111 Figure 6-11: Evolution of Fairness(Run 2) 96 Fairness Allele By Generation 50 40 Generation 1 Generation 2 30 Generation 3 Generation 4 20 Generation 5 10 Generation 6 Generation 7 0 Generation 8 0000 Generation 9 Generation 10 0011 0110 1001 1100 1111 Figure 6-12: Evolution of Fairness(Run 1) 6.7 Analysis The lessons in these results are minimal. However, the concept and framework for abstracting and optimizing design of experiments strategies here presented can provide a springboard for further work, and can be run in the Comparator against other data sets or with other abstractions and codings. 97 Chapter 7 Conclusion Statistically driven experimental strategies display a healthy diversity in practice. The relative qualities of experimental strategies are quantifiable within the supplied framework. We propose that it is possible to optimize an experimental design itself. 7.1 7.1.1 Summary Problem Statement Design of Experiments (DoE) is a fundamental part of Robust Engineering. Deciding which prototypes to build, how many of each to test and how to interpret the results is part of what makes Product Development an art. The goal of Robust Engineering is to minimize the effect of variability on performance and thus to optimize quality. The goal of DoE is to do this at the least possible cost by choosing and interpreting experiments wisely. However, it is not at all obvious what the correct set of configurations to select for experimentation are, or how to select the number of experiments to do on each one. With the rise of the use of statistics in product design through the twentieth century, and especially with Taguchi methods in the last half, the practice of fitting a statistical model to the reality of product performance, doing a predicted number 98 of experiments, and selecting a best configuration based on the model coefficients has become widely adopted for design projects. As with any business transformation movement, there is a risk that the core messages will be lost and the selection of configurations will become rote. Or that configurations selected in one product performance space will be inappropriately applied to another. So, we ask, ”what are people doing when they do design of experiments?” Further, ”How is management trying to influence such choices?” And, ”How effective are management interventions?” We also partially develop a method for asking how Design of Experiments can be improved to create good answers on an arbitrary design surface. To do this, we have to develop a way for objectively evaluating one strategy against another in a particular design space. Further, we have executed an optimization algorithm against that framework to demonstrate how one might characterize and perfect a design of experiments strategy. 7.2 Conclusion This thesis is organized in 7 chapters, dealing with three broad themes. The first is the contextualized prototype testing interviews; the second is the comparator; and the third is the optimization algorithm. In the first section, we discussed the practice of Design of Experiments in Robust Engineering. From previous work, we had some idea of the popularity of the concept, and we pursued a few in depth interviews to get some insight into the structure of prototype testing decisions made in individual companies. This has some implications. Our expectation was that the position of prototype testing in validated design, the last phase of detailed design, itself the last phase of the design phase of the product development time line, that is, the very last thing you do before implementing, would create a pressure to cut into the time spent carefully assembling and testing the prototypes. We did not find that practitioners tended to return to earlier parts of the design ef99 fort in response to information generated during the Design of Experiments endeavor. This suggests that the parameter design was done carefully enough to isolate it from the system design, and that the parameters themselves were not pushed up against their constraints. Analytic models provide some DoE opportunity at the system level. They are useful when the governing equations are well known and the statistical model has a good fit to the data, and can give valuable guidance around sensitivity. We were impressed by the confidence shown by analytic prototypers, and the clarity and definition around their belief they were saving their organizations time and money. We acknowledge system level mathematical analysis is difficult, and that the machines and expertise they require can be prohibitively expensive. Yet, as tools are always improving and computers becoming more powerful[31, fig. 1-1], we urge designers of complex product systems to periodically revisit the issue of whether to create such a system. We discovered a very real difference between product owners and subcontractors that indicates a supply chain issue with quality engineering. Neither the defense contractor in section ch4:prof-nmp nor the parts manufacturer in section ch4:profapm seem particularly concerned with delivering the best possible product, because there is no contractual reward for overdelivering on performance or variability. In section 4.3.1, we recommend that proper incentives for quality be written into parts and subsystem contracting. Design of experiments in robust engineering has come to be closely linked with design for six sigma (DFSS). DFSS would be considered a framework in the five-layer model, and the warning explicated in section D.5 may pertain. DFSS may start to lose its core meaning, and seems to lack coordination mechanisms such that any one implementation will comport with every other. This puts DFSS in danger of being used as a generic label for any kind of statistical control, which will eventually discredit it. An industry movement to create consistency in the certification of belted staff would be helpful in delaying that. In the second section, we showed that a comparator could be built, and would give 100 expected and consistent results for particular payoff structures and response surfaces. The comparator is adaptable to different response surfaces, and can operate as a data picker or a number generator, with or without underlying data. We showed that it was possible to code a variety of DoE strategies into this framework, and that it would generate a metric with which strategies could be compared objectively. In the third section, we illustrated how one might abstract the differences between design of experiments strategies, abstracting generic concepts in design of experiments. We use a Genetic Algorithm model to optimize a DoE strategy for a particular response surface. 7.3 Further Work Our intention here has been to articulate and develop ideas for further rigorous investigations into design of experiments, and we have some suggestions for how that work can proceed. 1. Ahlman[32] discusses some ways to apply design of experiments in robust engineering. We have discovered a variety of methods of creating statistically driven experimental designs, including Uniform Design, D-Optima, Latin Hypercubes, Orthogonal Arrays and other Fractional Factorials, as well as Full Factorial and One-At-A-Time methods. It may be of interest to find out how popular each of these methods are and what factors outside of the statistical characteristics of the system of interest – like DFSS vendor, industry, educational institution or region – drive the selection of a particular DoE strategy. 2. The model for process support introduced in section 4.3 can be further elucidated and applied to various situations. It could perhaps be recast as a prescriptive or diagnostic model. 3. We have suggested that the practice of DoE is generally adopted locally by design engineers when the corporation enrolls them in a design for six sigma 101 (DFSS) initiative. The mechanisms by which this occurs could be examined and codified, either as case studies or within an industry. 4. The framework for comparing design of experiments strategies can be employed to seek nonlinear optimization schemes in DoE that can beat or compare to the optimal statistical scheme in a number of different circumstances. 5. Frey, et al.[12] suggests that there is a time to use more search-oriented experimental designs over statistically oriented ones. What the nature and characteristics of this boundary are can be investigated using the Comparator, with the intent of developing a practical heuristic for strategy switching. 6. The optimization scheme for DoE strategies themselves outlined in Chapter 6 gave us answers limited in scope, as elucidated in section 6.6. However, we believe that this thesis does show the concept of abstracting design of experiments strategies and creating optimization schemes in the supplied framework is essentially sound, and we encourage others to follow this work by exploring different response and payoff spaces. 102 Appendix A Grewen Data On the following pages are the data from the paper airplane game as they appeared in Grewen[29]. This was used in chapters 5 and 7. The data measures flights of airplanes with different configurations and is discussed in section 5.3.1. 103 PLANE # Throw 1 Throw2 12 23 15 19 11.5 13 15.5 12 13 14 18.5 13 14 15 13 13 14 15 12 13 14.5 14 10.5 12 13 17 14 17 16 15.5 15.5 16.5 14 11 21 20 15.5 17.5 12 15 13 12 13 13 15 1121 1122 1123 1131 1132 1133 1211 1212 1213 1221 1222 1223 1231 1232 1233 1311 1312 1313 1321 1322 1323 1331 1332 1333 2111 2112 2113 2121 2122 2123 2131 2132 2133 2211 2212 15 18 9 13 15 19 20 16 14.5 19 13 12 12.5 11 16 15 11 8 14.5 13 14 13.5 15 13.5 17.5 15 14 1113 19 15.5 1112 2.20 Table A.1: 10 throws for each of 1111 81 configurations 18 104 15 16 15 14 14.5 11 11.5 16.5 17 20 22 13 13 13 14 18 24 18 16.5 19 13 11 13.5 10 16.5 13 9 7.5 14.5 12.5 14 15 16 14.5 16 14.5 17 16.5 Throw 3 14.5 15 13 14 14 14 14.5 16 17 20.5 23 16 12 14 16 17 23 20 16 18 13.5 12 20 12.5 14 14 9 14.5 13 16 15 15 16 13.5 15 14 18 16.5 Throw 4 12 15 15 13 15 13 12 15 16 17.5 22 11 14 15 16 17 20 18.5 15 20 13 12 10.5 10.5 14 14 10 15 13 14 15 18 16.5 15.5 17 14 14.5 17.5 Throw 5 11 13 14 15 14 11 13.5 18 16 20 21 12 15.5 13.5 15 18 21 19 16 21 13 12 13.5 9 13 14 11.5 15 14 13 13.5 16 16 15 14.5 15.5 13 17.5 Throw 6 11 14 13.5 13.5 15 13 12 17 17 21 19 14 14 15.5 15 14 21 17 17 20 12.5 14 11.5 11 14 15 10 13 14 16 15 16.5 15.5 15 16 15 18 17.5 Throw 7 13 14.5 15 15 15 11.5 12 19 16.5 23 21 10 14 14 16 18 18 18 15 18 13 13 16.5 9 14 14 5 13.5 16 16.5 14.5 14.5 15 14 15 15 17.5 18 Throw 8 12 10.5 17 13 14 11.5 14 19 18 22 22 13 12 13 15 17 19 19 17 22 12.5 14 14.5 10.5 13 16 7 15.5 16 15 15 14.5 14 15 16 14 16 15 Throw 9 16 12.5 15 16 13 12 13 15 15 19.5 24 12 12 15.5 16.5 20 14 16.5 16 21 13 15 12.5 12 13.5 16.5 11.5 13 14 19 14 15 19 14.5 17 25 14.5 17.5 Throw 10 19 13.35 13.65 14.55 13.75 14.3 12.5 12.6 17.2 16.3 20.15 21.8 12.4 12.95 14.3 15.4 17.35 19.6 17.9 15.7 19.5 12.95 12.7 13.55 10.95 14.25 14.45 9.6 13 14.3 14.8 14.3 15.3 15.7 14.35 16.25 15.6 16.3 Mean Distance 1.80 1.58 1.21 1.32 0.86 1.33 1.07 1.57 0.98 1.65 1.40 1.71 1.80 1.25 0.74 1.70 3.03 1.26 1.03 1.58 0.28 1.25 2.91 1.55 1.16 1.17 2.21 2.91 1.03 2.08 0.71 1.25 1.44 0.82 1.25 3.35 1.92 17.3 Std Dev 3.23 2.50 1.47 1.74 0.73 1.78 1.16 2.46 0.96 2.73 1.96 2.93 3.25 1.57 0.54 2.89 9.16 1.60 1.07 2.50 0.08 1.57 8.47 2.41 1.35 1.36 4.88 8.44 1.07 4.34 0.51 1.57 2.07 0.67 1.57 11.21 3.68 1.09 Variance 0.42 0.42 0.11 0.55 0.56 1.24 1.83 1.01 1.66 2.00 3.62 0.79 0.54 0.26 1.26 0.91 0.53 1.99 0.93 1.91 21.88 1.28 0.14 1.56 0.34 0.19 1.05 0.20 0.39 0.02 0.81 0.37 0.48 0.54 0.98 0.08 0.43 1.18 S/N Ratio 105 14 14 14 13 21 11 14 18 22 18 17 13 13 11 11 12 12 17 16 12 13 14 14.5 9 10 12 15 17 18 19 15.5 13 15 14 13 14 8.5 9 14.5 13.5 12 20 12 11 13.5 15 22 13 15 18 2221 2222 2223 2231 2232 2233 2311 2312 2313 2321 2322 2323 2331 2332 2333 3111 3112 3113 3121 3122 3123 3131 3132 3133 3211 3212 3213 3221 3222 3223 3231 3232 3233 3311 3312 3313 3321 3322 3323 18.8 13.5 7 11.5 14 13 14.5 13 14.5 15 19 17 19 14 14 13 11.5 14.5 14 14 10.5 15 19 11.5 12 11 13 14 11 7.5 2213 Throw2 Throw 1 PLANE # 18 18 20 22 16 13 11 11 14 13.5 13 15 9 9 16 12 14.5 16 14.5 17.5 20.5 16 20 13 12 11 10.5 15 13 14.5 13 17 15 16.5 12 10 10 14 14 6 Throw 3 17 15 18 20 12 14 12 11 18 11 12 15 5 9 14 13 16 17.5 15 17 18.5 17 20 16 15 12.5 11 13.5 14 18.5 12 24 16 11.5 11.5 9 12 14 13.5 12.5 Throw 4 19 17 18 20 17 14.5 12 13 14 11 12.5 15.5 8.5 14 16.5 14 15 14.5 15.5 17 19.5 19 22 17 14 16 9.5 16.5 13 14.5 14 16 18 10.5 12.5 11 13 14.5 12.5 12 Throw 5 17 17.5 20 19.5 16 14 12 10 14 13 12 11.5 5 14 16 14 16 14.5 16 16.5 22.5 14 20.5 15 14 12 11 17.5 12 14.5 14 18 18 13 12 11.5 9 15.5 13.5 7 Throw 6 16.5 17.5 19 20 17 14 12 11.5 19 14 11 13 9 12 13 13.5 14 13 17 17.5 19 16 20 20 13.5 13 12 15.5 15.5 15.5 7 27 17.5 13 15.5 11 13.5 15 15 12 Throw 7 14 14 17 22 18 14 14 12 17 13.5 10 15 9 14 15 12.5 16 15 14 17 20.5 17 21 22 14 14 14 16 12 13 12 26 18 12.5 10.5 12.5 12 12 15 6 Throw 8 17 16.5 18 21 16.5 13 14 11 19 13.5 11 14 10 14 14 13 15 15 16 17.5 21 20 20 20 16 13 12 17 13 15 12 17 18.5 11 13.5 12 11 15 14 8 Throw 9 15.5 14 21 17 13 8 13 17 20 14 12 14.5 10 10 13 13.5 16 15.5 22 18 17 19 20 15 13 12 10.5 16 14 16 14 16 12.5 16 12 10 10 13.5 14 12 Throw 10 17.08 16.15 18.2 20.55 15.85 13.15 12.5 11.95 17.6 12.85 12.1 14.2 8.15 11.6 14.55 13.15 15.1 14.9 15.75 16.85 19.65 17.3 19.95 16.7 13.75 12.65 11.1 15.6 13.45 14.85 12.05 19.2 16.95 12.75 12.35 10.9 11.45 14.05 13.85 9.4 Mean Distance 1.52 1.51 2.20 1.61 2.00 1.87 1.18 1.95 2.72 1.13 1.22 1.18 1.86 2.33 1.26 0.63 0.84 1.33 2.47 0.94 1.53 1.77 1.30 2.98 1.23 1.63 1.41 1.24 1.07 1.60 2.11 4.59 1.96 2.02 1.33 1.02 1.50 1.04 0.78 2.73 Std Dev 2.30 2.28 4.84 2.58 4.00 3.50 1.39 3.80 7.38 1.28 1.49 1.40 3.45 5.43 1.58 0.39 0.71 1.77 6.13 0.89 2.34 3.12 1.69 8.90 1.51 2.67 1.99 1.54 1.14 2.56 4.47 21.07 3.86 4.07 1.78 1.04 2.25 1.08 0.61 7.43 Variance 1.03 0.63 0.72 2.26 0.28 0.45 1.59 0.73 0.39 1.45 1.75 0.37 1.90 0.57 0.10 3.99 0.54 0.11 0.17 2.40 2.11 0.83 3.10 0.22 0.64 0.77 1.82 0.57 1.11 0.05 0.60 0.21 0.58 0.48 1.33 3.65 1.45 0.61 1.41 0.71 S/N Ratio 106 14 12 13 3332 3333 16.5 12.5 12.5 3331 Throw2 Throw 1 PLANE # 14 13 13 Throw 3 14.5 10.5 11 Throw 4 16 13.5 12 Throw 5 18 10.5 14.5 Throw 6 17 7.5 13 Throw 7 13.5 13.5 13 Throw 8 17 12.5 16 Throw 9 16.5 13 13 Throw 10 15.6 12 13.05 Mean Distance 1.71 1.99 1.36 Std Dev 2.93 3.94 1.86 Variance 0.30 0.69 0.89 S/N Ratio A.1 The Contributions This data was extracted in preparation for calculating the components in section 5.3.1. Whereas in table A.1 and in most discussions in this thesis the parameter levels are labled 1 to 3, two was subtracted from each level index to generate table A.2. 107 Dif measure A 0.106 14.713 Y A Constant term 19.50 15.70 17.90 19.60 17.35 15.40 14.30 12.95 12.40 21.80 20.15 16.30 17.20 12.60 12.50 14.30 0.58 3.97 3.52 2.68 2.14 0.77 0.11 0.18 3.91 4.18 2.24 0.88 2.41 1.20 0.45 1.26 12.95 9.60 14.45 4.05 0.87 13.00 2.73 0.39 14.30 3.51 12.70 14.80 0.70 13.55 14.30 0.39 0.78 15.30 1.58 0.07 15.70 1.96 10.95 14.35 0.20 14.25 16.25 0.90 2.55 15.60 2.22 0.36 16.30 1.12 1 14.75 13.70 15.01 16.32 14.06 15.97 17.89 12.58 13.06 13.53 13.26 14.67 16.08 13.93 16.28 18.63 13.34 13.48 13.62 13.50 14.61 15.71 13.65 15.73 17.81 14.10 13.91 13.72 13.74 14.55 15.35 13.38 15.18 0 0 0 0 0 0 0 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 -1 Table A.2: Variance due to linear components 0.32 in 17.30 the Grewen Data 16.98 108 C -1 -1 -1 -1 -1 -1 -1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 -1 -1 -1 -1 -1 -1 -1 -1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 B -1 -1.012 C -0.298 B -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 D -0.825 D 0 0 0 0 0 0 0 -1 -1 -1 -1 -1 -1 -1 -1 -1 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 -1 A*B -0.359 A*B 0 0 0 0 0 0 0 -1 -1 -1 0 0 0 1 1 1 -1 -1 -1 0 0 0 1 1 1 -1 -1 -1 0 0 0 1 1 1 A*C 0.111 A*C 0 0 0 0 0 0 0 -1 0 1 -1 0 1 -1 0 1 -1 0 1 -1 0 1 -1 0 1 -1 0 1 -1 0 1 -1 0 1 A*D 0.283 A*D -1 0 0 0 1 1 1 1 1 1 0 0 0 -1 -1 -1 0 0 0 0 0 0 0 0 0 -1 -1 -1 0 0 0 1 1 1 B*C -0.050 B*C 1 -1 0 1 -1 0 1 1 0 -1 1 0 -1 1 0 -1 0 0 0 0 0 0 0 0 0 -1 0 1 -1 0 1 -1 0 1 B*D 0.483 B*D -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 C*D 0.514 C*D 0 0 0 0 0 0 0 -1 -1 -1 0 0 0 1 1 1 0 0 0 0 0 0 0 0 0 1 1 1 0 0 0 -1 -1 1 A*B*C 0.440 A*B*C 0 0 0 0 0 0 0 -1 0 1 -1 0 1 -1 0 1 0 0 0 0 0 0 0 0 0 1 0 -1 1 0 -1 1 0 -1 A*B*D 0.785 A*B*D 0 0 0 0 0 0 0 -1 0 1 0 0 0 1 0 -1 -1 0 1 0 0 0 1 0 -1 -1 0 1 0 0 0 1 0 -1 A*C*D -0.454 A*C*D 1 0 0 0 1 0 -1 1 0 -1 0 0 0 -1 0 1 0 0 0 0 0 0 0 0 0 -1 0 1 0 0 0 1 0 -1 B*C*D -0.092 B*C*D 0 0 0 0 0 0 0 -1 0 1 0 0 0 1 0 -1 0 0 0 0 0 0 0 0 0 1 0 -1 0 0 0 -1 0 -1 A*B*C*D -0.063 A*B*C*D 109 11.10 12.65 13.75 16.70 19.95 17.30 19.65 16.85 15.75 14.90 15.10 13.15 14.55 11.60 2.17 0.70 0.32 2.09 3.18 2.56 2.36 1.37 2.08 0.88 0.92 0.56 1.77 4.12 15.85 15.60 1.53 1.18 13.45 0.97 13.15 14.85 0.09 12.50 12.05 2.66 0.70 19.20 3.72 0.94 16.95 0.71 11.95 12.75 0.64 1.97 12.35 1.35 17.60 10.90 3.11 3.20 11.45 2.44 12.85 14.05 0.66 12.10 13.85 1.69 1.43 9.40 4.99 2.72 13.35 2.37 8.15 13.65 3.41 14.20 14.55 1.20 1.16 13.75 6.97 measure Dif 0.30 Y 14.67 13.85 13.44 13.92 14.40 14.28 14.82 15.36 15.12 15.72 16.32 12.59 14.18 15.78 13.67 15.48 17.29 14.74 16.77 18.79 13.43 13.35 13.27 14.07 14.42 14.76 14.71 15.48 16.24 13.39 13.70 14.01 13.89 14.71 15.54 14.39 15.72 17.06 13.35 14.05 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A B 1 1 0 0 0 0 0 0 0 0 0 -1 -1 -1 -1 -1 -1 -1 -1 -1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 -1 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 C 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 D 1 1 0 0 0 0 0 0 0 0 0 -1 -1 -1 -1 -1 -1 -1 -1 -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*B -1 -1 1 1 1 0 0 0 -1 -1 -1 1 1 1 0 0 0 -1 -1 -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*C 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 1 0 -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*D -1 -1 0 0 0 0 0 0 0 0 0 -1 -1 -1 0 0 0 1 1 1 1 1 1 0 0 0 -1 -1 -1 0 0 0 0 0 0 0 0 0 -1 -1 B*C 0 -1 0 0 0 0 0 0 0 0 0 -1 0 1 -1 0 1 -1 0 1 1 0 -1 1 0 -1 1 0 -1 0 0 0 0 0 0 0 0 0 -1 0 B*D 0 1 1 0 -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 1 0 C*D -1 -1 0 0 0 0 0 0 0 0 0 -1 -1 -1 0 0 0 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*B*C 0 -1 0 0 0 0 0 0 0 0 0 -1 0 1 -1 0 1 -1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*B*D 0 1 1 0 -1 0 0 0 -1 0 1 1 0 -1 0 0 0 -1 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*C*D 0 1 0 0 0 0 0 0 0 0 0 -1 0 1 0 0 0 1 0 -1 1 0 -1 0 0 0 -1 0 1 0 0 0 0 0 0 0 0 0 -1 0 B*C*D 0 1 0 0 0 0 0 0 0 0 0 -1 0 1 0 0 0 1 0 -1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 A*B*C*D 110 measure 20.55 18.20 16.15 17.08 13.05 12.00 15.60 Dif 5.06 4.76 1.99 2.19 0.03 1.65 1.32 Y 14.28 13.65 13.02 14.89 14.16 13.44 15.49 1 1 1 1 1 1 1 A 1 1 1 1 1 1 1 B C 1 1 1 0 0 0 -1 1 0 -1 1 0 -1 1 D 1 1 1 1 1 1 1 A*B 1 1 1 0 0 0 -1 A*C 1 0 -1 1 0 -1 1 A*D 1 1 1 0 0 0 -1 B*C 1 0 -1 1 0 -1 1 B*D 1 0 -1 0 0 0 -1 C*D 1 1 1 0 0 0 -1 A*B*C 1 0 -1 1 0 -1 1 A*B*D 1 0 -1 0 0 0 -1 A*C*D 1 0 -1 0 0 0 -1 B*C*D 1 0 -1 0 0 0 -1 A*B*C*D This table was generated by listing the means for a configuration in the column ’Y’, and then indicating which elements were present. Let us illustrate the approach with the A2B1C3D1 configuration. We subtract two from each level so that A = 0, B = −1, C = 1 and D = −1. Then, for example, B × D = −1 · −1 = 1 and A × B × C = 0 · −1 · 1 = 0, meaning that this configuration shows the effect of parameter B on the effect of parameter D, but does not contain A’s effect on B’s effect on C. On the first page of the table, the contribution terms are along the top. To compose the mean of each configuration, these are added, subtracted or not depending on whether the corresponding column contains a 1, -1 or 0 for that configuration’s row. The values in the header for each component are XLMiner’s best solution for this system of 81 equations for 16 unknowns. The composed value is in the Y column, the actual mean for the configuration in in the ’measure’ column, and the difference – resulting from higher order terms like A2 C – is in the ’Dif’ column. This table can then be used to generate table 5.3.1, discussed in section 5.3.1. 111 Appendix B Interview Questions in Design of Experiments ”Design of Experiments” (DoE) herein refers to the approach your design group takes to deciding on which experiments to perform in parameterized product design. 1. Do you have a formal methodology in Design of Experiments? If so, please answer each of the following questions for each methodology, indicating where your practice deviates from the formal methodology. If not, please answer each question according to your tendency. (a) How did you come up with your methodology? (b) Have you validated your methodology? 2. How do modelling and prototyping balance in your methodology? 3. How do you determine your testing budget? (a) How does this vary between different types of product design endeavors? (b) How flexible is your testing budget? (c) How often do you come in under budget? Over? (d) How much does each of these experiments cost you? 112 4. How do you choose your parameter levels? (a) Do you generally work with product iterations, or with novel products? (b) How is your system design accomplished? Is it performed with parameter design in mind? (c) How well do you feel you know the approximate optimal parameter settings before you begin? 5. Please describe your level setting (a) How many parameters can your methodology handle? (b) Does each parameter need to have the same number of levels? (c) How many levels can each parameter have? (d) How do you proceed to fine tune your parameter settings? 6. How do you navigate through the configurations? (a) Do you know which configurations you will test before you begin? (b) Under what conditions will you add configurations to test? (c) Under what conditions will you not test configurations you had originally planned to? 7. Please annunciate your methodology’s philosophy (a) Does it try to balance particular groupings, for example, to make sure that every pair of settings is represented equally? Why? (b) Is it concerned with covering a large area of the design space? 8. How many experiments do you do on a configuration? (a) Do you do the same number of experiments on each configuration? Is there a minimum number that you will do? 113 (b) How do you decide when to stop doing experiments on a particular configuration? 9. Target and variance (a) How many measurements does your response surface have? What are they? (b) Do you combine these into a single metric? How? (c) Are you primarily concerned with value, variance, or some combination of the two? (d) Do do use DOE to minimize a response value, maximize a response value, or try to bring the response value to a particular measure? Do you have different goals for different responses? 10. Taguchi methods (a) How do you compare different deviations from an optimal response? Do you use a parabolic quality loss function to compare different response values? 114 Appendix C Tables 115 C.1 Selected Configura- PlayerID Configuration Mu Frequency Sigma 10826 2223 19.75543 7 0.337543 10826 3132 61.14312 7 0.095518 10826 3211 57.37536 7 0.170007 10826 3212 44.72732 7 0.776722 10826 3213 36.18656 7 1.127652 10826 3221 51.27432 7 0.196692 10826 3222 38.43142 7 0.325064 10826 3223 30.07707 7 0.202549 Representatives of the four canonical DoE 10826 3231 31.79037 7 0.637585 10826 3232 18.6933 7 0.462875 types were run against the Grewen Data 10826 3233 10.32495 7 0.229383 10826 3123 72.37437 7 0.630645 described in Appendix A, as described in 10826 1222 35.67104 7 0.180602 10826 1211 54.81157 7 0.257605 section 5.5. The Guess strategy had no 10826 3133 52.87849 7 0.13108 10826 2222 27.91374 7 0.228272 data, as it does no experiments before try- 10826 1212 41.28829 7 0.817457 10826 1221 48.50138 7 0.262994 ing for the reward. But, listed below are 10826 1223 27.45756 7 0.417111 10826 1231 28.63735 7 0.400477 ten exemplar players for each strategy. 10826 1232 16.21678 7 0.202462 10826 1233 7.730179 7 0.077896 10826 1311 89.73102 7 0.401164 10826 1312 76.82089 7 0.188109 10826 1313 68.44617 7 0.322122 10826 1321 83.28073 7 0.498183 10826 1322 70.52987 7 0.216 10826 1213 33.80961 7 0.371004 tions for the Strategies C.1.1 Full Factorial The list of full factorial agents illustrates that each such agent visits every configuration. 10826 1331 63.6701 7 0.261361 [toll]wunderbar height 10826 3111 100.2155 7 0.476937 10826 1332 50.57786 7 0.310212 10826 2113 68.36128 7 0.222182 10826 2221 40.88418 7 0.15749 10826 2121 83.38096 7 0.458565 10826 2213 26.3259 7 0.727618 10826 3131 74.11078 7 0.321494 10826 2212 34.10137 7 0.301946 10826 2122 70.4573 7 0.318062 10826 1323 62.13096 7 0.135455 10826 2123 62.07166 7 0.212633 10826 2211 47.18546 7 0.43047 10826 2131 63.77643 7 0.214037 10826 1333 42.36598 7 0.311944 10826 2132 50.75678 7 0.479546 10826 3312 79.60071 7 0.467187 10826 2111 89.48107 7 0.174434 10826 3313 71.20297 7 0.188753 10826 2112 76.64826 7 0.186993 10826 3322 73.26717 7 0.444756 10826 3121 93.72876 7 0.259797 10826 3332 53.23598 7 0.603127 10826 3112 87.21328 7 0.132382 10826 3321 86.05909 7 0.677151 10826 2133 42.46152 7 0.267301 10826 3333 45.3722 7 0.371612 10826 1111 97.23322 7 0.149476 10826 3331 66.5451 7 0.214546 10826 1133 50.24109 7 0.685311 10826 2331 56.25953 7 0.490401 10826 1132 58.50793 7 0.119559 10826 2333 34.94829 7 0.436146 10826 1131 71.34159 7 0.198802 10826 3323 64.78235 7 0.226452 10826 1123 69.87565 7 0.16048 10826 3311 92.23012 7 0.433479 10826 1122 78.12189 7 0.238391 10826 2332 43.09297 7 0.329486 10826 1121 91.31952 7 0.373196 10826 2323 54.5053 7 0.250316 10826 1113 76.08479 7 0.819885 10826 2322 62.72095 7 0.158909 10826 1112 84.30915 7 0.468066 10826 2313 60.53102 7 0.334882 10826 3122 80.83533 7 0.071054 10826 2312 68.63959 7 0.93714 10826 2233 -0.06581 7 0.321205 10827 3111 100.0464 7 1.072997 10826 2232 8.258003 7 0.365448 10827 2112 76.67161 7 0.265895 10826 2231 21.09231 7 0.333574 10827 3112 87.02136 7 0.225619 10826 2311 82.10878 7 0.219928 10826 2321 75.74433 7 0.64096 10826 3113 78.66255 7 0.203463 116 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10827 3122 80.8941 7 0.207776 10827 1333 42.38289 7 0.493227 10827 2111 89.38625 7 0.295761 10827 2211 47.14543 7 0.234344 10827 3131 74.08672 7 0.36862 10827 2223 19.70233 7 0.334384 10827 3121 93.76518 7 0.430431 10827 2222 28.02398 7 0.122266 10827 2113 68.52522 7 0.207912 10827 2221 40.77921 7 0.191947 10827 2121 83.42163 7 0.373431 10827 1332 50.58789 7 0.436083 10827 2122 70.51149 7 0.392169 10827 2213 25.57597 7 0.746787 10827 2131 63.59153 7 0.10322 10827 1221 48.4994 7 0.173895 10827 2132 50.86116 7 0.287367 10827 3313 71.16184 7 0.480815 10827 2133 42.39767 7 0.154328 10827 3322 73.41985 7 0.385665 10827 1111 97.16582 7 0.267775 10827 3312 79.38481 7 0.417309 10827 3132 61.21471 7 0.096123 10827 3332 53.70172 7 0.598834 10827 2123 62.22394 7 0.373695 10827 3321 86.17372 7 0.555143 10827 1133 50.03784 7 0.36937 10827 3333 45.09119 7 0.482313 10827 1112 84.61602 7 0.37374 10827 3331 66.32618 7 0.332175 10827 1121 91.0857 7 0.259083 10827 2333 34.88251 7 0.255912 10827 1122 78.24737 7 0.142405 10827 2312 69.17097 7 1.114411 10827 1123 69.82908 7 0.19136 10827 2313 60.81978 7 0.639111 10827 1113 76.29016 7 0.527992 10827 2322 62.69747 7 0.252641 10827 1132 58.27875 7 0.250715 10827 2323 54.59292 7 0.254655 10827 1131 71.2232 7 0.245375 10827 2332 43.00685 7 0.285261 10827 3221 51.26542 7 0.186864 10827 3311 92.30759 7 0.540487 10827 2233 -0.10512 7 0.246222 10827 3323 64.86583 7 0.26247 10827 2311 82.08967 7 0.28914 10828 3121 93.80603 7 0.393306 10827 2321 75.79582 7 0.39998 10828 3112 87.00746 7 0.236622 10827 2331 56.14681 7 0.510558 10828 2111 89.35407 7 0.253764 10827 1312 76.71937 7 0.308011 10828 3111 100.1738 7 0.615781 10827 2232 8.127993 7 0.198358 10828 2112 76.51951 7 0.246493 10827 3212 44.69136 7 0.629669 10828 2113 68.3949 7 0.263706 10827 3133 52.87122 7 0.125717 10828 2121 83.18535 7 0.383105 10827 3223 30.17953 7 0.121166 10828 2122 70.37558 7 0.276277 10827 3222 38.6346 7 0.150281 10828 2131 63.54766 7 0.206926 10827 3123 72.88921 7 0.509285 10828 2132 50.5979 7 0.270097 10827 3213 36.48044 7 0.572734 10828 2133 42.42251 7 0.216862 10827 3233 10.47756 7 0.168211 10828 1111 97.33272 7 0.225144 10827 3232 18.58516 7 0.673925 10828 3131 74.33652 7 0.34517 10827 3113 78.79769 7 0.535579 10828 2123 62.25202 7 0.387976 10827 3231 31.33284 7 0.321116 10828 1133 50.08459 7 0.494421 10827 3211 57.40009 7 0.318674 10828 1112 84.64597 7 0.659324 10827 1223 27.59778 7 0.334483 10828 1121 90.83905 7 0.313651 10827 1211 54.80719 7 0.286653 10828 1122 78.14566 7 0.241066 10827 1321 83.62698 7 0.562736 10828 1123 69.92576 7 0.276111 10827 2231 21.1476 7 0.17727 10828 1131 71.36943 7 0.184849 10827 1212 42.16889 7 0.473908 10828 1132 58.42131 7 0.13331 10827 1213 33.3326 7 0.289037 10828 1113 76.22013 7 0.641218 10827 1222 35.56694 7 0.110984 10828 2232 8.26113 7 0.210221 10827 1231 28.55128 7 0.474492 10828 2233 -0.09573 7 0.378231 10827 1232 16.00803 7 0.27603 10828 2311 81.83687 7 0.356894 10827 1233 7.687431 7 0.043323 10828 2231 21.24329 7 0.222391 10827 1311 89.66518 7 0.337454 10828 2221 40.88772 7 0.147584 10827 1313 68.50299 7 0.196558 10828 3232 18.47211 7 0.502869 10827 1322 70.6245 7 0.330486 10828 3222 38.54215 7 0.271337 10827 1323 62.24637 7 0.119609 10828 3113 78.79086 7 0.397972 10827 1331 63.72681 7 0.310501 10828 3231 31.39952 7 0.462361 10827 2212 34.3596 7 0.45069 10828 3122 80.91971 7 0.296848 117 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10828 3123 72.05493 7 0.433959 10829 2113 68.42635 7 0.229825 10828 3223 30.1779 7 0.190135 10829 2121 83.37988 7 0.396071 10828 3132 61.1019 7 0.136813 10829 2122 70.53217 7 0.241134 10828 3133 52.81496 7 0.094685 10829 2131 63.57518 7 0.190191 10828 3211 57.46619 7 0.154391 10829 2133 42.20438 7 0.298835 10828 3212 44.4349 7 0.666139 10829 2123 62.11723 7 0.15984 10828 3213 36.1785 7 0.613405 10829 1133 49.84422 7 0.465432 10828 3221 51.37692 7 0.161595 10829 3131 74.14424 7 0.208033 10828 3233 10.3248 7 0.259028 10829 1111 97.13857 7 0.349898 10828 1231 29.0472 7 0.614265 10829 1113 75.93712 7 1.038271 10828 1212 42.22875 7 0.653827 10829 1121 91.0866 7 0.317662 10828 1213 33.74085 7 0.417505 10829 1122 78.10229 7 0.138469 10828 1221 48.53656 7 0.277772 10829 1123 69.71775 7 0.340329 10828 2223 19.82986 7 0.290246 10829 1131 71.36426 7 0.16032 10828 1223 27.54512 7 0.406632 10829 1112 84.16204 7 0.47462 10828 1232 15.96498 7 0.234367 10829 1132 58.31867 7 0.18258 10828 1233 7.747209 7 0.036397 10829 3211 57.60604 7 0.30808 10828 1311 89.748 7 0.425798 10829 2232 8.220903 7 0.318173 10828 1312 76.75586 7 0.189069 10829 2233 -0.10856 7 0.297907 10828 1313 68.43218 7 0.195884 10829 2311 81.84451 7 0.289558 10828 1321 83.60116 7 0.916878 10829 2321 75.88627 7 0.295308 10828 1322 70.74817 7 0.188007 10829 3222 38.54431 7 0.312567 10828 1323 62.25103 7 0.11926 10829 3221 51.13753 7 0.236433 10828 1331 63.43309 7 0.140925 10829 2231 21.22055 7 0.162575 10828 1333 42.60045 7 0.378945 10829 2331 55.8943 7 0.323035 10828 2211 46.91662 7 0.363682 10829 3212 44.62272 7 0.376303 10828 1222 35.90616 7 0.206562 10829 3232 18.91144 7 0.675194 10828 1332 50.60954 7 0.534786 10829 3231 31.73651 7 0.415566 10828 2222 27.90685 7 0.127609 10829 3132 61.25321 7 0.166006 10828 2212 34.49342 7 0.399072 10829 3123 72.1469 7 0.55068 10828 2213 26.12851 7 0.544093 10829 3122 80.72006 7 0.14568 10828 1211 54.77262 7 0.175844 10829 3223 30.07654 7 0.136968 10828 3312 79.13236 7 0.486374 10829 3233 10.46892 7 0.173252 10828 3332 53.35021 7 0.650831 10829 3113 78.73872 7 0.301163 10828 3331 66.39347 7 0.385022 10829 3213 36.13676 7 0.445381 10828 3323 65.02283 7 0.305242 10829 1213 33.84972 7 0.378237 10828 3322 73.2647 7 0.25665 10829 2223 19.85847 7 0.48785 10828 3313 71.19716 7 0.273125 10829 3133 52.93364 7 0.068246 10828 3311 92.3657 7 0.55173 10829 1212 42.03442 7 0.340972 10828 2332 42.78661 7 0.467654 10829 1221 48.71138 7 0.223077 10828 3321 86.07349 7 0.340321 10829 1222 35.76385 7 0.263479 10828 2312 69.14748 7 0.812801 10829 1223 27.4276 7 0.48903 10828 2333 34.97653 7 0.212578 10829 1231 29.30435 7 0.352877 10828 2331 56.08963 7 0.339541 10829 1232 15.84157 7 0.115863 10828 2323 54.53473 7 0.300318 10829 1233 7.710956 7 0.05459 10828 2322 62.69022 7 0.183232 10829 1311 89.56654 7 0.241554 10828 2321 75.44436 7 0.247539 10829 1312 76.86509 7 0.114859 10828 2313 60.70046 7 0.385388 10829 1313 68.3949 7 0.254993 10828 3333 45.46196 7 0.415543 10829 1321 83.40976 7 0.559438 10829 2132 50.54586 7 0.355816 10829 1322 70.49447 7 0.208862 10829 2112 76.63599 7 0.122751 10829 1211 54.85568 7 0.272788 10829 3112 87.04453 7 0.17298 10829 2213 26.45385 7 0.663329 10829 3111 100.1194 7 0.831769 10829 2212 34.15749 7 0.201556 10829 2111 89.45248 7 0.179542 10829 1323 62.13889 7 0.137433 10829 3121 93.73225 7 0.238512 10829 2211 47.27901 7 0.267671 118 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10829 1333 42.61804 7 0.338711 10830 3212 44.70443 7 0.672588 10829 2221 40.95322 7 0.143111 10830 3213 36.48177 7 0.88677 10829 2222 27.99571 7 0.321374 10830 3221 51.46459 7 0.211801 10829 1332 50.63676 7 0.386874 10830 3133 52.91907 7 0.129992 10829 1331 63.61836 7 0.242451 10830 3222 38.39098 7 0.256834 10829 3333 45.27511 7 0.357936 10830 3223 30.23152 7 0.281034 10829 3322 73.19034 7 0.310072 10830 3232 18.72676 7 0.367823 10829 3312 79.48028 7 0.215265 10830 1312 76.77289 7 0.285286 10829 3323 64.94426 7 0.421552 10830 1311 89.87053 7 0.28437 10829 3321 85.81243 7 0.443739 10830 1211 54.74304 7 0.176669 10829 3313 71.22746 7 0.18962 10830 1212 42.17279 7 0.532576 10829 3332 53.67342 7 0.427327 10830 1213 33.58316 7 0.454091 10829 2332 43.20728 7 0.54548 10830 1221 48.59795 7 0.243622 10829 2333 34.69185 7 0.315795 10830 1222 35.60726 7 0.18594 10829 3331 66.23613 7 0.284805 10830 1223 27.45589 7 0.465687 10829 2312 69.41299 7 0.575349 10830 1231 28.59167 7 0.635704 10829 2313 60.68389 7 0.56372 10830 1321 83.40703 7 0.300325 10829 2322 62.81876 7 0.22435 10830 1233 7.729514 7 0.054726 10829 2323 54.39 7 0.35928 10830 2232 8.345623 7 0.220114 10829 3311 92.34354 7 0.461426 10830 1313 68.56356 7 0.213635 10830 3113 78.90729 7 0.380806 10830 1322 70.30972 7 0.230426 10830 3121 93.8101 7 0.231744 10830 1323 62.16175 7 0.194553 10830 3122 80.9111 7 0.158111 10830 2222 27.9744 7 0.15567 10830 3112 87.11371 7 0.28043 10830 1232 16.11223 7 0.248567 10830 3111 99.61813 7 0.339405 10830 2223 19.61998 7 0.28086 10830 2122 70.20121 7 0.1364 10830 1331 63.81635 7 0.195803 10830 3131 73.8601 7 0.353376 10830 2221 40.69039 7 0.042032 10830 3123 72.38722 7 0.730463 10830 2213 25.70403 7 0.603318 10830 2111 89.54611 7 0.176062 10830 2211 47.24354 7 0.290535 10830 2112 76.79443 7 0.189098 10830 1333 42.67951 7 0.653917 10830 2121 83.26451 7 0.267939 10830 1332 50.82532 7 0.588542 10830 2123 62.16327 7 0.226129 10830 2212 34.12938 7 0.208798 10830 2131 63.65349 7 0.208887 10830 2231 21.28191 7 0.173132 10830 2132 50.82017 7 0.436155 10830 3332 53.79562 7 0.501553 10830 2133 42.61168 7 0.268116 10830 3333 45.07963 7 0.138849 10830 1111 97.40122 7 0.169438 10830 3331 66.31069 7 0.352753 10830 3132 61.32208 7 0.121381 10830 3323 64.84398 7 0.217084 10830 2113 68.32832 7 0.283481 10830 3322 73.25379 7 0.299733 10830 1131 71.42875 7 0.137415 10830 3321 86.22778 7 0.614224 10830 1112 84.71107 7 0.325278 10830 3313 71.27266 7 0.277575 10830 1121 90.86902 7 0.20009 10830 3311 92.20105 7 0.573016 10830 1113 76.55897 7 0.835328 10830 2313 60.43956 7 0.640373 10830 1123 69.97334 7 0.417367 10830 2333 34.90762 7 0.33017 10830 1132 58.44314 7 0.144604 10830 2332 42.92384 7 0.187344 10830 1133 50.43431 7 0.409837 10830 2323 54.44863 7 0.209853 10830 1122 78.14103 7 0.213604 10830 3312 79.69797 7 0.33297 10830 2233 -0.27198 7 0.32512 10831 3112 87.17229 7 0.243995 10830 2311 81.38113 7 0.49345 10831 2112 76.80838 7 0.395032 10830 2312 68.84609 7 0.875206 10831 2113 68.3733 7 0.18972 10830 2321 75.78298 7 0.366654 10831 2121 83.28729 7 0.479507 10830 2322 62.92185 7 0.225048 10831 2122 70.55648 7 0.261219 10830 2331 55.87395 7 0.392158 10831 2123 62.19016 7 0.247272 10830 3233 10.33219 7 0.191124 10831 2131 63.64033 7 0.249082 10830 3211 57.49945 7 0.255353 10831 2132 50.6995 7 0.296479 10830 3231 31.37577 7 0.529126 10831 2111 89.60247 7 0.176781 119 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10831 3111 99.90105 7 0.459171 10831 2213 25.21279 7 0.529769 10831 3132 60.9975 7 0.134099 10831 1221 48.62773 7 0.211897 10831 3113 78.71602 7 0.274139 10831 2211 47.11778 7 0.380225 10831 3121 93.69053 7 0.212757 10831 1212 41.52864 7 0.757695 10831 3122 80.79605 7 0.281617 10831 1213 33.12627 7 0.316219 10831 3123 72.82056 7 0.430675 10831 2221 40.82144 7 0.175991 10831 3131 73.99756 7 0.367367 10831 3333 45.4242 7 0.391126 10831 3133 52.81248 7 0.156604 10831 3331 66.01542 7 0.427685 10831 1111 97.11059 7 0.225245 10831 3322 73.26866 7 0.353002 10831 2133 42.29832 7 0.277542 10831 3332 53.68726 7 0.544591 10831 1121 91.09289 7 0.317869 10831 3323 64.7581 7 0.242555 10831 1112 84.22855 7 0.265823 10831 3313 71.14196 7 0.206847 10831 1133 50.26167 7 0.33258 10831 3311 92.27003 7 0.619633 10831 1132 58.50069 7 0.140939 10831 3312 79.43226 7 0.428234 10831 1131 71.33755 7 0.23996 10831 2333 34.90447 7 0.26302 10831 1122 78.08654 7 0.17796 10831 2323 54.63933 7 0.172822 10831 1113 76.21231 7 0.76633 10831 3321 86.1577 7 0.318422 10831 1123 69.85782 7 0.191814 10832 2113 68.40362 7 0.222663 10831 3232 18.69926 7 0.225855 10832 3132 61.08162 7 0.120964 10831 3233 10.38109 7 0.247021 10832 3131 74.02567 7 0.30336 10831 2313 60.83324 7 0.389832 10832 3123 72.80607 7 0.313241 10831 2322 62.91726 7 0.419552 10832 3122 80.81831 7 0.204578 10831 3211 57.66638 7 0.320898 10832 3121 93.85092 7 0.30128 10831 2331 55.86751 7 0.478578 10832 3113 78.96974 7 0.322489 10831 2332 43.33577 7 0.358462 10832 3112 86.99945 7 0.225746 10831 1312 76.73959 7 0.188589 10832 2112 76.66304 7 0.346459 10831 3221 51.17442 7 0.30447 10832 3133 52.89075 7 0.139325 10831 3231 31.48594 7 0.379468 10832 2121 83.18633 7 0.12304 10831 2311 81.8816 7 0.51378 10832 2122 70.42481 7 0.179653 10831 3223 30.31445 7 0.270962 10832 2123 62.18343 7 0.26331 10831 2312 69.65491 7 0.982871 10832 2131 63.64576 7 0.133321 10831 3222 38.2463 7 0.195386 10832 2132 50.7137 7 0.374002 10831 3212 44.61051 7 0.35207 10832 2133 42.29165 7 0.44963 10831 3213 36.46266 7 0.787308 10832 3111 100.1509 7 0.695568 10831 1211 54.9416 7 0.106248 10832 2111 89.46871 7 0.215786 10831 1222 35.58253 7 0.087267 10832 1122 78.18839 7 0.187005 10831 1313 68.67613 7 0.254182 10832 1111 97.12239 7 0.189915 10831 1321 83.10537 7 0.347521 10832 1112 84.52212 7 0.17442 10831 1322 70.33526 7 0.405192 10832 1121 91.30848 7 0.316979 10831 1323 62.20039 7 0.106325 10832 1123 69.98372 7 0.287424 10831 1331 63.87246 7 0.193005 10832 1131 71.37551 7 0.45588 10831 1332 50.92549 7 0.336416 10832 1132 58.40558 7 0.117805 10831 1333 42.46745 7 0.406972 10832 1133 50.09111 7 0.382325 10831 1311 89.5734 7 0.23989 10832 1113 75.89634 7 0.925779 10831 1233 7.73327 7 0.031245 10832 3213 36.68833 7 0.530683 10831 1232 16.0395 7 0.271401 10832 2312 68.97422 7 1.051159 10831 2321 75.98762 7 0.514487 10832 2313 60.66234 7 0.435669 10831 1223 27.75266 7 0.280185 10832 2321 75.67904 7 0.307874 10831 2233 0.054588 7 0.21107 10832 2322 62.66666 7 0.241763 10831 2212 34.03542 7 0.489673 10832 2331 56.15659 7 0.279158 10831 2232 8.333177 7 0.249077 10832 2332 42.65186 7 0.611532 10831 2231 21.05024 7 0.20004 10832 1313 68.44726 7 0.207236 10831 2223 19.72947 7 0.562634 10832 3233 10.53975 7 0.214409 10831 2222 27.98408 7 0.137417 10832 3221 51.2957 7 0.281915 10831 1231 29.32489 7 0.586065 10832 3211 57.60323 7 0.147392 120 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10832 3222 38.293 7 0.279266 10833 2123 62.07111 7 0.17471 10832 3212 44.80703 7 0.292171 10833 2131 63.76487 7 0.145416 10832 3223 30.13561 7 0.177837 10833 2132 50.74604 7 0.234017 10832 3231 31.11738 7 0.212423 10833 1113 76.26339 7 0.517793 10832 3232 18.78078 7 0.273744 10833 3132 61.16893 7 0.158128 10832 2311 82.03939 7 0.337814 10833 2113 68.27166 7 0.190204 10832 1213 33.62377 7 0.505646 10833 1132 58.34217 7 0.155048 10832 1322 70.27236 7 0.349613 10833 1111 97.24384 7 0.294656 10832 2233 0.065081 7 0.340894 10833 1112 84.12903 7 0.393489 10832 1212 41.74627 7 0.842067 10833 1121 91.05544 7 0.195962 10832 1221 48.67998 7 0.217601 10833 1122 78.12577 7 0.113184 10832 1222 35.78949 7 0.205518 10833 1131 71.39219 7 0.38571 10832 1223 27.23152 7 0.359362 10833 3133 52.884 7 0.156001 10832 1231 28.99931 7 0.658745 10833 1133 50.10009 7 0.472719 10832 1232 16.00144 7 0.208288 10833 1123 69.76405 7 0.241678 10832 1233 7.724811 7 0.045514 10833 3212 45.17632 7 0.600919 10832 1311 89.69058 7 0.255808 10833 2233 0.182081 7 0.191368 10832 1312 76.84608 7 0.243965 10833 2311 82.02853 7 0.477471 10832 1321 83.85058 7 0.496037 10833 2312 68.85326 7 1.144537 10832 1211 54.78781 7 0.264928 10833 2321 75.76204 7 0.15949 10832 2212 34.13162 7 0.465116 10833 2322 62.77925 7 0.221258 10832 1323 62.18919 7 0.095341 10833 2331 56.16179 7 0.19951 10832 2223 19.64114 7 0.347986 10833 3211 57.45451 7 0.162099 10832 2222 28.08657 7 0.128733 10833 3232 18.92404 7 0.241477 10832 2221 40.75434 7 0.241415 10833 3213 36.35159 7 0.557376 10832 2213 25.85545 7 0.75291 10833 3221 51.20824 7 0.364657 10832 2231 21.30451 7 0.294104 10833 3222 38.30105 7 0.400999 10832 2211 46.98158 7 0.501255 10833 3223 30.33347 7 0.358597 10832 1333 42.23854 7 0.316322 10833 3231 31.5103 7 0.248989 10832 2232 8.289994 7 0.326262 10833 3233 10.50977 7 0.172747 10832 1332 50.83936 7 0.227835 10833 2221 40.91567 7 0.149656 10832 1331 64.04377 7 0.30812 10833 1231 28.99633 7 0.697732 10832 3323 65.0787 7 0.134032 10833 2232 8.042673 7 0.311906 10832 3322 73.23531 7 0.182539 10833 1211 54.84414 7 0.194406 10832 3311 91.90997 7 0.530978 10833 1212 42.10362 7 0.863173 10832 3331 66.37592 7 0.446575 10833 1213 33.49769 7 0.571707 10832 3333 45.29214 7 0.21529 10833 1221 48.73291 7 0.107133 10832 3313 71.31025 7 0.160091 10833 2223 20.14907 7 0.228818 10832 3312 79.53823 7 0.475524 10833 1223 27.44008 7 0.390827 10832 2323 54.66835 7 0.323337 10833 1232 16.17081 7 0.454539 10832 2333 34.6482 7 0.152965 10833 1233 7.71829 7 0.071014 10832 3321 86.01485 7 0.750701 10833 1311 89.42849 7 0.351894 10832 3332 53.58568 7 0.390011 10833 1312 76.77124 7 0.216688 10833 2122 70.49502 7 0.190054 10833 1313 68.49481 7 0.254006 10833 3131 73.85473 7 0.150772 10833 1321 83.47631 7 0.343992 10833 3123 72.68801 7 0.66082 10833 1322 70.56938 7 0.226542 10833 3122 80.86396 7 0.134326 10833 2212 34.13141 7 0.428131 10833 3121 93.79636 7 0.264531 10833 1331 63.62496 7 0.353116 10833 3113 78.79256 7 0.385325 10833 1332 50.82781 7 0.311623 10833 3112 87.13972 7 0.239028 10833 1333 42.62422 7 0.675669 10833 3111 99.70079 7 1.091695 10833 2231 21.1088 7 0.121161 10833 2111 89.47625 7 0.275939 10833 1323 62.25266 7 0.166031 10833 2112 77.15942 7 0.238663 10833 2211 46.84528 7 0.350518 10833 2121 83.35256 7 0.299531 10833 2222 27.95342 7 0.102299 10833 2133 42.41301 7 0.194516 10833 1222 35.77136 7 0.142757 121 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10833 2213 25.96627 7 0.792677 10834 1331 63.6641 7 0.098315 10833 3321 85.94551 7 0.535432 10834 1222 35.62074 7 0.116474 10833 3322 73.2008 7 0.319693 10834 1321 83.31855 7 0.677845 10833 3331 66.43243 7 0.219449 10834 1221 48.53687 7 0.280819 10833 3311 92.38447 7 0.660527 10834 1312 76.62685 7 0.205917 10833 3332 53.43254 7 0.524678 10834 1311 89.65325 7 0.167469 10833 3323 65.03515 7 0.262094 10834 1213 33.47576 7 0.882419 10833 3312 79.82158 7 0.250803 10834 1232 15.89561 7 0.243786 10833 2313 60.76944 7 0.472018 10834 1211 54.75498 7 0.232955 10833 2323 54.30314 7 0.250908 10834 1223 27.49401 7 0.505131 10833 3333 45.44838 7 0.294586 10834 2221 40.86389 7 0.151335 10833 2332 43.25981 7 0.246818 10834 3211 57.69107 7 0.344703 10833 2333 34.8617 7 0.227242 10834 2133 42.44379 7 0.130485 10833 3313 71.29952 7 0.14146 10834 2211 47.27825 7 0.224434 10834 2132 50.47305 7 0.290073 10834 2212 34.25534 7 0.491531 10834 2131 63.75272 7 0.284692 10834 2213 26.34719 7 0.845452 10834 2123 62.33169 7 0.30957 10834 1233 7.725874 7 0.039263 10834 2122 70.2387 7 0.35904 10834 3312 79.20107 7 0.439196 10834 2121 83.67291 7 0.488455 10834 3332 53.62585 7 0.478848 10834 2113 68.27001 7 0.190733 10834 3313 71.18868 7 0.295293 10834 2112 76.4726 7 0.339104 10834 3321 85.95688 7 0.417492 10834 2111 89.56639 7 0.222745 10834 3322 73.36442 7 0.238717 10834 1132 58.42755 7 0.168134 10834 3323 64.88466 7 0.242456 10834 1133 49.85813 7 0.23641 10834 3331 66.40392 7 0.305236 10834 1123 69.88329 7 0.368646 10834 3333 45.28919 7 0.212368 10834 1122 78.18725 7 0.169759 10834 3311 92.34938 7 0.588727 10834 1121 91.06208 7 0.228141 10834 2321 75.69878 7 0.479295 10834 1131 71.269 7 0.275228 10834 1313 68.46498 7 0.340697 10834 1112 84.34074 7 0.375115 10834 2311 81.79343 7 0.644058 10834 1113 76.07886 7 0.342506 10834 2312 68.93652 7 0.812724 10834 1111 97.19336 7 0.252561 10834 2313 60.58484 7 0.28656 10834 3121 93.62454 7 0.287439 10834 2322 62.57677 7 0.178891 10834 3132 61.28121 7 0.169288 10834 2323 54.49472 7 0.29715 10834 3233 10.3918 7 0.2704 10834 1323 62.25024 7 0.097494 10834 2232 8.09371 7 0.176603 10834 2332 43.09377 7 0.242704 10834 2231 21.08036 7 0.237552 10834 2333 34.90591 7 0.384819 10834 2223 19.62078 7 0.201046 10834 1322 70.60084 7 0.409037 10834 2233 0.14219 7 0.25399 10834 1332 50.752 7 0.355351 10834 2222 27.86845 7 0.216297 10834 2331 56.17172 7 0.353474 10834 3111 100.1123 7 0.718791 10834 1333 42.46641 7 0.399426 10834 3113 78.81596 7 0.321198 10835 2131 63.72573 7 0.129522 10834 3122 80.83402 7 0.264625 10835 1111 97.19196 7 0.309009 10834 3123 72.69082 7 0.494675 10835 2122 70.46117 7 0.462251 10834 3131 73.85627 7 0.294792 10835 2121 83.51424 7 0.384618 10834 3133 52.91208 7 0.197442 10835 2112 76.79217 7 0.316283 10834 3212 44.29851 7 0.348922 10835 2132 50.71526 7 0.436743 10834 3213 36.39924 7 0.744555 10835 2111 89.49915 7 0.211209 10834 3221 51.32401 7 0.331497 10835 1122 78.15615 7 0.132919 10834 3222 38.46995 7 0.264423 10835 2113 68.43544 7 0.220246 10834 3223 30.23315 7 0.233302 10835 1133 49.91835 7 0.646261 10834 3231 31.39133 7 0.421542 10835 1132 58.45909 7 0.090398 10834 3232 18.34314 7 0.173939 10835 1123 69.92408 7 0.31911 10834 3112 86.97728 7 0.231497 10835 1121 90.83871 7 0.29813 10834 1231 28.90641 7 0.709323 10835 1113 76.34321 7 0.793536 10834 1212 42.29036 7 0.80796 10835 1112 84.49454 7 0.350642 122 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 10835 1131 71.20642 7 0.353861 10835 1323 62.24549 7 0.104837 10835 2233 0.029805 7 0.315164 10835 2332 43.44515 7 0.298474 10835 3113 78.57753 7 0.181977 10835 2333 34.75154 7 0.274637 10835 3111 99.79826 7 0.593261 10835 2312 69.13031 7 0.72645 10835 3222 38.59023 7 0.247788 10835 2313 60.94599 7 0.678969 10835 2232 8.211241 7 0.268513 10835 2321 75.83721 7 0.480822 10835 2231 21.16143 7 0.260698 10835 2322 62.83502 7 0.25058 10835 2223 19.54255 7 0.378813 10835 1313 68.50339 7 0.179811 10835 3112 87.15465 7 0.146613 10835 2323 54.59581 7 0.301123 10835 3212 44.67289 7 0.409404 10835 2331 56.20111 7 0.402349 10835 3233 10.48446 7 0.273738 10835 1333 42.52949 7 0.32294 10835 1311 89.92522 7 0.435716 10835 2222 28.02773 7 0.148226 10835 3232 18.55074 7 0.289976 10835 3231 31.89949 7 0.34886 10835 3213 36.41008 7 0.616274 10835 3221 51.46614 7 0.193247 10835 3121 93.8832 7 0.221719 10835 3211 57.66559 7 0.278383 10835 3133 52.78071 7 0.108418 10835 3132 61.17678 7 0.195756 10835 3131 73.72182 7 0.33517 10835 3123 72.24507 7 0.254928 10835 3122 80.89652 7 0.244182 10835 3223 30.24506 7 0.345469 10835 1321 83.1899 7 0.366674 10835 1211 54.85667 7 0.19959 10835 1212 41.45912 7 0.56364 10835 1213 33.59543 7 1.299008 10835 1221 48.53083 7 0.311036 10835 1222 35.69783 7 0.21123 10835 1223 27.52718 7 0.398143 10835 1231 28.92846 7 0.613124 10835 1232 16.06236 7 0.185072 10835 1233 7.728351 7 0.053098 10835 1312 76.98637 7 0.143635 10835 2221 40.92842 7 0.105723 10835 1331 63.84923 7 0.38561 10835 2123 62.4187 7 0.328936 10835 2133 42.59919 7 0.295678 10835 2213 26.11802 7 0.442408 10835 2212 34.40497 7 0.271632 10835 2211 46.96041 7 0.362813 10835 1322 70.39979 7 0.284561 10835 3332 53.53088 7 0.426573 10835 3331 66.3836 7 0.413966 10835 3323 64.93469 7 0.33279 10835 3322 73.08386 7 0.303778 10835 3321 86.14012 7 0.522964 10835 3313 71.17992 7 0.303282 10835 3312 79.5723 7 0.666377 10835 3311 92.88952 7 0.464714 10835 3333 45.27531 7 0.204391 10835 2311 81.92651 7 0.459555 10835 1332 50.82785 7 0.580373 C.1.2 Adaptive One At A Time The Adaptive One At A Time players may not list their configurations in order. This is because they store their flights in local memory, and put them into the database with a Java Iterator which does not guarantee order. However, it should be possible to recreate the path along with they optimized response. 123 PlayerID Configuration Mu Frequency Sigma PlayerID Configuration Mu Frequency Sigma 26590 2212 34.19665 7 0.519765 26602 1222 30.90775 7 2.714101 26590 2222 28.10085 7 0.226637 26602 1312 60.45085 7 2.409952 26590 2232 8.464541 7 0.212538 26602 1321 84.62143 7 8.885429 26590 2211 46.92297 7 0.404632 26602 1322 67.19152 7 3.88737 26590 2213 25.84148 7 0.569421 26602 1323 51.90408 7 0.946307 26590 3111 100.0967 7 0.655665 26602 1332 28.42597 7 4.235913 26590 3211 57.5896 7 0.161021 26602 3321 79.27343 7 2.914844 26590 3311 92.44719 7 0.418913 26605 1221 45.98021 7 0.941654 26590 1211 54.80807 7 0.174184 26605 1311 83.06292 7 1.916766 26593 1312 61.104 7 0.201656 26605 1321 85.32993 7 5.175007 26593 3131 55.43405 7 0.246968 26605 1322 67.47119 7 2.562254 26593 3121 90.89119 7 0.199314 26605 1323 53.32884 7 0.636825 26593 3113 70.71259 7 0.253346 26605 1331 44.06813 7 2.275396 26593 3112 88.8697 7 0.281097 26605 2311 65.37017 7 3.71897 26593 3111 78.69616 7 0.375592 26605 3311 34.85719 7 2.800078 26593 2112 84.66041 7 0.338924 26605 1121 59.22335 7 1.518183 26593 1112 67.99615 7 0.242706 26608 1311 90.02925 7 12.96746 26593 1212 28.8151 7 0.672718 26608 1322 71.31015 7 18.92207 26596 3223 34.49221 7 0.163734 26608 1313 73.01483 7 8.06266 26596 3212 25.34481 7 0.296593 26608 1312 80.58723 7 8.076769 26596 3321 73.66032 7 0.127792 26608 1122 79.03247 7 6.862447 26596 3232 27.74172 7 0.156968 26608 1332 70.88285 7 24.04639 26596 1121 59.47275 7 0.17884 26608 2122 69.59454 7 17.44098 26596 3121 84.14812 7 0.106297 26608 3122 76.65061 7 13.89588 26596 2121 66.26704 7 0.218082 26608 1222 37.63453 7 12.85921 26596 3222 28.95659 7 0.110123 26611 2312 41.717 7 41.13201 26596 3221 44.3419 7 0.112176 26611 3322 63.03962 7 11.73747 26599 2131 63.68947 7 2.454898 26611 2332 19.68985 7 20.21053 26599 3121 93.55434 7 3.469804 26611 2323 49.97717 7 10.24318 26599 3331 66.60938 7 4.232919 26611 1322 66.08525 7 11.54335 26599 3131 72.55791 7 2.831388 26611 1222 25.78537 7 11.46969 26599 3111 103.3269 7 4.660926 26611 1122 62.20121 7 7.462624 26599 2133 41.71143 7 1.825812 26611 2322 34.39164 7 13.89453 26599 2132 52.06717 7 2.577272 26611 2321 63.98955 7 20.08557 26599 1131 69.82673 7 1.910776 26614 3121 86.8485 7 9.007366 26599 3231 27.80773 7 6.182551 26614 3131 48.49215 7 7.218236 26602 2321 58.78123 7 1.63597 26614 3122 64.8474 7 8.678333 26602 1122 55.81358 7 2.09833 26614 3111 52.71932 7 10.68481 26614 2122 28.73908 7 9.368227 26614 1322 68.64058 7 15.54882 26614 1222 47.18818 7 4.566915 26614 1122 43.21876 7 5.933681 26614 3123 68.0132 7 6.043964 26617 3322 73.29396 7 0.386114 26617 1322 70.35325 7 0.19215 26617 2322 62.83205 7 0.369178 26617 3111 100.2234 7 0.684279 26617 3112 86.94305 7 0.211954 26617 3113 78.82093 7 0.457979 26617 3122 80.8174 7 0.207597 26617 3132 61.25669 7 0.201653 26617 3222 38.54138 7 0.326939 124 C.1.3 Orthogonal Array Each of the following orthogonal array agents will execute an orthogonal array of order two. The experimental design is created swapping parameters and levels in the orthogonal array in section 5.5.5 125 PlayerID Configuration Mu Frequency Sigma 10946 1221 48.60667 7 0.219172 10946 1333 42.30141 7 0.341348 10946 3123 72.31511 7 0.370889 10946 3232 18.77591 7 0.288191 10946 3311 92.3299 7 0.702735 10946 1112 84.55836 7 0.557416 10946 2131 63.69005 7 0.195152 10946 2213 25.86446 7 1.035251 10946 2322 62.48865 7 0.138834 10947 1211 54.63951 7 0.163444 10947 3331 66.47012 7 0.193775 10947 3223 30.18015 7 0.210722 10947 3112 87.15736 7 0.304219 10947 2313 60.9307 7 0.593196 10947 2232 8.270327 7 0.307079 10947 1322 70.56702 7 0.275226 10947 1133 50.12974 7 0.32591 10947 2121 83.32415 7 0.366318 10948 2322 62.95481 7 0.276815 10948 1333 42.64687 7 0.460137 10948 3311 92.52408 7 0.416048 10948 3232 18.78873 7 0.486353 10948 2213 26.07395 7 0.9538 10948 1112 84.37748 7 0.330648 10948 1221 48.53873 7 0.336891 10948 3123 72.73286 7 0.610879 10948 2131 63.7177 7 0.083133 10949 2212 33.94948 7 0.518402 10949 3311 92.70265 7 0.146939 10949 3223 30.0827 7 0.229913 10949 2333 34.91587 7 0.212555 10949 2121 83.24281 7 0.274399 10949 1322 70.40214 7 0.196062 10949 1231 29.20707 7 0.447071 10949 1113 76.23515 7 0.654299 10949 3132 61.13782 7 0.111627 10950 3211 57.49271 7 0.232589 10950 1223 27.44253 7 0.374749 10950 1312 76.68832 7 0.21716 10950 2113 68.31138 7 0.147938 10950 2232 8.243985 7 0.324666 10950 3122 80.89701 7 0.049578 10950 3333 45.09368 7 0.399337 10950 2321 75.41604 7 0.301748 10950 1131 71.29367 7 0.207039 10951 1212 41.70221 7 0.911915 10951 1331 63.54337 7 0.30642 10951 2132 50.79972 7 0.309857 10951 2221 40.77909 7 0.157527 10951 2313 60.86375 7 0.759014 10951 3111 100.0469 7 1.012175 10951 3233 10.43045 7 0.34522 10951 3322 73.23975 7 0.399345 10951 1123 69.75229 7 0.312474 PlayerID Configuration Mu Frequency Sigma 10952 1131 71.12201 7 0.301926 10952 3332 53.33533 7 0.603991 10952 3221 51.11489 7 0.288656 10952 3113 78.96729 7 0.33525 10952 1212 41.93867 7 0.472821 10952 1323 62.31452 7 0.234483 10952 2233 -0.00602 7 0.304873 10952 2311 82.18706 7 0.390574 10952 2122 70.41951 7 0.124714 10953 1331 63.79408 7 0.266561 10953 3211 57.37337 7 0.200819 10953 3133 52.98102 7 0.112555 10953 3322 73.25708 7 0.20161 10953 2313 60.72863 7 0.485394 10953 1223 27.61456 7 0.271328 10953 1112 84.29851 7 0.445592 10953 2232 8.271014 7 0.299866 10953 2121 83.4801 7 0.353013 10954 3121 93.79588 7 0.301625 10954 3213 36.15413 7 0.613488 10954 2323 54.48621 7 0.346442 10954 2112 76.64522 7 0.242826 10954 1311 89.58607 7 0.281955 10954 1222 35.84078 7 0.397913 10954 1133 50.406 7 0.421711 10954 3332 53.4614 7 0.544836 10954 2231 21.19276 7 0.18202 10955 3311 92.52495 7 0.455538 10955 1121 91.15273 7 0.282324 10955 1213 33.22885 7 0.647565 10955 1332 50.68105 7 0.348434 10955 2112 76.55203 7 0.312418 10955 2231 21.30874 7 0.322138 10955 2323 54.63653 7 0.325799 10955 3133 52.93525 7 0.117066 10955 3222 38.24516 7 0.198092 126 C.2 Performance of the Strategies Listed below are the results of listing 5.5. Each instance of the parabolic form of each of the strategies has some payoff value (payoff equals 1 at the target [80] and decays parabolically to 0 at the minimum of means [0]) and wealth. The wealth is determined by the payoff count, the game’s total payoff, and the number of experiments the strategy ran. C.2.1 Interpreting this Table This table displays the effects of all the varied parameters for the ParabolicDistanceGoal, target = 80 games. The ’number of payoffs’ for these agents is not integral. For every flight of distance x, they get some payoff p of p(x) = 1 − (80 − x)2 /(80 − 0)2 (C.1) which 1 at a scaled length of 80 and less everywhere else. The coefficient was chosen such that p for the mean of any configuration would be zero or greater, making negative results relatively rare. As each player gets three tries t at the target, every attempt has a possible payout P/3. Every agent a has an experimental cost ca determined by its strategy, so ends up with a wealth wa of wa = −ca + 3 pX p(xt ) 3 t=1 (C.2) The first and second statistical moments are displayed in the table. While changing values of payoff should not effect the performance of a strategy, they will balance the experimental cost in different ways and have a direct effect on the wealth. 127 128 Frequency 400 400 400 200 400 400 400 200 1400 400 400 200 400 400 400 200 400 400 400 200 1400 400 400 200 400 400 400 200 400 400 400 200 1400 400 400 200 400 Payoff 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 567 2 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 2 TrialCount Noise 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 Nonlinearity AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic Strategy 476.3025 -5.555 -6.215 -5.095 -2.865 -6.31 -5.9175 -5.1525 -3.3175 -7.15 -7.1725 -9.375 -7.0525 19.015 19.765 19.5575 22.39714286 20.51 19.48 19.4325 22.4625 17.915 17.835 15.8525 17.4925 36.89 36.625 36.96 39.16 36.765 35.7975 37.43 39.1225 34.89 34.885 32.155 34.375 µW ealth 96.84720437 7.811976382 8.982693082 10.43819788 6.668769698 9.067188098 9.309709649 10.05928644 6.960365921 9.817713583 9.729991971 13.45843137 11.14561545 8.121870166 8.155045984 9.798045405 5.857058536 7.794863693 8.565313771 10.73896847 5.492594446 8.875684481 9.30874723 13.7965845 11.11710141 8.521613697 7.664814088 8.679769582 6.095089358 6.758681454 9.118744088 8.459024766 6.276742288 9.280511839 9.348089377 14.47483938 12.41871068 σW ealth 4.918082077 -0.71108766 -0.69188605 -0.48811108 -0.42961448 -0.69591586 -0.6356267 -0.51221327 -0.47662724 -0.72827547 -0.73715374 -0.69658935 -0.63276003 2.34120955 2.423652796 1.996061377 3.82395749 2.63121984 2.274289129 1.80953134 4.089597406 2.018435878 1.915939874 1.149016266 1.573476696 4.328992291 4.778328552 4.258177553 6.42484428 5.439670481 3.925705081 4.424859961 6.232930747 3.759490921 3.731778612 2.221440885 2.768000711 ST NW ealth Table C.1: Wealth and Payoff for players; ParabolicDistanceGoal, Target = 80 µP ayof f 2.62363458 2.67504289 2.64891529 2.71657199 2.82329195 2.64153377 2.66506928 2.71209731 2.80004863 2.60110788 2.60007 2.49472974 2.61011981 2.63489766 2.68011732 2.68510038 2.82789197 2.71830791 2.66246956 2.6797797 2.83179582 2.58340687 2.57726554 2.48707621 2.57090649 2.6752263 2.67333297 2.70467926 2.81770641 2.67150466 2.63042875 2.72715265 2.81213163 2.5837261 2.58430361 2.45217077 2.56238783 σP ayof f 0.513582552 0.381883033 0.438102414 0.513337095 0.333833055 0.441422933 0.453480766 0.493288983 0.345320874 0.476233939 0.470253021 0.648333149 0.539500023 0.424653347 0.424970903 0.512800285 0.313359291 0.402366796 0.441940656 0.553156383 0.291555118 0.453913598 0.476115982 0.687866154 0.561994708 0.442855816 0.401874046 0.456065235 0.323378895 0.353648475 0.47026118 0.437796115 0.331597686 0.471567179 0.471499567 0.72405687 0.622323784 5.10849633 7.00487493 6.04633803 5.29198458 8.45719711 5.98413352 5.87691801 5.49798881 8.10854148 5.46182803 5.52908729 3.8479133 4.83803466 6.2048202 6.30659018 5.23615227 9.02443951 6.75579581 6.02449566 4.84452459 9.7127289 5.69140665 5.41310445 3.61563975 4.57460979 6.04085168 6.65216627 5.93046576 8.71332811 7.55412464 5.59354856 6.22927558 8.48055265 5.47902019 5.48103071 3.38671017 4.11745123 ST NP ayof f 129 Frequency 400 400 200 400 400 400 200 400 400 400 200 400 400 400 200 400 400 400 200 400 400 400 200 400 400 400 200 400 400 400 200 400 400 400 200 400 400 200 Payoff 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 3000 3000 3000 2 2 2 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 TrialCount Noise 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 Nonlinearity Strategy AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic 2561.58 2418.865 2559.41 442.89 441.1175 438.6625 464.245 443.125 440.005 454.365 465.3275 421.19 431.5825 415.425 430.245 470.72 469.2175 470.285 497.35 472.475 462.46 471.345 494.4125 463.155 437.205 434.46 461.015 487.405 495.7975 500.1275 514.3975 478.305 484.17 485.575 508.535 472.805 464.6675 445.09 µW ealth σW ealth 470.3871635 660.6695935 550.2143191 73.45902191 84.38426212 103.8806459 67.44456965 77.39579688 83.83128876 78.41719693 59.63275311 94.02677225 74.93365862 113.3271564 100.7834559 71.78838068 69.51510767 86.47877644 59.21952803 74.00161738 82.11119534 89.84089255 58.89072375 75.46867546 111.1768095 121.4493244 84.01145621 69.14087774 61.91454994 81.84101199 62.00281843 84.47929909 84.03366052 99.6516401 68.26517249 85.07771139 104.7097748 160.7063841 5.445684319 3.661232519 4.65166011 6.029075646 5.227485421 4.222754837 6.88335625 5.725440112 5.248696597 5.794200989 7.803220139 4.479468878 5.759527934 3.665714496 4.269004236 6.557049979 6.749863674 5.43815511 8.398412087 6.384657751 5.632118715 5.246441644 8.395422378 6.13704954 3.932519757 3.577294498 5.487525402 7.049447677 8.007770395 6.110964269 8.29635673 5.661801236 5.761619773 4.872724619 7.449406212 5.557330966 4.437670702 2.769585057 ST NW ealth µP ayof f 2.58116897 2.43836926 2.57893256 2.68580218 2.67544655 2.66373053 2.79882785 2.68625359 2.66970579 2.74682221 2.80426845 2.56989102 2.62498628 2.53929052 2.6179536 2.68974609 2.68105537 2.68784416 2.83104306 2.69892549 2.64576266 2.69344843 2.81578157 2.64916238 2.51199368 2.49755792 2.63820822 2.68270269 2.72672671 2.75103341 2.82630733 2.63423506 2.66510351 2.67354263 2.79492152 2.6051544 2.56198407 2.45817791 σP ayof f 0.470455696 0.660845665 0.550421557 0.389650264 0.447096808 0.552075146 0.357612671 0.410181468 0.444562418 0.415875787 0.31634861 0.499166077 0.397334669 0.601130391 0.534643741 0.380819992 0.367935611 0.458682133 0.314322893 0.392086307 0.43522616 0.476431043 0.312237236 0.399936592 0.589655176 0.643879133 0.445439239 0.366464239 0.327967395 0.434225717 0.328687323 0.447654639 0.445310172 0.528344749 0.361751849 0.451092481 0.555092198 0.852125158 5.48652932 3.68977114 4.6853771 6.89285347 5.98404304 4.82494196 7.8264225 6.54893943 6.005244 6.60491015 8.86448797 5.14836872 6.60648689 4.22419255 4.89663191 7.06303805 7.28675151 5.85992776 9.00679881 6.88349847 6.07905247 5.65338567 9.01808384 6.62395598 4.26010621 3.87892354 5.92271175 7.32050335 8.31401764 6.33549166 8.59877193 5.88452534 5.98482514 5.06022371 7.7260739 5.77521133 4.61542079 2.88476157 ST NP ayof f 130 Frequency 400 400 400 200 400 400 400 200 400 400 200 400 400 400 200 400 400 400 200 400 400 200 400 400 400 200 400 400 400 200 720 240 240 240 200 280 280 240 Payoff 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 63 63 63 63 63 63 63 63 2 2 2 2 2 2 2 2 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 TrialCount 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 Noise 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 Nonlinearity Strategy FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic AdaptiveOneAtATimeParabolic -99.0041667 -99.0071429 -99.025 -100.145 -100.975 -102.479167 -102.254167 -99.9708333 2650.26 2598.2025 2653.095 2757.0525 2609.965 2599.025 2696.5775 2764.4525 2490.155 2462.325 2527.755 2652.065 2623.9525 2669.935 2794.57 2654.175 2637 2660.745 2777.765 2540.4 2409.265 2552.415 2567.775 2686.09 2669.51 2795.5625 2669.675 2686.68 2674.9775 2791.755 µW ealth σW ealth 0.064415103 0.084213044 0.23010091 1.908919852 3.350279839 5.490862043 6.908779388 3.726106752 334.5310335 455.0747647 498.6711301 330.2867311 328.9568418 431.3485243 418.2387464 279.8372701 539.6477101 701.7661572 557.3101515 355.8488595 445.0630408 520.5337028 339.8408379 341.5968594 437.1715224 532.1510828 312.4615013 484.8968859 675.4838153 625.5024682 532.6394225 413.4654543 500.6959156 340.4722398 346.9817133 347.0739152 541.4255138 330.2423958 -1536.97132 -1175.67468 -430.354666 -52.4616054 -30.1392734 -18.6635843 -14.8006125 -26.8298361 7.922314328 5.709397008 5.320330053 8.347451594 7.934065106 6.025348074 6.44745979 9.878785978 4.614408536 3.508754269 4.53563423 7.452784882 5.895687262 5.129225995 8.223172993 7.769904572 6.031957401 4.99998043 8.889943205 5.239052 3.566724983 4.080583418 4.820850451 6.49652824 5.331599314 8.21083828 7.693993365 7.740944745 4.94061959 8.453654151 ST NW ealth µP ayof f 2.98695531 2.99016041 2.98981249 2.90281027 2.86122315 2.7816386 2.79377746 2.93378972 2.71471676 2.66272178 2.71785457 2.82163547 2.67452996 2.66353655 2.76118564 2.82904827 2.55456867 2.52678239 2.59233892 2.68960538 2.66148654 2.70777213 2.83217426 2.69172692 2.67455135 2.69832986 2.81537941 2.57787841 2.44677236 2.58991377 2.58723575 2.70559755 2.68929207 2.81518326 2.68923209 2.70625057 2.6945721 2.81137996 σP ayof f 0.011121619 0.012768112 0.017801708 0.097871026 0.167165642 0.268928109 0.335776302 0.189179608 0.334482666 0.455176999 0.49889514 0.330330726 0.329015936 0.43146534 0.41835193 0.279928888 0.539852356 0.701990504 0.55744793 0.355842454 0.445149597 0.520747595 0.339902644 0.341685987 0.437285868 0.532319937 0.312536326 0.485104035 0.675688557 0.625655902 0.532869294 0.41354223 0.500885657 0.340551499 0.347068543 0.347082488 0.541611586 0.330272338 268.571993 234.189713 167.950877 29.6595469 17.1160959 10.3434283 8.32035331 15.5079596 8.11616578 5.84986013 5.44774714 8.5418499 8.12887665 6.17323409 6.60015036 10.1063105 4.73197652 3.5994538 4.65036961 7.55841625 5.9788587 5.19977847 8.33231017 7.87777967 6.1162538 5.06900018 9.00816698 5.31407332 3.62115406 4.13951785 4.85529149 6.54249397 5.36907382 8.26654197 7.74841782 7.79713947 4.97510055 8.51230827 ST NP ayof f 131 240 240 240 200 280 280 240 200 1280 280 280 200 360 240 240 240 200 240 240 63 63 63 63 63 63 63 63 63 63 63 567 567 567 567 567 567 567 240 63 63 280 63 720 280 63 63 200 63 200 240 63 63 240 63 280 240 63 63 720 63 280 200 63 63 280 63 200 280 63 1280 1280 63 63 200 63 63 Frequency Payoff 2 2 2 2 2 2 2 7 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 TrialCount 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 Noise 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 Nonlinearity Strategy FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic 400.7625 401.1458333 385.74 378.0833333 363.5791667 360.725 390.5611111 -504 -504 -504 -504 -504 -504 -504 -504 -504.38 -504.820833 -505.941667 -505.4 -504.433333 -261 -261 -261 -261 -261 -261 -261.003571 -261 -261.585 -262.170833 -263.404167 -263.05 -261.501389 -99 -99 -99 -99 -99 µW ealth 2.621595014 2.051316481 24.37462615 30.30761272 46.71895118 57.65334083 31.97910149 0 0 0 0 0 0 0 0 0.797245257 1.086461829 2.590996698 2.658006772 1.749920633 0 0 0 0 0 0 0.059654618 0 1.316349118 2.250921879 3.529515921 3.025860759 1.443375005 0 0 0 0 0 σW ealth 152.8697216 195.5553114 15.82547349 12.47486355 7.782263032 6.256792665 12.21301078 -632.653497 -464.646635 -195.269128 -190.14248 -288.260692 -4375.24503 -198.720078 -116.472649 -74.6289782 -86.9339408 -181.173561 ST NW ealth µP ayof f 2.98941121 2.99128497 2.90700248 2.86619062 2.78881039 2.7737703 2.93527203 2.99780184 2.99937156 2.99987537 2.99974065 2.99546546 2.99148535 2.99625768 2.9961965 2.94515923 2.91767876 2.85911194 2.89248767 2.96276033 2.99785707 2.99942802 2.99985849 2.9997255 2.99495592 2.98945554 2.99387245 2.99417776 2.93616751 2.90205125 2.83656857 2.85654692 2.95854764 2.99782265 2.9993801 2.99981298 2.9997066 2.99335242 σP ayof f 0.01512462 0.011681085 0.12985362 0.161133095 0.248411023 0.306284565 0.171119237 0.000461392 0.000388759 6.79129E-05 0.000160268 0.003461784 0.006301019 0.003835565 0.004622597 0.046171356 0.060800979 0.131175611 0.134582083 0.094207114 0.000425703 0.000328385 0.000105357 0.000168727 0.004794955 0.008616128 0.010119251 0.007466256 0.071732424 0.114617036 0.175462549 0.154341755 0.084507306 0.000411179 0.00034316 0.000214824 0.000190741 0.007111573 197.651983 256.079382 22.3867651 17.7877215 11.2265968 9.05618703 17.1533726 6497.30058 7715.25317 44172.3916 18717.0165 865.295389 474.762128 781.177701 648.16305 63.7875838 47.9873648 21.7960633 21.492368 31.4494332 7042.14079 9133.86338 28473.2622 17778.5606 624.605682 346.960438 295.85911 401.028007 40.932222 25.3195454 16.1662337 18.5079334 35.0093713 7290.79049 8740.47214 13964.0365 15726.5577 420.912853 ST NP ayof f 132 Frequency 240 200 240 240 240 200 150 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 40 40 40 40 40 40 40 Payoff 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 972 972 972 972 972 972 972 4 4 4 4 4 4 4 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 TrialCount 100 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 Noise 0 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 Nonlinearity Strategy FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic 646.05 644.3 645.125 644.2 601.35 581.75 601.525 0 0 0 0 -0.105 -0.7125 -0.12083333 -0.17083333 -10.86 -13.3375 -22.7625 -15.7416667 241 241.0041667 241.0166667 241.0125 240.905 240.1541667 240.4291667 240.6125 228.83 226.4958333 211.7666667 217.1791667 240.9666667 402 402.0083333 402.025 402.0208333 401.56 400.9125 µW ealth σW ealth 0.217944947 1.913112647 1.552216158 3.05122926 100.6224503 92.20486701 42.94647104 0 0 0 0 0.306553421 1.315235245 0.454128439 0.532274339 12.15279392 12.36555405 26.08651172 16.54891529 0 0.064415103 0.128019096 0.11110243 0.382066748 1.401332451 1.867257161 1.030700611 18.23406428 13.44029945 27.14950378 26.32911062 0.354338194 0 0.090905934 0.15612495 0.142826138 1.013114011 1.817922922 2964.280697 336.781005 415.6154391 211.1280225 5.976300501 6.309319875 14.00638948 -0.34251779 -0.54172818 -0.26607744 -0.32094978 -0.89362167 -1.07860108 -0.87257738 -0.95122045 3741.423264 1882.661842 2169.281982 630.5311865 171.3755836 128.7606076 233.4455781 12.54958832 16.85199308 7.800019786 8.248632846 680.0471158 4422.24522 2575.020841 2814.756741 396.3621029 220.5332774 ST NW ealth µP ayof f 2.99970702 2.99129848 2.99529526 2.99154167 2.85814922 2.7980277 2.85853348 2.9978092 2.99938172 2.99987096 2.99975062 2.99570884 2.99111857 2.99636827 2.99609186 2.93618743 2.92250757 2.87252098 2.90972136 2.99777706 2.99935997 2.99986292 2.99969524 2.9955078 2.99003297 2.99386938 2.99438336 2.92906309 2.91615787 2.83782743 2.86681673 2.99881662 2.99777125 2.9994044 2.99982006 2.99970038 2.99324905 2.98888085 σP ayof f 0.000162138 0.006424679 0.005200894 0.009760888 0.31101822 0.284975352 0.132706428 0.00048549 0.000352148 7.17655E-05 0.000174764 0.002822918 0.007566641 0.003833669 0.003806988 0.064471457 0.065946765 0.138099081 0.087989119 0.000456597 0.000355399 0.000102284 0.000185598 0.003429336 0.008127034 0.010740175 0.006702987 0.097237962 0.071618641 0.144702298 0.140811179 0.002679595 0.000449691 0.00035501 0.000208166 0.000186167 0.006592456 0.010221742 18500.9899 465.595027 575.919341 306.48252 9.18965206 9.81849017 21.5402789 6174.80855 8517.39224 41801.0207 17164.5746 1061.20984 395.303359 781.592838 786.998049 45.5424395 44.3161625 20.8004351 33.0691043 6565.47157 8439.41782 29328.8406 16162.3683 873.494966 367.911952 278.754234 446.723749 30.1226292 40.717861 19.6114883 20.3592979 1119.13071 6666.29479 8448.7835 14410.7115 16112.9553 454.041593 292.404249 ST NP ayof f 133 Frequency 40 40 240 240 40 240 240 240 200 240 240 240 200 240 240 40 240 240 240 200 240 240 240 200 240 113 40 240 240 240 200 240 240 240 200 360 240 240 Payoff 972 972 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 63 63 63 2 2 2 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 4 4 TrialCount 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 Noise 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.527051027 0.1 Nonlinearity Strategy Guess Guess Guess FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic FullFactorialParabolic 38.42916667 45.6875 46.56944444 2429.745 2429.995833 2430.016667 2430.041667 2426.44 2422.554167 2426.629167 2427.620833 2327.5 2314.345133 2336.8 2672.81 2673.004167 2673.033333 2673.016667 2668.795 2664.5875 2668.404167 2668.691667 2591.875 2497 2526.975 2834.62 2834.991667 2835.033333 2835.0125 2828.05 2823.945833 2825.616667 2825.375 2721.15 2607.595833 2612.683333 646 646.025 µW ealth σW ealth 22.79499176 17.10282171 14.4399392 0.583073752 0.193604336 0.128019096 0.199826313 3.526528038 6.61037563 6.338373817 2.990607403 106.3628225 108.1315791 99.47035572 0.483632092 0.193604336 0.179505494 0.128019096 5.626986316 7.657284794 8.758280804 6.55400111 67.85948257 243.729341 205.450905 0.682348884 0.223451461 0.2013841 0.11110243 10.0407918 11.41751137 12.92393344 20.26822739 128.0764518 223.2707045 261.8619065 0 0.15612495 1.685860082 2.671342821 3.225044358 4167.131504 12551.35027 18981.67341 12160.76914 688.0535116 366.477535 382.8472786 811.7484197 21.88264608 21.40304573 23.49242629 5526.535659 13806.53049 14891.09486 20879.82773 474.2849636 347.9807232 304.6721413 407.1851105 38.19473568 10.24497088 12.29965378 4154.209181 12687.28184 14077.74169 25517.1061 281.6560741 247.3346197 218.6344181 139.3992156 21.2462944 11.67907738 9.97733259 4137.871622 ST NW ealth µP ayof f 1.89749394 2.25212464 2.29395421 2.99785554 2.99936418 2.99987793 2.99973081 2.99529356 2.99119745 2.99552725 2.99648947 2.89597913 2.88287942 2.90542937 2.99790871 2.99939922 2.99986832 2.99971501 2.99462283 2.99025118 2.99434387 2.9945582 2.91751326 2.8225315 2.85250233 2.99781583 2.9993491 2.99982244 2.99968639 2.99186885 2.98761067 2.98938995 2.98926033 2.88475414 2.77114531 2.77628428 2.99934774 2.99987317 σP ayof f 1.115038951 0.833117085 0.701119432 0.000425526 0.000385781 7.01218E-05 0.000168855 0.003573669 0.006666972 0.006475695 0.003157087 0.106378724 0.108240328 0.099554579 0.000399672 0.000354815 9.09848E-05 0.000192909 0.005729981 0.007737358 0.008904183 0.006713692 0.067937253 0.243879937 0.205442216 0.000501175 0.000367287 0.000181018 0.000182243 0.010162416 0.011491916 0.013025445 0.020320879 0.128270786 0.223330482 0.26189141 0.000388675 7.63476E-05 1.70172884 2.70325105 3.27184515 7045.06407 7774.77897 42780.9622 17765.148 838.156511 448.659099 462.580005 949.131033 27.2232925 26.6340604 29.1842866 7500.92898 8453.42338 32971.0796 15549.8872 522.623557 386.469298 336.285061 446.03744 42.9442338 11.5734469 13.8846942 5981.57111 8166.23784 16571.9191 16459.8173 294.405283 259.974984 229.503856 147.102907 22.4895646 12.4082718 10.6008986 7716.86004 39292.2936 ST NP ayof f 134 Frequency 240 200 280 240 240 200 280 280 280 200 360 240 240 240 200 280 240 240 200 280 280 280 200 360 240 240 240 200 280 240 240 200 280 280 280 200 240 240 Payoff 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 567 567 2 2 7 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 TrialCount Noise 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 Nonlinearity Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Strategy µW ealth 382.7 416.8416667 47.785 46.82857143 44.01428571 50.625 48.215 46.275 44.8625 49.74285714 45.53 45.54583333 37.01666667 46.38333333 45.76944444 48.265 48.25357143 43.57857143 50.12857143 48.435 47.12916667 45.5875 49.78928571 47.315 46.33333333 40.34583333 46.92916667 45.96388889 48.24 46.34642857 43.50357143 50.66428571 47.78 47.27916667 43.64583333 50.58928571 47.115 45.61666667 σW ealth 166.2212431 164.0933167 11.85701375 12.74751676 16.98612279 13.01312867 11.12873645 12.5289415 16.2332558 14.66671228 13.97208288 14.3232759 24.15194379 16.48645067 16.75193719 11.20601513 11.86787326 17.41574323 12.94236833 10.77616699 12.66704975 16.7886572 14.31913902 12.51861714 14.20993393 20.9127449 15.50024137 14.83881271 12.25407687 12.28666473 17.18034554 13.61570035 9.677892333 11.85648483 17.4022144 12.71722227 11.87315354 13.06023439 2.302353134 2.540272053 4.030104124 3.673544604 2.591190837 3.890301961 4.332477475 3.693448486 2.763616896 3.391547895 3.258640848 3.179847518 1.532657867 2.813421413 2.732188159 4.307061829 4.06589878 2.502251604 3.873214713 4.494640817 3.720611161 2.715374998 3.477114485 3.779570817 3.260629752 1.929246186 3.027641024 3.097544917 3.936649044 3.772091906 2.532170923 3.721019443 4.937025373 3.987620895 2.508062039 3.978013801 3.968195967 3.492790812 ST NW ealth µP ayof f 2.03250238 2.21320573 2.35013686 2.30377368 2.17306373 2.49565799 2.37021575 2.27592251 2.21729089 2.45195837 2.24193965 2.24286252 1.83431429 2.28290499 2.25710058 2.37549285 2.37610148 2.15330607 2.47065018 2.38029294 2.31841767 2.2496404 2.45865442 2.32917633 2.28360857 1.99175083 2.31402349 2.26703408 2.37002414 2.28581009 2.15167494 2.4981975 2.35092705 2.32451061 2.15557371 2.49265261 2.31728381 2.24728687 σP ayof f 0.881358075 0.869768142 0.577466979 0.61928836 0.818887049 0.63359647 0.537893373 0.604192843 0.782255137 0.716560431 0.682566722 0.697120891 1.174766962 0.802555339 0.810170434 0.544438708 0.572422513 0.841650767 0.631602577 0.520652487 0.610791042 0.813492963 0.694028384 0.607400157 0.690212104 1.021593409 0.75287173 0.717409454 0.593978023 0.596835282 0.828260498 0.657857815 0.464044981 0.5767566 0.841439836 0.620793636 0.575563347 0.630977916 2.30610286 2.54459278 4.06973377 3.72003387 2.65367945 3.9388761 4.40647881 3.76688095 2.83448556 3.42184451 3.28457215 3.2173222 1.56142822 2.84454526 2.78595773 4.3631961 4.15095742 2.55843178 3.91171644 4.57174987 3.7957624 2.76540856 3.54258482 3.83466533 3.3085606 1.94965121 3.07359594 3.16002817 3.99008726 3.82988432 2.59782392 3.79747333 5.066162 4.03031471 2.56176807 4.01526765 4.02611428 3.56159354 ST NP ayof f 135 Frequency 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 40 40 40 40 Payoff 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 972 972 972 972 4 4 4 4 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 TrialCount Noise 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 Nonlinearity Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Strategy 799.975 795.825 639.175 752.95 433.27 449.525 404.7208333 459.9666667 447.885 441.8958333 417.4458333 470.4625 430.135 430.1375 386.7583333 426.4291667 448.375 436.9083333 400.0958333 477.7458333 435.685 436.8541667 410.4833333 454.9041667 457.425 430.8083333 367.3083333 423.0208333 433.545 450.7875 387.7166667 477.925 449.335 433.025 401.5333333 472.55 441.74 439.775 µW ealth σW ealth 212.4249853 126.1823854 299.7336224 299.3309999 109.6447313 98.43106408 158.7883063 130.2939263 92.79688451 109.266692 151.959168 123.9702865 117.7542219 117.4000011 176.3891293 180.9552292 106.6775252 115.6304816 164.6007644 112.7842242 113.430092 116.2384026 154.8926232 128.7118033 94.25085875 119.2577947 198.5822582 149.3201384 118.5368212 111.0240995 160.0399785 121.8218072 105.0206778 119.9441163 151.7370221 129.5710455 104.6869734 110.6413849 3.765917643 6.306942112 2.132476814 2.515442772 3.951580663 4.566901762 2.548807545 3.530223394 4.826509019 4.044195218 2.747092123 3.794961788 3.652820197 3.663862827 2.192642681 2.356545144 4.203087757 3.778487536 2.430704588 4.235927823 3.841000146 3.758260239 2.650115447 3.534284773 4.85327143 3.612412374 1.849653321 2.832979114 3.657471119 4.060267112 2.422623836 3.92314817 4.278538372 3.610222938 2.646244981 3.647033935 4.219627196 3.974778521 ST NW ealth µP ayof f 2.47358433 2.46135546 1.97696835 2.32798815 2.30018634 2.38707999 2.1493462 2.44193504 2.37800668 2.34632279 2.21702891 2.49758372 2.28358663 2.28411077 2.05347117 2.26418438 2.38041615 2.31980475 2.12488479 2.53594271 2.31309765 2.31900678 2.1795769 2.41533925 2.4281511 2.28731203 1.95074074 2.24608073 2.30182524 2.39350444 2.05923576 2.53698133 2.38562563 2.29914637 2.13252861 2.50870798 2.34511553 2.33482762 σP ayof f 0.655484714 0.389698448 0.92604765 0.926017508 0.580502521 0.521490144 0.841022398 0.689849819 0.491484917 0.578412317 0.804691745 0.656905188 0.624110695 0.622199137 0.935239552 0.959531129 0.564673383 0.612451682 0.870990383 0.596921921 0.600894773 0.615723626 0.820398182 0.682013454 0.499494391 0.632395999 1.052912555 0.791185088 0.628383798 0.588080296 0.847823759 0.64535175 0.55610983 0.635439132 0.803551288 0.686124256 0.555224695 0.586429953 3.77367203 6.31605149 2.13484516 2.51397854 3.96240543 4.5774211 2.55563491 3.53980674 4.83841233 4.05648829 2.75512819 3.80204596 3.65894489 3.67102851 2.19566331 2.35967788 4.21556288 3.78773513 2.43961911 4.24836585 3.84942215 3.76631119 2.65673053 3.54148329 4.86121795 3.61689833 1.85270917 2.83888153 3.66308815 4.07003 2.42884885 3.93116054 4.28984618 3.61820078 2.65387989 3.65634644 4.22372339 3.98142628 ST NP ayof f 136 Frequency 40 40 40 40 40 240 40 40 240 240 240 200 240 240 240 200 240 40 40 240 240 240 200 240 240 240 200 240 40 40 240 240 240 200 240 240 240 200 Payoff 972 972 972 972 972 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 4 4 4 4 4 TrialCount 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 10 10 Noise 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.527051027 0.1 0 0.527051027 0.1 Nonlinearity Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Guess Strategy 2320.05 2301.108333 2128.633333 2550.925 2371.14 2302.683333 2183.65 2464.358333 2269.7 2198.85 2255.5125 2401.81 2345.6 2085.495833 2497.645833 2369.815 2304.745833 2202.0875 2479.883333 2037 2205.525 2257.741667 2359.385 2342.679167 2164.058333 2524.304167 2384.93 2270.016667 2236.1125 2457.591667 2228.775 1987.9 2286.929167 738.9 715.3 822.65 720.2 684.425 µW ealth σW ealth 500.763245 556.346599 847.8570431 596.5506218 545.4923376 577.7423299 827.7871068 699.5246516 613.5645932 846.4155466 839.1423398 470.7621309 568.6841889 900.8633637 694.9932521 526.0697014 558.6667294 834.9165007 673.3411986 983.9329245 887.8823117 836.3098458 608.5787433 549.234529 822.5631769 647.9497113 503.583732 573.8096735 726.3098511 710.2148442 619.6429814 785.1950968 784.7505755 227.2265169 263.6784216 171.0847086 189.8867031 267.10802 4.63302773 4.136105689 2.510604058 4.276124954 4.34678883 3.98565799 2.637936714 3.5229042 3.699203026 2.597837444 2.687878317 5.101960932 4.124609134 2.314996832 3.593769905 4.504754776 4.125439572 2.637494286 3.682952029 2.070263073 2.484028537 2.699647359 3.876877111 4.265353037 2.630871882 3.895833462 4.735915497 3.956044611 3.078730788 3.460349621 3.596869596 2.531727475 2.914211519 3.251821178 2.712774127 4.80843675 3.792787954 2.562352863 ST NW ealth µP ayof f 2.32156232 2.30258761 2.13016689 2.55243868 2.37274957 2.30423128 2.18517854 2.46593037 2.27109475 2.20014493 2.2569835 2.40333583 2.34706946 2.08691571 2.49911949 2.371294 2.30626088 2.20353459 2.48140903 2.03825747 2.20682684 2.25919054 2.36091181 2.34411844 2.16557588 2.52576966 2.38639513 2.27148496 2.2376022 2.45914163 2.23020199 1.98932941 2.28832789 2.28504773 2.21237778 2.54359762 2.22734138 2.11748343 σP ayof f 0.500810923 0.556289354 0.848006405 0.59662397 0.545443796 0.577788418 0.827844162 0.6996072 0.613778982 0.846861028 0.839425546 0.470794795 0.568769019 0.901012161 0.69516821 0.526146899 0.558703553 0.834952483 0.6734074 0.984786127 0.888276537 0.836547984 0.608649111 0.549309782 0.822617006 0.648013942 0.50354122 0.573841085 0.726396872 0.710327415 0.619880221 0.785226932 0.785003242 0.701305774 0.813765811 0.527960292 0.586197493 0.824781971 4.63560641 4.1391905 2.51197028 4.27813633 4.35012661 3.98801915 2.63960132 3.52473555 3.70018331 2.59799997 2.68872386 5.10484792 4.12657754 2.31619039 3.59498529 4.50690482 4.12787939 2.63911376 3.6848556 2.06974633 2.48439168 2.70061083 3.87893742 4.26738884 2.63254451 3.89770883 4.73922499 3.95838677 3.08041276 3.46198328 3.59779505 2.53344521 2.91505533 3.25827594 2.71869099 4.81778206 3.79964331 2.56732506 ST NP ayof f 137 Frequency 720 240 240 240 200 280 280 240 200 280 280 280 200 720 240 240 240 200 280 265 240 200 280 280 280 200 720 240 240 240 200 280 240 240 200 280 280 280 Payoff 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 63 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 2 2 TrialCount 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 0 Noise Nonlinearity 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 1.7977E+308 Strategy OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic -4.425 -1.20714286 0 -4.485 -4.2625 -0.94583333 0 -6.735 -5.66666667 -5.48333333 -4 -3.09444444 21.11 21.87857143 23.975 25.02142857 21.285 21.09583333 23.64150943 24.99642857 18.24 19.03333333 19.17916667 21.3 22.39861111 37.28 38.95357143 40.79642857 42.025 37.045 38.18333333 40.63571429 42.00357143 34.465 36.17083333 35.44166667 37.43333333 38.9375 µW ealth σW ealth 6.863054557 2.45093693 0 6.090137519 6.034091515 2.338353974 0 7.764971024 6.736509152 6.644525733 5.800862005 5.493882681 6.117017247 4.984214879 2.007864004 0.144808107 5.269134179 5.039343473 3.017523285 0.179248716 7.200166665 7.172788083 7.73770418 5.625240736 4.753741014 7.655821315 5.379324316 2.415276107 0.15612495 6.405698635 4.96820446 2.992158118 0.059654618 8.846964169 7.860236381 8.394736281 7.049507469 5.8266852 -0.64475664 -0.49252302 -0.73643657 -0.70640294 -0.40448681 -0.86735675 -0.84118741 -0.82524074 -0.68955269 -0.56325273 3.45102836 4.389572272 11.94054974 172.7902474 4.0395631 4.186226529 7.834739685 139.4510889 2.533274693 2.653547423 2.478663725 3.786504614 4.711786159 4.869497141 7.24135024 16.89099994 269.1754265 5.783131882 7.685539845 13.58073761 704.1126524 3.895686627 4.6017488 4.221891609 5.310063646 6.682616044 ST NW ealth µP ayof f 2.72476534 2.90535461 2.99972906 2.72628953 2.73774554 2.91215617 2.99118522 2.6187764 2.67355167 2.68534266 2.75656001 2.80991822 2.74619109 2.78271336 2.9106239 2.99969989 2.7593282 2.74831234 2.89142532 2.98856826 2.60079363 2.64383883 2.65272222 2.76868227 2.82833479 2.70346049 2.78782393 2.90195038 2.9997119 2.68973127 2.75062018 2.89081239 2.98764558 2.55911036 2.64524257 2.61192184 2.71921195 2.80428014 σP ayof f 0.33516027 0.142533747 0.000158308 0.297140859 0.300474688 0.123663392 0.010574527 0.375623614 0.329297536 0.32170976 0.282417322 0.278061917 0.315134118 0.264108481 0.125039143 0.000179199 0.279035441 0.263945132 0.167193724 0.01526028 0.373644596 0.367398099 0.391074314 0.283159695 0.255705578 0.394590944 0.281899759 0.147114163 0.000216917 0.329396326 0.263679426 0.164641009 0.011716893 0.45220936 0.399726776 0.424446465 0.35339794 0.306900324 8.12973847 20.3836261 18948.7232 9.17507454 9.11140156 23.5490563 282.867043 6.97180982 8.11895438 8.34709727 9.76059113 10.1053688 8.71435661 10.5362514 23.2777018 16739.4483 9.88880907 10.4124381 17.2938628 195.839674 6.96060819 7.19611461 6.78316658 9.77781202 11.0609038 6.85129887 9.88941581 19.7258395 13828.87 8.16563836 10.431683 17.5582766 254.986159 5.65912737 6.61762667 6.15371325 7.69447596 9.1374297 ST NP ayof f 138 Frequency 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 240 240 240 200 40 Payoff 63 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 567 972 4 7 7 7 7 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 2 7 TrialCount 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 1.9371775 100 Noise 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 Nonlinearity Strategy OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic 869.2 458.975 457.4958333 483.3375 501.0375 448.52 458.2125 484.2833333 500.0375 427.89 435.3458333 437.2125 454.45 477.48 486.1333333 512.0791667 528.0333333 474.565 482.8333333 507.8708333 526.8541667 456.395 457.5041667 460.25 480.7916667 498.115 501.4208333 529.5458333 546.0333333 493.005 504.8833333 527.0958333 544.5166667 470.32 474.7416667 480.8 496.9791667 -4.58 µW ealth σW ealth 78.22889492 47.24356438 57.32240675 29.21255884 0.210777331 74.70735974 47.35786465 25.14845367 1.99024967 77.87597768 63.30739477 59.81276907 59.4484861 67.57336458 55.05102683 27.51904612 0.2013841 56.83023645 47.54547531 32.10445165 1.991540268 81.19894688 76.03491511 77.35128204 54.32278464 56.27398844 54.55969574 24.93136377 0.179505494 69.55591258 50.74998905 27.63789637 2.173642616 76.42805506 79.18890872 60.33208102 59.84719765 6.934233916 11.110984 9.715079843 7.981099525 16.54553792 2377.093861 6.003692294 9.675531265 19.25698254 251.243604 5.494505658 6.876697974 7.309684985 7.644433522 7.066097759 8.8305952 18.60817284 2622.020976 8.35057233 10.15518996 15.81932745 264.5460778 5.620700976 6.017027388 5.950127624 8.850644714 8.851602913 9.190315791 21.24014708 3041.875335 7.087894928 9.948442212 19.07148888 250.5088291 6.153761202 5.995052519 7.969226188 8.304134299 -0.66049113 ST NW ealth µP ayof f 2.799298 2.76978542 2.7620232 2.90146497 2.99973312 2.71425235 2.76607019 2.9047195 2.99072942 2.60594054 2.64532021 2.65532089 2.74635682 2.72498132 2.77075034 2.91035586 2.99973331 2.70965572 2.75349318 2.88640034 2.9893442 2.61345425 2.61968071 2.63427613 2.74294249 2.73916609 2.7562923 2.90786966 2.9997274 2.71190174 2.774912 2.89324477 2.98735944 2.5918442 2.61504961 2.64759324 2.73343084 2.72434992 σP ayof f 0.241311487 0.25023487 0.303591313 0.156321741 0.000174393 0.396987836 0.250950509 0.133766585 0.011781835 0.412426107 0.335529444 0.317066867 0.315275491 0.358302179 0.291062317 0.147058003 0.000145311 0.301172332 0.251805328 0.170400721 0.011855443 0.430694305 0.402975627 0.409862701 0.287785424 0.297995618 0.288791143 0.133742207 0.000159737 0.368973817 0.268880826 0.147097372 0.013198658 0.404613015 0.419544812 0.319561187 0.317388424 0.339388206 11.6003512 11.0687428 9.09783345 18.5608537 17200.9974 6.83711718 11.0223733 21.7148364 253.84242 6.31856348 7.88401812 8.37464007 8.71097468 7.60526027 9.51944025 19.7905302 20643.4883 8.9970274 10.9350076 16.9388974 252.149506 6.06800281 6.50084158 6.42721605 9.53120715 9.19196768 9.54424111 21.7423484 18779.2159 7.34984873 10.3202301 19.6689086 226.33812 6.4057361 6.23306386 8.28509015 8.61225752 8.02723804 ST NP ayof f 139 Frequency 40 40 40 40 40 40 40 40 240 43 40 240 240 240 200 240 240 240 200 240 40 40 240 240 240 200 240 240 240 200 240 40 40 240 240 240 200 240 Payoff 972 972 972 972 972 972 972 972 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 3000 7 7 7 7 7 7 7 7 4 4 4 4 4 4 4 4 4 4 4 2 2 2 2 2 2 2 2 2 2 2 4 4 4 4 4 4 4 4 TrialCount Noise 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 100 10 10 10 10 1.9371775 1.9371775 1.9371775 100 100 100 10 10 10 1.9371775 1.9371775 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.8 0.527051027 0.1 0 0.527051027 0.1 0 0.527051027 0.1 0 0.527051027 0.1 0 0.527051027 0.1 Nonlinearity Strategy OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic 2934.033333 2629.35 2674.754167 2831.720833 2926.0625 2555.975 2649.75 2694.025 2646.64 2693.479167 2871.095833 2961.020833 2673.575 2705.954167 2865.708333 2951.154167 2572.625 2597.675 2724.858333 2714.3 2787.55 2898.341667 2979.029167 2706.195 2769.979167 2881.245833 2966.945833 2634.775 2498.093023 2736.020833 835.7 914 933.025 873.65 914.375 930.025 811.25 799.175 µW ealth σW ealth 0.179505494 344.6716807 284.4157498 144.4046729 11.40980253 376.0750249 201.650905 329.5359607 446.197199 299.3575892 139.8720367 0.142826138 303.3748908 274.8882689 132.5452373 13.90283182 349.5953152 428.5359604 246.4828864 322.2342781 248.8143836 134.4859036 0.191440083 336.6273265 245.2297978 157.8836926 13.05793115 401.7544329 524.2303689 259.3416095 96.149935 39.34399573 0.15612495 69.00998116 24.89245619 4.251396829 128.2592589 134.014717 16345.08936 7.628564071 9.404381328 19.60962049 256.4516338 6.796449726 13.1402832 8.175207933 5.931547768 8.997530925 20.52658917 20731.64538 8.812776143 9.843832831 21.62060585 212.270004 7.358865776 6.061743331 11.05495953 8.423374496 11.20333141 21.55126738 15561.15685 8.039142359 11.29544285 18.24916675 227.2140815 6.558172814 4.765258122 10.54987219 8.691633541 23.23099073 5976.142828 12.65976291 36.73301635 218.7575137 6.325079427 5.963337593 ST NW ealth µP ayof f 2.99972945 2.6938528 2.73920285 2.89633402 2.99089427 2.62050613 2.71434214 2.75854174 2.68408486 2.7309589 2.9089622 2.99972407 2.71114221 2.7434739 2.90325738 2.98896765 2.61003307 2.63506937 2.76240818 2.73380361 2.80702245 2.91821121 2.99970718 2.72572058 2.78953604 2.9008534 2.98671176 2.65434198 2.51778574 2.75561643 2.69490795 2.93821309 2.99973677 2.81265888 2.93851068 2.9880099 2.61985055 2.58209804 σP ayof f 0.000166752 0.344781948 0.284406153 0.144446961 0.011521721 0.376109276 0.201680101 0.329702081 0.446463284 0.299394804 0.140160275 0.000159453 0.303384618 0.274897456 0.132609341 0.014017915 0.349641017 0.428657568 0.246517859 0.322399835 0.248897283 0.134774811 0.000175984 0.336663922 0.245281622 0.157920098 0.013175291 0.401873062 0.524310921 0.259394762 0.296885806 0.122220451 0.000128274 0.213230472 0.076812827 0.013892916 0.395638129 0.414151133 17989.1602 7.8132072 9.63130658 20.0511938 259.587463 6.96740626 13.4586512 8.36677079 6.01188263 9.1215975 20.7545411 18812.578 8.93632059 9.97999015 21.8933098 213.224841 7.46489382 6.14725963 11.2057122 8.47954408 11.2778349 21.6524971 17045.3151 8.09626575 11.3727886 18.369121 226.690391 6.60492637 4.80208524 10.6232539 9.07725427 24.0402737 23385.3391 13.1906986 38.2554686 215.074351 6.62183536 6.23467577 ST NP ayof f 140 Frequency 240 240 200 Payoff 3000 3000 3000 7 7 7 TrialCount 100 100 100 Noise 0.8 0.527051027 0.1 Nonlinearity Strategy OrthogonalArrayParabolic OrthogonalArrayParabolic OrthogonalArrayParabolic 2646.14 2660.370833 2839.325 µW ealth σW ealth 393.8919146 329.3002809 136.7570633 6.717934291 8.078859897 20.7618161 ST NW ealth µP ayof f 2.71059652 2.7248916 2.9041494 σP ayof f 0.394082534 0.329348037 0.137023061 6.87824575 8.27359298 21.1946032 ST NP ayof f Appendix D Five Layer Process Support Model To attempt to understand the connection between the individual or local adoption of design of experiments and quality engineering practices, division-level DFSS implementations and a top-level corporate push to improve quality and lower costs due to detailed design prototype testing and warrantee returns, we propose in section 4.3 a five-layer process support model. We take this opportunity to lay out some of our thinking about it in an appendix. The model is nascent, and our investigations in Chapter ch4 did look at the organizations from enough perspectives to apply it fully. However, it does reconcile their apparent autonomy with their alignment with a corporate goal, and explains why some practitioners have faltered in their development of DoE practices while others have flourished. D.1 Layers The layers as introduced in section 4.3 were Environment The environment is the responsibility of the top management in an organization. While a process can survive in a hostile environment, it would typically need a strong champion, and would evanesce should that champion leave. 141 An environment is created when the senior management decides to devote significant resources to a goal, and effectively communicates that decision to those who create frameworks for polity. Framework A framework is the bridge between environments and policies. The framework happens in accordance with the set environment, but typically closer to the policy making level. A framework specifies and articulates corporate goals vaguely rendered by the environment, and describes necessary policies by type. Policy A policy is a bundled set of rules that are intended to direct instruction toward meeting the goals set up in a framework. A policy might indicate who does what when. Instruction Instruction enacts policy. Only when an order is given is it clear that a policy must be followed. Without a reinforcing environment, instructions are unlikely to follow policy. Action On the operational level, actions, of course, must complete the intent. Actions without instructions can be helpful to an enterprise, but will not generally be inspired by appropriate context, and will suffer from all the ill effects of poor coordination. The levels can crowd together somewhat. If a software lead sets a coding standard, and follows it when he writes software elements, then he embodies three levels. Without a framework, however, that policy will be moribund even with an auspicious environment, and it may be difficult enrolling other team members into abiding by it. This is explicitly a layered architecture. There are conflicting drives and diverse needs at the corporate level, warring frameworks trying to fill some number of those needs (for example, moving the corporate headquarters while implementing Business Process Restructuring,) attempts to fit favored policies into as many current frameworks as possible, instructions that balance the demands of various policies with 142 available resources, and actions driven at least in part by the instructions of every person with instructing authority over a particular actor. D.2 Effecting Process Change Clearly, this model can be applied in process change as a top down approach, or even with a leading from within method. A new process can start at any layer. However, the central point of this appendix is that all five layers must be whole and active for a process to persist. We therefore recommend that process change identify what the five layers would be for a particular process change, and start campaigns on all of those levels. The five-layer model can be compared to a theoretical treatment of Policy Deployment or Hoshin Management[33, Chapter 14]. The toolset built around Policy Deployment may be useful in adopting the five-layer model, but there are some differences, notably that processes in the five layer model can exist comfortably with other processes. The layered architecture of the five-layer model allows influences of one layer on another to balance each other somewhat, whereas Policy Deployment argues for monolithic action. D.3 Describing Current Processes The true power, however, of the model is in describing current processes. If you know what your process is, you can tend to its health on all five levels, and coordinate among the five levels to retire it. Particularly, obsolesced frameworks create a danger of undercutting internal authority. This model may help a process owner detect such things, and preserve corporate resources for current intents. 143 D.4 Accountability This concept of the five-layer model came out of a discussion of the accountability toplevel decision makers have in the moral failings of ground-level workers. We proposed that each act had a certain moral weight, that would accrue to the actor only insofar as he had the option of not doing it. The remainder of the act would flow up the chain of command. The use of the model for partial and accretive accountability will have to wait until reliable methods of measuring moral weights and willfulness have been established. D.5 Fads Over the course of the practice of scientific managment, innovations have grown into movements, evolved into fads, and then been discredited in the public perception. The five-layer process support model suggests that when a corporate commitment is not made to the underlying principles of a movement, it is being introduced into a hostile environment. Currently, DFSS shows some signs of turning into a fad. The five-layer model suggests that this happens when DFSS is adopted as a policy framework with a corporate need of saving money or staying abreast of industry – needs for which other frameworks will yield better metrics – not of improving quality through early analysis. D.6 An Example: Balancing the Registers In one over the counter tax preparation service, a new process was put in place whereby the registers would be fully closed out every night with particular documentation. 144 D.6.1 A process All of the layers worked together to create this new process. Once a corporate need for tighter accounting was annunciated, a framework was created to try to satisfy it. This daily-register-closing policy came out of that framework. The instructions were left up to the branch managers. The business of tax preparation requires mathematically competent seasonal specialists to work odd hours for little pay. Some of these specialists balked. One specialist refused to close out the registers, but the instruction was flexible enough that she could perform all the activities required to close out the register without calling it by its proper name. Another maintained that she was, in fact, closing out the registers, but pushed the action back on her manager. As the actions were largely competent and successful, this was, for the 2004 tax season, a successful process implementation. However, if the second tax preparers failure to comply with the new policy had been more common, the manager would have been overwhelmed in trying to perform the functions left by her direct reports, and the process would have failed. D.6.2 Conflict with other corporate needs Another corporate need is to reduce costs. The environment around this can engender a framework of cost-reduction, and particular cost-cutting policies such as the reduction of overtime. This policy conflicts with the proper register closing policy, as the latter will sometimes push the operator into overtime. In this kind of situation, both environments have to be maintained and balanced, and management has to enforce both needs at the operational level. 145 Bibliography [1] J. D. Musa and A. F. Ackerman Quantifying software validation: when to stop testing? Software, IEEE, vol. 6, no. 3, May, pp. 19-27, 1989. [2] E. F. Crawley, System Architecture, MIT course ESD.34J, Lecture 1, Fall, 2003. [3] W. J. Hopp, Factory Physics : Foundations of Manufacturing Management, Boston, Massachusetts: Irwin/McGraw-Hill, 2001. [4] M. S. Phadke, Quality Engineering using Robust Design, Engelwood Cliffs, New Jersey: Prentice Hall, 1989. [5] J. A. Morrow, An investigation of the penetration of Taguchi Quality Engineering in five U.S. companies Thesis of Master of Science in the Management of Technology, MIT, June 1988. [6] N.W. Hirschi and D.D. Frey, Cognition and complexity: An experiment on the effect of coupling in parameter design, Research in Engineering Design vol. 13, pp 123-131, 2002. [7] A. Peters, An investigation of the implementation and application of Taguchi’s Quality Engineering method in US corporations Thesis for the Diplomarbeit, TU Berlin, October 1994. [8] G. Taguchi, Quality Engineering in Production Systems, New York, New York: McGraw-Hill, 1988. [9] C. F. J. Wu, Experiments: Planning, Analysis, and Parameter Design Optimization, New York, New York: J. Wiley, 2000. 146 [10] QFD Institute, Design for Six Sigma and QFD Workshop by the QFD Institute, [Online document], 2004, Available HTTP: http://www.qfdi.org/tut dfss.htm [11] Six Sigma Qualtec, Six Sigma Products and Services web site [Online document], 2004, Available HTTP: http://www.sixsigmaqualtec.com/six-sigma-products/ [12] D. D. Frey, F. Engelhardt, E. M. Greitzer, A role for ”one-factor-at-a-time” experimentation in parameter design, Research in Engineering Design, Volume 14, Number 2, pp. 65 - 74, 2003. [13] D. D. Frey and Rajesh Jugulum, How one-factor-at-a-time experimentation can lead to greater improvement than orthogonal arrays, Proceedings of DETC ’03, ASME 2003 Design Engineering Technical Conference and Computers and Information in Engineering Conference, DETC2003/DTM-48646, pp. 505 - 513. [14] D. D. Frey Personal Conversation, June, 2003. [15] S. H. Thomke, The economics of experimentation in the design of new products and processes, Thesis for Doctor of Philosophy in Electrical Engineering and Management, MIT, February 1995. [16] W. Middendorf, Design of Devices and Systems, New York: Marcel Decker, 1986. [17] S. H. Thomke and A. Nimgade, IDEO Product Development, Harvard Business School Case No. 600-143 (2000), 21 p. [18] S. H. Thomke, E. von Hippel and R. Franke, Modes of experimentation: an innovation process–and competitive–variable, Research Policy, Volume 27, Issue 3, July 1998, Pages 315-332. [19] N. A. Barricelli, Symbiogenetic evolution processes realized by artificial methods, Methodos, no. 9, pp. 35-36, 1957. 147 [20] G. E. P. Box, Evolutionary operation: A method for increasing industrial productivity, Applied Statistics, vol. 6, no. 2, pp. 81–101, 1957. [21] A. Marczyk, Genetic Algorithms and Evolutionary Computation, [Online document], 2004, Available HTTP: http://www.talkorigins.org/faqs/genalg/genalg.html#history [22] D. E. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning, Reading, Massachusetts: Addison-Wesley, 1989. [23] K. A. De Jong, An analysis of the behavior of a class of genetic adaptive systems, Thesis for Doctor of Philosophy in Computer Science, University of Michigan, 1975. [24] I. Faybish, Applying the genetic algorithm to the game of Othello, Thesis for Master of Science degree in Computer Science, Vrije Universiteit Brussel, 1999. Available HTTP: http://www.geocities.com/Athens/Delphi/2402/othello.html [25] D. Beasley, D. R. Bull and R. R. Martin, An Overview of Genetic Algorithms: Part 1, Fundamentals, University Computing, vol. 15, no. 2, pp. 58 - 69. [26] M. G. Neubauer, W. Watkins and J. Zeitlin, Notes on D-optimal designs, Linear Algebra and its Applications, vol. 280, no. 2-3, Sep., pp. 109-127, 1998. [27] D. Clausing, personal conversation, March, 2004. [28] K. T. Ulrich, and Steven D. Eppinger Product Design and Development, Boston, Massachusetts: Irwin/McGraw-Hill, 2000. [29] J. T. Grewen, An experimental study of the effects of interactions upon parameter design with Taguchi Methods, Thesis of Bachelor of Science in Mechanical Engineering, MIT, 1999. 148 [30] H. Chipman, M. Hamada and C. F. J. Wu, A Bayesian variable selection approach for analyzing designed experiments with complex aliasing, Technometrics,vol. 39, Feb., pp. 372-381, 1997. [31] S. H. Thomke, Experimentation matters: unlocking the potential of new technologies for innovation, Boston, Massachusetts: Harvard Business School Press, 2003. [32] S. M. Ahlman, Complex dynamic system architecture evaluation through a hierarchical synthesis of tools and methods, Thesis of Master of Science in Engineering and Management, MIT, 2003. [33] S. Shiba, D. Walden, A. Graham, A new American TQM: four practical revolutions in management, Norwalk, Connecticut: Productivity Press, 1993 [34] Epstein, Joshua M., Growing Artificial Societies: Social Science from the Bottom Up, Washington, D.C.: Brookings Institution Press, 1996. 149