CEC99 SPECIAL SESSIONS Session: Theory and Foundation of Evolutionary Computation – 5 Sessions Organizer: David Fogel Natural Selection Inc., La Jolla, CA 92037 fogel@ece.ucsd.edu 1. "Characterizations of Trajectory Structure of Fitness Landscapes Based on Pairwise Transition Probabilities of Solutions" Mark Jelasity, Boglarka Toth, and Tamas Vinko Characterization of trajectory structure of fitness landscapes is a major problem of EC theory. In this paper a hardness measure of fitness landscapes will be introduced which is based on statistical properties of trajectories. These properties are approximated with the help of a heuristic based on the transition probabilities between the elements of the search space. This makes it possible to compute the measure for some well known function: a ridge function, a long path function, a fully deceptive function, and a combinatorial problem: the subset sum problem. Using the same transition probabilities the expected number of evaluations needed to reach the global optimum from any point in the space is approximated and examined for the above problems. 2. "Aggregating Models of Evolutionary Algorithms" William M. Spears The standard difficulty in modeling evolutionary algorithms (EAs) is in finding the correct level of granularity. If the granularity is too coarse the model has poor predictive ability. If the granularity is too find, the model is generally intractable. A solution to this dilemma is to find methods for aggregating finegrained models of EAs -- the aggregation removes unnecessary detail from models, producing simpler models with good predictive accuracy. This paper summarizes two useful aggregation techniques for EA models that introduce little or no error in accuracy. 3. "Some Observations on the Interaction of Recombination and Self-Adaptation in Evolution Strategies" Lothar Grunz and Hans-Georg Beyer The performance of the multirecombinant (mu/mu,lambda)-ES with sigma-self-adaptation (sSA) is investigated on the sphere model. The investigation includes the computation of the maximal performance of an ES using recombination can reach when the mutation strength is optimally adjusted during the whole evolution. The comparison between the strategies with and without sSA shows that selfadaptation is not always able to drive the ES in its optimal working regime, although it still guarantees linear convergence order. The static and dynamic aspects of self-adaptation are discussed and it is shown that the learning parameter has a sensible influence on the progress rate. 4. "Self-Adaptation and Global Convergence: A Counter-Example" Guenter Rudolph The self-adaptation of the mutation distribution is a distinguishing feature of evolutionary algorithms that optimize over continuous variables. It is widely recognized that self-adaptation accelerates the search for optima and enhances the ability to locate optima accurately, but it is generally unclear whether these optima are global or not. Here, it is proven that the probability of convergence to the global optimum is less than one in general -- even if the objective function is continuous. 5. "Emphasizing Extinction in Evolutionary Programming" Garrison Greenwood, Gary B. Fogel, and Manuel Ciobanu Evolutionary programming uses tournament selection to choose parents for reproduction. Tournaments naturally emphasize survival. However, a natural opposite of survival is extinction, and a study of the fossil record shows extinction plays a key role in the evolutionary process. This paper presents a new evolutionary algorithm that emphasizes extinction to conduct search operations over a fitness landscape. 6. "A Performance Analysis of Evolutionary Pattern Search with Generalized Mutation Steps" William Hart and Keith Hunter Evolutionary pattern search algorithms (EPSAs) are a class of evolutionary algorithms (EAs) that have convergence guarantees on a broad class of nonconvex continuous problems. In previous work we have analyzed the empirical performance of EPSAs. This paper revisits that analysis and extends it to a more general model of mutation. We experimentally evaluate how the choice of the set of mutation offsets affects optimization performance for EPSAs. Additionally, we compare EPSAs to self-adaptive Eas with respect to robustness and rate of optimization. All experiments employ a suite of test functions representing a range of modality and number of multiple minima. 7. "On the Real Arity of Multiparent Recombination" I. G. Sprinkhuizen-Kuyper, C. A. Schippers, and A .E. Eiben In the last years several papers reported experimental results for multi-parent recombination operators, looking at the effects of using more parents. Tacitly, these studies assume that the number of parents (the arity of the given recombination operator) tells how many old individuals contribute to a new one by passing their genetic information to it. In this article we point out that this assumption is not valid for a number of well-known recombination operators and distinguish parents and donors, the latter ones being those parents that really deliver information to the offspring. We perform a mainly theoretical analysis on the number of donors. Experimental results are provided to support theoretical estimates and predictions. 8. "Learning Landscapes: Regression on Discrete Spaces" William G. Macready and Bennett S. Levitan It is often useful to be able to reconstruct landscapes from a set of data points sampled from the landscape. Neural networks and other supervised learning techniques can accomplish this task but typically do not exploit the metric structure of discrete input spaces. In this paper we propose a new method based on Gaussian processes which reconstructs landscapes over discrete spaces from data sampled from the landscape and the optional prior beliefs about the correlation structure of the landscape. In addition to speeding up costly fitness evaluations, the methods can be used to characterize landscapes in terms of a small set of easily interpretable quantities. 9. "The Deterministic Genetic Algorithm: Implementation Details and Some Results" Ralf Salomon Recent literature on genetic algorithms provides a controversial discussion on the efficiency of this particular class of randomized optimization procedures; despite several encouraging empirical results, recent theoretical analyses have argued that in most cases, the runtime behavior of genetic algorithms is increased by at least a factor of ln(n) with n denoting the number of parameters to be optimized. It has been argued that these inefficiencies are due to intrinsic resampling effects. As a result of these theoretical considerations, a deterministic genetic algorithm has been suggested as a theoretical concept. Since its proposition, information discussions have been raised concerning some implementation details as well as efficacy issues. Since some implementation details are a bit tricky, this paper discusses some of them in a pseudo programming language similar to PASCAL or C. In addition, this paper presents two possible variants in detail and compares their runtime behavior with another fairly-established procedure, the breeder genetic algorithm. It turns out that on widely-used test functions, the deterministic variants scale strictly better. Furthermore, this paper discusses some specific fitness functions on which random algorithms yield better worst-case expectations than deterministic algorithms; but both types require constant time on average, i.e., one function evaluation. 10. "Effective Fitness Landscapes for Evolutionary Systems" Chris Stephens In evolution theory the concept of a fitness landscape has played an important role, evolution itself being portrayed as a hill-climbing process on a rugged landscape. In this article it is shown that in general, in the presence of other genetic operators such as mutation and recombination, hill-climbing is the exception rather than the rule. This discrepancy can be traced to the different ways that the concept of fitness appears -- as a measure of the number of fit offspring, or as a measure of the probability to reach reproductive age. Effective fitness models the former not the latter and gives an intuitive way to understand population dynamics as flows on an effective fitness landscape when genetic operators other than selection play an important role. The efficacy of the concept is shown using several simple analytic examples and also some more complicated cases illustrated by simulations. 11. "An Examination of Building Block Dynamics in Different Representations" Annie S. Wu and Kenneth A. De Jong We compare the traditional, fixed genetic algorithm (GA) representation scheme with a floating representation scheme and examine the differences in building block dynamics and how these differences affect a GA's ability to balance exploration and exploitation of building blocks. This study examines both the overall performance of a GA and the detailed events that contribute to overall behavior. Results indicate that the floating representation allows a GA to maintain a higher level of construction which results in a more diverse population from which to build solutions. 12. "The Futility of Programming Computers by Means of Natural Selection" Terry Fogarty Abstract Forthcoming 13. "The Time Complexity of Maximum Matching by Evolutionary Programming" Xin Yao Abstract Forthcoming 14. "Adaptive Genetic Algorithms -- Modeling and Convergence" Alexandru Agapie Head of Computational Intelligence Laboratory Institute of Microtechnology Bucharest, 72225, P.O. Box 38-160, Romania e-mail: agapie@oblio.imt.pub.ro agapie_a@hotmail.com http://www.imt.ro/ The paper presents a new mathematical analysis of GAs; we propose the use of random systems with complete connections (RSCC), a non-trivial extension of the Markovian dependence, accounting for the complete, rather than recent, history of a stochastic evolution. As far as we know, this is the first theoretical modeling of an adaptive GA. First we introduce the RSCC model of an p(m)-adaptive GA, then we prove that a "classification of states" is still valid for our model, and finally we derive a convergence condition for the algorithm. 15. Colin Reeves, Title, Abstract, Forthcoming 16. "Problem Perturbation: Implications on Fitness Landscapes" Worthy Martin Abstract Forthcoming 17. "FDA - An Evolutionary Algorithm for Additively Decomposed Functions" Heinz Muehlenbein and Thilo Mahnig FDA - the Factorized Distribution Algorithm - is an evolutionary algorithm which combines mutation and recombination by using a distribution instead. First the distribution is estimated from a set of selected points. It is then used to generate new points for the next generation. In general a distribution defined for n binary variables has 2^n parameters. Therefore it is too expensive to compute. For additively decomposed discrete function (ADFs) there exists an algorithm which factors the distribution into conditional and marginal distributions, each of which can be computed in polynomial time. The scaling of FDA is investigated theoretically and numerically. The scaling depends on the ADF structure and the specific assignment of function values. Difficult functions on a chain or a tree structure are optimized in about O(n*sqrt(n)) function evaluations. More standard genetic algorithms are not able to optimize these functions. FDA is not restricted to exact factorizations. It also works for approximate factorizations. 18. Emanuel Falkenauer, Title, Abstract, Forthcoming. 19. "On Some Difficulties in Local Evolutionary Search" Hans-Michael Voigt In this paper we consider the very simple problem of optimizing a stationary unimodal function over R^n without using analytical gradient information. There exist numerous algorithms from mathematical programming to evolutionary algorithms for this problem. Here we have a closer look at advanced evolution strategies (GSA, CMA), the evolutionary gradient search algorithm (EGS), local search enhancement by random memorizing (LSERM), and the simple (1+1)-evolution strategy. These approaches show different problem solving capabilities for different test functions. We introduce different measures which reflect certain aspects of what might be seen as the problem difficulty. Based on these measures it is possible to characterize the weak and strong points of the mentioned approaches which may lead to even more advanced algorithms. 20. Some Information Theoretic Results on Evolutionary Optimization Thomas M. English Tom English Computing Innovations 2401 45th Street, Apt. 30 Lubbock, Texas 79412 USA The body of theoretical results regarding conservation of information ("no free lunch") in optimization has not related directly to evolutionary computation. Prior work has assumed that an optimizer traverses a sequence of points in the domain of a function without revisiting points. The present work reduces the difference between theory and practice by a) allowing points to be revisited, b) reasoning about the set of visited points instead of the sequence, and c) considering the impact of bounded memory and revisited points upon optimizer performance. Fortuitously, this leads to clarification of the fundamental results in conservation of information. Although most work in this area emphasizes the futility of attempting to design a generally superior optimizer, the present work highlights possible constructive use of the theory in restricted problem domains. Session: Time Series Prediction – 3 Sessions Organizer: Byoung - Tak Zhang Artificial Intelligence Lab (SCAI) Department of Computer Engineering Seoul National University Seoul 151-742, Korea Phone: +82-2-880-1833; fax: +82-2-883-3595 btzhang@scai.snu.ac.kr http://scai.snu.ac.kr/~btzhang http://scai.snu.ac.kr/~btzhang/conferences/cec99/index.html 1. Shu-Heng Chen and Chun-Fen Lu: Would Evolutionary Computation Help in Designs of Artificial Neural Nets in Forecasting Financial Time Series? 2. Adelino R. Ferreira da Silva: Evolving Best-Basis Representations 3. E. Gomez-Ramirez, A. Gonzalez-Yunes, and M. Avila-Alvarez: Adaptive Architecture of Polynomial Artificial Neural Network to Forecast Nonlinear Time Series 4. Ch. Hafner and J. Froehlich: Generalized Function Analysis Using Hybrid Evolutionary Algorithms 5. Hitoshi Iba: Using Genetic Programming to Prediction Financial Data 6. Je-Gun Joung and Boung-Tak Zhang: Time Series Prediction Using Committee Machines of Evolutionary Neural Trees 7. Jung-Jip Kim and Boung-Tak Zhang: Heritability of Selection Schemes in GB-Based Time Series Analysis 8. Dong-Wook Lee and Kwee-Bo Sim: Evolving Chaotic Neural Systems for Time Series Prediction 9. Helmut A. Mayer and Roland Schwaiger: Evolutionary and Coevolutionary Approaches to Time Series Prediction Using Generalized Muti-Layer Perceptrons 10. Sathyanarayan S. Rao, Hemanth K. Birru, and Kumar H. Chellapilla: Evolving Nonlinear Time-Series Models Using Evolutionary Programming 11. Bernhard Sendhoff and Martin Kreutz: Variable Encoding of Modular Neural Networks for Time Series Prediction 12. Alaa F. Sheta: Nonlinear Channel Equalization Using Evolutionary Strategy 13. Ikuo Yoshihara, M. Numata, K. Sugawara, S. Yamada, and K. Abe: Time Series Prediction Model Building with BP-like Parameter Optimization Session: Dynamically Changing Fitness Landscapes – 1 Session Organizers: Kenneth DeJong George Mason University Fairfax, VA kdejong@cs.gmu.edu Ron Morrison George Mason University Fairfax, VA rmorrison@grci.com Ron Morrison 1. Experience With A Problem Generator For Changing Landscapes John Grefenstette 2. On Evolution Strategy Optimization In Dynamic Environments Karsten & Nicole Weicker 3. A Test Problem Generator for Non-stationary Environments Ron Morrison and Kenneth DeJong 4. The Usefulness Of Tag Bits In Changing Environments William Liles and Kenneth DeJong Session: Coevolution – 1 Session Organizer: Brian Mayoh Aarhus University Computer Science Department NyMunkegade Bldg. 504 DK-8000 Aarhus C, Denmark Phone: +45 8942 3373; Fax: +45 8942 3255 brian@daimi.aau.dk www.daimi.au.dk/~brian/cec99.html 1. "On a Coevolutionary Genetic Algorithm for Constrained Optimization" Helio Barbosa (LNCC.Petropolis,Brazil) hcbm@fluidyn.lncc.br 2. "Curiosity through cooperating/competing algorithmic predictors" Jurgen Schmidhuber(IDSIA,Lugano,Switzerland) juergen@idsia.ch 3. "Learning Nash Equilibria by Coevolving Distributed Classifier Systems" F. Seredynski (IPIPAN,Warsaw,Poland) sered@ipipan.waw.pl Cezary Z. Janikow 4. On preliminary studies of learning variable interdependencies in coevolutionary optimizers" Karsten & Nicole Weicker (Un.Tubingen,Un.Stuttgart,Germany) weicker@Informatik.Uni-Tuebingen.De 5. "Multinational evolutionary algorithms" Rasmus K. Ursem Since practical problems often are very complex with a large number of objectives it can be difficult or impossible to create an objective function expressing all the criterias of good solutions. Sometimes a simpler function can be used where local optimas could be both valid and interesting. Because evolutionary algorithms are population-based they have the best potential for finding more of the best solutions among the possible solutions. However, standard EAs often converge to one solution and leave therefore only this single option for a final human selection. So far at least two methods, sharing and tagging, have been proposed to solve the problem. This paper presents a new method for finding more quality solutions, not only global optimas but local as well. The method tries to adapt it's search strategy to the problem by taking the topology of the fitness landscape into account. The idea is to use the topology of the fitness landscape to group the individuals into sub-populations each covering a part of the fitness landscape. Session: Evolutionary Programming and Neural Networks Applied to Breast Cancer Research – 1 Session Organizers: Walker Land Binghamton University Computer Science Department wland@binghamton.edu Dr. Barbara Croft Diagnostic Imaging Program National Cancer Institute Breast Cancer is second only to lung cancer as a tumor-related cause of death in women. More that 180,000 new cases are reported annually in the US alone and, of these, 43,900 women and 290 men died last year. Furthermore, the American Cancer Society estimates that at least 25% of these deaths could be prevented if all women in the appropriate age groups were regularly screened. Although there exists reasonable agreement on the criteria for benign/malignant diagnoses using fine needle aspirate (FNA) and mammogram data, the application of these criteria are often quite subjective. Additionally, proper evaluation of FNA and mammogram sensor data is a time consuming task for the physician. Inter-and-inter-observer disagreement and/or inconsistencies in the FNA and mammogram interpretation further exacerbate the problem. Consequently, Computer Aided Diagnostics (CAD) in the form of Evolutionary Programming derived neural networks, neural network hybrids and neural networks operationg alone (utilized for pattern recognition and classification) offer significant potential to provide an accurate and early automated diagnostic technology. This automated technology may well be useful in further assisting with other problems resulting from physical fatigue, poor mammogram image quality, inconsistent FNA discriminator numerical assignments, as well as other possible sensor interpretation problems. The purpose of this proposed session is to present current and ongoing research in the CAD of breast carcinoma. Specifically, this session has the following objectives: To show the advantages of using Evolutionary Programming and Neural Networks as an aid in the breast cancer diagnostic process To demonstrate the application of EP and neural network hybrid systems in solving the InterObservability problem. To establish that CAD tools are simple and economical to implement in the clinical setting To demonstrate that CAD tools can provide the cytophathogists, radiologists and neurosurgeons with an early diagnosis of breast cancer that is accurate, consistent and efficient as well as accessible. Some practical results of CAD of breast cancer sensor data using neural networks are expected to be: Operational software which will aid the physician in making the diagnosis, quite possibly in real time, and once formulated and tested, they are always consistent, not prone to human fatigue or bias. Providing diagnostic assistance for the intra-and-inter-observability problems by ultimately minimizing the subjective component of the diagnostic process Providing an initial detection and/or classification process in the absence of a qualified physician Providing possible (and probably currently unknown) relationships between sensor environment discriminators and a correct diagnosis. This session is comprised of the following five invited papers: 1. A Status Report on Identifying Important Features for Mammogram Classification D.B. Fogel, (Natural Selection, Inc.) E.C. Wasson (Maui Memorial Hospital) E.M. Boughton, (Hawaii Industrial Laboratories) V.W. Porto and P.J. Angeline (Natural Selection, Inc.) Disagreement or inconsistencies in mammographic interpretation motivates utilizing computerized pattern recognition algorithms to aid the assessment of radiographic features. We have studied the potential for using artificial neural networks (ANNs) to analyze interpreted radiographic features from film screen mammograms. Attention was given to 216 cases (mammogram series) that presented suspicious characteristics. The domain expert (Wasson) quantified up to 12 radiographic features for each case based on guidelines from previous literature. Patient age was also included. The existence or absence of malignancy was confirmed in each case via open surgical biopsy. The ANNs were trained using evolutionary algorithms in a leave-one-out cross validation procedure. Results indicate the ability for small linear models to also provide reasonable discrimination. Sensitivity analysis also indicates the potential for understanding the networks’ response to various input features. 2. Application of Artificial Neural Networks for Diagnosis of Breast Cancer J.Y. Lo and C.E. Floyd (Digital Imaging Research Division, Dept. of Radiology, Duke Univ. Medical Center, and Dept. of Biomedical Engineering, Duke Univ.) We will present several current projects pertaining to artificial neural networks (ANN) computer models that merge radiologist-extracted findings to perform computer-aided diagnostics (CADx) of breast cancer. These projects include (1) the prediction of breast lesion malignancy using mammographic and patient history findings, (2) the further classification of malignant lesions as in situ carcinoma vs. invasive cancer, (3) the prediction of breast cancer utilizing ultrasound findings, and (4) the customization and evaluation of CADx models in a multi-institution study. Methods: These projects share in common the use of feedforward, error backpropagation ANNs. Inputs to the ANNs are medical findings such as mammographic or ultrasound lesion descriptors and patient history data. The output to the ANN is the biopsy outcome (benign vs. malignant, or in situ vs. invasive cancer) which is being predicted. All ANNs undergo supervised training using actual patient data. Performance is evaluated by ROC area, specificity for a given high sensitivity, and/or positive predictive value (PPV). Results: We have developed ANNs, which have the ability to predict the outcome of breast biopsy at a level comparable or better than expert radiologists. For example, using only 10 mammographic findings and patient age, the ANN predicted malignancy with a ROC area of 0.86 = B1 0.02, a specificity of 42% at a given sensitivity of 98%, and a 43% PPV. onclusion: These ANN decision models may assist in the management of patients with breast lesions. By providing information which was previously available only through biopsy, these ANNs may help to reduce the number of unnecessary surgical procedures and their associated cost. Contributions made by this abstract: This abstract describes the application of simple backprop ANNs to a wide range of predictive modeling tasks in the diagnosis of breast cancer. The group is one of the most authoritative in the field of computer-aided diagnosis, with a tract record that encompasses many radiological imaging modalities and engineering disciplines. 3. Optimizing the Effective Number of Parameters in Neural Network Ensembles: Application to Breast Cancer Diagnosis Y. Lin and X. Yao, ( Dept of Computer Science, Australian Defense Force Academy, Canberra) The idea of negative correlation learning is to encourage different individual networks in an ensemble to learn different parts or aspects of the training data so that the ensemble can learn the whole training data better. This paper develops a technique of optimizing the effective number of parameters in a neural network ensemble by negative correlation learning. The technique has been applied to the problem of breast cancer diagnosis. 4. Artificial Neural Networks in Breast Cancer Diagnosis: Merging of Computer-Extracted Features from Breast Images Maryellen L. Giger Kurt Rossmann Laboratories for Radiologic Image Research The University of Chicago Department of Radiology MC2026 5841 S. Maryland Ave. Chicago, IL 60637 m-giger@uchicago.edu Phone: 773-702-6778 Fax: 773-702-0371 Computer-aided diagnosis (CAD) can be defined as a diagnosis made by a radiologist who takes into consideration the results of a computerized analysis of radiographic images and uses them as a "second opinion" in detecting lesions, assessing disease, and/or in making diagnostic decisions. The final decision regarding diagnosis and patient management would be made by the radiologist. CAD is analogous to a spell checker. In mammography, computerized methods for the detection and characterization of lesions are being developed to aid in the early detection of breast cancer. Since mammography is becoming a high volume x-ray procedure routinely interpreted by radiologists and since radiologists do not detect all cancers that are visible on images in retrospect, it is expected that the efficiency and effectiveness of screening could be increased by CAD. In addition, computerized methods are being developed to aid in the characterization of lesions in order to potentially reduce the number of benign cases sent to biopsy. Once a lesion is detected, the radiologist's decision concerns patient management -- that is, return to routine screening, perform follow-up mammograms, or send to biopsy? This differs from a purely benign versus malignant determination. Many breast cancers are detected and referred for surgical biopsy on the basis of a radiographically detected mass lesion or cluster of microcalcifications. Although general rules for the differentiation between benign and malignant mammographically identified breast lesions exist, considerable misclassification of lesions can occur. On average, less than 30% of masses referred for surgical breast biopsy are actually malignant. We have been developing computerized analysis schemes to aid in distinguishing between malignant and benign lesions. Such methods can use features extracted either by computer or by radiologists. These features are then merged by classifiers, such as linear discriminant functions or artificial neural networks, to yield estimate of the likelihood of malignancy. These computerized methods for the characterization of breast lesions are being developed for mammography, sonography, and magnetic resonance imaging. 5. Investigation of and Preliminary Results for the Solution of the Inter-Observability Problem using Fine Needle Aspirate (FNA) Data W. H. Land, JR and L. Loren (Computer Science Dept., Binghamton Univ. and T. Masters (TMAIC, Vestal, NY) This paper provides a preliminary evaluation of the accuracy of Computer Aided Diagnostics (CAD) in addressing the inconsistencies of Inter-Observability scoring. The Inter-Observability problem generally relates to different cytopathologists and radiologists, etc. at seperate locations scoring the same type of samples differently using the same methodologies and environment discriminates. Two different approaches are currently being investigated: (1) a recently developed Evolutionary Programming (EP) / Probabilistic Neural Network (PNN) hybrid, and (2) a classification model based on the thresholding of means of all predictors called the “mean of predictors” model. Method: Two distinctly different FNA data sets were used. The first was the data set collected at the Univ. of Wisconsin (Wolberg data set) while the other was a completely independent one defined and processed at the Breast Care Center, Syracuse University (Syracuse dataset). Results of several experiments performed using the EP/PNN hybrid (which provided several unique network configurations) are first summarized. The EP/PNN hybrid was trained on the Wolberg dataset and the resultant models evaluated on the Syracuse dataset. For comparative purposes, these same hybrid architectures which were trained on the Wolberg set were also evaluated on the Wolberg validation set. The “mean of predictors” method first trained the thresholds using the original Wolberg training set. This model was then tested on the Wolberg test and validation sets, and on the Syracuse set. All three Wolberg datasets (train, test, validate) were then used to train the threshold, and this model was applied to the Syracuse data Results: Initial results using the EP/PNN hybrid showed a 85.2% correct classification accuracy with a 2.6% Type II (classifying malignant as benign) error averaged over five experiments when trained on the Wolberg data set and validated on the Syracuse data set. Training and validating on the Wolberg data set resulted in a 97% correct classification accuracy and a < 0.2% Type II error. These results are preliminary in that no attempt has been made to optimize the threshold setting. The paper will include several additional EP/PNN hybrid experimental results as well as optimum threshold settings and an ROC analysis. The “mean of predictors method” analysis produced the following preliminary results. Training the thresholds on the first 349 Wolberg samples resulted in a CAD model which provided: (1) a 98.8% classification accuracy and a 0% Type II error, (2) a 96% classification accuracy with a 1.7% Type II error when using the Wolberg test and validation sets respectively which confirms the EP/PNN preliminary results. Using the Syracuse validation set yielded a 96% classification accuracy and a 1% Type II error which is improved performance when compared with the EP/PNN results. Training the “mean of predictors” model on all 699 Wolberg samples and validating on the Syracuse dataset resulted in a 86% classification accuracy and a 1% Type II error. Again, these results match well with the EP/PNN results. We again emphasize these results are preliminary but very promising. Conclusions: Preliminary results using both the newly developed EP/PNN hybrid and the “mean of predictors” methods are very encouraging. We believe that both of these CAD tools will, with additional research and development effort, be useful additions to our growing group of CAD tools being developed at Binghamton University. Session: Engineering Design – 2 Sessions Organizers: Ian Parmee Plymouth Engineering Design Centre Reader in Evolutionary/Adaptive Computing in Design and Manufacture School of Computing University of Plymouth Drakes Circus Plymouth PL4 8AA Devon, UK Phone: 01752 233509; Fax: 01752 233529 I.parmee@plymouth.ac.uk Prabhat Hajela RPI hajela@rpi.edu Mark Jakiela Hunter Professor of Mechanical Design Washington University in St. Louis mjj@mecf.wustl.edu 1. Marc Schoenauear Ecole Polytechnique, Paris 2. Eric Goodman Michigan State 3. Ian Parmee EDC, Plymouth 4. Chris Bonham EDC, Plymouth 6. communications 7. antenna design 8. building structures 9. 10. Session: Data Mining – 2 Sessions Organizer: Jan Zytkow zytkow@uncc.edu 1. Clustering TV Prefererences Using Genetic Algorithms Teresa Goncalves & Fernando Moura-Pires tcg@di.fct.unl.pt 2. A Comparison of Two Feature Selection Algorithms Kan Chen & Huan Liu liuh@comp.nus.edu.sg 3. A Survey of Genetic Feature Selection in Mining Issues Maria Jose Martin-Bautista and Maria-Amparo Vila mbautis@decsai.ugr.es 4. Discovering interesting prediction rules with genetic algorithms Edgar Noda, Alex Freitas & Heitor Lopes alex@dainf.cefetpr.br 5. Learning Classifications using Evolutionary Programming Michael Cavaretta & Kumar Chellapilla mcavaret@ford.com 6. Evolution of Logic Programs: Part-of-Speech Tagging Philip G. K. Reiser & Patricia Riddle philip@cs.auckland.ac.nz 7. GA-based Hierarchical Rule Learning Jukka Hekanaho hekanaho@ra.abo.fi Session: Scheduling – 2 Sessions Organizers: Jeff Herrmann University of Maryland jwh2@glue.umd.edu Edmund Burke University of Nottingham ekb@cs.nott.ac.uk Bryan Norman University of Pittsburgh banorman@engrng.pitt.edu 1. Jeff Herrmann 2. Bryan Norman 3. Edmund Burke 4. Jeff Joines 5. X. Q. Cai 6. D. Montana Session: Teaching of Evolutionary Computation – 1 Session Organizer: Xin Yao Department of Computer Science Australian Defense Force Academy Canberra xin@cs.adfa.oz.au 1. Sung-Bae Cho (Korea) 2. Riccardo Poli (UK) 3. Alice E. Smith (USA) 4. Xin Yao (Australia) 5. Zbigniew Michalewicz (USA) Session: DNA Computing – 2 Sessions Organizer: Jan Mulawka Institute of Electronics Fundamentals Warsaw University of Technology ul. Nowowiejska 15/19 00-665 Warsaw, Poland Phone: (+48 22) 660 53 19; Fax: (+48 22) 825 23 00 jml@ipe.pw.edu.pl http://www.ipe.pw.edu.pl/~gair 1. Towards a System for Simulating DNA Computing Using Secondary Structure Akio Nishikawa, Masami Hagiya University of Tokyo, Japan nisikawa@is.s.u-tokyo.ac.jp Whiplash PCR is a useful method for analyzing DNA. For example, state transitions can be implemented by this method. In normal PCR, extension is guided by the annealed complementary DNA sequence. Whiplash PCR is a modification of normal PCR in which the 3'-end of the target single-stranded DNA molecule is extended by polymerase when it anneals to a complementary sequence in the target DNA molecule, so that the secondary structure of the target single-stranded DNA forms the guiding sequence. When the annealed sequences are extended by PCR, the guiding sequence and the composition of the PCR reaction buffer control termination. If the PCR buffer lacks T and the guiding sequence contains an AAA sequence, the PCR extension will stop at the sequence AAA. In Whiplash PCR, like normal PCR, the molecule is subsequently denatured by high temperature, the guiding sequence re-anneals when the temperature is lowered for the next extension step and the cycle repeats. Whiplash PCR is a very powerful technique for solving certain types of problems, but the feasibility of using it in a specific instance should be carefully considered. The target molecule must form a secondary structure by annealing to itself, and it must be capable of melting and reforming a similar secondary structure. To implement this, the sequences must be carefully designed. The simulation system we are constructing checks whether the sequences can form such secondary structures by examining the nucleotide sequence. It also simulates sequence extension by ligation, PCR and whiplash PCR, mishybridization, affinity separation, and restriction enzyme digests. By combining all these types of reaction simulations, it can simulate a series of DNA computing processes, which will aid the designers of DNA computing reactions. We are now constructing such simulation software based on sequence comparison, and will report our current stage of development. In addition to developing the simulation methods based on sequence comparison, we have investigated possible enhancements of our system. Simply using a method based on sequence comparison to examine the feasibility of DNA computing using secondary structures is inadequate. We plan to enhance our system in two ways. First, it will check the conditions for each reaction step with reaction parameters such as the PH, temperature, and makeup of the PCR reaction buffer. Second, it will calculate and examine the physico-chemical parameters affecting the secondary structure necessary for whiplash PCR. In other words, the sequences can form such secondary structures by examining the nucleotide sequence. It also simulates sequence extension by ligation, PCR and whiplash PCR, mishybridization, affinity separation, and restriction enzyme digests. By combining all these types of reaction simulations, it can simulate a series of DNA computing processes, which will aid the designers of DNA computing reactions. We are now constructing such simulation software based on sequence comparison, and will report our current stage of development. In addition to developing the simulation methods based on sequence comparison, we have investigated possible enhancements of our system. Simply using a method based on sequence comparison to examine the feasibility of DNA computing using secondary structures is inadequate. We plan to enhance our system in two ways. First, it will check the conditions for each reaction step with reaction parameters such as the PH, temperature, and makeup of the PCR reaction buffer. Second, it will calculate and examine the physico-chemical parameters affecting the secondary structure necessary for whiplash PCR. In other words, it will check the feasibility from a physico-chemical perspective. We would like to discuss these enhancements and the feasibility of our simulation system. Furthermore, if possible, we would like to compare the results of our simulation system with in vitro experiments. 2. PNA-enhanced DNA Computing J. Miro-Julia, F. Rossello UIB University, Spain joe@ipc4.uib.es Since Adleman's seminal paper, DNA computing is emerging as a very promising alternative to electronic computing. It s massive parallelism, its low energy needs, even its possible ability of breaking the NP barrier make its study irresistible. But for all its promise DNA computing has few results to show. This is in because there are many biochemical problems in handling DNA that are still to be solved. Peptyde Nucleic Acid (PNA) is a new class of informational biomolecule, a nucleic acid analogue with a backbone consisting of a peptide-like chain formed of glycine units to which the usual nucleobases are attached. DNA molecules have some properties, relevant to the area of computation with biomolecules, unique among nucleic acids: among others, they hybridize to both complementary DNA and RNA with higher affinity that DNA itself. They can be used to enhance PCR amplification, and they can be used to restrict the action of restriction enzymes on DNA strands. The goal of this paper is to introduce PNA in the area of computation with biomolecules, and to discuss its possible role in overcoming some of the drawbacks that appear with DNA. 3. Efficient DNA Computation of Spanning Trees Martyn Amos, Shinichiro Yoshii, Paul E. Dunne and Alan Gibbons University of Liverpool, UK martyn@csc.liv.ac.uk In this paper we demonstrate the translation of a high-level algorithm down to the level of molecular operations on DNA. For DNA computation to be a viable and competitive paradigm, we need algorithms that require viable resources (time and volume). We describe a translation of a parallel algorithm for the computation of spanning trees on graphs that is resource-efficient. 4. The Analysis of a Simple Programmed Mutagenesis System. Julia Khodor and David K. Gifford Laboratory for Computer Science, MIT, USA Jkhodor@MIT.EDU We present the experimental results of the advanced cycles of string rewrite system based on programmed mutagenesis, a technique for rewriting DNA strands according to programmed rules. We have previously shown that formation of the second-cycle product depends on successful completion of the first cycle, and demonstrated that two oligonucleotides annealing to the template next to each other can ligate and extend to form the correct product. We now report our findings on the behavior of the advanced cycles of the system. We discuss several factors influencing the behavior of the system, such as the choice of enzymes, temperatures, and concentrations. The efficiency of the system and the potential significance of the results to the field of DNA-computing are also discussed. 5. Executing Parallel Logical Operations with DNA Mitsunori Ogihara and Animesh Ray University of Rochester, USA ogihara@cs.rochester.edu DNA computation investigates the potential of DNA as a massively parallel computing device. Research is focused on designing parallel computation models executable by DNA-based chemical processes and on developing algorithms in the models. In 1994 Leonard Adleman initiated this area of research by presenting a DNA-based method for solving the Hamilton Path Problem. That contribution raised the hope that parallel computation by DNA could be used to tackle NP-complete problems which are thought of as intractable. The current realization, however, is that NP-complete problems may not be best suited for DNA-based (more generally, molecule-based) computing. A better subject for DNA computing could be large-scale simulation of parallel computation models. Several proposals have been made in this direction. We overview those methods, discuss technical and theoretical issues involved, and present some possible applications of those methods. 6. Aspects of Evolutionary DNA Computing Thomas Baeck , Joost Kok, and Grzegorz Rozenberg Leiden University, The Netherlandes baeck@icd.de Evolutionary Computation focuses on probabilistic search and optimization methods gleaned from the model of organic evolution. Genetic algorithms, evolution strategies and evolutionary programming are three independently developed representatives of this class of algorithms, with genetic programming and classifier systems as additional paradigms in the field. After giving a general introduction to evolutionary algorithms and demonstrating their practical usefulness as heuristics for finding good solutions in combinatorial and continuous problem domains with a few application examples, the presentation emphasizes on certain features of genetic algorithms and evolution strategies: Namely, the mutation and recombination operators from genetic algorithms and the selection operator from evolution strategies are discussed in more detail. Furthermore, some theoretical results concerning the convergence velocity of certain variants of genetic algorithms are presented, and in particular the question for optimal mutation rates in genetic algorithms is discussed. Based on this detailed background information about evolutionary algorithms, it is then outlined how these algorithms relate to DNA based computing. 7. The Inference via DNA Computing J. Mulawka, P. Wasiewicz, and A. Plucienniczak Warsaw University of Technology, Poland jml@ipe.pw.edu.pl Most of research on DNA computing concerns solving NP-complete problems, programming in logic as well as purely mathematical consideration of splicing theory. There is no doubt that these works have given new impetus to find another areas for DNA computing. In proposed paper we show that selfassembly of DNA strands may be applied to implement the inference process. The ability of such an application was provided for the first time by J.J. Mulawka et al. At the International Conference on Evolutionary Computation in Anchorage, 1998. In this paper we would like to report another achievements in this field. The inference process can be implemented either by backward or forward chaining. We have developed new methods of inference based on another concept of genetic engineering. The primary objective of our paper is to describe these methods and procedures. After short introduction to the subject an overview of the problem will be provided. It will be shown how knowledge can be structured, and stored by means of DNA strands. Next, we describe an implementation of DNA inference engine. Such system can store knowledge for a narrowly defined subject area and solve problems by making logical deductions. The inference mechanism performs these tasks by manipulating on DNA strands. It provides the problem-solving methods by which the rules are processed. Our approach uses standard operations of genetic engineering. To check correctness of our method a number of laboratory tests have been carried out. The results of these experiments as well as concluding remarks will be provided in final version of the paper. 8. Byoun-Tak Zhang from Korea 9. H. Rubin rubinh@mail.med.upenn.edu 10. T. Head tom@math.binghamton.edu Session: Ant Colony Methods – 1 Session Organizer: Marco Dorigo Chercheur Qualifie' du FNRS IRIDIA CP 194/6 Universite' Libre de Bruxelles Avenue Franklin Roosevelt 50 1050 Bruxelles Belgium Phone: +32-2-6503169; Fax: +32-2-6502715 mdorigo@ulb.ac.be http://iridia.ulb.ac.be/dorigo/dorigo.html 1. An Ant Colony Optimization Approach for the Single Machine Total Tardiness Problem Bauer A., Bullnheimer B., Hartl R.F., Strauss C. Machine Scheduling is a central task in production planning. In general it consists of the problem of scheduling job operations on a given number of available machines. In this paper we consider a machine scheduling problem with one machine, the Single Machine Total Tardiness Problem. As this problem is NP-hard, we apply the ant colony optimization metaphor, a recently developed meta-heuristic that has proven its potential for various other combinatorial optimization problems, and present computational results. Keywords Ant Colony Optimization, Machine Scheduling, Meta-Heuristics, Single Machine Total Tardiness Machine Scheduling using ACO 2. A Cooperative Ant Algorithm for Dynamic Routing and Load Balancing Martin Heusse and Dominique Snyers ENST de Bretagne, Brest, France snyers@ieee.org This talk presents a new family of completely distributed algorithms for routing in dynamic communication networks inspired by insect collective behaviors. These algorithms are compared with others on simulated networks under various traffic loads. Early results using the realistic Opnet simulator will also be presented. The connection between these ``ant algorithms'' and dynamic programming is discussed. These new ``ant algorithms'' can be described as an online asynchronous version of dynamic programming combined with reinforcement learning. Estimates of the current load and link costs are measured by sending routing agents (or artificial ants) in the network that mix with the regular information packets and keep track of the costs encountered during their journey. The routing tables are then regularly updated based on that information without any central control nor complete knowledge of the network topology. Two new algorithms are proposed here. The first one is based on round trip routing agents that update the routing tables by backtracking their way after having reached the destination. The second one relies on forward agents that update the routing tables directly as they move toward their destination. An efficient co-operative scheme is proposed to deal with asymmetric network connections. 3. A New Version of Ant System for Subset Problems Guillermo Leguizamon Interest Group on Computer Systems Universidad Nacional de San Luis Argentina legui@unsl.edu.ar Zbigniew Michalewicz Department of Computer Science University of North Carolina at Charlotte USA zbyszek@uncc.edu Ant Colony Optimization (ACO) is a relatively new meta-heuristic for hard combinatorial optimization problems. This meta-heuristic belongs to the class of heuristics derived from nature, which includes Evolutionary Algorithms, Neural Networks, Simulated Annealing, Tabu Search. An ACO algorithm is based on the result of low-level interaction among many cooperating simple agents that are not explicitly aware of their cooperative behavior. Each simple agent is called ant and the ACO algorithm (a distributed algorithm) is based on a set of ants working independently and cooperating sporadically in a common problem solving activity. Since the design of Ant System, the first example of ACO algorithm, plenty of work has been done in this area with applications to problems such as the Traveling Salesman Problem, the Bin Packing Problem, and the Quadratic Assignment Problem. In this paper we present the original Ant System as well as a new version applied to three subset problems: the Multiple Knapsack Problem (MKP), the Set Covering Problem (SCP), and the Maximum Independent Set Problem (MISP). The results show the potential power of the ACO approach for solving subset problems. 4. Hamiltonian(t) – An Ant-Inspired Heuristic for Recognizing Hamiltonian Graphs Israel A. Wagner (1,2) Alfred M. Bruckstein (2) (1) IBM Haifa Research Lab, Matam, Haifa 31905, Israel (2) Department of Computer Science, Technion City, Haifa 32000, Israel israelw@vnet.ibm.com, freddy@cs.technion.ac.il Given a graph G(V,E), we consider the problem of deciding whether G is Hamiltonian, i.e., whether or not there is a simple cycle in E spanning all vertices in V. This problem is known to be NP-complete, hence cannot be solved in time polynomial in |V| unless P=NP. The problem is a special case of the Travelling Salesperson Problem (TSP), that was extensively studied in the literature, and has recently been attacked by various ant-colony methods. We address the Hamiltonian cycle problem using a new antinspired approach, based on repeated covering of the graph. Our method is based on a process in which an ant traverses the graph by moving from vertex to vertex along the edges, occasionally leaving traces in the vertices, and deciding on the next step according to the level of traces in the surrounding neighborhood. We show that Hamiltonian cycles are limit cycles of the process, and investigate the average time needed by our ant process to recognize a Hamiltonian graph, on the basis of simulations made over large samples of random graphs with varying structure and density. 5. Adapted versions of the ABC algorithm in network control L.J.M. Rothkrantz, J. Wojdel, A. Wojdel roth@kgs.twi.tudelft.nl University of Delft, The Netherlands Recently agents algorithms based on the natural behavior of ants were successfully applied to route data in telecommunication networks. Mobile agents created the routing tables in the simulated network with behaviors modeled on the trail-laying abilities of agents. This paper describes some adapted versions of the Ant Based Control algorithm. To test the algorithms in different network topologies, a simulation environment was designed and implemented. The simulation software, algorithms and results of experiments will be presented in this paper. Session: Evolutionary Computation and Biological Modeling – 2 Sessions Organizers: Kumar Chellapilla H. Dept. of ECE, Univ. of California, San Diego 9500 Gilman Drive La Jolla, CA 92093-4007 Ph: (619) 534-5935; Fax: (619) 534-1004 kchellap@ece.ucsd.edu http://vision.ucsd.edu/~kchellap Gary Fogel Natural Selection Inc., La Jolla, CA 92037 1. Computer Experiments on the Development of Niche Specialization in an Artificial Ecosystem Jon Brewster and Michael Conrad Dept. of Computer Science Wayne State University Detroit, MI 48202 Evolve IV is the most recent of the EVOLVE series of evolutionary ecosystem models. The key new feature is all interactions among organisms are mediated by metabolic transformations. The model is thus well suited to the study of niche proliferation since evolutionary pressures lead to specialization and mutually stimulatory relationships in a natural way. 2. Multiple RNA Sequence Alignment Using Evolutionary Programming Kumar Chellapilla and Gary Fogel Multiple sequence alignment and comparison are popularly used techniques for the identification of common structure among ordered strings of nucleotides (as is the case with DNA or RNA) or amino acids (as is the case with proteins). Current multiple sequence alignment algorithms are characterized by great computational complexity. The focus of this paper is to use evolutionary programming as the basis for developing an efficient multiple sequence alignment algorithm for RNA sequences. An evolutionary programming (EP) based multiple sequence alignment algorithm is presented. Variants on the basic algorithm are also presented and their trade-offs are discussed. The appropriate weights associated with matches, mismatches and gaps are critical for the success of any algorithm. Simulation results on the canonical version indicate that the proposed EP method is not only viable but offers a much more robust alternative to conventional methods as the differences in the structure increase. 3. When Metaphors Collide: Biological Underpinnings To Genetic Programming Theory Jason Daida University of Michigan, Artificial Intelligence Lab & Space Physics Research Lab, 2455 Hayward Ave, Ann Arbor, MI 48109-2143 USA, (734) 647-4581 (wk), (734) 764-5137 (fax), daida@eecs.umich.edu (email), http://www.sprl.umich.edu/acers/ (www) Current theoretical research in genetic programming (GP) owes part of its heritage to the use of biological metaphor. Indeed, one could go even further by stating that biological metaphors have significantly shaped the direction of theoretical development in GP. Whether in the particulars of what in GP maps to genotype and to phenotype, or in the wholes of what in GP maps to a set of dynamics that is reflective of some biological phenomena, the use of biology in framing GP theory has been pervasive. However, current research also shows that descriptions of GP dynamics can deviate significantly from what one would expect in biology. Therefore, our investigations have lead us to rethink how the biology maps to processes and artifacts in GP. Consequently, this paper discusses how the use of biological metaphor can over constrain critical portions in the development of GP theory. 4. Simulated Sequencing by Hybridization Using Evolutionary Programming Gary Fogel and Kumar Chellapilla Sequencing of DNA is among the most important tasks in molecular biology. DNA chips are considered to be a more rapid alternative to more common gel-based methods for sequencing. Previously, we demonstrated the reconstruction of DNA sequence information from a simulated DNA chip using evolutionary programming. The research presented here extends this work by relaxing several assumptions required in our initial investigation. We also examine the relationship between base composition of the target sequence and the useful set of probes required to decipher the target on a DNA chip. Comments regarding the nature of the optimal ratio for the target and probe lengths are offered. Our results go further to suggest that evolutionary computation is well suited for the sequence reconstruction problem. 5. A Survey of Recent Work on Evolutionary Approaches to the Protein Folding Problem. Garrison Greenwood* , Byungkook Lee**, Jae-Min Shin**, and Gary Fogel*** *Dept. of Electrical & Computer Engineering, Western Michigan University, Kalamazoo, MI 49008 ** Laboratory of Molecular Biology, National Cancer Institute, Bethesda, MD 20892 *** Natural Selection Inc., La Jolla, CA 92037 A problem of immense importance in computational biology is the determination of the function conformations of protein molecules. With the advent of faster computers, it is now possible to use heuristic algorithms to search conformation space for protein structures that have minimal free-energy. Surveys work done in the last five years using evolutionary search algorithms to find low energy protein conformations. In particular, a detailed description is included of some work recently started at the National Cancer Institute, which uses evolution strategies. 6. Creation Of A Biomimetic Dolphin Hearing Model Through The Use Of Evolutionary Computation D. S. Houser1, D. A. Helweg, and P. W. B. Moore 1SPAWARSYSCEN-San Diego, Code D351, 49620 Beluga Road, San Diego, CA 92152-5435 Niche exploitation by an organism cumulatively results from its existing adaptations and phylogenetic history. The biological sonar of dolphins is an adaptation for object (e.g. prey or obstacle) detection and classification in visually limited environments. The unparalleled echo discrimination capability of the dolphin provides an excellent model for investigating similar synthetic systems. Current biomimetic modeling of dolphin echo discrimination emphasizes the mechanical and neurological filtering of the peripheral auditory system prior to central nervous system processing of echoes. Psychoacoustic, anatomical, and neurophysiological data collected from the bottlenose dolphin (Tursiops truncatus) indicate the structure of some auditory tuning curves. However, an optimal filter set has yet to be developed that demonstrates comparable frequency-dependent sensitivity. Evolutionary computation techniques have been employed to optimize the sensitivity of filters to that observed in the bottlenose dolphin by seeding the population with known filter parameters and evolving the number, frequency distribution, and shape of individual filters. Comparisons of evolved and known biological tuning curves will be discussed. 7. Evolutionary Computation Enhancement of Olfactory System Model Mary Lou Padgett and Gerry V. Dozier, Auburn University Recent electron microscopy work on the anatomy of the olfactory system in the rat has suggested a structural basis for grouping input stimuli before processing to classify odors. In the construction of a simulated nose, the appropriate number of inputs per group is a design parameter which can be finetuned using evolutionary computation. Previous results indicate that improvements to classification accuracy can be made by grouping inputs. On the other hand, the cost of such grouping can potentially increase the number of samples required for each "sniff". This increase can be expensive in terms of hardware and processing time. This paper suggests a methodology for selecting a size range based on improvement in accuracy and cost of grouping inputs. Session: Fitness Distributions: Tools for Designing Intelligent Operators and Efficient Evolutionary Computations – 1 Session Organizer: Kumar Chellapilla Dept. of ECE, Univ. of California, San Diego 9500 Gilman Drive La Jolla, CA 92093-4007 Ph: (619) 534-5935; Fax: (619) 534-1004 kchellap@ece.ucsd.edu http://vision.ucsd.edu/~kchellap 1. Fitness Distributions in Evolutionary Computation: Analysis of Local Extrema in the Continuous Domain Kumar Chellapilla and David B. Fogel The design of evolutionary computations based on schema processing, minimizing expected losses, and emphasizing certain genetic operators has failed to provide robust optimization performance. Recently, fitness distribution analysis have been proposed as an alternative tool for exploring operator behavior and designing efficient evolutionary computations. For example, the step size of a single parent variation operator determines the corresponding probability of finding better solutions and the expected improvement that will be obtained. This paper analyses the utility of Gaussian, Cauchy, and Mean mutation operators when a parent is located on a local extrema of a continuous objective function that is to be optimized. 2. Towards Evolution of Modular Finite State Machines using Fitness Distributions David Czarnecki For a given representations in an evolutionary computation, there are a number of variation operators that can be applied to existing solutions to create new solutions. These variation operators can be classified into two broad categories, exploratory and exploitative operators. Exploratory operators allow for the traversal of a given search space. Exploitative operators may induce behavior that uses the current location in the fitness landscape to move a solution towards a more optimal region in the neighboring search space. Fitness Distributions is a recent technique to assess the viability and quality of such variation operators. This technique is applied to the evolution of modular and non-modular finite state machines. Experiments are conducted on a number of tracking and control problems. Discussion is directed towards assessing the overall effectiveness of operators for such machines. 3. Using Fitness Distributions to Improve the Evolution of Learning Structures Christian Igel, Martin Kreutz, and Peter Stagge The analysis of features of the fitness distribution (FD) [3,4] is used to improve the performance of three different structure optimization algorithms by guiding the choice and design of the used variation operators. All three algorithms employ hybrid approaches where an evolutionary algorithm (EA) is used for structure optimization in conjunction with gradient based learning strategies for parameter adaptation. 4. On Functions With A Given Fitness--Distance Relation Leila Kallel, Bart Naudts and Marc Schoenauer Recent work stresses the limitations of fitness distance correlation (FDC) as an indicator of landscape difficulty for genetic algorithms (GAs). Realizing that the fitness distance correlation (FDC) value cannot be reliably related to landscape difficulty, this paper investigates whether an interpretation of the correlation plot can yield reliable information about the behavior of the GA. Our approach is as follows: A generic method for constructing fitness functions which share the same fitness versus distance-tooptimum relation (FD relation) is presented. Special attention is given to FD relations which show no local optimum in the correlation plot, as is the case for the relation induced by Horn's longpath. A tentative taxonomy is proposed for he different types of GA behavior found within a class of fitness functions with a common correlation plot. Finally, the behavior of the GA is shown to be very sensitive to small modifications of the fitness--distance relation. 5. Empirical Study Of Two Classes Of Bit Mutation Operators In Evolutionary Computation Hemanth Birru Bit mutation operators are widely used in evolutionary computations (EC) that adopt a binary representation. Most commonly, one or more randomly selected bits are flipped with some probability. The success of such a variation operator invariably depends on the location and number of bits being varied in the chromosome. Given an objective function and a candidate parent, bit mutations in some locations will provide an improvement in fitness whereas no improvement will result when mutations are applied to other locations. Two classes of bit mutation operators are defined and the relationship between the probability of improvement in fitness and the expected improvement obtained is studied. Such a study would help in designing better mutation operators to obtain faster convergence in EC methods using a binary representation. Session: Knowledge - Based Approaches to Evolutionary Computation Using Cultural Algorithms, Case - Injected Genetic Algorithms, and Related Techniques for Real World Engineering Applications – 2 Sessions Organizer: Robert G. Reynolds Department of Computer Science 431 State Hall Wayne State University Detroit, MI 48202 reynolds@CS.Wayne.EDU 1. Using Cultural Algorithms to Improve Performance in Semantic Networks Nestor Rychtyckyj Manufacturing and Engineering Development AI and Expert Systems Group Ford Motor Company Detroit Evolutionary computation has been successfully applied in a variety of problem domains and applications. In this paper we describe the use of a specific form of evolutionary computation known as cultural algorithms to improve the efficiency of the subsumption algorithm in semantic networks. Subsumption is the process that determines if one node in the network is a child of another node. As such, it is utilized as part of the node classification algorithm within semantic network-based applications. One method of improving subsumption efficiency is to reduce the number of attributes that need to be compared for every node without impacting the results. We suggest that a cultural algorithm approach can be used to identify these defining attributes that are most significant for node retrieval. These results can then be utilized within an existing vehicle assembly process planning application that utilizes a semantic network based knowledge base to improve the performance and reduce complexity of the network. 2. Performing Fault detection within a Complex Engineering Environment through the utilization of Chained Cultural Algorithms. David Ostrowski Science Laboratories Ford Motor Company Detroit Software testing is extremely difficult in the context of engineering applications. Biezer made a distinction between a functional approach to testing software as opposed to a structural approach. The functional approach is considered Black Box testing and the structural White Box testing. We feel that these methods are difficult to apply since they represent deterministic approaches to complex problems which have been known to be evaluated to NP-hard. Since constantly changing environments are heuristical in nature, we suggest that the application of the White and Black Box testing methods within a Cultural Algorithm environment will present a successful approach to fault detection. In order to utilize both a functional and structural approach, two Cultural Algorithms will be applied within this tool. The first algorithm will utilize the Black Box testing by establishing an equivalence class of data through the means of maintaining a belief space over a number of populations. the equivalence class will then be passed over to the second Cultural Algorithm that will apply program slicing techniques to determine program slices from the data established within the first algorithm. The goal will be to pinpoint faults within the program design. Through the searching of the program code this approach can be considered behavioral mining of a program. 3. Learning to Assess the Quality of Genetic Programs Using Cultural Algorithms George Cowan Park Davis Laboratories Michigan The total program length of Genetic Programming (GP) solution programs can be partitioned into effective program length and several types of excess code length for which, following Angeline, we use the term "bloat". Local Bloat is the excess code length attributable to local introns, sections of code in the solution program that do not affect the results for any training input. If the training data is highly representative of the problem domain, then it is likely that local intron removal does not affect the results for any possible input. Considering code after the removal of introns, we define Global Bloat to be the excess length of code which can be represented more succinctly while using the same operators that were allowed during evolution. Finally we distinguish Representational Bloat, the excess length, after the removal of Global Bloat, of code which can be represented more succinctly using operators from an expanded set. The remaining code is the Effective Program. We explore the relationships found between these types of code length and three GP process modifications purported to reduce bloat. 4. A Cultural Algorithm Application Robert G. Reynolds 5. Solving High Dimensional Real-valued Function Optimization Problems With Cultural Algorithms Chan – Jin Chung Lawrence Technological University Department of Math and Computer Science Southfield, Michigan 6. Learning to Identify Objects with Case-Injected Genetic Algorithms Sushil J. Louis Department of Computer Science University of Nevada Reno Reno, Nevada 89557-0148 Session: Particle Swarm Optimization – 1 Session Organizer: Russ Eberhart Director of the Advanced Vehicle Technology Institute Purdue School of Engineering and Technology at IUPUI 799 West Michigan Street Indianapolis, IN 46202, USA Phone: (317) 278-0255; Fax: (317) 274-4567 eberhart@engr.iupui.edu 1. Russ Eberhart 2. Jim Kennedy 3. Yuhui Shi 4. Pete Angeline Session: Multi - Objective Optimization With Evolutionary Algorithms – 2 Sessions Organizer: El - Ghazali Talbi University of Lille Laboratoire d'Informatique Fondamentale de Lille Bat. M3 59655 Villeneuve d'Aseq Cedex France Phone: 03 20 43 45 13; Fax: 03 20 43 65 66 talbi@lifl.fr http://www.lifl.fr/~talbi/cec99.html 1. An Updated Survey Of Evolutionary Multiobjective Optimization Techniques: State Of The Art And Future Trends Carlos A. Coello Coello Laboratorio Nacional de Informatica Avanzada (LANIA) P.O. Box 60326-394 Houston, Texas 77205 USA e-mail: ccoello@xalapa.lania.mx After using evolutionary techniques for single-objective optimization during more than two decades, the incorporation of more than one objective in the fitness function has finally become a popular area of research. As a consequence, many new evolutionary-based approaches and variations of existing techniques have been recently published in the technical literature. The purpose of this paper is to provide a critical review of the existing approaches, their limitations, advantages, and degree of applicability. Finally, the future trends in this discipline and some of the most promising areas of future research will be described. 2. Multicriteria Optimization and Decision Engineering of an Extrusion Process Aided by a Diploid Genetic Algorithm S. Massbeuf, C. Fonteix (1), L.N. Kiss (2), I. Marc, F. Pla and K. Zaras (1) Laboratoire des Sciences du Genie Chimique - UPR CNRS 6811 ENSAIA, 2 avenue de la foret de Haye, BP 172, 54505 Vandoeuvre-les-Nancy Cedex, France (2) Universite Laval - Faculte des Sciences de l'Administration Bureau 3202A/PAP, Sainte Foy G1K 7P4, Quebec, Canada 03 83 59 58 42, fax 03 83 59 58 04, perso 03 83 96 37 04 In many, if not most, optimization problems, industrials are often confronted with multiobjective decision problems. For example, in manufacturing processes, it can be necessary to optimize several criteria to take into account all the market constraints. So, the aim is to choose the best tradeoffs among all the defined and conflicting objectives. In multicriteria optimization, after the decision maker has chosen all his objectives, he has to determine the multicriteria optimal zone by using the concept of domination criterion called Pareto domination. Two points, in the research domain, are compared. If one is better for all attributes, it is a non dominated solution. All the non dominated points form the Pareto's region. Two multiobjective optimization algorithms are used to obtain the Pareto's zone. These methods are based on a diploid genetic algorithm and are compared on an industrial application : food's granulation. Some constraints are applicated to reduce the zone if some points are not interesting for the maker. In the remaining zone, the decision maker has to choose the best solution after he has made a ranking with all potential solutions. 3. Distributed Multi-Objective Optimisation for Complex Aerospace Engineering problems using Genetic Algorithms and Game Theory B. Mantel (Dassault-Aviation), J. Periaux (Dassault-Aviation) and M. Sefrioui (Univ. Paris 6, LIP6). Direction de la Prospective Computational Applied Mathematics periaux@rascasse.inria.fr 4. Multi-objective Genetic Programming For Dynamic Systems Modelling Katya Rodríguez-Vázquez and Peter J. Fleming Dept. of Automatic Control and Systems Engineering University of Sheffield, Mappin Street S1 3JD, U.K. e-mail: katya@acse.shef.ac.uk, P. Fleming@sheffield.ac.uk This work presents an investigation into the use of Genetic Programming (GP) applied to chaotic systems modelling. A difference equation model representation, the NARMAX (Non-Linear AutoRegressive Moving Average with eXtra inputs) model is proposed for being the basis of the hierarchical tree encoding in GP. Based upon this difference equation model representation and formulating the identification as a multiobjective optimisation problem, the well-known benchmark problem, the Chua's circuit, is studied. The formulation of the GP fitness function, multicriteria cost function based upon the definition of Paretooptimality, generated a set of non-dominated chaotic models. This approach considered criteria related to the complexity, performance and also statistical validation of the models in the fitness evaluation. The final set of non-dominated model solutions are able to capture the dynamic characteristics of the system and reproduce the chaotic motion of the double scroll attractor. 5. Assessing the Performance of Multiobjective Genetic Algorithms for Optimisation of a Batch Process Scheduling Problem K. J. Shaw, C. M. Fonseca [1], A. L. Nortcliffe, M. Thompson, J. Love, and P. J. Fleming. Department of Automatic Control & Systems Engineering, University of Sheffield, Mappin Street, S1 3JD, UK. Fax: +44 (0) 114 273 1729 Email: shaw@acse.shef.ac.uk [1] ADEEC/UCEH, Universidade do Algarve, Campus de Gambelas, 8000 Faro, Portugal Email: cmfonsec@ualg.pt Scheduling provides a source of many problems suitable for solution by genetic algorithms (Davis, 1985; Bagchi et, al, 1991; Cleveland and Smith, 1993; Lee et al., 1993; Shaw and Fleming, 1996; Löhl et al., 1998). The complexities, constraints and practicalities involved in the field motivate the development of genetic algorithm techniques to allow innovative and flexible scheduling solutions. Multiobjective genetic algorithms (MOGAs) extend the standard evolutionary-based optimisation genetic algorithm optimisation technique to allow individual treatment of several objectives simultaneously (Schaffer, 1985; Goldberg, 1989, Fonseca and Fleming, 1993, 1998, Horn, Nafpliotis, and Goldberg, 1994, Srivinas and Deb, 1995). This allows the user to attempt to optimise several conflicting objectives, and to explore the trade-offs, conflicts and constraints inherent in this process. The area of MOGA performance assessment and comparison is a relatively new field, as much research concentrates on applications rather than the theory behind MOGA capabilities (e.g., Fonseca and Fleming, 1996; Van Veldhuizen and Lamont, 1998; Zitzler and Thiele, 1998). However, the theoretical exploration of these issues can have tangible effects on the development of highly practical applications, such as the process plant scheduling system under development in this work (Shaw, et. al., 1999). By assessing and comparing their strengths, variations and limitations with a quantitative method, the user can develop a highly efficient MOGA to suit the application, and gain insight into behaviour the application itself. In this work, two MOGAs are implemented to solve a process scheduling optimisation problem. A quantitative comparison of their performances is conducted, and the results and implications of this comparison are then examined within the context of the problem. Finally, a discussion of how this work may be applied beyond this field, to many other types of multiobjective optimisation techniques and applications is provided. This demonstrates the utility of this study to many other MOGA users. 6. A Genetic Algorithm Approach to Multi-Objective Scheduling Problems with Earliness and Tardiness Penalties Hisashi Tamaki, Etsuo Nishino and Shigeo Abe. Dept. of Electrical and Electronics Engr. Kobe University Rokko-dai, Nada-ku, Kobe 657-8501, Japan Tel/Fax: +81-78-803-1063 tamaki@chevrolet.eedept.kobe-u.ac.jp This paper deals with identical parallel machine scheduling problems with two kinds of objective functions, i.e., both regular and non-regular objective functions, and proposes a genetic algorithm approach in which (a) the sequence of jobs on each machine as well as the assignment of jobs to machines are determined directly by referring to a string (genotype), and then (b) the start time of each job is fixed by solving the linear programming problem. As for (b), we newly introduce a method of representing the problem to determine the start time of each job as a linear programming problem whose objective function is formed as a weighted sum of the original multiple objective functions. This method enables to generate a lot of potential schedules. Moreover, through computational experiments by using our genetic algorithm approach, the effectiveness for generating a variety of Pareto-optimal schedules is investigated. 7. Cooperative Strategies for Solving the Bicriteria Sparse MultipleKnapsack Problem F. S. Salman, GSIA, CMU, Pittsburgh, PA 15213, fs2c@andrew.cmu.edu J. Kalagnanam, S. Murthy, IBM T.J. Watson Research Center, PO Box 218, Yorktown Hts, NY 10591 jayant@us.ibm.com, 914 945 1039 speaker: Jayant Kalagnanam The problem of assigning a given set of orders to the production units in the inventory arises frequently in production planning and scheduling. Typically, this problem can be formulated as a variation of the multiple knapsack problem where items (orders) are packed into knapsacks (production units). In our model we consider variations such as multiple objectives and additional assignment restrictions. We consider the objectives of both maximizing total amount of assignment and minimizing total waste due to unused knapsack capacity, as well as assignment restrictions which can be represented by sparse bipartite graphs. The focus of this paper is on obtaining non-dominated solutions in reasonably short computational time. This paper studies real instances from the steel industry which proved very hard to solve using conventional integer programming techniques. A cooperative organization of heuristics was used to solve these instances with surprising effectiveness. Such an approach can generalized for use for other hard optimization problems. 8. Multi-objective design of finite word-length controllers R.S.H. Istepanian Dept of Electrical and Computer Engineering, Ryerson Polytechnic University, 350 Victoria St, Toronto M5B 2K3, Canada. istepanr@ee.port.ac.uk J.F. Whidborne Division of Engineering, Kings College London, Strand, London WC2R 2LS, U.K. james.whidborne@kcl.ac.uk Feedback controller implementations with fixed-point arithmetic offer advantages of speed, memory space, cost and simplicity compared to floating-point arithmetic. Thus, to date, fixed-point processors dedicated for digital controller implementations are still the dominant architecture in many modern digital control engineering applications, particularly for automotive, consumer, military and medical applications. However, most control design methodologies assume that the designed controller will be implemented with infinite precision. Thus, when the controller is implemented, rounding errors in the controller parameters can cause a performance degradation and even instability. However, to directly design the controller with finite-precision control parameters can result in lower performance and stability robustness. Hence, there is a trade-off between the cost of the controller (in terms of word-length and memory requirements) and the performance of the system. This paper will present an approach to designing finite word-length feedback controllers by means of a multi-objective genetic algorithm. The approach provides a set of solutions that are (near) Pareto-optimal with respect to minimal word-length and memory requirements and to closed loop performance. The method is applied to the design of a finite word-length controller for the IFAC93 benchmark problem. Session: Quantum Computing – 2 Sessions Organizer: Ajit Narayanan Department of Computer Science University of Exeter Exeter EX4 4PT, UK ajit@dcs.exeter.ac.uk 1. Introduction to Quantum Mechanics Ajit Narayanan University of Exeter UK This talk will briefly introduce the basic concepts of quantum computing, such as superpositions and quantum Turing Machines, before describing Shor's quantum cryptanalysis algorithm. 2. Quantum Computing Richard Hughes Los Alamos National Laboratory NM This talk will bring the audience up-to-date on the most recent developments in quantum computing and will describe current interest and research at LANL, which included the use of trapped ions for undertaking quantum computation. 3. Quantum Genetic Programming Lee Spector Hampshire College Amherst, MA Methods for discovering algorithms for quantum computers are described where such methods are more efficient than any classical computer algorithms for the same problems. 4. Quantum Cryptography Anders Karlsson Royal Institute of Technology Sweden Current research in quantum information theory and cryptography will by described. 5. Quantum Algorithms for Search Lov Grover Bell Labs NJ Grover's algorithm for database search will be described. 6. Quantum Search For NP - Complete Problems Colin P. Williams Jet Propulsion Laboratory California Institute of Technology CA Current work based on Grover's algorithm for solving NP - complete problems through quantum search will be described. Session: Advanced Adaptive Algorithms for IE/OR Problems– 2 Sessions Organizers: Mitsuo Gen Intelligent Systems Engg. Lab. Dept. of Indust. & Inform. Systems Engg. Graduate School of Engineering Ashikaga Inst. of Technology Ashikaga, 326-8558 Japan Phone: +81(284)62-0605 ext. 376 or +81(284)62-2985 direct Fax: +81(284)64-1071 Email: gen@ashitech.ac.jp or gen@genlab.ashitech.ac.jp Gursel A. Suer Univ. of Puerto Rico - Mayaguez, P.R. a_suer@rumac.upr.clu.edu 1. Spanning Tree-based Genetic Algorithm for Bicriteria Fixed Charge Transportation Problem Mitsuo Gen, Yinzen Li, and Jong Ryul Kim Dept. of Indust. & Inform. Systems Engg. Ashikaga Inst. of Tech. Ashikaga 326-8558, Japan 2. Genetic Algorithm for Solving Bicriteria Network Topology Design Problem Jong Ryul Kim and Mitsuo Gen Dept. of Indust. & Inform. Systems Engg. Ashikaga Inst. of Tech. Ashikaga 326-8558, Japan 3. Genetic Algorithms in Lot Sizing Decisions William Hernandez and Gursel A. Suer Dept. of Industrial Engg. University of Puerto Rico Mayaguez, PR 00681, USA The work presented in this paper focuses on the application of genetic algorithms to obtain the order quantities for an uncapacitated, no shortages allowed, single-item, single-level lot sizing problem. The problem is tested under various conditions with different lot sizing and genetic algorithm parameters. The impact of scaling on the fitness function is also explored. The results indicate that genetic algorithms are promising for this area as well and more complex lot sizing problems can be tackled. 4. A Hybrid Approach of Genetic Algorithms and Local Optimizers in Cell Loading Gursel A. Suer *, Ramon Vazquez +, and Miguel Cortes ++ * Dept. of Industrial Engg. + Depts. of Electrical Engg. & Civil Engg. University of Puerto Rico-Mayaguez Mayaguez, PR 00681, USA and ++ Dept. of Science and Technology Interamerican University-Aguadilla Aguadilla, PR 00605, USA In this paper, we explore the potential application of evolutionary programming to cell loading problem. The objective is to minimize the number of tardy jobs. The proposed approach is a hybrid three phase approach; 1) Evolutionary programming is used to generate a job sequence, 2) Minimum load rule is applied to assign jobs to cells and 3) Moore's Algorithm is applied to each cell independently. The experimentation results show that the inclusion of the local optimizing procedure mentioned in step 3 accelarates reaching a good solution. 5. Adaptive Penalty Methods for Reliability Optimization of Series-Parallel Systems Using an Ant System Approach Yun-Chia Liang and Alice E. Smith Department of Industrial Engineering, University of Pittsburgh, Pittsburgh, PA 15261 aesmith@engrng.pitt.edu This paper solves the redundancy allocation problem of a series-parallel system by developing and demonstrating a problem-specific Ant System. The problem is to select components and redundancylevels to optimize some objective function, given system-level constraints on reliability, cost, and/or weight. The Ant System algorithm presented in this paper is combined with adaptive penalty methods to deal with the highly constrained problem. Experiments on well-known problems from the literature were used to test this approach. 6. A Hybrid Genetic Algorithm Approach for Backbone Design of Communication Networks Abdullah Konak and Alice E. Smith Department of Industrial Engineering, University of Pittsburgh, Pittsburgh, PA 15261 aesmith@engrng.pitt.edu This paper presents a hybrid approach of a genetic algorithm (GA) and local search algorithms for the backbone design of communication networks. The backbone network design problem is defined as finding the network topology minimizing the design/operating cost of a network under performance and survivability considerations. This problem is known to be NP-hard. In the hybrid approach, the local search algorithm efficiently improves the solutions in the population by using domain-specific information while the GA recombines good solutions in order to investigate different regions of the solution space. The results of the test problems show that the hybrid methodology improves on previous approaches.