Hieu Le Novelty search – a promising approach for constructing Ensemble system. Introduction In the field of Artificial Intelligence, Classification problem is one of the most popular problems with wide applications in various fields including healthcare and agriculture. One of the classic examples is probably disease detection. Given a report with set of values regarding a patient health (blood pressure, height, blood type etc.) detect whether the patient has a particular disease. More formally, a classification problem is given a set of values regarding some particular fields; determine whether the set of values (input) belongs to a particular class. The set of values is often denoted in vector form as <x1, x2, x3, … xn> where xi is a value of corresponding criteria (height, weight etc.). The classifier is trained beforehand with a set of training examples and returns a hypothesis/class about the input vector. Improving accuracy of the classifier is an interesting task which attracts a lot of researchers. One well-known approach is Ensemble methods. According to Dietterich (2000), Ensemble methods conduct a hypothesis based on average vote from an ensemble/set of classifiers. The idea behind ensemble methods is quite intuitive: an ensemble with many classifiers may have higher possibility of giving a correct hypothesis than a single classifier. Even when one classifier returns a wrong hypothesis, the ensemble may still give correct hypothesis if the majority of classifiers work correctly. However, there is also a chance that the majority if not all of the classifiers make the same mistakes, which make the ensemble methods fall short. This give the notion that classifiers should behaves differently to avoid the case which many classifiers make the same mistakes. Since we want an ensemble systems consisted of classifiers with different behaviors, an interesting question is how to construct such ensemble system. In other words, how should we pick classifiers such that they can form an ensemble with a very broad range of classifiers’ behaviors? This question will be my main research topic for the honor project under the guidance of professor Braught. Some methods have been applied to construct the desired ensemble systems such as bagging or ADABoost (Dietterich, 2000). On the other hand, the proposed method for this research question will be novelty search, which will be described in the next section. The goal of the research will be applying novelty search to the ensemble construction process and compared the result with other methods used to construct ensemble system. Method for research. Novelty search is a new approach to Evolutionary computation. The intuitive and widely applied approach is evolving toward the goal of the search space (Lehman, 2011). More specifically, with some possible next steps, the system will choose the step that is closest to the goal using an objective function. However, the evolving with objective approach can still fail if objective function deceptively chooses a step that seems closer to the goal but actually leads to local optima. Novelty search effectively avoids local optima because it doesn’t explicitly look for an object but instead look for behavioral novelty (Lehman, 2000). Novelty search looks for new behaviors by keeping a collection of known behaviors and comparing new behaviors with the known ones to find out the most radical behavior. According to Lehman (2000), a simple novelty metric is calculating the “average distance to the k-nearest neighbors” of a given point. The further one point is from other known behaviors, the more novel it is. This approach may seem less effective compared to objective search but experiments on problems with high level of deception (more likely to fall into local optima) have shown that novelty search can outperform objective search in some cases. With the promising result from other researches, this research project will attempt to apply novelty search to classifiers space to select classifiers that behave differently to construct ensemble system. References Dietterich, T. G. (2000). Ensemble Methods in Machine Learning. In J. Kittler and F. Roli (Ed.) First International Workshop on Multiple Classifier Systems, Lecture Notes in Computer Science (pp. 1-15). New York: Springer Verlag. Lehman, J. Stanley, K. (2011). Abandoning Objectives: Evolution through the Search for Novelty Alone. Evolutionary Computation (19):2, pages 189-223, Cambridge, MA: MIT Press Lehman, J. Stanley, K. (2010). Revising the Evolutionary Computation Abstraction: Minimal Criteria Novelty Search. Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2010). New York, NY:ACM Opitz, D. Shavlik, J (1996). Actively Searching for an Effective Neural-Network Ensemble. Connection Science, 8, pp. 337-353. Sallam, H, Regazzoni, C. (2008). Evolving neural networks ensembles NNEs. Cognitive Information Processing.