From: FLAIRS-01 Proceedings. Copyright © 2001, AAAI (www.aaai.org). All rights reserved. Skill Refinement Through Competence Feedback David Patterson wd.patterson@ulst.ac.ukSchoolof Informationand SoftwareEngineering, University of Ulster, NorthernIreland BT370QB Sarabjot S. Anand ss.anand@ulst.ac.ukSchoolof Informationand SoftwareEngineering, University of Ulster, NorthernIreland BT370QB John Hughes jg.hughes@ulst.ac.ukSchoolof Informationand SoftwareEngineering, University of Ulster, NorthernIreland BT370QB Abstract Learningis generally performedin two stages, knowledge acquisition and skill refinement. Developments within machinelearning have tended to concentrate on knowledgeacquisition as opposedto skill refinement. In this paper we develop mechanismsfor skill refinementwithin the context of lazy learning by incorporating competencefeedback into the exemplar base. The extent of competencefeedback dependson the richness of the data collected during task execution. Wepresent two techniques for competencefeedback, exception spaces and knowledgeintensive exception spaces (KINS).The techniques differ in the extent of competencefeedback and the resulting degree of skill refinement. Previousdefinitions of exception spaces and KINSare extendedand the resulting improvementis evaluated using six data sets. A genetic algorithmis utilised to optimise the definitions of the exception spaces and KINSfor each exemplarin the exemplarbase. Wealso provide a visualisation of KINSand exceptionspaces with an aimto exemplify howthese mechanismsfor skill refinement affect the structure of the exemplarbase. 1 Introduction A general model of learning consists for four main components: the environment, knowledgebase, learning element and performance element [1]. The environment provides the inputs (raw data) into the leaming task through various sensors. The environment may also take on the role of teacher and provide the expected outcome for a set of inputs, during learning. Based on the inputs from the environment, the performance element is expected to perform a task of some description with some degree of competence using knowledge it has access to within the knowledgebase. The learning element is central to this system and aims to bridge the gap between the level of information available from the environment and that required by the performance element. Thus, the learning element converts inputs from the environment into knowledge stored within a predefined representation, a process known as knowledge acquisition. Based on the 394 FLAIRS-2001 amount and complexity of the computation that the learning task must undertake, it may be considered as a lazy or eager learner. The latter characteristically generalize/abstract their input values into more general representations such as rules or decision trees which are subsequently stored in the knowledge base and used by the performance element. On the other hand lazy learners, such as the nearest neighbor (k-NN), perform no processing of inputs initially and wait until a request is made for task execution. Whencalled upon to perform a task, the performance element utilises the previously stored inputs to perform the task. Apart from performing the task itself, the performance element must also perform the function of learning. A characteristic of the lazy learners is that any learning carried out during task execution is discarded after the task has been performed. Other characteristics include a large memoryrequirement, low training costs, efficient solution reuse, high perspicuity and inefficiency in knowledgedeployment. In addition to performing the task, whose performance enhancement is the goal of learning, the performance element may also provide feedback on its own competence. This feedback maybe utilised to refine the learnt knowledge, leading to improved competence and goaloriented learning, knownas skill refinement. For example, the backpropagation algorithm used in learning neural network weights uses competence (measured by the observed error in prediction/classification) as a driver of search within the hypothesis space. In comparison, decision tree induction algorithms do not utilise such feedback while performing a greedy search of the search space. The goal of this paper is to present two techniques for incorporating competencefeedback into a nearest neighbor model with the aim of improving its competence on future tasks. The rest of the paper is in the following format. The next section (Section 2) presents a summaryof previous related work that forms the motivation for the research preCopyright ©2001, AAAI. Allrights reserved. sented in this paper. In Section 3 we describe the implementation of competence feedback within the context of the k-NN algorithm using exception spaces and KINS. This is followed in Section 4, by a visual interpretation of KINSthat demonstrates how they generalize the definition of exception spaces and alter the structure of the exemplar base. The empirical evaluation of this implementation of competence feedback using six data sets from disparate domains is presented in Sections 5 and 6. Finally, Section 7 draws conclusions from the evaluation and suggests future work in the development of KINS and exception spaces. 2 Related work Whenpresented with a target example, the 1-Nearest Neighbour (I-NN) [3] retrieves the most similar exemplar from the exemplar base (based on the Euclidean distance metric) and allocates the class of the retrieved exemplar to the target example. The 1-NNhas been criticized on a number of accounts: expensive due to their large storage requirements, sensitive to the choice of similarity function, cannot easily work with missing attribute values, cannot easily work with nominal attributes and does not yield concise summaries of the concept [4]. The IB series [5] of enhancements to the basic nearest neighbour algorithm addressed a number of these criticisms, as did work by Cost and Salzberg [6]. Anand et al. [2] classify the extensions to the nearest neighbour along the following four dimensions: kdimension, attribute weight dimension, exemplar weight dimension, distance metric dimension. The development of exception spaces and KINS is along the exemplar weight dimension and thus we shall concentrate our discussion primarily on extensions proposed along this dimension. Cost and Salzberg [6] weighted an exemplar according to howreliable a classifier it proved itself to be. Thus, implementing the notion of competencefeedback for classification goals, resulting in the restructuring of the exemplar base. Goodclassifiers were assigned a small weight (close to 1) which had the effect of making them appear closer to a target exemplar and therefore more likely to be retrieved from the exemplar base. Poorer classifiers were assigned larger weights (>1) which has the opposite effect on them. The rationale behind this is that poor classifiers are either noise or exceptions within the exemplar base describing regions of the instance space where normal rules are not applicable. Geometrically exemplar weights define circular regions around exemplars, called exception spaces The size of the exception space is inversely proportional to the size of the weight. For an exemplar to be retrieved from the exemplar base when presented with a target exemplar, the target must fall within the region bounded by the exemplar’s exception space (see Section 4). In IB3, Aha et al. [5] first introduced the notion of competence feedback. The principle focus of this feedback was to make the algorithm more tolerant to noisy exemplars. IB3 applies a selective filter to prevent noisy exemplars from being used to misclassify new exemplars. Exemplarsare excluded based on their classification record. A poor classification record indicates that an exemplar’s most similar neighbours in the exemplar space have a different classification to itself and that it is therefore a noisy exemplar. Exemplarsare acceptable if their classification accuracy is significantly greater (statistically) than their class’s observed frequency and excluded if it is significantly less. System competence is improved through time by applying this filter through continual feedback. In this way an exemplar already included within the exemplar base can be excluded as the system changes dynamically with the addition of new exemplars. Note that as opposed to Cost and Salzberg’s exception spaces, exemplars within IB3 are either in the exemplar base or out, there is no grading based on competence. IB4, in addressing the issue of attribute relevancy under differing contexts, also uses competence feedback. The weights reflect the relevance of the attributes used to describe each exemplar. 3 Competence Feedback using Exception Spaces and Knowledge Intensive Exception Spaces Anandet al. [2] extended the exception space definition to apply not only to classification problems (Cost & Salzberg model) but also to regression problems through the use of a goodness membership function (GMF) which redefines the measure of usefulness (goodness) of an exemplar. The GMFprovides a methodology whereby exemplars can be weighted according to their competency despite having a continuous output field. The global weight of an exemplar is defined ultimately by a weighted quartile measure formula (see later), which uses the GMF,and is defined as the ratio of the uses of the exemplar to the correct uses of the exemplar. For classification problems an exemplar is said to have been retrieved correctly if it is used to classify a target problem into its correct class. For regression problemsit is a bit moredifficult to define. Cross validation using the training data and employing the nearest neighbor algorithm results in a distribution of predictive errors of individual exemplar retrievals which can be described using the statistical measure of spread namely quartiles. Those exemplars resulting in a low predictive error will fall into quartile I(Q1), those with the worst errors will fall into Q 4, with exemplars exhibiting errors between these extremes falling into Q 2 & Q3 respectively. A GMFis defined on the quartiles according to the degree to which a retrieval of an exemplar, producing an error in that quartile, is considered a good use MACHINELEARNING 39S of the exemplar. Therefore a GMFof {1.0, 0.7, 0.3, 0} implies that the retrieval of an exemplarwhenit produces an error falling into Q 1 is good, with a gradual deterioration in goodness downto Q 4 where the retrieval of an exemplar producing an error in this quartile is poor (a goodness membershipvalue of 0). As an exemplar can be retrieved a number of times throughout the cross validation process it may produce errors in more than one quartile. Therefore the weighted quartile measure formula is used to calculate its overall goodness and hence its global weight as shownbelow. 4 iE nx i=1 Wx 4 i=| where, P.i is the membershipvalue for quartile i as defined by th¢ GMF and n i is the no. of retrievals of x in quartile i x As defined by Cost and Salzberg, a small global weight has the effect of makingan exemplarappear closer to a target. A large global weight has the opposite effect and makesit appear to be further awayand therefore less likely to be retrieved. Theseweights can be visualized (section 4) circular boundaries around exemplars, knownas exception spaces. The radius of the exceptionspace is inversely proportional to the size of the global weight. Smallweights therefore produce large exception spaces. For an exemplar to be retrieved for a target the target mustfall into the exception space of the exemplar. Exemplarswith small exception spaces (defined by large global weights) are therefore less likely to be usedin retrieval. Exemplar # I I 2 3 3 Error Quartile I 3 I 3 4 Frequency. 3 2 4 4 I Table I Competence Feedbackdata used to define Exception Spaces The competencefeedbackdata utilized within the definition of exception spaces takes the form shownin Table 1. Here, examplefeedback on the competenceof the first three exemplars from one of the housing data sets (see Section 5 for a brief description) is shown.This data is interpreted as follows: Exemplar1 was retrieved five times in total. Of these, three times the error resulting from the retrieval was in the first quartile of the error distribution resulting from the use of the exemplarbase as a whole, while twice the retrieval resulted in errors in the third quartile of the error distribution. 396 FLAIRS-2001 Usingthe GMF defined previously of { 1.0, 0.7, 0.3, 0} and the weighted quartile measureformula this will producea global weight for exemplarl as follows. w] = 5/(3"1 + 2*0.3)= 1.38 If exemplar1 was retrieved again, this time producingan error in Q4its newglobal weight wouldbe wl -6/(3"1 + 2*0.3 + 1"0)= 1.67 In this wayfeedbackfromretrievals is used to constantly refine the size of the exceptionspaces and therefore the structure of the exemplarbase. While some exemplars always produce poor predictions (error in Q4)for a range of targets and others alwaysproduce good predictions (error in Q1), moreoften exemplars display a range of goodand bad usage for different target exemplars. That is, exemplarsare generally not universally goodor bad predictors. The reason for this mayeither be noise or that these differing results occur underdifferent retrieval circumstances. Onewayof describing the retrieval circumstancesis in terms of the individual attribute distanees betweenthe feature vectors describing the exemplar and target example.Oneshortfall in the reasoning behind exception spaces is that they do not take the circumstances of retrieval into consideration. That is, evenif an exemplar producesgood predictions sometimesit will be discriminated against by the times it producespoor predictions. WhatKINSattempts to accomplish is to removethis bias by attaching a weight to an exemplarwhich is determined by the circumstancesof its retrieval. Thereforeundercertain circumstances KINSwill attach a small weight to an exemplar and under other circumstances a larger weight. As more knowledgeis being used to define these new exception spaces, they are knownas KnowledgeINtensive exception Spaces (KINS). To enable this dynamicallocation weights, whichare dependenton the retrieval circumstances, additional competencefeedbackdata is generated for each exemplarduring cross validation including: 1) the individual attribute distances betweenthe exemplar and the target 2) the quartile into whichits error in prediction fell. Examplecompetencefeedback data for a single exemplaris shownin Table 2 which was generated using a data set from the colorectal cancer domain(see Section 5 for a brief description of the data set). Notethat only the last rowin Table 2 is used whendefining exception spaces. As can be seen from Table 2, KINSutilizes moredetailed competencefeedbackthan exception spaces. A classifier is built using this data for each exemplar.The individual attribute distances are the independentattributes that are input to the classifier and the error quartiles are the classification labels that the classifier must learn to discriminate between. For example let us assume that the classifier for a particular exemplar is defined as KINS1. This classifier maybe interpreted as follows: if the distance between the values of attribute Age of the target exemplar and the retrieved exemplar is less then 0.2 then assign a weighting of I(Q1). If the distance between them for attribute Age is >= 0.2 and the distance between them for attribute Dukes Stage < 0.6 then assign a weight of 2 (Q2). If the distances for Age is >= 0.2 and for Dukes Stage is >= 0.6 then assign a weight of 3 (Q3). the structure of the exemplar base. Note that using a rule based classifier is not necessary for defining KINS. Any type of classifier may be used. Weuse rules here as it makes the visualization of KINSmore intuitive. KINSI: if d(Age) <0.2 then weight = 1 else if d(Age) >=0.2 & d(Dukes Stage) < then weight = 2 else if d(Age) >=0.2 & d(Dukes Stage) >= 0.6 then weight =3 4 Attribute Sex Pathological Type Polarity Tubule Confi~uration Tumour Pattern Lymphocytic Infiltration Fibrosis Venous Invasion Mitotic Count Penetration Differentiation DukesStage A/~e Obstruction Site Error Quartile Use#1 0 0 Use#2 0 0 Use#3 0 0 Use#4 0 0 Use#S 0 0 0 0 0 0 0 0.111 0 0 0 0 0 0 0 0 0 0 0 0 0.25 0 0.25 0 0 0 0.25 0 0 0.56 0 0 0.027 0 0 0.062 0.0008 1 0 3 0.25 0.062 0 0.25 0.0008 0 0 4 0.25 0 0.25 0.062 0.02 0 0 3 0.11 0 0 0.062 0.016 0 0 3 0 0 0 0.062 0.058 0 0 3 Now, when a new target example is encountered the KINS rules for an exemplar are used to classify the current retrieval circumstances (individual attribute distances) into one of four error categories (quartiles). The exemplar weight associated with the relevant quartile is then used to compute the "true" distance of the target from the exemplar. Visualizing Exception Boundaries Exception spaces and KINS effectively incorporate competence feedback into the exemplar base by altering its structure. While exception spaces define circular boundaries (assuming a two dimensional space) around an exemplar, KINS define more complex spaces. As described earlier, in exception spaces the weight assigned to the exemplar defines the radius of the boundary using an inverse relationship whereby the larger the weight the smaller the boundary. Figure I shows the exception spaces defined for exemplar X using exemplar weights of 1,2 and 3. In this example, exemplar X will only be retrieved for a target, T, if the exemplar weight associated with X, at the time of retrieval, is less than or equal to 2. o Table 2 CompetenceFeedbackdata utilised by KINSshowing the difference in attribute values betweentarget exemplarsand a retrieved exemplar. KINS, as defined above, removes the bias introduced by exception spaces which associate only one global weight with an exemplar. The definition of K1NSis a dynamic process where the competency of the system improves over time through introspective feedback from exemplar retrievals. Each time an exemplar is retrieved, competence feedback data in the form shown in Table 2 is presented to the learning element and the classifier is used to refine the learned knowledge that defines the KINS. This dynamically modifies the shape of the exception boundary around an exemplar through introspective feedback, thus modifying Age Figure 1: Visualising Exception Spaces defined on exemplarX KINS enable a less rigid and more realistic boundary to be defined around exemplars as seen in Figures 2 and 3. To help visualize how KINS changes the structure of the exemplar base let us consider the previous KINS rule (KINS1) and the one defined as KINS2 both based on the coloreetal cancer data. For simplifying the discussion, we assume here that the only attributes found to be significant in defining the KINS are Age and Dukes Stage. KINS2: if d(Age) <0.2 then weight = 1 else if d(Age) > =0.2 & d(Dukes Stage) > then weight = 2 else if d(Age) >=0.2 & d(Dukes Stage) <= 0.6 then weight =3 MACHINE LEARNING397 KINS2can be interpreted in a similar way as KINS1. Note here that the weights used in defining KINSincrease as the estimated error increases. On the other hand, in the case of the GMF,as the defining weights are measures of goodness, they reduce as the estimated error ¯ increases. Nowlet us see how these two KINScan be visualized. Figures 2 and 3 show two exemplars, X and T. Here, X is an exemplar within the exemplar base and T is the target exemplar, for which a prediction is to be made. Weassume that KINSI and KINS2 are alternative KINS defined around exemplar X. The three concentric circles around exemplar X correspond to the exception boundaries around the exemplar, defined by the three weights l, 2 and 3 within the definitions of KINS1and KINS2.The smallest weighting of 1 corresponds to the largest circle. Notice that up until this point we have not considered antecedents of the rules in the KINSdefinition. The two rectangular areas bounding the exemplar correspond to the region on the Age axis where the distance between the target and retrieved examplesis less then 0.2 and on the Dukes Stage axis where the distance is less then 0.6. These areas correspond to the antecedent expressions in the KINSdefinitions. Age Figure2: Visualizing KINSIdefined on Exemplar X Figure 2 shows the exemplar with the regions defined by KINSI shaded while Figure 3 illustrates KINS2. If exemplar X had KINS1 defined on it, exemplar X would not be retrieved when attempting to make a prediction for the target exemplar as it lies outside the KINSboundary defined. On the other hand, as T falls within the KINS boundary region defined by KINS2, X would be retrieved in this case. Thus, KINSIand KINS2result in different restructuring of the original exemplarbase. Age Figure 3: Visualizing KINS2defined on Exemplar X 5 Competence Methodology Feedback Evaluation To evaluate the effectiveness of the proposed competence feedback mechanismswe empirically evaluate them using the followingsix data sets. 1. Housing1 consists of 565 records and I0 attributes taken from a housing database supplied by the Valuation and Lands Agencyof Northern Ireland. The goal is to build a modelfor predicting house price. 2. Housing2 consists of 584 records and 10 attributes taken from a different housingdatabase. It also is a regression problemaimedat predicting house price 3. Loansconsists of 430 records each with 23 attributes taken from a lending institution database. This is a binary classification learning problemof predicting whetherto give a customera loan or not. For the purposes of this paper we transformed the probleminto one of predicting the propensity of the applicant being a good investment. Thus the outcomeattribute was a continuousattribute with domain[0,1 ]. 4. Car Insuranceconsists of 600 records and 19 attributes. The aimis to predict the propensity of a customer to renewa ear insurance. 5. Breast Cancerconsists of 683 records and 9 attributes taken from at the UCIMachineLearning Repository [13]. The data was originally part of the Wisconsin Breast CancerDatabases.It is a binary classification learning problem,the aim being to predict whether a tumor is benign or malignant. Onceagain the problem was transformedto that of predicting the likelihood of the tumor being malignant. 6. Colorectal Cancerconsists of 188 records each with 15 attributes taken from a database of patients from the RoyalVictoria Hospital and the Belfast City Hospital, Belfast, NorthernIreland. This is a regression problem, the aim being to predict the numberof monthsthe patient is expected to survive after being diagnosedwith colorectal cancer. In the original definition of exception spaces for regression problems, Anand et al. suggested defining the GMF 398 FLAIRS-2001 using error quartiles [2]. The exemplar weight that defines the exception space around an exemplar was defined by the weighted quartile measure formula. A drawback of the initial GMF definition, (an integral part of the formula) was that the GMF(each of the ~i’s) is user fined. To rectify this we use a genetic algorithm to search the space of all possible GMF definitions. Additionally the original definition of exception spaces was based on quartiles of the error distribution. This was chosen arbitrarily as an initial basis of investigating the validity of the technique in extending the definition of exception spaces for regression models. In this paper we investigate the effects of extending this work through the implementation of a continuous definition of the GMF (which therefore no longer requires the weighted quartile measure formula to define exemplar weights). The continuous GMFis defined as: /a(e) emax - e ernax - emi n where, e is the actual error value and emax and emi n are the maximumand minimumerror produced by all retrievals This definition of the GMFis a simple linear function of the error due to exemplar retrieval. Wealso investigate the effects of dividing the error distribution into eight and twelve intervals to determine if this improves the model further. Clearly, the numberof error intervals on which the GMFis defined will define the complexity of the search space to be searched by the genetic algorithm. For example, if we assume the membership values, P-i, all have a precision of two decimal points, the size of the search space would be 100", where n is the number of intervals that the error distribution has been split into within the GMFdefinition. Thus, the more complex the GMFdefinition, the more computationally expensive will be the search undertaken by the genetic algorithm. Initial results reported by Anandet al. using KINSwere encouraging [2]. In this paper we also address some of the deficiencies within the preliminary definition of KINS. The deficiencies within KINSdefinition are the same as those with the GMFdefinition in relation to exception spaces. In the original definition of KINSthe quartile interval weights were assigned arbitrarily as was the choice of the numberof intervals into which the error distribution was split. In this paper we use a GAto optimise the interval weights and we investigate the effects on competency of splitting the error distribution into eight and twelve intervals instead of using quartiles. The numberof intervals into which the error predictions must be split clearly has a direct effect on the sparsity of the competence feedback data (Table 2) used to create the classification function that defines the KINSfor an exemplar. Increasing the numberof intervals would clearly have a negative influence on the accuracy of the discovered classifier as less data wouldbe available to achieve a useful generalisation. The improvement in competence of the nearest neighbor model when using KINS would be affected by the accuracy of the classifier defining the KINS. So any reduction in competence of the classifier will have a direct, derogatory impact on the value of using the KINS.Additionally, the complexity of the search for the optimal weights will also dependon this choice of numberof intervals, as in the ease of the GMF. 6 Results of Evaluation The k-NN algorithm [10] employed in the experiments used the Euclidean distance metric, retrieved five neighbors for each target and arrived at a prediction using a weighted sum of the output values for the retrieved neighbors. The weights used in the prediction were based on the distance of the neighbors from the target. To evaluate the competence improvement achieved by using exception spaces and KINS, we used three fold cross validation. Thus, three models were generated for each data set and the results shown in Table 3 are the average mean absolute errors for the three training and test data sets for each of the data sets. Three-fold cross validation was used to minimize any sampling errors in the reported results. To obtain optimal KINSand exception space definitions, a genetic algorithm was employed to search the search space of all GMFsand KINS. The genetic algorithm used a population size of 50, a crossover probability of 0.8, a mutation probability of 0.1, linear sealing of fitness, 2nd level of incest prohibition [11] (i.e. crossover between two chromosomesthat shared either parents or grandparents was prohibited) and uniform crossover. Selection was based on survival of the fittest using the stochastic remainder sampling with replacement method [ 12]. Tenfold cross-validation employing the kNNalgorithm on the training data, with the parameter settings discussed above, was used as the fitness evaluation function. The results presented here were the optimal solution found by the genetic algorithm in 50 generations. For the GMF,it clearly makes sense for the goodness value associated with intervals pertaining to large errors to be smaller than that associated with an interval pertaining to smaller errors. Thus, the genetic algorithm was provided with the domain knowledge that the optimal solution will be a monotonically decreasing function. For KINS,the exemplar weights increase as the error quartile increases. Thus, in this ease, the genetic algorithm was provided with the domain knowledge that the optimal solution will be a monotonicallyincreasing function. As can be seen from Table 3, competence feedback improves the competence of the k-NNmodel for five out of the six data sets. Also, in general, in cases where competence feedback produces worse results on the test data, it also does so on the training data set - clearly a desirable MACHINELEARNING 399 feature for any model. Twoexceptions to the rule are K.INSin the Housing1 data and exception spaces in the colorectal cancerdata set. Competence MAEon Data Set Feedback Training data Housing1 None 4751.76 Exception 4777.22 ~paces Continuous 4962.67 KINS 4550.34 MAEon ImproveTest ment on data Test data 5209.31 5393.93 -3.54% Exception Spaces KINS HousingI. ¯ 1,0,96~0.92,0~92.~(.’,.,. i:I~,:~’8~8~~~.. Housing 2 0.93,0.93,0.9,0.87 1.85,1.98,1.99,1.99 Loans ¯ 0.99, 0.99,0.:i,0 ) li~l~O~l!i’9~~i CarInsurance 0.95,0.95,0.89,0.15 1,1.14,1.2,1.61 Breast Cancer 0.88,0.75;0.75,0.g51:2~ i~-52~1~3,~(~f8 Colorectal 0.7,0.64,0.64,0.15 1.3,1.4,1.67,1.98 Cancer ¯ " .’: .~ ~... " ~ ’ :’ r,.: .... ,. "~ ",’;’~ Loans Breast Cancer C~tiil-tioUs 7.5’04"~3!-:~:’: KINS 7146;55 None 0.665 Exception 0.446 spaces Continuous 0.441 0.465 KINS 0.465 0.497 None 0.068 Exception 0.039 spaces 0.12 0.097 expected to suffer if the numberof intervals was increased due to the competencefeedback data becoming too sparse for buildingcompetentclassifiers for the individual exemplars.To test these hypothesesweused eight and twelve intervals on the colorectal cancer data. The results of these experimentsare summarizedin Table 5. As can be seen fromTable5, using eight or twelve intervals reduces the gain in competenceexperienced when usingonly four intervals. Table 4: OptimalGMF and KINSdefinitions for 4 Intervals 29.4% 24.5% 4 Intervals 8 Intervals 12 Intervals Test Train Test Train Test Train Exception 26.36 29.96 26.5430.03 28.51 30.79 Space KINS 21.35 27.96 21.63 28.08 21.36 28.84 Table5: Mean AbsoluteErrorwhenusingmorethan4 intervals for definingKINS andException Spaces One surprising outcomefrom the experiments was that the KINSmodel of competence feedback didn’t show consistent improvement over the exception spaces model. This is not an unreasonable assumptionto makeas KINS is utilising moredetailed knowledgeto calculate exemplar weights. Out of the five data sets wherecompetence feedback improved the competencyof the k-NNmodel KINSproduced modelsof similar competencyto exception spacesin three data sets, it faired worsein one data set and improvedthe competencyin one data set. Obviously there is an increase in the computationaloverheads Table3: Effect of Competence Feedback on modelcompetence associated with the KINSmodelwhichdecreases the efficiency of the system. This overhead is acceptable if Table 4 shows the optimal definitions for the GMF and there is an associated benefit in terms of competency.If KINSweights as discovered by the genetic algorithm. Interestingly, there is little variationin the weightsacross there is no benefit then the exercise is a futile one. There are a numberof possibilities, whichmayexplain this. The the four quartiles for the two housing data sets. This six data sets used to evaluate the KINSand exception wouldsuggest that the competencefeedbackdata in these space modelsof competencefeedback were quite small. sets could not improve much upon the competence This in turn wouldmeanthat during the initial ten-fold achieved by the unweightedk-NNmodel. This is clearly cross validation used to produce the competencefeedreflected within the results in Table3 for these two data back data, each exemplarwouldonly be retrieved a limsets. ited numberof times. Thefeedbackdata for an individual exemplaris used to define its KINSthroughthe building In Section 5 we discussed howincreasing the numberof of a classification function for the exemplar.This lack of intervals into whichthe initial error distribution is split affects the complexityof the search space for the genetic data mayhavelimited the effectiveness of the classifier to define accurate KINSfrom the introspective feedback algorithm. Wesuggested that the increased size of the search space mayresult in the genetic algorithmreturning process. Additionallythe choiceof classifier, particularly its ability to avoid over-fitting, mayalso havehadan efsub optimal solutions. Wealso noted that KINSwouldbe 4OO FLAIRS-2001 16.7% feet on the effectiveness of KINS.The results in Table 3 suggest that over-fining is occurring when using KINS, especially in the case of the Colorectal Cancer and Breast Cancerdata sets. Conclusions & future work 7 In this paper, we introduced two mechanisms for skill refinement through competence feedback within the context of lazy learning. These techniques, exception spaces and KINS,differ in the extent to which they use the available feedback data. The techniques are independent of the distance metric employed and can potentially be extended to handle structured representations of exemplars. The techniques have been validated using six data sets. There is significant improvement in competence of the exemplar base when using competence feedback. In this paper, we assumedall independent attributes to have equal relevance to the task of predicting their relevant output fields. However,Anandet al. have shown previously that KINSand exception spaces are effective techniques even whenusing attribute weightings [2]. A further extension to this evaluation of the two competence feedback models based on synthetic data sets will be carried out to attempt to characterize circumstances where KINSimproves over exception spaces and when it is not beneficial to use competencefeedback at all. Oneextension to this work is to investigate the effects of larger data sets on defining KINSto see if the provision of more examples to the classifier during training improves the KINSdefinition leading in turn to more competent results. Related to this is the need to investigate the use of different classifiers. A direct extension of the work presented in this paper is an investigation into whether the use of characteristics of the error distribution in defining the continuous GMFcan add value. Also, the effectiveness of the genetic algorithm’s parameters in improving its convergence characteristics needs to be undertaken. Techniques for reducing over-fitting by the genetic algorithm of the GMFand KINSdefinitions, to the training data also need to be investigated. The authors have previously suggested the use of a fitness function that takes into account not only the mean absolute error but also the variance of the model across the cross-validation folds [10] with a view to controlling over-fitting. This approach needs to be evaluated in this context. The competence feedback mechanisms defined in this paper assumethat the exemplars are represented as feature vectors and that the features are independent of each other. An extension to these mechanisms would be to incorporate structured exemplars and exemplars with interacting features [8]. matching case that cannot be adapted, adaptation knowledge should be considered during retrieval to choose the most closely matching case which can be adapted [9]. Therefore, to be used effectively in CBR, these competence feedback mechanisms must be extended to take adaptation knowledgeinto consideration within its definition. 8 References [1] Barr and Feigenbaum. The Handbook of Artificial Intelligence, vol. 3. pp 326-333, 1982. [2] S. S. Anand, D. Patterson, J. G. Hughes. Knowledge Intensive Exception Spaces, Proceedings of AAAI-98,pp 574-579 [3] Cover, T.; and Hart, P. Nearest Neighbour Pattern Classification, IEEETransactions on Information Theory, 13(1): 21-27, 1967. [4] L. Breiman, J.H. Friedman, R.A. Olshen, C.J. Stone. Classification and Regression Trees, Chapmanand Hall, 1990 [5] Aha, D.; and Kibler, D. Noise Tolerant instance-based learning algorithms. In Proceedings of the 1 l th International Joint Conference on Artificial Intelligence, 794799. MelnoPark, Calif., 1989 [6] Cost, S.; and Salzberg, S. A Weighted Nearest Neighbour Algorithm for Learning with Symbolic Features. MachineLearning 10: 57-78, 1993. [7] Clementine User Guide, Integral Solution Limited, 1998. [8] D. Aha, D. Wettschereck (ed.). Proceedings of the ECML-97 Workshop on Case-Based Learning: Beyond Classification of Feature Vectors, 1997 [9] Smyth, B. & Keane, M. T. (1998) Adaptation-Guided Retrieval: Questioning the similarity Assumptionin Reasoning. Artificial Intelligence, (102)2, pp. 249-293 [I0] S. S. Anand, J. G. Hughes. Hybrid Data Mining Systems: The Nest Generation, Proceedings of 2nd Pacific-Asia Conference in KnowledgeDiscovery and Data Mining, 1998. [11] R. Craighurst, W. Martin. Enhancing GAPerformance through Crossover Prohibitions Based on Ancestry, Proc. of the 6th International Conference on Genetic Algorithms and Applications, pp. 130 - 135, 1995. [12] M. Srinivas, L. M. Patnaik. Genetic Algorithms: A Survey, IEEEComputer,Vol 27, No. 6, June, 1994. [13] C. L. Blake, C. J. Merz. UCIRepository of machine learning databases, Irvine, CA:University of California, Department of Information and ComputerScience, 1998. The technologies developed within this paper have important implications to retrieval in case-base reasoning (CBR).As there is little advantage on retrieving a closely MACHINELEARNING 401