Classification Classification Classification Classification Features and labels (Green, 6, 4, 4.5) (Green, 7, 4.5, 5) (Red, 6, 3, 3.5) (Red, 4.5, 4, 4.5) (Yellow, 1.5, 8, 2) (Yellow, 1.5, 7, 2.5) Features and labels Objects Features (X) Labels (Y) Classification (Train) Objects Features (X) Labels (Y) Classification (Test) h(Red, 7, 4, 4.5) (Red, 7, 4, 4.5) ? Classification (Test) h(Red, 5, 3, 4.5) (Red, 5, 3, 4.5) ? Classification (Test) h(Green, 6, 4, 4.5) (Green, 6, 4, 4.5) ? Classification (Test) h(Yellow, 2, 10, 1.2) (Yellow, 2, 10, 1.2) ? Dictionary • Classification, Pattern recognition, Discrimination • X: Features, Explanatory variables, Covariates • Y: Labels, Response variables, Outcome • h(X) : Classifier, hypothesis Example • Cancer classification: – Y: the tumor is benign or malignant – X: information from the tumor as well as the patient, e.g. size of tumor, shape of tumor, other measurements, etc. – Objective: from what we observe, what is the probability that the tumor is malignant? Examples More science and technology applications: – handwritten zip code identification – drug discovery (to identify the biological activity of chemical compounds using features describing the chemical structures) – Gene expression analysis ( thousands of features with only dozens of observations) Support Vector Machine • If the data is linearly separable then there exist infinitely many solutions. • Which solution is the best? Support Vector Machine Choose the line farthest from both classes. i.e. Choose the line which has the maximum distance from the closest point. Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Support Vector Machine Kernel Methods Kernel Trick Observed and Feature Tasks • Supervised Learning: given examples of inputs and corresponding desired outputs, predict outputs on future inputs. e.g.: classification, regression • Unsupervised Learning: given only inputs, automatically discover representations, features, structure, etc. e.g.: clustering, dimensionality reduction, Feature extraction Example • Load the sample data, which includes digits 2 and 3 of 64 measurements on a sample of 400. load 2_3.mat • Extract appropriate features by PCA [u s v]=svd(X','econ'); • Create data data=u(:,1:2); • Create a new vector, groups, to classify data into two groups groups=[zeros(1,200) ones(1,200)]; • Randomly select training and test sets. [train, test] = crossvalind('holdOut',groups); Example • Train an SVM classifier using a linear kernel function and plot the grouped data. svmStruct = svmtrain(data(train,:),groups(train),'showplot',true); Example Example • Use the svmclassify function to classify the test set. classes = svmclassify(svmStruct,data(test,:),'showplot',true); Example Example • Use Kernel function svmStruct = svmtrain(data(train,:),groups(train),'showplot',true,'Kernel _Function','rbf') Example Example • Use the svmclassify function to classify the test set. classes = svmclassify(svmStruct,data(test,:),'showplot',true); Example