Machine Learning Methods for Data Modeling, Decision Support and Discovery Constantin Aliferis M.D., Ph.D., Ioannis Tsamardinos Ph.D. Discovery Systems Laboratory, Department of Biomedical Informatics, Vanderbilt University 2003 Fall AMIA Conference Tutorial 7-8 November 2003 1 Acknowledgments Alexander Statnikov for code and for putting together the Resource Web page and CD Yindalon Aphinyanaphongs for access to data and results for case study #1 Laura E. Brown for access to data and results for case study #4 Doug Hardin and Pierre Massion, collaborators in the original studies from which case #3 is inspired 2 Goal The purpose of this tutorial is: To help participants develop a solid understanding of some of the most useful machine learning methods. To give several examples of how these methods can be applied in practice, and To provide resources for expanding the knowledge gained in the tutorial. 3 Outline Part I: Overview and Foundations 1. Tutorial Overview and goals 2. Importance of Machine Learning for discovery and decision-support system construction 3. A framework for inductive Machine Learning 4. Generalization and Over-fitting 5. Quick review of data preparation and model evaluation 6. Families of methods a. Bayesian classifiers break b. Neural Networks (including relationship to regression) c. Support Vector Machines break 7. Quick Review of Additional families a. K-Nearest Neighborhs, b. Clustering, c. Decision Tree Induction, d. Genetic Algorithms 4 Outline (cont’d) Part II.More Advanced Methods and Case Studies 1. More Advanced Methods a. Causal Discovery methods using Causal Probabilistic Networks b. Feature selection break 2: Case Studies a. Categorizing text into content categories b. Recognising invasive breast cancer c. Building a diagnostic model from gene expression data & addressing the missing values problem break d. Discovery of causal structure using Causal Probabilistic Network induction 3. Conclusions and wrap-up a. Topics not covered b. Resources for machine learning c. Questions & feedback 5 Definitions & Importance of Machine Learning 6 What is Machine Learning (ML)? How is it different than Statistics and Data Mining? Machine Learning is the branch of Computer Science (Artificial Intelligence in particular) that studies systems that learn. Systems that learn = systems that improve their performance with experience. 7 What is Machine Learning (ML)? How is it different than Statistics and Data Mining? Typical tasks: image recognition, Diagnosis, elicitation of possible causal structure of problem domain, game playing, solving optimization problems, prediction of structure or function of biomolecules, text categorization, identification of relevant variables, etc. 8 Indicative Example applications of ML in Biomedicine Bioinformatics • Prediction of Protein Secondary Structure • Prediction of Signal Peptides • Gene Finding and Intron/Exon Splice Site Prediction • Diagnosis using cDNA and oligonucleotide array gene expression data • Identification of molecular subtypes of patients with various forms of cancer Clinical problem areas • Survival after Pneumonia (CAP) • Survival after Syncope • Diagnosis of Acute M.I. • Diagnosis of Prostate Cancer • Diagnosis of Breast Cancer • Prescription and monitoring in hemodialysis • Prediction of renal transplant graft failure 9 Importance of ML: Task Types Diagnosis (what is the most likely disease given a set of clinical findings?), Prognosis (what will be the outcome after a certain treatment has been given to a patient?), Treatment selection (what treatment to give to a specific patient?), Prevention (what is the likelihood that a specific patient will develop disease X if preventable risk factor Y is present?). ML has practically replaced Knowledge Acquisition for building Decision Support (“Expert”) Systems. 10 Importance of ML: Task Types (cont’d) Discovery – Feature selection (e.g., what is a minimal set of laboratory values needed for pneumonia diagnosis?); – Concept formation (e.g., what are patterns of genomic instability as measured by array CGH that constitute molecular subtypes of lung cancer capable of guiding development of new treatments?); – Feature construction (e.g., how can mass-spectrometry signals be decomposed into individual variables that are highly predictive for detection of cancer and can be traced back to individual proteins that may play important roles in carcinogensis?); information retrieval query construction (e.g., what are PubMed Mesh terms that predict with high sensitivity and specificity whether medical journals talk about treatment?); – Questions about function, interactions, and structure (e.g., how do genes and proteins regulate each other in the cells of lower and higher organisms? what is the most likely function of a protein given the sequence of its aminoacids?), etc. 11 What is Machine Learning (ML)? How is it different than Statistics and Data Mining? Broadly speaking ML, DM, and Statistics have similar goals (modeling for classification and hypothesis generation or testing). Statistics has traditionally emphasized models that can be solved analytically (for example various versions of the Generalized Linear Model – GLM). To achieve this both restrictions in the expressive power of models and their parametric distributions are heavily used. Data Mining emphasizes very large-scale data storage, integration, retrieval and analysis (typically the last one as a secondary focus). Machine Learning seeks to use computationally powerful approaches to learn very complex non- or quasi-parametric models of the data. Some of these models are closer to human representations of the problem domain per se (or of problem solving in the domain) 12 Importance of ML: Data Types and Volume Overwhelming production of data: – Bioinformatics (mass-throughput assays for gene expression, protein abundance, SNPs…) – Clinical Systems (EPR, POE) – Bibliographic collections – The Web: web pages, transaction records,… 13 Importance of ML: Reliance on Hard data and evidence Machine learning has become critical for Decision Support System Construction given extensive cognitive biases and the corresponding need to base MDSSs on hard scientific evidence and highquality data 14 Cognitive Biases Main thesis: – human cognitive abilities are tailored to support instinctive, reflexive, life-preserving reactions traced back in our evolution as species. They are not designed for rational, rigorous reasoning such as the reasoning needed in science and engineering. In other words, there is a disconnect between our innate cognitive ability and the complexity of reasoning tasks required by the explosive advances in science and technology in the last few hundred years. 15 But is the Cognitive Biases Thesis Correct? Psychology of Judgment and Decision Making (Plous) Tversky and Kahneman (Judgment under uncertainty: Heuristics and Biases) Methods of Influence (Cialdini) And highly-recommended supplementary information can be found in: Professional Judgment (Elstein) Institute of Medicine’s Report in Medical Errors (1999) 16 Tversky and Kahneman “Judgment under uncertainty: Heuristics and Biases” This work (a constellation of psychological studies converging to a description of human decision making under uncertainty) is very highly regarded and influential It was recently (2002) awarded the Nobel Prize of Economics. Main points: – – – People use a few simple heuristics when making judgments under uncertainty These heuristics sometimes are useful and other times lead to severe and systematic errors These heuristics are: representativeness, availability and anchoring 17 Representativeness E.g., : the probability P that patient X has disease D given that she has findings F is assessed by the similarity of X to a prototypical description of D (found in a textbook, or recalled from earlier practice and training). Why is this wrong? – – – – Reason #1: similarity ignores base-rate of D Reason #2: similarity ignores sample size Reason #3: similarity ignores predictability Reason #4: similarity is affected by redundant features 18 Availability • E.g., : the probability P that patient X with disease D given that she is given treatment T will become healthy is assessed by recalling such occurrences in one’s prior experience • Why is this wrong? – Reason #1: availability is influenced by familiarity – Reason #2: availability is influenced by salience – Reason #3: availability is influenced by elapsed time – Reason #4: availability is influenced by rate of abstract terms – Reason #5: availability is influenced by imaginability – Reason #6: availability is influenced by perceived association 19 Adjustment and Anchoring • E.g., : the probability P that patient X has disease D given that she has findings F is assessed by making an initial estimate P1 for findings F1 and updating it when new evidence F2, F3, …, and so on, becomes available. • What goes wrong? – Problem #1: initial estimate over-influences the final estimate – Problem #2: initial estimate is often based on quick and then extrapolated calculations – Problem #3: people overestimate the probability of conjunctive events – Problem #4: according to initial anchor, people’s predictions are calibrated differently 20 Additional • Methods of Influence (Cialdini, 1993): – Reciprocation – Commitment & Consistency – Social Proof – Liking – Authority – Scarcity • Professional Judgment (Dowie and Elstein 1988) • Institute of Medicine’s Report in Medical Errors (1999) 21 Putting MDSSs and Machine Learning in Historical Context 40s – Foundations of Formal Decision-Making Theory by VonNeuman and Morgerstern 50s – Ledley and Lusted lay out how logic and probabilistic reasoning can help in diagnosis and treatment selection in medicine 60s – Applications of Bayes theorem for diagnosis and treatment selection pioneered by Warner and DeDombal – Medline (NLM) Early 70s – Ad-hoc systems (Myers et al; Pauker et al) – Study of Cognitive Biases (Kahneman, Tversky) Late 70s – Rule-based systems (Buchanan & Shortliffe) 22 Milestones in MDSSs • 80s – Analysis of ad-hoc and RBSs (Heckerman et al.) – Bayesian Networks (Pearl, Cooper, Heckerman et al.) – Medical Decision Making as discipline (Pauker) – Literature-driven decision support (Renels & Shortliffe) • Early 90s – Web-enabled decision support & wide-spread information retrieval – Computational Causal Discovery (Pearl, Spirtes et al. Cooper et al.) – Sound re-formulation of very large ad-hoc systems (Shwe et al) – Analysis of Bayesian systems (Domingos et al, Henrion et al.) – Proliferation of focused Statistics and Machine Learning MDDSs – First-order Logics that combine classical FOL with probabilistic reasoning, causation and planning (Haddaway) 23 Milestones in MDSSs • Late 90s – Efficient Inference for very large probabilistic systems (Jordan et al) – Kernel-based methods for sample-efficient learning (Vapnik) – Evidence-Based Medicine (Haynes et al) • 21st Century – Diagnosis, Prognosis and Treatment selection (a.k.a. “Personalized medicine” or “Pharmacogenomics”) based on molecular information (proteomic spectra, gene expression arrays, SNPs) collected via massthroughput assaying technology, and modeleld using machine learning methods – Provide-order entry delivery of advanced decision support – Advanced representation, storage, retrieval and application of EBM information (guidelines, journals, meta-analyses, clinical bioinformatics models) 24 Importance of ML How often ML techniques are being used? #Articles in Medline (in parentheses last 2 years): – Artificial Intelligence: – Expert systems: – Neural Networks – Support Vector Machines – Clustering – Genetic Algorithms – Decision Trees – Bayesian (Belief) Networks – Bayes (Bayesian Statistics + Nets) Compare to: – Regression – Knowledge acquisition – Knowledge representation – 4 major Symbolic DSS (Internist-I, QMR, ILIAD, DxPlain) – Rule-based systems 12,441 2,271 5,403 163 17,937 2,798 4,958 1,627 4,369 (2,358) (121) (1,158) (121) (4,080) (969) (752) (585) (561) 164,305 310 227 145 (28,134) (56) (27) (10) 802 (151) 25 Importance of ML – Importance of ML becomes very evident in cases where: » data analysis is too time consuming (e.g., classify web pages or medline documents into content or quality categories) » There is little or no domain theory What is the diagnosis? Is this an early cancer? 26 A Framework for Inductive ML & Related Introductory Concepts 27 What is the difference between supervised and unsupervised ML methods? Supervised learning: - Give to the learning algorithm several instances of input-output pairs; the algorithm learns to predict the correct output that corresponds to some inputs (not only previously seen but also previously unseen ones (“generalization”)). - Example: show to learning algorithm patient cases (i.e., findings vector and a correct diagnosis for each case); then the algorithm induces a classifier that can classify a previously unseen patient to the correct diagnostic category given the findings observed in that patient) 28 Classification A B C D TRAIN INSTANCES CLASSIFIERINDUCTIVE ALGORITHM CLASSIFIER E APPLICATION INSTANCES A1, B1, C1, D1, E1 A2, B2, C2, D2, E2 An, Bn, Cn, Dn, En CLASSIFICATION PERFORMANCE 29 What is the difference between supervised and unsupervised ML methods? Unsupervised learning: - Discover the categories (or other structural properties of the domain) - Example: give the learning algorithm gene expression measurements of patients with Lung Cancer; the algorithm finds sub-types (“molecular profiles”) of patients that are very similar to each other, and different to the rest of the types. Or another algorithm may discover how various genes interact among themselves to determine development of cancer. 30 Discovery A B C D TRAIN INSTANCES E STRUCTUREINDUCTION ALGORITHM A B C E D A1, B1, C1, D1, E1 A2, B2, C2, D2, E2 PERFORMANCE An, Bn, Cn, Dn, En 31 What is the theoretical basis of supervised Inductive ML? Inductive Machine Learning algorithms can be designed and analyzed using the following framework: A language L in which we express models. The set of all possible models expressible in L constitutes our hypothesis space H A scoring metric M tells us how good is a particular model A search procedure S helps us identify the best model in H Space of all possible models x x x x x x x x x x x x xx Models in H 32 How can ML methods fail? Wrong language Bias: best model is not in H Example: we look for linear models, and the domain is non-linear x x x x x x x x x x x x x x Space of all possible models Models in H 33 How can ML methods fail? Search Failure: best model is in H but search fails to examine it Example we use a steepest-ascent search in a multi-modal fitness landscape Space of all possible models xx x x x x x x x x x x x x Models in H 34 How can ML methods fail? Metric Failure: best model is in H, search finds it but is deemed to be worse than an inferior model Example: we do not incorporate prior information for models in M Space of all possible models x x x x x x x x x x x x x Models in H 35 Generalization & Over-fitting 36 Generalization & Over-fitting It was mentioned previously that a good learning program learns something about the data beyond the specific cases that have been presented to it. Indeed, it is trivial to just store and retrieve the cases that have been seen in the past (“rote learning” implemented as a lookup table). This does not address the problem of how to handle new cases, however. In supervised learning we typically seek to minimize “i.i.d.” error, that is error over future cases (not used in training). Such cases contain both previously encountered as well as new cases. 37 Generalization & Over-fitting “i.i.d.” context = independently sampled and identically distributed problem instances. In other words, the training and application samples come from the same population (distribution) with identical probability to be selected for inclusion and this population/distribution is time-invariant. (Note: if not time invariant then by incorporating time as independent variable or by other appropriate transformations we restore the i.i.d. condition) 38 Generalization & Over-fitting Consider now the following simplified diagnostic classification problem: classify patients into cancer (red) versus normal (green) on the basis of the values of two lab test values (test1, test2) Test1 Test2 39 Generalization & Over-fitting The diagonal line represents a perfect classifier for this problem (do not worry for the time being how to mathematically represent or computationally implement the line – we will see how to do so in the Neural Network and Support Vector Machine segments): Test1 Test2 40 Generalization & Over-fitting Let’s solve the same problem from a small sample; one such possible small sample is: Test1 Test2 41 Generalization & Over-fitting We may be tempted to solve the problem with a fairly complicated line: Test1 Test2 42 Generalization & Over-fitting In which case we get several errors: Test1 Test2 43 Generalization & Over-fitting …whereas with a simpler line…: Test1 Test2 44 Generalization & Over-fitting …a much smaller error: Test1 Test2 45 Generalization & Over-fitting In general, over-fitting a model to the data means that instead of general properties of the population from which the data is sampled we learn idiosyncracies (i.e., non-representative properties) of the sample data. Over-fitting and poor generalization (i.e., the error in the overall population (“true error”) is large) are synonymous as long as we have learned the training data well (i.e., “small apparent error”). Over-fitting is not only affected by the “simplicity” of the classifier (e.g., straight vs wiggly line) but also by: – – – – the size of the sample, the complexity of the function we wish to learn from data, the amount of noise, and the number and nature (continuous discrete, ordered, distribution, etc.) of the variables. 46 Generalization & Over-fitting We wish to particularly emphasize the danger of grossly over-fitting the learning when the number of predictive variables is large relative to the available sample. Consider for example the following situation: Assume we have 5 binary predictors and two samples, and wish to classify instances into two classes The predictors can encode 2^5=32 possible distinct patterns. Assume all patterns are equally probable. Hence the chances of the two cases having different predictive patterns are 31/32=97%. Thus in 97% of our samples of size two, the five variables are sufficient to identify perfectly the case. Combine this with a powerful enough learning algorithm (i.e., one that can effectively associate any pattern with the desired outcome) and it follows that in 97% of samples, one gets optimal apparent error even when there is no relationship between the target variable and the predictive variables! 47 Generalization & Over-fitting This situation is particularly relevant in bioinformatics in which we routinely have >10,000 continuous variables, noise, and <500 samples. Under these conditions every training instance has almost always a unique value set of the predictive variables; thus if one is not careful, the learning algorithm can simply learn what amounts to a lookup table (i.e., by associating the unique predictor signature with the outcome of that case for every case). 48 Generalization & Over-fitting So how does one avoid over-fitting? Via a variety of approaches: – Use learning algorithms that intrinsically (by design) generalize well – Pursue simple (“highly biased”) classifiers for small samples – Choose unbiased and low-variance statistical estimators of the true error and employ them sparingly Very important rule: Estimate the performance (true error) of a model with data you did not use to construct the model 49 Generalization & Over-fitting Avoiding over-fitting will be a primary concern of ours in this tutorial We will outline here some specific cross-validation procedures and use them to build models in the case studies segment 50 Generalization & Over-fitting Hold-out cross-validation method: – Split data in Train and Test data – Learn with Train and estimate true error with Test test data train 51 Generalization & Over-fitting N-fold Cross-validation: – Split data in Train and Test data n times such that union of test sets is full data set – Learn with Train and estimate true error with Test in each split separately – Average test performance test test test test train train train train train train train test test 52 Generalization & Over-fitting Leave-one-out = n-fold C.V. where n is equal to the number of data instances 53 Generalization & Over-fitting Stratified (balanced) Cross-validation: An n-fold C.V. in which (by design) the target class has the same distribution as in the full dataset 54 Generalization & Over-fitting Nested Cross-validation: – Assume we wish to apply cross-validation to find the best parameter values for parameter C for a classifier from parameter value set [1,..,100]. – One way to use C.V. to select the best values for C is to apply the holdout method 100 times, one for each value of C and select the value that gives the best error in the test sample. – The problem with this approach is that the true error estimate is not reliable since it is produced by running the best model on a test set that was used to derive the best model. – In other words, our data used to estimate the true error can no longer be used to produce unbiased estimates since it also guided the selection of the model. 55 Generalization & Over-fitting solution: – Split the Train data into two (Traintrain and Validation), – Use the validation set to find the best parameters, – Use the test set to estimate the true error test Validation data Traintrain 56 Generalization & Over-fitting If the sample is small, the nesting can be repeated with different assignment of the test set (i.e., nested n-fold C.V.): Te data V TT V TT TT V TT TT V TT V V Te One can also nest LOO with n-fold C.V. or LOO with LOO 57 Generalization & Over-fitting Important notes: – Estimating the true error of the best model is a separate procedure than generating the best model; the former requires an additional layer of nesting our cross-validation – When there are several types of parameters to be selected (e.g., normalization, discretization, classifier parameters) one can: » do one n-fold cross-validation using the cartesian product of all parameters which uses more sample but yields more conservative true error estimates, or one can » nest the cross-validation to as many nesting levels as the number of distinct parameters that need optimization, which yields more unbiased true error estimates but uses less sample 58 Data Preparation 59 Data preparation Non-specific – Is the data lawfully in our disposal? – Are there issues that deal with patient privacy and confidentiality as well as intellectual property issues that need be resolved? – How were the data produced, by whom, when, with what purpose in mind? – Any known or plausible biases present? – References in the literature? – Is there a codebook with clear definitions of variables location, date of creation, method of creation, value list, value meaning, missing value codes and meanings, history of the database and its elements? 60 Data preparation Data specific – – – – Valid values? Variable distributions? Descriptive statistics? Imputation 61 Data preparation Learner specific – – – – – – – – De-noising Scaling/Normalization Discretization Transforming variable distributions Co-linearities Homoskedasticity Outliers Feature selection 62 Data preparation Task specific – Reconstruct hidden or distorted signals from observed ones – Infer presence of hidden variables, determine their cardinality and values – Stem, normalize, extract terms – Weight or Project variables 63 Evaluation Metrics 64 Evaluation Metrics T: test D: disease T+ T- D+ D- a b a+b c d c+d a+c b+d a+b+ c+d Accuracy (0/1 loss): Number of correct classifications Number of total classifications a+d a+b+c+d 65 Evaluation Metrics T: test D: disease T+ T- D+ D- a b a+b c d c+d a+c b+d a+b+ c+d Sensitivity: proportion of true positives identified by test a a+c 66 Evaluation Metrics T: test D: disease T+ T- D+ D- a b a+b c d c+d a+c b+d a+b+ c+d Specificity: proportion of true negatives identified by test d b+d 67 Evaluation Metrics T: test D: disease T+ T- D+ D- a b a+b c d c+d a+c b+d a+b+ c+d Positive predictive value (PPV): proportion of true positives over test positives a a+b 68 Evaluation Metrics T: test D: disease T+ T- D+ D- a b a+b c d c+d a+c b+d a+b+ c+d Negative predictive value (NPV): proportion of true negatives over test negatives d c+d 69 Evaluation Metrics Mean squared error (MSE) (“Quadratic loss”): |D| 1/|D| Si (predicted_value(i)-true_value(i)) 2 |D| = cardinality of test dataset Suitable for continuous outputs 70 Evaluation Metrics ROC area 1 Sensitivity 0 1 1-Specificity 71 Evaluation Metrics In information retrieval: – “Precision” is the name for “PPV” and – “Recall” is the name for “Sensitivity” 72 Evaluation Metrics Recall-precision curve (and area under it): 100% Recall 0 100% Precision 73 Bayesian Classifiers Note: we will be discussing Bayesian classifiers using the diagnostic context, (which in terms of applications and historical development of the related ideas is representative). However the ideas discussed readily translate to any type of learning for classification and concomitant decision support function. 74 Bayesian Classifiers Bayes’ Theorem (or formula) says that: P (D) * P(F| D) P (D | F) = P(F) Where: – P(D) is the probability of some disease D in the general population (i.e., before obtaining some evidence F), a.k.a. as “disease prior probability” – P(F) is the probability of some evidence in the form of findings such as lab tests, physical examination findings etc. – P(F | D) is the probability of the same findings given that someone has disease D – P(D | F) is the probability of disease D given that someone has the findings F (i.e., after obtaining some evidence F), a.k.a. as “disease posterior probability” 75 Bayesian Classifiers Since the most likely diagnosis is the one with the maximum a posteriori probability, Bayes’ formula allows one to solve the differential diagnosis problem, as well as as any classification learning problem that can be cast as supervised learning Indeed, in the sample limit, there cannot be a better way to infer the most likely diagnosis than Bayes’ theorem and thus it serves as the theoretical gold standard against which statistical and machine learning classifiers are measured in terms of true error. In that context it is referenced as the “Bayes Optimal Classifier” 76 Bayesian Classifiers Note that Bayes’ formula can be applied to diagnosis of multiple possibly inter-depended diseases and non-independent findings since where there is “F” one can place a vector of findings (e.g., F1+, F2-, F3-,…,Fn+) and where there is “D” one can put a vector of diseases (e.g., D1-, D2-, D3+,…,Dm+) . 77 Bayesian Classifiers Further, the intuitive interpretation of Bayes’ rule is that of updating belief about the patient’s true state: before seeing F we have some prior belief (measured as probability) that the patient has disease(s) D. After seeing F we update the prior belief (diagnosis) to reflect (incorporate) the new evidence F; the new belief is the posterior produced by Bayes’s rule 78 Bayesian Classifiers Unfortunately there is a significant drawback with straightforward Bayes rule: we need number of probabilities, storage and computational time that is exponential to the number of findings (i.e., |F|) and the number of diseases (i.e., |D|). This means that for any diagnostic or other classification problem of non-trivial size (measured in terms of |F| and |D|) straight Bayes is not feasible 79 Bayesian Classifiers This has led to a simplified version in which we disallow multiple diseases (i.e., require that the patient may have only one disease at a time) and we require that findings are independent conditioned on the disease states (note: this does not mean that the findings are independent in general, but rather, that they are conditionally independent). The combination of these two assumptions yields required number of probabilities, storage and computational time that is linear to the number of findings and the number of diseases. 80 Simple (a.k.a. “Naive”) Bayes Application of Bayes’ rule with the Mutual Disease Exclusivity assumption (MEE) and the Conditional Independence assumption (FCID) is known as “Simple Bayes’ Rule”, “Naïve Bayes”, or, rather non-tastefully, as “Idiot’s Bayes”. Simple Bayes can be implemented by plugging in the main formula: P(F | D) = P P(Fi | Dj) i,j and P( F) = S P(Fi, Dj) = S [P(Fi | Dj) * P(Dj)] i,j i,j where Fi is the ith (singular) finding and Dj the jth (singular) disease. Several other (mathematically) equivalent formulations exist using sensitivities and specificities, likelihood ratios or other convenient building blocks 81 Simple (a.k.a. “Naive”) Bayes Simple Bayes was applied very early (from the early 60’s and on) in Medical Informatics for diagnosis and optimal treatment selection as well as sequential testing. See for example the classic papers by Warner et al (1961), DeDombal (1972), Leaper (1972), Gorry and Barnett (1968) 82 Variants of Simple Bayes Since the MEE and FCID assumptions clearly are violated in many medical contexts, researchers early on sought to relax them and created modified Bayesian classifiers that approximated P(F |D) (Fryback 1978) or assumed independent diseases and multiple diagnoses (“Multi-membership model” of Ben-Basat, 1980). These models (and many others not mentioned here) have primarily historical significance currently, because: – (a) It was shown (1997, Domingos and Pazzani) that the MEE and FCID assumptions are not necessary but sufficient conditions for a wide variety of target functions under 0/1 loss – (b) Bayesian Networks were invented and as we will see next they allow flexible representation of dependencies so that parsimony and tractability is maintained without compromising soundness 83 Bayesian Networks 84 Bayesian Networks: Overview A Note On Terminology Brief Historical Perspective The Bayesian Network Model and Its Uses Learning BNs Reference & Resources 85 Bayesian Networks: A Note On Terminology Bayesian Networks (or “Nets”): generic name Belief Networks: subjective probability-based, non-causal Causal Probabilistic Networks: frequentist probability-based, causal 86 Bayesian Networks: A Note On Terminology Various other names for special model classes: – Influence Diagrams (Howard and Mathesson): incorporate decision and utility nodes. Used for decision analyses – Dynamic Bayesian Networks (Dagum et al.): temporal semantics. Used as alternatives to multivariate time series models and dynamic control – Markov Decision Processes (Dean et al.): for decision policy formulation in temporally-evolving domains – Modifiable Temporal Belief Networks (Aliferis et al.): for wellstructured and very large problem models that involve time and causation and cannot be stored explicitly 87 Bayesian Networks: Historical Perspective Naïve Bayesian Model (mutually exclusive diseases, findings independent given diseases) predominant model for medical decision support systems in the 60’s and early 70’s because it requires linear number of parameters and computational steps (to total findings and diseases) Theorem 1 (Minsky, Peot): Naïve Bayes heuristic usefulness (expected classification performance) over all domains gets exponentially worse as number of variables increases Theorem 2 (see Mitchell): Full Bayesian classifier=perfect classifier However FBC impractical and serves as analytical tool only 88 Bayesian Networks: Historical Perspective In the late 70’s and up to mid-80’s this led to: Production Systems (i.e., rule-based systems, that is simplifications of first-order logic). The most influential version of PSs (Shortliffe, Buchanan) handled uncertainty through a modular account of subjective belief (the Certainty Factor Calculus) Theorem 3 (Heckerman): The CFC is inconsistent with probability theory unless rule-space search graph is a tree. Consequently, forward and backward reasoning cannot be combined in a CFC PS and still produce valid results 89 Bayesian Networks: Historical Perspective That led to research (late 80s) in Bayesian Networks which can vary expressiveness between the full dependency (or even the full Bayesian classifier) and the Naïve Bayes model (Pearl, Cooper) Variables Conditionally Independent Given Categories & Bayesian Networks Variables Conditionally Dependent Categories Mutually Exclusive 90 Bayesian Networks: Historical Perspective In the early 90’s researchers developed the first algorithms for learning BNs from data (Herskovits, Cooper, Heckerman) In the mid 90’s researchers (Spirtes, Glymour, Sheines, Pearl, Verma) discovered methods to learn CPNs from observational data(!). We will cover the foundations of this in the causal discovery segment. Overall BNs is the brain child of computer scientists, medical informaticians, artificial intelligence researchers, and industrial engineers and is considered to be the representation language of choice for most biomedical Decision Support Systems today 91 Bayesian Networks: The Bayesian Network Model and Its Uses BN=Graph (Variables (nodes), dependencies (arcs)) + Joint Probability Distribution + Markov Property Graph has to be DAG (directed acyclic) in the standard BN model A B C P(A+, P(A+, P(A+, P(A+, P(A-, P(A-, P(A-, P(A-, JPD B+, C+)=0.006 B+, C-)=0.014 B-, C+)=0.054 B-, C-)=0.126 B+, C+)=0.240 B+, C-)=0.160 B-, C+)=0.240 B-, C-)=0.160 Theorem 4 (Neapolitan): any JPD can be represented in BN form 92 Bayesian Networks: The Bayesian Network Model and Its Uses Markov Property: the probability distribution of any node N given its parents P is independent of any subset of the non-descendent nodes W of N A e.g., : B C D ^ {B,C,E,F,G | A} D F ^ {A,D,E,F,G,H,I,J | B, C } E F G I H J 93 Bayesian Networks: The Bayesian Network Model and Its Uses Theorem 5 (Pearl): the Markov property enables us to decompose (factor) the joint probability distribution into a product of prior and conditional probability distributions The original JPD: P(A+, B+, C+)=0.006 P(A+, B+, C-)=0.014 A P(A+, B-, C+)=0.054 P(A+, B-, C-)=0.126 P(A-, B+, C+)=0.240 Up to P(A-, B+, C-)=0.160 P(A-, B-, C+)=0.240 Exponential B C P(A-, B-, C-)=0.160 Saving in P(V) = P p(V |Pa(V )) i i i Becomes: P(A+)=0.8 P(B+ | A+)=0.1 P(B+ | A-)=0.5 P(C+ | A+)=0.3 P(C+ | A-)=0.6 Number of Parameters! 94 Bayesian Networks: The Bayesian Network Model and Its Uses BNs can help us learn causal relationships without doing experiments! Smoking (unmeasured) Lung Ca Heart Disease Lung Ca Heart Disease But Fisher says these two causal graphs are not distinguishable without doing an experiment (!?) 95 Bayesian Networks: The Bayesian Network Model and Its Uses BNs can help us learn causal relationships without doing experiments! Smoking (unmeasured) Family Hx Lung Ca Family Hx Lung Ca Diet Heart Disease Diet Heart Disease Fisher is right of course; however if we know a cause of each variable of interest then, in many cases, we can derive causal associations without an experiment 96 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property captures causality: – Revealing confounders Smoking Lung Ca Heart Disease 97 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property captures causality: – Modeling “explaining away” – Modeling/understanding selection bias Lung Ca Tuberculosis Haemoptysis 98 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property captures causality: – Modeling causal pathways Smoking Lung Ca Haemoptysis 99 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property captures causality: – Manipulation in the presence of confounders Smoking Lung Ca (target) Effective Manipulation! Heart Disease Ineffective Manipulation! 100 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property captures causality: – Manipulation in the presence of selection bias Lung Ca (target) Tuberculosis Ineffective Manipulation! Haemoptysis 101 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property captures causality: – Identifying targets for manipulation in causal chains G1 Ineffective Manipulation once we set G2 G2! More Effective Manipulation than manipulating G1! Disease (target) 102 Bayesian Networks: The Bayesian Network Model and Its Uses Once we have a BN model of some domain we can ask questions: A • Forward: P(D+,I-| A+)=? • Backward: P(A+| C+, D+)=? B C D • Forward & Backward: P(D+,C-| I+, E+)=? E F G I H • Arbitrary abstraction/Arbitrary predictors/predicted variables J 103 Bayesian Networks: The Bayesian Network Model and Its Uses The Markov property tells us which variables are important to predict a variable (Markov Blanket), thus providing a principled way to reduce variable dimensionality A B E C F D G I H J 104 Bayesian Networks: The Bayesian Network Model and Its Uses BNs can serve as sound (i.e., non-heuristic) alternatives to associative (i.e., non-similarity-based) clustering A B D I C E J F G L H K M N O P 105 Bayesian Networks: Demonstration of Flexible Representation A BN in which FCID holds D1 F1 D2 F2 D3 F3 F4 106 Bayesian Networks: Demonstration of Flexible Representation A BN in which MEE holds D1 F1 D2 F2 D3 F3 F4 107 Bayesian Networks: Demonstration of Flexible Representation A BN in which MEE and FCID hold D1 F1 D2 F2 D3 F3 F4 108 Bayesian Networks: Demonstration of Flexible Representation Hybrid assumptions D1 F1 D2 D3 F2 F3 F5 F6 F4 D2 F7 D3 F8 F9 109 Bayesian Networks: Practical Considerations Learning time tractability (as in genomic datasets) – – – – How good is learning with BNs? “Sparse candidate” algorithm Learn partial models Divide and conquer Inference tractability Learning sample size tractability & parameter estimation 110 Bayesian Networks: How Good Is learning? In discovering causal structure (Aliferis and Cooper, simulated data): – – – – K2 algorithm discovers >70% of arcs 94% of the time 94% of the time K2 does not add more than 10% superfluous arcs Mean correctly identied arcs=94% Mean superfluous arcs=4.7% In predicting outcomes & reducing number of predictors (Cooper, Aliferis et al., M. Fine pneumonia PORT data): K2 and Tetrad algorithms almost as good as best algorithm for domain, but requiring 6 instead of >200 variables 111 Pathfinder Heckerman et al early 90s Diagnosis and test selection of lymph-node pathology Assumes MEE but not FICD Similarity networks (special enhancement to BNs) allow more efficient Knowledge acquisition Myopic test selection strategy (similar to Gorry and Barnett) & combined monetary cost/expected survival utility measure Led to Intellipath commercial product Reference: Heckerman DE, Horvitz EJ, Nathwani BN. Toward normative expert systems: Part I. The Pathfinder project. Methods Inf Med. 1992 Jun;31(2):90-105. Heckerman DE, Nathwani BN. Toward normative expert systems: Part II. Probability-based representations for efficient knowledge acquisition and inference.Methods Inf Med. 1992 Jun;31(2):106-16. Heckerman DE, Nathwani BN. An evaluation of the diagnostic accuracy of Pathfinder. Comput Biomed Res. 1992 Feb;25(1):56-74. 112 QMR-DT Stanford, late 80s, early 90s Probabilistic formulation of QMR KB (subsequent version of INTERNIST I) Full-scope of INTERNIST/QMR Uses – two-layered BN representation, – No MEE or FICD assumptions – Stochastic inference Reference: Shwe M, Cooper G. An empirical analysis of likelihood-weighting simulation on a large, multiply connected medical belief network. Comput Biomed Res. 1991 Oct;24(5):45375. 113 Automatic Construction of Bayesian Networks from Data For causal discovery: – Perl, Verma (1988) – Spirtes, Glymour, Scheines, (1991) For classification/automatic DSS construction – Herskovits, Cooper (1991): Kutato (entropy-based) – Cooper, Herskovits (1992): K2 (Bayesian) [to be discussed at length in second part] Reference: Computation, Causation, and Discovery by Clark Glymour (Preface), Gregory F. Cooper (Editor), 2000, AAAI Press 114 Inference Algorithms Exact – Lauritzen & Spigelhalter – Cooper: Recursive decomposition Stochastic-approximate – Likelihood weighting – Dagum and Luby Variational (approximate but not stochastic) – Jordan et al. (1998): solves queries in QMR-DT in seconds Reference: An Introduction to Variational Methods for Graphical Methods (1998) Michael I. Jordan, Zoubin Ghahramani, Tommi S. Jaakkola, Lawrence K. Saul. Machine Learning An Optimal Approximation Algorithm For Bayesian Inference (1997) Paul Dagum, Michael Luby. Artificial Intelligence Probabilistic Reasoning in Expert Systems: Theory and Algorithms by Richard E. Neapolitan. Kohn Wiley 1990 115 Theoretical Complexity Inference is NP-hard (Cooper (exact, 1990) Dagum and Luby (stochastic, 1993)) Learning is NP-hard (Chickering, 1994, Bukaert, 1995) However: Many widely-applicable algorithms are very efficient (allowing up to thousands of variables for inference and up to >100,000 variables for focused learning) 116 Simple Bayes Revisited Domingos and pazzani 1997: – Naïve Bayes assumptions are sufficient for accurate probability estimates in the sample limit but not necessary for a wide variety of learning problems when accuracy is the evaluation metric – In small samples even if the assumptions are violated SB can do better than more expressive representations due to the biasvariance decomposition of the error – The best way to correct (extend) SB is not to join highly-associated findings These results explain the excellent performance of SB in text categorization with thousands of variables (words) and many other learning/inference tasks even against more expressive representations Reference: On the Optimality of the Simple Bayesian Classifier under Zero-One Loss (1997) Pedro Domingos and Michael Pazzani. Machine Learning 117 Analysis of sensitivity of BNs to errors in probability specification Henrion et al. 1996 – System: CPCS (subset of QMR) – Results: average probabilities assigned to the actual diseases showed small sensitivity even to large amounts of noise. – Explanation: One reason is that the criterion for performance is average probability of the true hypotheses, which is insensitive to symmetric noise distributions. But, even asymmetric, logodds-normal noise has modest effects. A second reason is that the gold-standard posterior probabilities are often near zero or one, and are little disturbed by noise. Reference: Max Henrion, Malcolm Pradhan, Brendan Del Favero, Kurt Huang, Gregory Provan and Paul O'Rorke. Why is Diagnosis Using Belief Networks Insensitive to Imprecision in Probabilities? UAI, 1996 118 Temporal Causal and Spatial Reasoning with Probabilistic methods Haddaway 1995: temporal, causal and probabilistic FOL Aliferis (97, 98): temporal and causal Bayesian Networks with clear causal-temporal semantics Spatio-temporal BNs for GI endoscopy Reference: Aliferis CF, Cooper GF. Temporal representation design principles: an assessment in the domain of liver transplantation. Proc AMIA Symp. 1998;:170-4. Ngo L, Haddawy P, Krieger RA, Helwig J. Efficient temporal probabilistic reasoning via context-sensitive model construction. Comput Biol Med. 1997 Sep;27(5):453-76. 119 Dynamic construction of BNs from Knowledge Bases to solve problem instances of interest (KBMC) Haddaway 1995 probabilistic FOL KBBN Aliferis: Modifiable Temporal BNs: temporal and causal Bayesian Networks with adjustable hybrid granularities, variable time horizon, and interacting subnetworks (“contexts”) 1996-8 Koller et al. object-oriented BNs, 1997 Reference: Generating Bayesian Networks from Probability Logic Knowledge Bases P. Haddawy , Proceedings of the Tenth Conference on Uncertainty in Artificial Intelligence, July, 1994. Daphne Koller and Avi Pfeffer. Object-Oriented Bayesian Networks , UAI, 1997 Aliferis C.F. A Temporal representation and Reasoning Model for Medical Decision Support Systems, Doctoral Thesis, 1998 120 Other applications Parsing of natural language with BNs – Haug et al 1999 – Charniak et al. Extensive applications for classification and discovery Margaritis et al (1999), Aliferis & Tsamardinos et al. (2001,2,3): – focused causal discovery (parents-children or Markov Blankets) – feature selection [to be discussed at length in second part] Reference: Fiszman M, Chapman WW, Evans SR, Haug PJ. Automatic identification of pneumonia related concepts on chest x-ray reports.Proc AMIA Symp. 1999;:67-71. Charniak E. Bayesian Networks Without Tears. AI Magazine 1991 121 Myths Surrounding Bayesian Decision Support In General and Classifiers in Particular Bayes’ Theorem requires that diseases are mutually exclusive and that findings are independent Corollary: Simple Bayes is wrong whenever the MEE and FICD assumptions do not hold Related: A good way to fix Simple Bayes is to “lamp together” highly-dependent findings Bayes’ probabilities are difficult to assess and because errors accumulate the final conclusions are wrong Related: To apply Bayesian inference you need an astronomical number of probabilities Related: Bayesian inference is too subjective because probabilities are not frequency-based 122 Conclusions Since many of the previous advances are very recent, the medical informatics community that does not specialize in probabilistic systems has not yet caught up with them Dissemination issues aside, the significant progress that has been accomplished in the theory and technology of Bayesian Networks in the 1990s has yielded: – Algorithms that allow efficient learning and inference with thousands of variables without unrealistic assumptions – Formal handling of temporal and causal reasoning – Decision-theory and probability theory compliant decision making – Well-understood properties – A plethora of tools for representation, inference , learning, and experimentation – Pioneering applications in many medical areas 123 Bayesian Networks: Sparse Candidate Algorithm Repeat Select candidate parents Ci for each variable Xi Set new best NW B to be Gn s.t. Gn maximizes a Bayesian score Score(G|D) where G is a member of class of BNs for which: PaG(Xi) PaBprev(Xi) " Xi Restriction Step Maximization Step Until Convergence Return B 124 Bayesian Networks: Sparse Candidate Algorithm SCA proceeds by selecting up to k candidate parents for each variable on the basis of pair-wise association Then search is performed for a best network within the space defined by the union of all potential parents identified in the previous step The procedure is iterated by feeding the parents in the currently best network to the restriction step Theorem 6 (Friedman) : SCA monotonically improves the quality of the examined networks Convergence criterion: no gain in score, and maximum number of cycles with no improvement in score 125 Bayesian Networks: Learning Partial Models Partial model: feature (Friedman et al.) Examples: – Order relation (is X an ascendant or descendent of Y?) – Markov Blanket membership (Is A in the MB of B?) We want: P(f(G|D) = S (f(G) * p(G|D)) G And we approximate it by: Conf(f) = 1 m S f(Gi) m * i=1 126 Bayesian Networks: Reference Simple Bayes weakness: – M. Peot, Proc. Proc. UAI 96 – M. Minsky, Transactions of IRE, 49:8-30, 1961 Simple Bayes application: – H. Warner et al. Annals of NYAS, 115:2-16, 1964 – F. de Dombal et al. BMJ, 1:376-380, 1972 Full Bayesian Classifier: – T. Mitchell, Machine Learning, McGraw Hill, 1997 Bayesian Networks as a knowledge representation: – J. Pearl, Probabilistic Reasoning in Expert Systems, Morgan Kaufmann, 1988 Certainty Factor/PSs weaknesses: – D. Heckerman et al., Proc. UAI 86 127 Bayesian Networks: Reference Causal discovery using BNs: – P. Spirtes et al. , Causation, Prediction and Search, MIT Press 2000 – C. Glymour, G. Cooper, Computation, Causation and Discovery, AAAI Press/MIT Press, 1999 – C. Aliferis, G. Cooper, Proc. UAI 94 Textbooks on BNs: – R. Neapolitan, Probabilistic Reasoning in Expert Systems, John Wiley, 1990 – F. Jensen, An Introduction to Bayesian Networks, UCL Press, 1996 – E. Castillo, et al. Expert Systems and Probabilistic Network Models, Springer 1997 Learning BNs: – – – – – G. Cooper et al. Machine Learning 9:309-347, 1992 E. Herskovits, Report No. STAN-CS-91-1367 (Thesis) D. Heckerman, Technical report Msr TR-95-06, 1995 J. Pearl, Causality, Gambridge University Press, 2001 N. Friedman et al. J Comput Biol, 7(3/4):601-620, 2000, and Proc. UAI 99 Comparison to other learning algorithms: – G. Cooper, C. Aliferis et al. Artificial Intelligence in Medicine, 9:107-138, 1997 128