Modeling Mistakes 2008 Palisade Risk and Decision Analysis Conference Terry Reilly Babson College reilly@babson.edu History ¾ Advanced Decision Making ¾Over 250 students ¾Project Based ¾Decision Analysis ¾ Financial Modeling ¾Over 500 students ¾Project Based ¾Simulation Job Decisions Eye Surgery Marriage Children Location Sports Valuations Social Security Distillery Housing Getting Started ¾ Overwhelming ¾Scope of project ¾Scope of tools ¾ Drive for Perfection ¾Limits ‘playing around’ ¾Modeling by numbers ¾No Mistakes Frozen Modeling All Models are Wrong! You can never replicate the real system exactly. Modeling is not linear ¾Start simple ¾Add complexity/realism as you cycle through ¾ Ignore dependency relations at first ¾ If difficult to model, then Don’t! ¾Purposefully make mistakes/Demented What-If Analysis ¾Play, Play, Play If Wrong, Why Model? ¾ Models do not solve the question/problem. ¾ Models provide insights into the problem. ¾ Expected Values ¾ Distribution of payoffs/costs ¾ Nuanced Risk Analysis ¾ What-If Analysis ¾ Sensitivity Analysis Know Objectives Thoroughly understand what the model is to be used for and by whom. Purpose ¾ What questions are to answered? ¾ What measures are to used? ¾ What inputs will be available? ¾ Avoid Type III errors Stock Option Model: Keep stock or sell? (narrow) Balanced Portfolio (broad) Housing Model: Keep rental or sell? (narrow) Balanced Portfolio (broad) Know your Audience One of the most difficult concepts for quant heads. Who will be using model or reading model results? ¾ What is their sophistication? ¾ What do they want to know first? ¾ What measures do they understand? ¾ Do they hate or embrace uncertainty? ¾ Do they even understand uncertainty measures? This impacts not only the final product, but also how the model is constructed, e.g. distribution choice, output templates. Communicate Effectively One of the most important aspects of modeling, if not the most important. ¾Gulf between modeler and end user ¾ Knowledge of system vs. model ¾ Expected/desired output vs. actual ¾Use tools, but easy to overwhelm ¾ Graphs ¾ @RISK Templates ¾ Sliders People Abhor Uncertainty ¾ NIMBY ¾ Uncertainty ≠ Randomness ¾ Need to develop probabilistic thinking ¾ Compare to Worst-Case/Best-Case Analysis ¾ More Nuanced, e.g., leverage variability Analysts need not only understand probabilistic thinking, but how to communicate the analysis results to the end user in a meaningful way. Distribution Choice ¾ Step away from the normal ¾ Infinite Tails? ¾ What happens if output distribution is unusual? ¾ Taking the choice too seriously ¾ Fitting procedures tell us the best fits. ¾ Changing the choice easy (SA on distribution choice). Bidding Example Distribution for Profit 0.200 0.180 0.160 0.140 0.120 @RISK Student Version 0.100 For Academic Use Only 0.080 0.060 0.040 0.020 0.000 -15 -10 -5 0 5 10 90% 15 20 How We Think What is the next number in the following sequence? 2, 4, 6, ? Pot smokers are unmotivated and likely to commit crimes. Pot Smoker Upstanding Degenerate √ No Toker Disconfirming Evidence When working on a project, we look for confirming evidence and weigh anything that supports our ideas heavily. We tend to underweight disconfirming evidence, to the point of ignoring it. Contrary information is typically never sought out. Presumed Associations When assessing probabilities, we tend to think back to similar events and the easier it is to recall, the higher the assessed probability. (Availability Heuristic) ¾ Works well except it can lead us to overestimate the likelihood of vivid or recent events and underestimate the likelihood of more commonly occurring bland events. ¾ When assessing the likelihood of two events occurring, we tend to recall similar events occurring together. We forget that there are always at least three other combinations to think through. This fact is universally ignored. Overconfidence Others fail, but I won’t. Novices are boldly confident. When assessing a probability distribution, we tend to derive a too narrow range. With experience comes a more nuanced and complete understanding of what could go wrong and the range of possible results/outcomes. Limitations ¾ Understand your model’s limitations. ¾ Communicate the limitations. ¾ Remember, all models are wrong! Conclusion ¾ Knowing the decision maker’s objectives, what questions the model is to answer and why, the more focused the model and easier it is to communicate the results. Avoid Type III errors. ¾ Knowing your audience allows you to choose the appropriate output to report and how to interpret the output for the end user. ¾ Knowing the heuristics people tend to use helps you guide probability assessments around pitfalls and mistakes common to all of us.