k k k k k Principles of Econometrics Fifth Edition k R. CARTER HILL Louisiana State University WILLIAM E. GRIFFITHS University of Melbourne GUAY C. LIM University of Melbourne k k k EDITORIAL DIRECTOR EXECUTIVE EDITOR CHANNEL MARKETING MANAGER CONTENT ENABLEMENT SENIOR MANAGER CONTENT MANAGEMENT DIRECTOR CONTENT MANAGER SENIOR CONTENT SPECIALIST PRODUCTION EDITOR COVER PHOTO CREDIT Michael McDonald Lise Johnson Michele Szczesniak Leah Michael Lisa Wojcik Nichole Urban Nicole Repasky Abidha Sulaiman © liuzishan/iStockphoto This book was set in STIX-Regular 10/12pt by SPi Global and printed and bound by Strategic Content Imaging. This book is printed on acid free paper. ∞ Founded in 1807, John Wiley & Sons, Inc. has been a valued source of knowledge and understanding for more than 200 years, helping people around the world meet their needs and fulfill their aspirations. Our company is built on a foundation of principles that include responsibility to the communities we serve and where we live and work. In 2008, we launched a Corporate Citizenship Initiative, a global effort to address the environmental, social, economic, and ethical challenges we face in our business. Among the issues we are addressing are carbon impact, paper specifications and procurement, ethical conduct within our business and among our vendors, and community and charitable support. For more information, please visit our website: www.wiley.com/go/citizenship. Copyright © 2018, 2011 and 2007 John Wiley & Sons, Inc. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923 (Web site: www.copyright.com). Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030-5774, (201) 748-6011, fax (201) 748-6008, or online at: www.wiley.com/go/permissions. k Evaluation copies are provided to qualified academics and professionals for review purposes only, for use in their courses during the next academic year. These copies are licensed and may not be sold or transferred to a third party. Upon completion of the review period, please return the evaluation copy to Wiley. Return instructions and a free of charge return shipping label are available at: www.wiley.com/go/returnlabel. If you have chosen to adopt this textbook for use in your course, please accept this book as your complimentary desk copy. Outside of the United States, please contact your local sales representative. ISBN: 9781118452271 (PBK) ISBN: 9781119320951 (EVALC) Library of Congress Cataloging-in-Publication Data Names: Hill, R. Carter, author. | Griffiths, William E., author. | Lim, G. C. (Guay C.), author. Title: Principles of econometrics / R. Carter Hill, Louisiana State University, William E. Griffiths, University of Melbourne, Guay C. Lim, University of Melbourn. Description: Fifth Edition. | Hoboken : Wiley, 2017. | Revised edition of the authors’ Principles of econometrics, c2011. | Includes bibliographical references and index. | Identifiers: LCCN 2017043094 (print) | LCCN 2017056927 (ebook) | ISBN 9781119342854 (pdf) | ISBN 9781119320944 (epub) | ISBN 9781118452271 (paperback) | ISBN 9781119320951 (eval copy) Subjects: LCSH: Econometrics. | BISAC: BUSINESS & ECONOMICS / Econometrics. Classification: LCC HB139 (ebook) | LCC HB139 .H548 2018 (print) | DDC 330.01/5195—dc23 LC record available at https://lccn.loc.gov/2017043094 The inside back cover will contain printing identification and country of origin if omitted from this page. In addition, if the ISBN on the back cover differs from the ISBN on this page, the one on the back cover is correct. 10 9 8 7 6 5 4 3 2 1 k k k Carter Hill dedicates this work to his wife, Melissa Waters Bill Griffiths dedicates this work to Jill, David, and Wendy Griffiths Guay Lim dedicates this work to Tony Meagher k k k k Brief Contents PREFACE v LIST OF EXAMPLES 1 An Introduction to Econometrics Probability Primer 2 3 4 10 Endogenous Regressors and Moment-Based Estimation 481 11 Simultaneous Equations Models 531 12 Regression with Time-Series Data: Nonstationary Variables 563 13 Vector Error Correction and Vector Autoregressive Models 597 14 Time-Varying Volatility and ARCH Models 614 15 Panel Data Models 634 16 Qualitative and Limited Dependent Variable Models 681 xxi 15 The Simple Linear Regression Model 46 Interval Estimation and Hypothesis Testing 112 Prediction, Goodness-of-Fit, and Modeling Issues 152 1 5 The Multiple Regression Model 196 6 Further Inference in the Multiple Regression Model 260 APPENDIX A Mathematical Tools 7 Using Indicator Variables 317 APPENDIXB Probability Concepts 8 Heteroskedasticity 368 APPENDIXC Review of Statistical Inference 9 Regression with Time-Series Data: Stationary Variables 417 k APPENDIXD Statistical Tables INDEX iv k 869 k 748 862 768 812 k Preface Principles of Econometrics, Fifth Edition, is an introductory book for undergraduate students in economics and finance, as well as first-year graduate students in economics, finance, accounting, agricultural economics, marketing, public policy, sociology, law, forestry, and political science. We assume that students have taken courses in the principles of economics and elementary statistics. Matrix algebra is not used, and we introduce and develop calculus concepts in an Appendix. The title Principles of Econometrics emphasizes our belief that econometrics should be part of the economics curriculum, in the same way as the principles of microeconomics and the principles of macroeconomics. Those who have been studying and teaching econometrics as long as we have will remember that Principles of Econometrics was the title that Henri Theil used for his 1971 classic, which was also published by John Wiley & Sons. Our choice of the same title is not intended to signal that our book is similar in level and content. Theil’s work was, and remains, a unique treatise on advanced graduate level econometrics. Our book is an introductory level econometrics text. Book Objectives Principles of Econometrics is designed to give students an understanding of why econometrics is necessary and to provide them with a working knowledge of basic econometric tools so that k i. They can apply these tools to modeling, estimation, inference, and forecasting in the context of real-world economic problems. ii. They can evaluate critically the results and conclusions from others who use basic econometric tools. iii. They have a foundation and understanding for further study of econometrics. iv. They have an appreciation of the range of more advanced techniques that exist and that may be covered in later econometric courses. k The book is neither an econometrics cookbook nor is it in a theorem-proof format. It emphasizes motivation, understanding, and implementation. Motivation is achieved by introducing very simple economic models and asking economic questions that the student can answer. Understanding is aided by lucid description of techniques, clear interpretation, and appropriate applications. Learning is reinforced by doing, with clear worked examples in the text and exercises at the end of each chapter. Overview of Contents This fifth edition is a major revision in format and content. The chapters contain core material and exercises, while appendices contain more advanced material. Chapter examples are now identified and separated from other content so that they may be easily referenced. From the beginning, we recognize the observational nature of most economic data and modify modeling assumptions accordingly. Chapter 1 introduces econometrics and gives general guidelines for writing an empirical research paper and locating economic data sources. The Probability Primer preceding Chapter 2 summarizes essential properties of random variables and their probability distributions and reviews summation notation. The simple linear regression model is covered in Chapters 2–4, while the multiple regression model is treated in Chapters 5–7. Chapters 8 and 9 introduce econometric problems that are unique to cross-sectional data (heteroskedasticity) and time-series data k v k vi Preface (dynamic models), respectively. Chapters 10 and 11 deal with endogenous regressors, the failure of least squares when a regressor is endogenous, and instrumental variables estimation, first in the general case, and then in the simultaneous equations model. In Chapter 12, the analysis of time-series data is extended to discussions of nonstationarity and cointegration. Chapter 13 introduces econometric issues specific to two special time-series models, the vector error correction, and vector autoregressive models, while Chapter 14 considers the analysis of volatility in data and the ARCH model. In Chapters 15 and 16, we introduce microeconometric models for panel data and qualitative and limited dependent variables. In appendices A, B, and C, we introduce math, probability, and statistical inference concepts that are used in the book. Summary of Changes and New Material This edition includes a great deal of new material, including new examples and exercises using real data and some significant reorganizations. In this edition, we number examples for easy reference and offer 25–30 new exercises in each chapter. Important new features include k • Chapter 1 includes a discussion of data types and sources of economic data on the Internet. Tips on writing a research paper are given “up front” so that students can form ideas for a paper as the course develops. • A Probability Primer precedes Chapter 2. This Primer reviews the concepts of random variables and how probabilities are calculated from probability density functions. Mathematical expectation and rules of expected values are summarized for discrete random variables. These rules are applied to develop the concepts of variance and covariance. Calculations of probabilities using the normal distribution are illustrated. New material includes sections on conditional expectation, conditional variance, iterated expectations, and the bivariate normal distribution. • Chapter 2 now starts with a discussion of causality. We define the population regression function and discuss exogeneity in considerable detail. The properties of the ordinary least squares (OLS) estimator are examined within the framework of the new assumptions. New appendices have been added on the independent variable, covering the various assumptions that might be made about the sampling process, derivations of the properties of the OLS estimator, and Monte Carlo experiments to numerically illustrate estimator properties. • In Chapter 3, we note that hypothesis test mechanics remain the same under the revised assumptions because test statistics are “pivotal,” meaning that their distributions under the null hypothesis do not depend on the data. In appendices, we add an extended discussion of test behavior under the alternative, introduce the noncentral t-distribution, and illustrate test power. We also include new Monte Carlo experiments illustrating test properties when the explanatory variable is random. • Chapter 4 discusses in detail nonlinear relationships such as the log-log, log-linear, linear-log, and polynomial models. We have expanded the discussion of diagnostic residual plots and added sections on identifying influential observations. The familiar concepts of compound interest are used to motivate several log-linear models. We add an appendix on the concept of mean squared error and the minimum mean squared error predictor. • Chapter 5 introduces multiple regression in the random-x framework. The Frisch–Waugh– Lovell (FWL) theorem is introduced as a way to help understand interpretation of the multiple regression model and used throughout the remainder of the book. Discussions of the properties of the OLS estimator, and interval estimates and t-tests, are updated. The large sample properties of the OLS estimator, and the delta method, are now introduced within the chapter rather than an appendix. Appendices provide further discussion and Monte Carlo k k k Preface vii properties to illustrate the delta method. We provide a new appendix on bootstrapping and its uses. • Chapter 6 adds a new section on large sample tests. We explain the use of control variables and the difference between causal and predictive models. We revise the discussion of collinearity and include a discussion of influential observations. We introduce nonlinear regression models and nonlinear least squares algorithms are discussed. Appendices are added to discuss the statistical power of F-tests and further uses of the Frisch–Waugh–Lovell theorem. • Chapter 7 now includes an extensive section on treatment effects and causal modeling in Rubin’s potential outcomes framework. We explain and illustrate the interesting regression discontinuity design. An appendix includes a discussion of the important “overlap” assumption. k • Chapter 8 has been reorganized so that the heteroskedasticity robust variance of the OLS estimator appears before testing. We add a section on how model specification can ameliorate heteroskedasticity in some applications. We add appendices to explain the properties of the OLS residuals and another to explain alternative robust sandwich variance estimators. We present Monte Carlo experiments to illustrate the differences. • Chapter 9 has been reorganized and streamlined. The initial section introduces the different ways that dynamic elements can be added to the regression model. These include using finite lag models, infinite lag models, and autoregressive errors. We carefully discuss autocorrelations, including testing for autocorrelation and representing autocorrelations using a correlogram. After introducing the concepts of stationarity and weak dependence, we discuss the general notions of forecasting and forecast intervals in the context of autoregressive distributed lag (ARDL) models. Following these introductory concepts, there are details of estimating and using alternative models, covering such topics as choosing lag lengths, testing for Granger causality, the Lagrange multiplier test for serial correlation, and using models for policy analysis. We provide very specific sets of assumptions for time-series regression models and outline how heteroskedastic and autocorrelation consistent, robust, standard errors are used. We discuss generalized least squares estimation of a time-series regression model and its relation to nonlinear least squares regression. A detailed discussion of the infinite lag model and how to use multiplier analysis is provided. An appendix contains details of the Durbin–Watson test. • Chapter 10 on endogeneity problems has been streamlined because the concept of random explanatory variables is now introduced much earlier in the book. We provide further analysis of weak instruments and how weak instruments adversely affect the precision of IV estimation. The details of the Hausman test are now included in the chapter. • Chapter 11 now adds Klein’s Model I as an example. • Chapter 12 includes more details of deterministic trends and unit roots. The section on unit root testing has been restructured so that each Dickey–Fuller test is more fully explained and illustrated with an example. Numerical examples of ARDL models with nonstationary variables that are, and are not, cointegrated have been added. • The data in Chapter 13 have been updated and new exercises added. • Chapter 14 mentions further extensions of ARCH volatility models. • Chapter 15 has been restructured to give priority to how panel data can be used to cope with the endogeneity caused by unobserved heterogeneity. We introduce the advantages of having panel data using the first difference estimator, and then discuss the within/fixed effects estimator. We provide an extended discussion of cluster robust standard errors in both the OLS and fixed effects model. We discuss the Mundlak version of the Hausman test for endogeneity. We give brief mention to how to extend the use of panel data in several ways. k k k viii Preface • The Chapter 16 discussion of binary choice models is reorganized and expanded. It now includes brief discussions of advanced topics such as binary choice models with endogenous explanatory variables and binary choice models with panel data. We add new appendices on random utility models and latent variable models. • Appendix A includes new sections on second derivatives and finding maxima and minima of univariate and bivariate functions. • Appendix B includes new material on conditional expectations and conditional variances, including several useful decompositions. We include new sections on truncated random variables, including the truncated normal and Poisson distributions. To facilitate discussions of test power, we have new sections on the noncentral t-distribution, the noncentral Chi-square distribution, and the noncentral F-distribution. We have included an expanded new section on the log-normal distribution. • Appendix C content does not change a great deal, but 20 new exercises are included. • Statistical Tables for the Standard Normal cumulative distribution function, the t-distribution and Chi-square distribution critical values for selected percentiles, the F-distribution critical values for the 95th and 99th percentiles, and the Standard Normal density function values appear in Appendix D. • A useful “cheat sheet” of essential formulas is provided at the authors’ website, www.principlesofeconometrics.com, rather than inside the covers as in the previous edition. k For Instructors: Suggested Course Plans Principles of Econometrics, Fifth Edition is suitable for one or two semester courses at the undergraduate or first year graduate level. Some suitable plans for alternative courses are as follows: • One-semester survey course: Sections P.1–P.6.2 and P.7; Sections 2.1–2.9; Chapters 3 and 4; Sections 5.1–5.6; Sections 6.1–6.5; Sections 7.1–7.3; Sections 8.1–8.4 and 8.6; Sections 9.1–9.4.2 and 9.5–9.5.1. • One-semester survey course enhancements for Master’s or Ph.D.: Include Appendices for Chapters 2–9. • Two-semester survey second course, cross-section emphasis: Section P.6; Section 2.10; Section 5.7; Section 6.6; Sections 7.4–7.6; Sections 8.5 and 8.6.3–8.6.5; Sections 10.1–10.4; Sections 15.1–15.4; Sections 16.1–16.2 and 16.6; • Two-semester survey second course, time series emphasis: Section P.6; Section 2.10; Section 5.7; Section 6.6; Sections 7.4–7.6; Sections 8.5 and 8.6.3–8.6.5; Section 9.5; Sections 10.1–10.4; Sections 12.1–12.5; Sections 13.1–13.5; Sections 14.1–14.4; • Two-semester survey course enhancements for Master’s or Ph.D.: Include Appendices from Chapters 10, Chapter 11, Appendices 15A–15B, Sections 16.3–16.5 and 16.7, Appendices 16A–16D, Book Appendices B and C. Computer Supplement Books There are several computer supplements to Principles of Econometrics, Fifth Edition. The supplements are not versions of the text and cannot substitute for the text. They use the examples in the text as a vehicle for learning the software. We show how to use the software to get the answers for each example in the text. k k k Preface k • Using EViews for the Principles of Econometrics, Fifth Edition, by William E. Griffiths, R. Carter Hill, and Guay C. Lim [ISBN 9781118469842]. This supplementary book presents the EViews 10 [www.eviews.com] software commands required for the examples in Principles of Econometrics in a clear and concise way. It includes many illustrations that are student friendly. It is useful not only for students and instructors who will be using this software as part of their econometrics course but also for those who wish to learn how to use EViews. • Using Stata for the Principles of Econometrics, Fifth Edition, by Lee C. Adkins and R. Carter Hill [ISBN 9781118469873]. This supplementary book presents the Stata 15.0 [www.stata.com] software commands required for the examples in Principles of Econometrics. It is useful not only for students and instructors who will be using this software as part of their econometrics course but also for those who wish to learn how to use Stata. Screen shots illustrate the use of Stata’s drop-down menus. Stata commands are explained and the use of “do-files” illustrated. • Using SAS for the Principles of Econometrics, Fifth Edition, by Randall C. Campbell and R. Carter Hill [ISBN 9781118469880]. This supplementary book gives SAS 9.4 [www.sas .com] software commands for econometric tasks, following the general outline of Principles of Econometrics, Fifth Edition. It includes enough background material on econometrics so that instructors using any textbook can easily use this book as a supplement. The volume spans several levels of econometrics. It is suitable for undergraduate students who will use “canned” SAS statistical procedures, and for graduate students who will use advanced procedures as well as direct programming in SAS’s matrix language; the latter is discussed in chapter appendices. • Using Excel for Principles of Econometrics, Fifth Edition, by Genevieve Briand and R. Carter Hill [ISBN 9781118469835]. This supplement explains how to use Excel to reproduce most of the examples in Principles of Econometrics. Detailed instructions and screen shots are provided explaining both the computations and clarifying the operations of Excel. Templates are developed for common tasks. • Using GRETL for Principles of Econometrics, Fifth Edition, by Lee C. Adkins. This free supplement, readable using Adobe Acrobat, explains how to use the freely available statistical software GRETL (download from http://gretl.sourceforge.net). Professor Adkins explains in detail, and using screen shots, how to use GRETL to replicate the examples in Principles of Econometrics. The manual is freely available at www.learneconometrics.com/gretl.html. • Using R for Principles of Econometrics, Fifth Edition, by Constantin Colonescu and R. Carter Hill. This free supplement, readable using Adobe Acrobat, explains how to use the freely available statistical software R (download from https://www.r-project.org/). The supplement explains in detail, and using screen shots, how to use R to replicate the examples in Principles of Econometrics, Fifth Edition. The manual is freely available at https://bookdown.org/ccolonescu/RPOE5/. Data Files Data files for the book are provided in a variety of formats at the book website www.wiley.com /college/hill. These include • ASCII format (*.dat). These are text files containing only data. • Definition files (*.def). These are text files describing the data file contents, with a listing of variable names, variable definitions, and summary statistics. • EViews (*.wf1) workfiles for each data file. • Excel (*.xls) workbooks for each data file, including variable names in the first row. k ix k k x Preface • • • • Comma separated values (*.csv) files that can be read into almost all software. Stata (*.dta) data files. SAS (*.sas7bdat) data files. GRETL (*.gdt) data files. • R (*.rdata) data files. The author website www.principlesofeconometrics.com includes a complete list of the data files and where they are used in the book. Additional Resources The book website www.principlesofeconometrics.com includes • Individual data files in each format as well as ZIP files containing data in compressed format. • Book errata. • Brief answers to odd number problems. These answers are also provided on the book website at www.wiley.com/college/hill. • Additional examples with solutions. Some extra examples come with complete solutions so that students will know what a good answer looks like. • Tips on writing research papers. k k Resources for Instructors For instructors, also available at the website www.wiley.com/college/hill are • Complete solutions, in both Microsoft Word and *.pdf formats, to all exercises in the text. • PowerPoint slides and PowerPoint Viewer. k k Preface xi Acknowledgments Authors Hill and Griffiths want to acknowledge the gifts given to them over the past 40 years by mentor, friend, and colleague George Judge. Neither this book nor any of the other books we have shared in the writing of would have ever seen the light of day without his vision and inspiration. We also wish to give thanks to the many students and faculty who have commented on the fourth edition of the text and contributed to the fifth edition. This list includes Alejandra Breve Ferrari, Alex James, Alyssa Wans, August Saibeni, Barry Rafferty, Bill Rising, Bob Martin, Brad Lewis, Bronson Fong, David Harris, David Iseral, Deborah Williams, Deokrye Baek, Diana Whistler, Emma Powers, Ercan Saridogan, Erdogan Cevher, Erika Haguette, Ethan Luedecke, Gareth Thomas, Gawon Yoon, Genevieve Briand, German Altgelt, Glenn Sueyoshi, Henry McCool, James Railton, Jana Ruimerman, Jeffery Parker, Joe Goss, John Jackson, Julie Leiby, Katharina Hauck, Katherine Ramirez, Kelley Pace, Lee Adkins, Matias Cattaneo, Max O’Krepki, Meagan McCollum, Micah West, Michelle Savolainen, Oystein Myrland, Patrick Scholten, Randy Campbell, Regina Riphahn, Sandamali Kankanamge, Sergio Pastorello, Shahrokh Towfighi, Tom Fomby, Tong Zeng, Victoria Pryor, Yann Nicolas, and Yuanbo Zhang. In the book errata we acknowledge those who have pointed out our errors. R. Carter Hill William E. Griffiths Guay C. Lim k k k k Table of Contents PREFACE v LIST OF EXAMPLES P.5.6 xxi P.6 1 An Introduction to Econometrics 1.1 1.2 Why Study Econometrics? 1.3 k 2 4 P.8 5 2.1 7 Time-Series Data 7 Cross-Section Data 8 Panel or Longitudinal Data 2.2 Writing an Empirical Research Paper 9 13 Probability Distributions P.3 Joint, Marginal, and Conditional Probabilities 20 17 2.4 22 Properties of Probability Distributions Expected Value of a Random Variable Conditional Expectation 25 Rules for Expected Values 25 Variance of a Random Variable 26 Expected Values of Several Random Variables 27 23 49 Data Generating Process 51 The Random Error and Strict Exogeneity 52 The Regression Function 53 Random Error Variation 54 Variation in x 56 Error Normality 56 Generalizing the Exogeneity Assumption 56 Error Correlation 57 Summarizing the Assumptions k 58 The Gauss–Markov Theorem 2.6 The Probability Distributions of the Least Squares Estimators 73 2.7 2.7.2 k 72 Estimating the Variance of the Error Term 2.7.1 66 The Estimator b2 67 68 The Expected Values of b1 and b2 Sampling Variation 69 The Variances and Covariance of b1 and b2 69 2.5 24 59 The Least Squares Principle 61 Other Economic Models 65 Assessing the Least Squares Estimators 2.4.1 2.4.2 2.4.3 2.4.4 Marginal Distributions 20 Conditional Probability 21 Statistical Independence 21 A Digression: Summation Notation P.5.1 P.5.2 P.5.3 P.5.4 P.5.5 xii 16 47 Estimating the Regression Parameters 2.3.1 2.3.2 P.2 P.5 2.2.8 2.2.9 2.3 Random Variables P.4 13 15 P.1 P.3.1 P.3.2 P.3.3 11 Links to Economic Data on the Internet Interpreting Economic Data 14 Obtaining the Data 14 Probability Primer 46 An Econometric Model 2.2.3 2.2.4 2.2.5 2.2.6 2.2.7 11 Writing a Research Proposal 11 A Format for Writing a Research Report Sources of Economic Data 39 An Economic Model 2.2.1 2.2.2 9 1.7 1.8.1 1.8.2 1.8.3 37 7 The Research Process 1.8 Exercises Model 6 1.6 1.7.1 1.7.2 34 The Bivariate Normal Distribution 2 The Simple Linear Regression 5 Experimental Data 6 Quasi-Experimental Data Nonexperimental Data 29 Conditional Expectation 30 Conditional Variance 31 Iterated Expectations 32 Variance Decomposition 33 Covariance Decomposition 34 The Normal Distribution P.7.1 Causality and Prediction Economic Data Types 1.5.1 1.5.2 1.5.3 P.7 3 How Are Data Generated? 1.4.1 1.4.2 1.4.3 1.5 Some Examples The Econometric Model 1.3.1 1.4 1 What Is Econometrics About? 1.2.1 Conditioning P.6.1 P.6.2 P.6.3 P.6.4 P.6.5 1 Covariance Between Two Random Variables 27 74 Estimating the Variances and Covariance of the Least Squares Estimators 74 Interpreting the Standard Errors 76 k Table of Contents 2.8 Estimating Nonlinear Relationships 2.8.1 2.8.2 2.8.3 2.8.4 2.8.5 2.9 77 Regression with Indicator Variables 2.10 The Independent Variable 2.11 Exercises 84 86 3.3.3 Derivation of the Least Squares Estimates 98 Appendix 2B Deviation from the Mean Form 99 of b2 Appendix 2C b2 Is a Linear Estimator Appendix 2D Derivation of Theoretical Expression for b2 100 Appendix 2E Deriving the Conditional Variance of b2 100 Proof of the Gauss–Markov Theorem 102 Appendix 2G Proofs of Results Introduced in Section 2.10 103 Appendix 2H 2H.1 2H.2 2H.3 2H.4 2H.5 2H.6 2H.7 The p-Value 3.6 Linear Combinations of Parameters Interval Estimation 126 129 133 Problems 133 Computer Exercises 139 Appendix 3A Derivation of the t-Distribution Appendix 3B Distribution of the t-Statistic under H1 145 Appendix 3C 3C.1 3C.2 3C.3 3C.4 Monte Carlo Simulation 144 147 Sampling Properties of Interval Estimators 148 Sampling Properties of Hypothesis Tests 149 Choosing the Number of Monte Carlo Samples 149 Random-x Monte Carlo Results 150 106 4 Prediction, Goodness-of-Fit, and Modeling Issues 4.1 4.2 Least Squares Prediction Measuring Goodness-of-Fit 4.2.1 4.2.2 4.3 113 115 k 152 153 156 Correlation Analysis 158 Correlation Analysis and R2 Modeling Issues 4.3.1 4.3.2 4.3.3 4.3.4 4.3.5 112 The t-Distribution 113 Obtaining Interval Estimates The Sampling Context 116 123 Testing a Linear Combination of Parameters 131 Exercises 3.7.1 3.7.2 The Regression Function 106 The Random Error 107 Theoretically True Values 107 Creating a Sample of Data 108 Monte Carlo Objectives 109 Monte Carlo Results 109 Random-x Monte Carlo Results 110 Hypothesis Testing 3.1.1 3.1.2 3.1.3 3.5 3.7 3 Interval Estimation and 3.1 Examples of Hypothesis Tests The Implications of Strict Exogeneity 103 The Random and Independent x Case 103 The Random and Strictly Exogenous x Case 105 Random Sampling 106 Monte Carlo Simulation One-Tail Tests with Alternative ‘‘Greater Than’’ (>) 120 One-Tail Tests with Alternative ‘‘Less Than’’ (<) 121 Two-Tail Tests with Alternative ‘‘Not Equal To’’ (≠) 122 3.4 3.6.1 100 Appendix 2F 2G.4 3.3.2 93 118 The Null Hypothesis 118 The Alternative Hypothesis 118 The Test Statistic 119 The Rejection Region 119 A Conclusion 120 Rejection Regions for Specific Alternatives 120 3.3.1 Appendix 2A 2G.1 2G.2 2G.3 3.3 82 89 2.11.1 Problems 89 2.11.2 Computer Exercises Hypothesis Tests 3.2.1 3.2.2 3.2.3 3.2.4 3.2.5 2.10.1 Random and Independent x 84 2.10.2 Random and Strictly Exogenous x 2.10.3 Random Sampling 87 k 3.2 Quadratic Functions 77 Using a Quadratic Model 77 A Log-Linear Function 79 Using a Log-Linear Model 80 Choosing a Functional Form 82 xiii 158 160 The Effects of Scaling the Data 160 Choosing a Functional Form 161 A Linear-Log Food Expenditure Model Using Diagnostic Residual Plots 165 Are the Regression Errors Normally Distributed? 167 163 k k xiv Table of Contents 4.3.6 4.4 Polynomial Models 4.4.1 4.5 4.7 Exercises 5.5.2 5.5.3 171 173 Prediction in the Log-Linear Model 175 A Generalized R2 Measure 176 Prediction Intervals in the Log-Linear Model 177 Log-Log Models 4.7.1 4.7.2 169 171 Quadratic and Cubic Equations Log-Linear Models 4.5.1 4.5.2 4.5.3 4.6 Identifying Influential Observations 5.6 Nonlinear Relationships 5.7 Large Sample Properties of the Least Squares Estimator 227 5.7.1 5.7.2 5.7.3 5.7.4 177 179 Problems 179 Computer Exercises 185 5.8 Development of a Prediction Interval 192 Appendix 4B The Sum of Squares Decomposition 193 Appendix 5A Appendix 4C Mean Squared Error: Estimation and Prediction 193 Appendix 5B 5.8.1 5.8.2 5B.1 5 The Multiple Regression 5B.2 Appendix 5C 196 5C.1 k 5.1 Introduction 5.1.1 5.1.2 5.1.3 5.1.4 5.2 5.3.2 Appendix 5D 5D.1 5D.2 5D.3 5D.4 5D.5 5.4.2 Hypothesis Testing 5.5.1 240 Derivation of Least Squares Estimators 247 The Delta Method 248 Nonlinear Function of a Single Parameter 248 Nonlinear Function of Two Parameters Monte Carlo Simulation 249 250 Least Squares Estimation with Chi-Square Errors 250 Monte Carlo Simulation of the Delta Method 252 Bootstrapping 254 Resampling 255 Bootstrap Bias Estimate 256 Bootstrap Standard Error 256 Bootstrap Percentile Interval Estimate Asymptotic Refinement 258 Regression Model 6.1 257 6.1.3 6.1.4 6.1.5 6.2 6.3 Testing the Significance of a Single Coefficient 219 k 261 Testing the Significance of the Model The Relationship Between t- and F-Tests 265 More General F-Tests 267 Using Computer Software 268 Large Sample Tests 269 The Use of Nonsample Information Model Specification 6.3.1 6.3.2 6.3.3 218 260 Testing Joint Hypotheses: The F-test 6.1.1 6.1.2 216 Interval Estimation for a Single Coefficient 216 Interval Estimation for a Linear Combination of Coefficients 217 234 Problems 234 Computer Exercises 6 Further Inference in the Multiple The Variances and Covariances of the Least Squares Estimators 212 The Distribution of the Least Squares Estimators 214 Interval Estimation 5.4.1 5.5 Least Squares Estimation Procedure 205 Estimating the Error Variance σ2 207 Measuring Goodness-of-Fit 208 Frisch–Waugh–Lovell (FWL) Theorem 209 Finite Sample Properties of the Least Squares Estimator 211 5.3.1 5.4 The Economic Model 197 The Econometric Model 198 The General Model 202 Assumptions of the Multiple Regression Model 203 5C.2 Estimating the Parameters of the Multiple Regression Model 205 5.2.1 5.2.2 5.2.3 5.2.4 5.3 197 222 Consistency 227 Asymptotic Normality 229 Relaxing Assumptions 230 Inference for a Nonlinear Function of Coefficients 232 Exercises Appendix 4A Model One-Tail Hypothesis Testing for a Single Coefficient 220 Hypothesis Testing for a Linear Combination of Coefficients 221 271 273 Causality versus Prediction Omitted Variables 275 Irrelevant Variables 277 273 264 k k Table of Contents 6.3.4 6.3.5 6.3.6 6.4 Prediction 6.4.1 6.5 Control Variables Choosing a Model RESET 281 7.7 278 280 282 Predictive Model Selection Criteria 285 The Consequences of Collinearity 289 Identifying and Mitigating Collinearity 290 Investigating Influential Observations 293 6.6 Nonlinear Least Squares 6.7 Exercises 6.7.1 6.7.2 294 303 Appendix 6A The Statistical Power of F-Tests Appendix 6B Further Results from the FWL Theorem 315 7 Using Indicator Variables Indicator Variables 7.1.1 7.1.2 k 7.2 7.2.2 7.2.3 7.2.4 7.3 7.5 7.6 Details of Log-Linear Model Interpretation 366 Appendix 7B Derivation of the Differences-in-Differences Estimator 366 Appendix 7C The Overlap Assumption: Details 367 317 Treatment Effects 8.2 Heteroskedasticity in the Multiple Regression Model 370 8.4 Generalized Least Squares: Known Form of Variance 375 8.4.1 8.4.2 8.5 8.6 8.6.4 8.6.5 332 8.7 8.8 342 k 381 383 Residual Plots 384 The Goldfeld–Quandt Test 384 A General Test for Conditional Heteroskedasticity 385 The White Test 387 Model Specification and Heteroskedasticity 388 Heteroskedasticity in the Linear Probability Model 390 Exercises 8.8.1 8.8.2 The Nature of Causal Effects 342 Treatment Effect Models 343 Decomposing the Treatment Effect 344 Introducing Control Variables 345 The Overlap Assumption 347 Regression Discontinuity Designs 347 Estimating the Multiplicative Model Detecting Heteroskedasticity 8.6.1 8.6.2 8.6.3 331 k Transforming the Model: Proportional Heteroskedasticity 375 Weighted Least Squares: Proportional Heteroskedasticity 377 Generalized Least Squares: Unknown Form of Variance 379 8.5.1 334 The Heteroskedastic Regression Model 371 Heteroskedasticity Consequences for the OLS Estimator 373 Heteroskedasticity Robust Variance Estimator 374 330 330 The Difference Estimator 334 Analysis of the Difference Estimator The Differences-in-Differences Estimator 338 369 8.3 323 Treatment Effects and Causal Modeling 7.6.1 7.6.2 7.6.3 7.6.4 7.6.5 7.6.6 The Nature of Heteroskedasticity 329 A Rough Calculation An Exact Calculation 368 8.1 8.2.1 8.2.2 318 The Linear Probability Model 7.5.1 7.5.2 7.5.3 311 Interactions Between Qualitative Factors 323 Qualitative Factors with Several Categories 324 Testing the Equivalence of Two Regressions 326 Controlling for Time 328 Log-Linear Models 7.3.1 7.3.2 7.4 Appendix 7A Intercept Indicator Variables 318 Slope-Indicator Variables 320 Applying Indicator Variables 7.2.1 358 8 Heteroskedasticity 297 Problems 297 Computer Exercises 351 Problems 351 Computer Exercises Poor Data, Collinearity, and Insignificance 288 6.5.1 6.5.2 6.5.3 7.1 Exercises 7.7.1 7.7.2 xv 391 Problems 391 Computer Exercises 401 Appendix 8A Properties of the Least Squares Estimator 407 Appendix 8B Lagrange Multiplier Tests for Heteroskedasticity 408 k xvi Table of Contents 8C.1 10.1.2 Why Least Squares Estimation Fails 484 10.1.3 Proving the Inconsistency of OLS 486 Properties of the Least Squares Residuals 410 Appendix 8C Details of Multiplicative Heteroskedasticity Model 411 10.2 Cases in Which x and e are Contemporaneously Correlated 487 Appendix 8D Alternative Robust Sandwich Estimators 411 Appendix 8E Monte Carlo Evidence: OLS, GLS, and FGLS 414 10.2.1 Measurement Error 487 10.2.2 Simultaneous Equations Bias 488 10.2.3 Lagged-Dependent Variable Models with Serial Correlation 489 10.2.4 Omitted Variables 489 9 Regression with Time-Series Data: Stationary Variables 9.1 Introduction 9.1.1 9.1.2 9.2 9.3 9.4 9.4.2 9.4.3 9.5 427 430 Forecast Intervals and Standard Errors Assumptions for Forecasting 435 Selecting Lag Lengths 436 Testing for Granger Causality 437 433 438 Checking the Correlogram of the Least Squares Residuals 439 Lagrange Multiplier Test 440 Durbin–Watson Test 443 Time-Series Regressions for Policy Analysis 443 9.5.1 9.5.2 9.5.3 9.5.4 9.6 420 Testing for Serially Correlated Errors 9.4.1 k Modeling Dynamic Relationships Autocorrelations 424 Forecasting 10.3.1 Method of Moments Estimation of a Population Mean and Variance 490 10.3.2 Method of Moments Estimation in the Simple Regression Model 491 10.3.3 Instrumental Variables Estimation in the Simple Regression Model 492 10.3.4 The Importance of Using Strong Instruments 493 10.3.5 Proving the Consistency of the IV Estimator 494 10.3.6 IV Estimation Using Two-Stage Least Squares (2SLS) 495 10.3.7 Using Surplus Moment Conditions 496 10.3.8 Instrumental Variables Estimation in the Multiple Regression Model 498 10.3.9 Assessing Instrument Strength Using the First-Stage Model 500 10.3.10 Instrumental Variables Estimation in a General Model 502 10.3.11 Additional Issues When Using IV Estimation 504 418 Stationarity and Weak Dependence 9.3.1 9.3.2 9.3.3 9.3.4 10.3 Estimators Based on the Method of Moments 490 417 Finite Distributed Lags 445 HAC Standard Errors 448 Estimation with AR(1) Errors 452 Infinite Distributed Lags 456 Exercises 9.6.1 9.6.2 Appendix 9A 9A.1 Appendix 9B 10.4 Specification Tests 463 Problems 463 Computer Exercises 468 The Durbin–Watson Test 505 10.4.1 The Hausman Test for Endogeneity 505 10.4.2 The Logic of the Hausman Test 507 10.4.3 Testing Instrument Validity 508 476 The Durbin–Watson Bounds Test Properties of an AR(1) Error 478 516 Appendix 10A Testing for Weak Instruments 10A.1 10A.2 481 10.1 Least Squares Estimation with Endogenous Regressors 482 510 10.5.1 Problems 510 10.5.2 Computer Exercises 479 10 Endogenous Regressors and Moment-Based Estimation 10.5 Exercises Appendix 10B Monte Carlo Simulation 10B.1 10B.2 10.1.1 Large Sample Properties of the OLS Estimator 483 k 520 A Test for Weak Identification 521 Testing for Weak Identification: Conclusions 525 525 Illustrations Using Simulated Data The Sampling Properties of IV/2SLS 528 526 k k Table of Contents 11 Simultaneous Equations Models 12.7 Exercises 11.1 A Supply and Demand Model 532 11.2 The Reduced-Form Equations 11.3 The Failure of Least Squares Estimation 11.4 The Identification Problem Vector Autoregressive Models 597 535 535 536 13.1 VEC and VAR Models 11.5 Two-Stage Least Squares Estimation 538 k 13.4.1 Impulse Response Functions 13.4.2 Forecast Error Variance Decompositions 605 557 The k-Class of Estimators 557 The LIML Estimator 558 Monte Carlo Simulation Results 13.5 Exercises 562 ARCH Models 564 12.1.1 Trend Stationary Variables 567 12.1.2 The First-Order Autoregressive Model 12.1.3 Random Walk Models 572 570 14.1 The ARCH Model 615 616 14.3 Testing, Estimating, and Forecasting 14.4 Extensions 574 620 622 14.4.1 The GARCH Model—Generalized ARCH 622 14.4.2 Allowing for an Asymmetric Effect 623 14.4.3 GARCH-in-Mean and Time-Varying Risk Premium 624 14.4.4 Other Developments 625 576 14.5 Exercises 626 14.5.1 Problems 626 14.5.2 Computer Exercises 15 Panel Data Models 582 627 634 15.1 The Panel Data Regression Function 584 636 15.1.1 Further Discussion of Unobserved Heterogeneity 638 15.1.2 The Panel Data Regression Exogeneity Assumption 639 12.5 Regression When There Is No Cointegration 585 12.6 Summary 614 14.2 Time-Varying Volatility 12.3.1 Unit Roots 576 12.3.2 Dickey–Fuller Tests 577 12.3.3 Dickey–Fuller Test with Intercept and No Trend 577 12.3.4 Dickey–Fuller Test with Intercept and Trend 579 12.3.5 Dickey–Fuller Test with No Intercept and No Trend 580 12.3.6 Order of Integration 581 12.3.7 Other Unit Root Tests 582 12.4.1 The Error Correction Model 612 14 Time-Varying Volatility and 12.1 Stationary and Nonstationary Variables 12.4 Cointegration 608 Appendix 13A The Identification Problem Data: Nonstationary Variables 563 12.2 Consequences of Stochastic Trends 603 607 13.5.1 Problems 607 13.5.2 Computer Exercises 12 Regression with Time-Series 12.3 Unit Root Tests for Stationarity 601 13.4 Impulse Responses and Variance Decompositions 603 551 Appendix 11A 2SLS Alternatives 11A.1 11A.2 11A.3 13.3 Estimating a VAR Model 545 11.6.1 Problems 545 11.6.2 Computer Exercises 598 13.2 Estimating a Vector Error Correction Model 600 11.5.1 The General Two-Stage Least Squares Estimation Procedure 539 11.5.2 The Properties of the Two-Stage Least Squares Estimator 540 11.6 Exercises 592 13 Vector Error Correction and 534 11.3.1 Proving the Failure of OLS 588 12.7.1 Problems 588 12.7.2 Computer Exercises 531 xvii 587 k k k xviii Table of Contents 15.1.3 Using OLS to Estimate the Panel Data Regression 639 15.2 The Fixed Effects Estimator 15.2.1 15.2.2 15.2.3 15.2.4 16.2.9 Binary Choice Models with a Binary Endogenous Variable 699 16.2.10 Binary Endogenous Explanatory Variables 700 16.2.11 Binary Choice Models and Panel Data 701 640 The Difference Estimator: T = 2 640 The Within Estimator: T = 2 642 The Within Estimator: T > 2 643 The Least Squares Dummy Variable Model 644 16.3 Multinomial Logit 15.3 Panel Data Regression Error Assumptions 646 15.3.1 OLS Estimation with Cluster-Robust Standard Errors 648 15.3.2 Fixed Effects Estimation with Cluster-Robust Standard Errors 650 15.4 The Random Effects Estimator 15.5 Exercises k 16.4 Conditional Logit 16.5 Ordered Choice Models 16.6 Models for Count Data Appendix 15A Cluster-Robust Standard Errors: Some Details 677 681 16.1 Introducing Models with Binary Dependent Variables 682 16.8 Exercises 685 16.2.1 The Probit Model for Binary Choice 686 16.2.2 Interpreting the Probit Model 687 16.2.3 Maximum Likelihood Estimation of the Probit Model 690 16.2.4 The Logit Model for Binary Choices 693 16.2.5 Wald Hypothesis Tests 695 16.2.6 Likelihood Ratio Hypothesis Tests 696 16.2.7 Robust Inference in Probit and Logit Models 698 16.2.8 Binary Choice Models with a Continuous Endogenous Variable 698 718 725 16.8.1 Problems 725 16.8.2 Computer Exercises 683 733 Appendix 16A Probit Marginal Effects: Details 16A.1 16A.2 741 Binary Choice Model 741 Probit or Logit? 742 Appendix 16C Using Latent Variables 16C.1 16C.2 743 Tobit (Tobit Type I) 743 Heckit (Tobit Type II) 744 Appendix 16D A Tobit Monte Carlo Experiment k 739 Standard Error of Marginal Effect at a Given Point 739 Standard Error of Average Marginal Effect 740 Appendix 16B Random Utility Models 16B.1 16B.2 k 717 16.7.1 Maximum Likelihood Estimation of the Simple Linear Regression Model 717 16.7.2 Truncated Regression 718 16.7.3 Censored Samples and Regression 16.7.4 Tobit Model Interpretation 720 16.7.5 Sample Selection 723 16 Qualitative and Limited 16.2 Modeling Binary Choices 713 16.7 Limited Dependent Variables Appendix 15B Estimation of Error Components 679 16.1.1 The Linear Probability Model 710 16.6.1 Maximum Likelihood Estimation of the Poisson Regression Model 713 16.6.2 Interpreting the Poisson Regression Model 714 670 Dependent Variable Models 709 16.5.1 Ordinal Probit Choice Probabilities 16.5.2 Ordered Probit Estimation and Interpretation 711 663 15.5.1 Problems 663 15.5.2 Computer Exercises 707 16.4.1 Conditional Logit Choice Probabilities 707 16.4.2 Conditional Logit Postestimation Analysis 708 651 15.4.1 Testing for Random Effects 653 15.4.2 A Hausman Test for Endogeneity in the Random Effects Model 654 15.4.3 A Regression-Based Hausman Test 656 15.4.4 The Hausman–Taylor Estimator 658 15.4.5 Summarizing Panel Data Assumptions 660 15.4.6 Summarizing and Extending Panel Data Model Estimation 661 702 16.3.1 Multinomial Logit Choice Probabilities 703 16.3.2 Maximum Likelihood Estimation 703 16.3.3 Multinomial Logit Postestimation Analysis 704 745 k Table of Contents A Mathematical Tools A.1 Some Basics A.1.1 A.1.2 A.1.3 A.1.4 A.1.5 A.1.6 A.2 A.3 k A.5 753 753 Rules for Derivatives 754 Elasticity of a Nonlinear Relationship Second Derivatives 757 Maxima and Minima 758 Partial Derivatives 759 Maxima and Minima of Bivariate Functions 760 B.1.2 B.1.3 B.1.4 B.1.5 B.1.6 B.1.7 B.1.8 B.1.9 B.2 B.2.3 B.2.4 805 806 A Sample of Data C.2 An Econometric Model C.3 Estimating the Mean of a Population C.3.1 C.3.2 C.3.3 C.3.4 C.3.5 769 Expected Value of a Discrete Random Variable 769 Variance of a Discrete Random Variable 770 Joint, Marginal, and Conditional Distributions 771 Expectations Involving Several Random Variables 772 Covariance and Correlation 773 Conditional Expectations 774 Iterated Expectations 774 Variance Decomposition 774 Covariance Decomposition 777 C.4 C.4.2 C.5 815 Estimating the Population Variance 821 Estimating Higher Moments 821 822 Interval Estimation: σ2 Known 822 Interval Estimation: σ2 Unknown 825 Hypothesis Tests About a Population Mean 826 C.6.1 C.6.2 k 814 The Expected Value of Y 816 The Variance of Y 817 The Sampling Distribution of Y 817 The Central Limit Theorem 818 Best Linear Unbiased Estimation 820 Interval Estimation C.5.1 C.5.2 C.6 813 Estimating the Population Variance and Other Moments 820 C.4.1 Probability Calculations 779 Properties of Continuous Random Variables 780 Joint, Marginal, and Conditional Probability Distributions 781 Using Iterated Expectations with Continuous Random Variables 785 812 C.1 768 Working with Continuous Random Variables 778 B.2.1 B.2.2 Exercises Inference 764 Discrete Random Variables 800 Uniform Random Numbers C Review of Statistical Computing the Area Under a Curve 762 Exercises B.1.1 B.5 The Bernoulli Distribution 790 The Binomial Distribution 790 The Poisson Distribution 791 The Uniform Distribution 792 The Normal Distribution 793 The Chi-Square Distribution 794 The t-Distribution 796 The F-Distribution 797 The Log-Normal Distribution 799 Random Numbers B.4.1 762 B Probability Concepts B.1 B.4 757 Distributions of Functions of Random Variables 787 Truncated Random Variables 789 Some Important Probability Distributions 789 B.3.1 B.3.2 B.3.3 B.3.4 B.3.5 B.3.6 B.3.7 B.3.8 B.3.9 752 Slopes and Derivatives Elasticity 753 Integrals A.4.1 B.3 Numbers 749 Exponents 749 Scientific Notation 749 Logarithms and the Number e 750 Decimals and Percentages 751 Logarithms and Percentages 751 Nonlinear Relationships A.3.1 A.3.2 A.3.3 A.3.4 A.3.5 A.3.6 A.4 B.2.6 749 Linear Relationships A.2.1 A.2.2 B.2.5 748 xix C.6.3 C.6.4 C.6.5 C.6.6 Components of Hypothesis Tests 826 One-Tail Tests with Alternative ‘‘Greater Than’’ (>) 828 One-Tail Tests with Alternative ‘‘Less Than’’ (<) 829 Two-Tail Tests with Alternative ‘‘Not Equal To’’ (≠) 829 The p-Value 831 A Comment on Stating Null and Alternative Hypotheses 832 k k xx Table of Contents C.6.7 C.6.8 C.7 Some Other Useful Tests C.7.1 C.7.2 C.7.3 C.7.4 C.8 C.8.2 C.8.3 C.8.4 Inference with Maximum Likelihood Estimators 840 The Variance of the Maximum Likelihood Estimator 841 The Distribution of the Sample Proportion 842 Asymptotic Test Procedures 843 Algebraic Supplements C.9.1 C.9.2 C.10 Kernel Density Estimator C.11 Exercises 848 851 854 C.11.1 Problems 854 C.11.2 Computer Exercises 857 D Statistical Tables 862 834 Testing the Population Variance 834 Testing the Equality of Two Population Means 834 Testing the Ratio of Two Population Variances 835 Testing the Normality of a Population 836 Introduction to Maximum Likelihood Estimation 837 C.8.1 C.9 Type I and Type II Errors 833 A Relationship Between Hypothesis Testing and Confidence Intervals 833 TableD.1 Cumulative Probabilities for the Standard Normal Distribution 𝚽(z) = P(Z ≤ z) 862 TableD.2 Percentiles of the t-distribution 863 TableD.3 Percentiles of the Chi-square Distribution 864 TableD.4 95th Percentile for the F-distribution 865 TableD.5 99th Percentile for the F-distribution 866 TableD.6 Standard Normal pdf Values 𝛟(z) INDEX 867 869 Derivation of Least Squares Estimator 848 Best Linear Unbiased Estimation 849 k k k k List of Examples Example P.1 Using a cdf Example P.2 Calculating a Conditional Probability 21 19 Example 3.4 (continued) p-Value for a Left-Tail Test 128 Example P.3 Calculating an Expected Value Example P.4 Calculating a Conditional Expectation 25 Example P.5 Calculating a Variance Example P.6 Calculating a Correlation 28 Example P.7 Conditional Expectation 30 Example P.8 Conditional Variance 31 Example P.9 Iterated Expectation 32 Example 3.5 (continued) p-Value for a Two-Tail Test 129 24 Example 3.6 (continued) p-Value for a Two-Tail Test of Significance 129 26 Example P.10 Covariance Decomposition 34 Example P.11 Normal Distribution Probability Calculation 36 k Example 3.7 Estimating Expected Food Expenditure 130 Example 3.8 An Interval Estimate of Expected Food Expenditure 131 Example 3.9 Testing Expected Food Expenditure 132 Example 4.1 Prediction in the Food Expenditure Model 156 Example 4.2 Goodness-of-Fit in the Food Expenditure Model 159 Example 2.1 A Failure of the Exogeneity Assumption 53 Example 4.3 Reporting Regression Results Example 2.2 Strict Exogeneity in the Household Food Expenditure Model 54 Example 4.4 Using the Linear-Log Model for Food Expenditure 164 Example 2.3 Food Expenditure Model Data 59 Example 4.5 Example 2.4a Estimates for the Food Expenditure Function 63 Heteroskedasticity in the Food Expenditure Model 167 Example 4.6 Testing Normality in the Food Expenditure Model 168 Example 4.7 Influential Observations in the Food Expenditure Data 171 Example 4.8 An Empirical Example of a Cubic Equation 172 Example 4.9 A Growth Model Example 2.4b Using the Estimates 64 Example 2.5 Calculations for the Food Expenditure Data 75 Example 2.6 Baton Rouge House Data Example 2.7 Baton Rouge House Data, Log-Linear Model 81 Example 3.1 78 Interval Estimate for Food Expenditure Data 116 Example 4.10 A Wage Equation 159 174 175 Example 4.11 Prediction in a Log-Linear Model 176 Example 3.2 Right-Tail Test of Significance Example 3.3 Right-Tail Test of an Economic Hypothesis 124 Example 4.12 Prediction Intervals for a Log-Linear Model 177 Example 3.4 Left-Tail Test of an Economic Hypothesis 125 Example 4.13 A Log-Log Poultry Demand Equation 178 Example 3.5 Two-Tail Test of an Economic Hypothesis 125 Example 5.1 Data for Hamburger Chain Example 5.2 Example 3.6 Two-Tail Test of Significance OLS Estimates for Hamburger Chain Data 206 Example 5.3 Error Variance Estimate for Hamburger Chain Data 208 123 126 Example 3.3 (continued) p-Value for a Right-Tail Test 127 k 200 xxi k k xxii List of Examples Example 5.4 R2 for Hamburger Chain Data Example 5.5 Variances, Covariances, and Standard Errors for Hamburger Chain Data 214 209 Example 5.6 Interval Estimates for Coefficients in Hamburger Sales Equation 216 Example 5.7 Interval Estimate for a Change in Sales 218 Example 5.8 Testing the Significance of Price 219 Example 5.9 Testing the Significance of Advertising Expenditure 220 Example 5.10 Testing for Elastic Demand 270 Example 6.8 A Nonlinear Hypothesis Example 6.9 Restricted Least Squares 272 Example 6.10 Family Income Equation 275 Example 6.11 Adding Children Aged Less Than 6 Years 277 277 Example 6.13 A Control Variable for Ability 279 Example 6.14 Applying RESET to Family Income Equation 282 Example 6.15 Forecasting SALES for the Burger Barn 284 220 Example 6.16 Predicting House Prices Example 5.12 Testing the Effect of Changes in Price and Advertising 222 Example 5.13 Cost and Product Curves 270 Example 6.12 Adding Irrelevant Variables Example 5.11 Testing Advertising Effectiveness 221 223 Example 5.14 Extending the Model for Burger Barn Sales 224 k Examples6.2and6.5 Revisited Example 5.16 A Log-Quadratic Wage Equation Example 6.18 Influential Observations in the House Price Equation 293 Example 6.19 Nonlinear Least Squares Estimates for Simple Model 295 Example 5.18 How Much Experience Maximizes Wages? 233 Example 5.19 An Interval Estimate for ( ) exp 𝛃2 ∕10 249 250 Example 5.21 Bootstrapping for Nonlinear ( ) ( ) Functions g1 𝛃2 = exp 𝛃2 ∕10 and ( ) g2 𝛃1 , 𝛃2 = 𝛃1 ∕𝛃2 258 296 Example 7.1 The University Effect on House Prices 321 Example 7.2 The Effects of Race and Sex on Wage 323 Example 7.3 A Wage Equation with Regional Indicators 325 Example 7.4 Testing the Equivalence of Two Regressions: The Chow Test 327 Example 7.5 Indicator Variables in a Log-Linear Model: The Rough Approximation 330 Example 7.6 Indicator Variables in a Log-Linear Model: An Exact Calculation 330 Example 7.7 The Linear Probability Model: An Example from Marketing 332 226 Example 5.17 The Optimal Level of Advertising 232 Example 5.20 An Interval Estimate for 𝛃1 ∕𝛃2 Example 6.17 Collinearity in a Rice Production Function 291 Example 6.20 A Logistic Growth Curve Example 5.15 An Interaction Variable in a Wage Equation 225 287 Example 6.1 Testing the Effect of Advertising Example 6.2 The F-Test Procedure Example 6.3 Overall Significance of Burger Barns Equation 265 Example 7.8 An Application of Difference Estimation: Project STAR 335 Example 6.4 When are t- and F-tests equivalent? 266 Example 7.9 The Difference Estimator with Additional Controls 336 Example 6.5 Testing Optimal Advertising Example 6.6 A One-Tail Test Example 6.7 Two (J = 2) Complex Hypotheses 268 268 262 263 267 Example 7.10 The Difference Estimator with Fixed Effects 337 Example 7.11 Linear Probability Model Check of Random Assignment 338 k k k List of Examples xxiii Example 7.12 Estimating the Effect of a Minimum Wage Change: The DID Estimator 340 Example 9.10 Checking the Residual Correlogram for the ARDL(2, 1) Unemployment Equation 439 Example 7.13 Estimating the Effect of a Minimum Wage Change: Using Panel Data 341 Example 9.11 Checking the Residual Correlogram for an ARDL(1, 1) Unemployment Equation 440 Example 8.1 Example 8.2 Heteroskedasticity in the Food Expenditure Model 372 Robust Standard Errors in the Food Expenditure Model 374 Example 8.3 Applying GLS/WLS to the Food Expenditure Data 378 Example 8.4 Multiplicative Heteroskedasticity in the Food Expenditure Model 382 Example 8.5 A Heteroskedastic Partition Example 8.6 The Goldfeld–Quandt Test with Partitioned Data 384 Example 8.7 The Goldfeld–Quandt Test in the Food Expenditure Model 385 Example 8.8 Variance Stabilizing Log-transformation 389 Example 8.9 The Marketing Example Revisited 391 k Example 9.12 LM Test for Serial Correlation in the ARDL Unemployment Equation 443 Example 9.13 Okun’s Law Example 9.14 A Phillips Curve 450 Example 9.15 The Phillips Curve with AR(1) Errors 455 Example 9.16 A Consumption Function 383 458 Example 9.17 Deriving Multipliers for an Infinite Lag Okun’s Law Model 459 Example 9.18 Computing the Multiplier Estimates for the Infinite Lag Okun’s Law Model 460 Example 9.19 Testing for Consistency of Least Squares Estimation of Okun’s Law 462 Example 8.10 Alternative Robust Standard Errors in the Food Expenditure Model 413 Example 9.1 446 Plotting the Unemployment Rate and the GDP Growth Rate for the United States 419 k Example 9.20 Durbin–Watson Bounds Test for Phillips Curve 479 Example 10.1 Least Squares Estimation of a Wage Equation 489 Example 10.2 IV Estimation of a Simple Wage Equation 495 Example 9.2 Sample Autocorrelations for Unemployment 426 Example 10.3 2SLS Estimation of a Simple Wage Equation 496 Example 9.3 Sample Autocorrelations for GDP Growth Rate 427 Example 10.4 Using Surplus Instruments in the Simple Wage Equation 497 Example 9.4 Are the Unemployment and Growth Rate Series Stationary and Weakly Dependent? 428 Example 10.5 IV/2SLS Estimation in the Wage Equation 499 Example 9.5 Forecasting Unemployment with an AR(2) Model 431 Example 9.6 Forecast Intervals for Unemployment from the AR(2) Model 434 Example 9.7 Forecasting Unemployment with an ARDL(2, 1) Model 434 Example 9.8 Choosing Lag Lengths in an ARDL(p, q) Unemployment Equation 437 Example 11.2 Supply and Demand at the Fulton Fish Market 542 Does the Growth Rate Granger Cause Unemployment? 438 Example 11.4 Testing for Weak Instruments Using LIML 560 Example 9.9 Example 10.6 Checking Instrument Strength in the Wage Equation 502 Example 10.7 Specification Tests for the Wage Equation 509 Example 10.8 Testing for Weak Instruments 523 Example 11.1 Supply and Demand for Truffles Example 11.3 Klein’s Model I k 541 544 k xxiv List of Examples Example 11.5 Testing for Weak Instruments with Fuller-Modified LIML 561 Example 15.3 Using T = 2 Differenced Observations for a Wage Equation 641 Example 12.1 Plots of Some U.S. Economic Time Series 564 Example 15.4 Using the Within Transformation with T = 2 Observations for a Production Function 642 Example 12.2 A Deterministic Trend for Wheat Yield 569 Example 12.3 A Regression with Two Random Walks 575 Example 12.4 Checking the Two Interest Rate Series for Stationarity 579 Example 12.5 Is GDP Trend Stationary? 580 Example 12.7 The Order of Integration of the Two Interest Rate Series 581 Example 12.8 Are the Federal Funds Rate and Bond Rate Cointegrated? 583 Example 12.9 An Error Correction Model for the Bond and Federal Funds Rates 585 Example12.10 A Consumption Function in First Differences 586 Example 13.1 VEC Model for GDP 600 Example 13.2 VAR Model for Consumption and Income 602 Example 14.1 Characteristics of Financial Variables 617 Example 14.3 Testing for ARCH in BrightenYourDay (BYD) Lighting 620 Example 14.4 ARCH Model Estimates for BrightenYourDay (BYD) Lighting 621 Example 14.6 A GARCH Model for BrightenYourDay 623 624 Example 14.8 A GARCH-in-Mean Model for BYD 625 Example 15.1 A Microeconometric Panel Example 15.9 Random Effects Estimation of a Production Function 652 Example15.10 Random Effects Estimation of a Wage Equation 652 Example15.11 Testing for Random Effects in a Production Function 654 Example15.12 Testing for Endogenous Random Effects in a Production Function 656 Example15.14 The Mundlak Approach for a Production Function 657 Example15.15 The Mundlak Approach for a Wage Equation 658 Example15.16 The Hausman–Taylor Estimator for a Wage Equation 659 Example 14.5 Forecasting BrightenYourDay (BYD) Volatility 621 Example 14.7 A T-GARCH Model for BYD Example 15.8 Using Fixed Effects and Cluster-Robust Standard Errors for a Production Function 651 Example15.13 Testing for Endogenous Random Effects in a Wage Equation 656 Example 14.2 Simulating Time-Varying Volatility 618 Example 15.1 Revisited Example 15.6 Using the Fixed Effects Estimator with T = 3 Observations for a Production Function 646 Example 15.7 Using Pooled OLS with Cluster-Robust Standard Errors for a Production Function 650 Example 12.6 Is Wheat Yield Trend Stationary? 580 k Example 15.5 Using the Within Transformation with T = 3 Observations for a Production Function 644 Example 16.1 A Transportation Problem 683 Example 16.2 A Transportation Problem: The Linear Probability Model 684 Example 16.3 Probit Maximum Likelihood: A Small Example 690 Example 16.4 The Transportation Data: Probit 691 636 637 Example 15.2 Using T = 2 Differenced Observations for a Production Function 641 Example 16.5 The Transportation Data: More Postestimation Analysis 692 Example 16.6 An Empirical Example from Marketing 694 k k k List of Examples Example 16.7 Coke Choice Model: Wald Hypothesis Tests 696 Computing a Joint Probability 783 Example B.7 Example 16.8 Coke Choice Model: Likelihood Ratio Hypothesis Tests 697 Another Joint Probability Calculation 783 Example B.8 Example 16.9 Estimating the Effect of Education on Labor Force Participation 699 Finding and Using a Marginal pdf 784 Example B.9 Finding and Using a Conditional pdf 784 Example16.10 Women’s Labor Force Participation and Having More Than Two Children 700 Example B.10 Computing a Correlation 786 Example B.12 Change of Variable: Continuous Case 788 Example16.12 Postsecondary Education Multinomial Choice 705 Example B.13 Change of Variable: Continuous Case 789 Example16.13 Conditional Logit Soft Drink Choice 708 Example B.14 An Inverse Transformation Example B.16 Linear Congruential Generator Example 805 Example16.15 A Count Model for the Number of Doctor Visits 715 Example16.16 Tobit Model of Hours Worked 801 Example B.15 The Inversion Method: An Example 802 Example16.14 Postsecondary Education Ordered Choice Model 712 Example16.17 Heckit Model of Wages 785 Example B.11 Using Iterated Expectation Example16.11 Effect of War Veteran Status on Wages 701 k Example B.6 xxv 721 724 Example C.1 Histogram of Hip Width Data Example C.2 Sample Mean of Hip Width Data Example C.3 Sampling Variation of Sample Means of Hip Width Data 816 Example C.4 The Effect of Sample Size on Sample Mean Precision 818 814 Example A.1 Slope of a Linear Function Example A.2 Slope of a Quadratic Function Example A.3 Taylor Series Approximation Example A.4 Second Derivative of a Linear Function 758 Example C.5 Illustrating the Central Limit Theorem 819 Example A.5 Second Derivative of a Quadratic Function 758 Example C.6 Sample Moments of the Hip Data 822 Example A.6 Finding the Minimum of a Quadratic Function 759 Example C.7 Using the Hip Data Estimates Example C.8 Example A.7 Maximizing a Profit Function Simulating the Hip Data: Interval Estimates 824 Example A.8 Minimizing a Sum of Squared Differences 761 Example C.9 Simulating the Hip Data: Continued 825 Example A.9 Area Under a Curve Example B.1 Variance Decomposition: Numerical Example 776 Example C.10 Interval Estimation Using the Hip Data 826 Example C.11 One-tail Test Using the Hip Data Example B.2 Probability Calculation Using Geometry 779 Example C.12 Two-tail Test Using the Hip Data 830 Example B.3 Probability Calculation Using Integration 780 Example C.13 One-tail Test p-value: The Hip Data 831 Example B.4 Expected Value of a Continuous Random Variable 780 Example C.14 Two-Tail Test p-Value: The Hip Data 832 Example B.5 Variance of a Continuous Random Variable 781 Example C.15 Testing the Normality of the Hip Data 836 755 755 756 761 763 k 815 822 830 k k xxvi List of Examples Example C.16 The ‘‘Wheel of Fortune’’ Game: p = 1∕4 or 3/4 837 Example C.21 Likelihood Ratio Test of the Population Proportion 845 Example C.17 The ‘‘Wheel of Fortune’’ Game: 0 < p < 1 838 Example C.22 Wald Test of the Population Proportion 846 Example C.18 The ‘‘Wheel of Fortune’’ Game: Maximizing the Log-likelihood 839 Example C.23 Lagrange Multiplier Test of the Population Proportion 848 Example C.19 Estimating a Population Proportion 840 Example C.24 Hip Data: Minimizing the Sum of Squares Function 849 Example C.20 Testing a Population Proportion 843 k k k k CHAPTER 1 An Introduction to Econometrics 1.1 k Why Study Econometrics? Econometrics is fundamental for economic measurement. However, its importance extends far beyond the discipline of economics. Econometrics is a set of research tools also employed in the business disciplines of accounting, finance, marketing, and management. It is used by social scientists, specifically researchers in history, political science, and sociology. Econometrics plays an important role in such diverse fields as forestry and agricultural economics. This breadth of interest in econometrics arises in part because economics is the foundation of business analysis and is the core social science. Thus, research methods employed by economists, which includes the field of econometrics, are useful to a broad spectrum of individuals. Econometrics plays a special role in the training of economists. As a student of economics, you are learning to “think like an economist.” You are learning economic concepts such as opportunity cost, scarcity, and comparative advantage. You are working with economic models of supply and demand, macroeconomic behavior, and international trade. Through this training you become a person who better understands the world in which we live; you become someone who understands how markets work, and the way in which government policies affect the marketplace. If economics is your major or minor field of study, a wide range of opportunities is open to you upon graduation. If you wish to enter the business world, your employer will want to know the answer to the question, “What can you do for me?” Students taking a traditional economics curriculum answer, “I can think like an economist.” While we may view such a response to be powerful, it is not very specific and may not be very satisfying to an employer who does not understand economics. The problem is that a gap exists between what you have learned as an economics student and what economists actually do. Very few economists make their livings by studying economic theory alone, and those who do are usually employed by universities. Most economists, whether they work in the business world or for the government, or teach in universities, engage in economic analysis that is in part “empirical.” By this we mean that they use economic data to estimate economic relationships, test economic hypotheses, and predict economic outcomes. Studying econometrics fills the gap between being “a student of economics” and being “a practicing economist.” With the econometric skills you will learn from this book, including how to work with econometric software, you will be able to elaborate on your answer to the employer’s question above by saying “I can predict the sales of your product.” “I can estimate the effect on your sales if your competition lowers its price by $1 per unit.” “I can test whether your new ad campaign is actually increasing your sales.” These answers are music to an employer’s ears, because they reflect your ability to think like an economist and to analyze economic data. k 1 k k 2 CHAPTER 1 An Introduction to Econometrics Such pieces of information are keys to good business decisions. Being able to provide your employer with useful information will make you a valuable employee and increase your odds of getting a desirable job. On the other hand, if you plan to continue your education by enrolling in graduate school or law school, you will find that this introduction to econometrics is invaluable. If your goal is to earn a master’s or Ph.D. degree in economics, finance, data analytics, data science, accounting, marketing, agricultural economics, sociology, political science, or forestry, you will encounter more econometrics in your future. The graduate courses tend to be quite technical and mathematical, and the forest often gets lost in studying the trees. By taking this introduction to econometrics you will gain an overview of what econometrics is about and develop some “intuition” about how things work before entering a technically oriented course. 1.2 What Is Econometrics About? At this point we need to describe the nature of econometrics. It all begins with a theory from your field of study—whether it is accounting, sociology, or economics—about how important variables are related to one another. In economics we express our ideas about relationships between economic variables using the mathematical concept of a function. For example, to express a relationship between income and consumption, we may write CONSUMPTION = 𝑓 (INCOME) which says that the level of consumption is some function, f (•), of income. The demand for an individual commodity—say, the Honda Accord—might be expressed as k Qd = 𝑓 (P, Ps , Pc , INC) which says that the quantity of Honda Accords demanded, Qd , is a function f (P, Ps , Pc , INC) of the price of Honda Accords P, the price of cars that are substitutes Ps , the price of items that are complements Pc (like gasoline), and the level of income INC. The supply of an agricultural commodity such as beef might be written as ) ( Qs = 𝑓 P, Pc , P𝑓 where Qs is the quantity supplied, P is the price of beef, Pc is the price of competitive products in production (e.g., the price of hogs), and Pf is the price of factors or inputs (e.g., the price of corn) used in the production process. Each of the above equations is a general economic model that describes how we visualize the way in which economic variables are interrelated. Economic models of this type guide our economic analysis. Econometrics allows us to go further than knowing that certain economic variables are interrelated, or even the direction of a relationship. Econometrics allows us to assign magnitudes to questions about the interrelationships between variables. One aspect of econometrics is prediction or forecasting. If we know the value of INCOME, what will be the magnitude of CONSUMPTION? If we have values for the prices of Honda Accords, their substitutes and complements, and income, how many Honda Accords will be sold? Similarly, we could ask how much beef would be supplied given values of the variables on which its supply depends. A second contribution of econometrics is to enable us to say how much a change in one variable affects another. If the price for Honda Accords is increased, by how much will quantity demanded decline? If the price of beef goes up, by how much will quantity supplied increase? Finally, econometrics contributes to our understanding of the interrelationships between variables by giving us the ability to test the validity of hypothesized relationships. k k k 1.2 What Is Econometrics About? 3 Econometrics is about how we can use theory and data from economics, business, and the social sciences, along with tools from statistics, to predict outcomes, answer “how much” type questions, and test hypotheses. 1.2.1 k Some Examples Consider the problem faced by decision makers in a central bank. In the United States, the Federal Reserve System and, in particular, the Chair of the Board of Governors of the FRB must make decisions about interest rates. When prices are observed to rise, suggesting an increase in the inflation rate, the FRB must make a decision about whether to dampen the rate of growth of the economy. It can do so by raising the interest rate it charges its member banks when they borrow money (the discount rate) or the rate on overnight loans between banks (the federal funds rate). Increasing these rates sends a ripple effect through the economy, causing increases in other interest rates, such as those faced by would-be investors, who may be firms seeking funds for capital expansion or individuals who wish to buy consumer durables like automobiles and refrigerators. This has the economic effect of increasing costs, and consumers react by reducing the quantity of the durable goods demanded. Overall, aggregate demand falls, which slows the rate of inflation. These relationships are suggested by economic theory. The real question facing the Chair is “How much should we increase the discount rate to slow inflation and yet maintain a stable and growing economy?” The answer will depend on the responsiveness of firms and individuals to increases in the interest rates and to the effects of reduced investment on gross national product (GNP). The key elasticities and multipliers are called parameters. The values of economic parameters are unknown and must be estimated using a sample of economic data when formulating economic policies. Econometrics is about how to best estimate economic parameters given the data we have. “Good” econometrics is important since errors in the estimates used by policymakers such as the FRB may lead to interest rate corrections that are too large or too small, which has consequences for all of us. Every day, decision-makers face “how much” questions similar to those facing the FRB Chair: • A city council ponders the question of how much violent crime will be reduced if an additional million dollars is spent putting uniformed police on the street. • The owner of a local Pizza Hut must decide how much advertising space to purchase in the local newspaper and thus must estimate the relationship between advertising and sales. • Louisiana State University must estimate how much enrollment will fall if tuition is raised by $300 per semester and thus whether its revenue from tuition will rise or fall. • The CEO of Proctor & Gamble must predict how much demand there will be in 10 years for the detergent Tide and how much to invest in new plant and equipment. • A real estate developer must predict by how much population and income will increase to the south of Baton Rouge, Louisiana, over the next few years and whether it will be profitable to begin construction of a gambling casino and golf course. • You must decide how much of your savings will go into a stock fund and how much into the money market. This requires you to make predictions of the level of economic activity, the rate of inflation, and interest rates over your planning horizon. • A public transportation council in Melbourne, Australia, must decide how an increase in fares for public transportation (trams, trains, and buses) will affect the number of travelers who switch to car or bike and the effect of this switch on revenue going to public transportation. k k k 4 CHAPTER 1 An Introduction to Econometrics To answer these questions of “how much,” decision-makers rely on information provided by empirical economic research. In such research, an economist uses economic theory and reasoning to construct relationships between the variables in question. Data on these variables are collected and econometric methods are used to estimate the key underlying parameters and to make predictions. The decision-makers in the above examples obtain their “estimates” and “predictions” in different ways. The FRB has a large staff of economists to carry out econometric analyses. The CEO of Proctor & Gamble may hire econometric consultants to provide the firm with projections of sales. You may get advice about investing from a stock broker, who in turn is provided with econometric projections made by economists working for the parent company. Whatever the source of your information about “how much” questions, it is a good bet that there is an economist involved who is using econometric methods to analyze data that yield the answers. In the next section, we show how to introduce parameters into an economic model and how to convert an economic model into an econometric model. 1.3 k The Econometric Model What is an econometric model, and where does it come from? We will give you a general overview, and we may use terms that are unfamiliar to you. Be assured that before you are too far into this book, all the terminology will be clearly defined. In an econometric model we must first realize that economic relations are not exact. Economic theory does not claim to be able to predict the specific behavior of any individual or firm, but rather describes the average or systematic behavior of many individuals or firms. When studying car sales we recognize that the actual number of Hondas sold is the sum of this systematic part and a random and unpredictable component e that we will call a random error. Thus, an econometric model representing the sales of Honda Accords is Qd = 𝑓 (P, Ps, Pc, INC) + e The random error e accounts for the many factors that affect sales that we have omitted from this simple model, and it also reflects the intrinsic uncertainty in economic activity. To complete the specification of the econometric model, we must also say something about the form of the algebraic relationship among our economic variables. For example, in your first economics courses quantity demanded was depicted as a linear function of price. We extend that assumption to the other variables as well, making the systematic part of the demand relation 𝑓 (P, Ps, Pc, INC) = β1 + β2 P + β3 Ps + β4 Pc + β5 INC The corresponding econometric model is Qd = β1 + β2 P + β3 Ps + β4 Pc + β5 INC + e The coefficients β1 , β2 ,…,β5 are unknown parameters of the model that we estimate using economic data and an econometric technique. The functional form represents a hypothesis about the relationship between the variables. In any particular problem, one challenge is to determine a functional form that is compatible with economic theory and the data. In every econometric model, whether it is a demand equation, a supply equation, or a production function, there is a systematic portion and an unobservable random component. The systematic portion is the part we obtain from economic theory, and includes an assumption about the functional form. The random component represents a “noise” component, which obscures our understanding of the relationship among variables, and which we represent using the random variable e. We use the econometric model as a basis for statistical inference. Using the econometric model and a sample of data, we make inferences concerning the real world, learning something in the process. The ways in which statistical inference are carried out include the following: • Estimating economic parameters, such as elasticities, using econometric methods k k k 1.4 How Are Data Generated? 5 • Predicting economic outcomes, such as the enrollment in two-year colleges in the United States for the next 10 years • Testing economic hypotheses, such as the question of whether newspaper advertising is better than store displays for increasing sales Econometrics includes all of these aspects of statistical inference. As we proceed through this book, you will learn how to properly estimate, predict, and test, given the characteristics of the data at hand. 1.3.1 Causality and Prediction A question that often arises when specifying an econometric model is whether a relationship can be viewed as both causal and predictive or only predictive. To appreciate the difference, consider an equation where a student’s grade in Econometrics GRADE is related to the proportion of class lectures that are skipped SKIP. GRADE = β1 + β2 SKIP + e k We would expect β2 to be negative: the greater the proportion of lectures that are skipped, the lower the grade. But, can we say that skipping lectures causes grades to be lower? If lectures are captured by video, they could be viewed at another time. Perhaps a student is skipping lectures because he or she has a demanding job, and the demanding job does not leave enough time for study, and this is the underlying cause of a poor grade. Or, it might be that skipping lectures comes from a general lack of commitment or motivation, and this is the cause of a poor grade. Under these circumstances, what can we say about the equation that relates GRADE to SKIP? We can still call it a predictive equation. GRADE and SKIP are (negatively) correlated and so information about SKIP can be used to help predict GRADE. However, we cannot call it a causal relationship. Skipping lectures does not cause a low grade. The parameter β2 does not convey the direct causal effect of skipping lectures on grade. It also includes the effect of other variables that are omitted from the equation and correlated with SKIP, such as hours spent studying or student motivation. Economists are frequently interested in parameters that can be interpreted as causal. Honda would like to know the direct effect of a price change on the sales of their Accords. When there is technological improvement in the beef industry, the price elasticities of demand and supply have important implications for changes in consumer and producer welfare. One of our tasks will be to see what assumptions are necessary for an econometric model to be interpreted as causal and to assess whether those assumptions hold. An area where predictive relationships are important is in the use of “big data.” Advances in computer technology have led to storage of massive amounts of information. Travel sites on the Internet keep track of destinations you have been looking at. Google targets you with advertisements based on sites that you have been surfing. Through their loyalty cards, supermarkets keep data on your purchases and identify sale items relevant for you. Data analysts use big data to identify predictive relationships that help predict our behavior. In general, the type of data we have impacts on the specification of an econometric model and the assumptions that we make about it. We turn now to a discussion of different types of data and where they can be found. 1.4 How Are Data Generated? In order to carry out statistical inference we must have data. Where do data come from? What type of real processes generate data? Economists and other social scientists work in a complex world in which data on variables are “observed” and rarely obtained from a controlled experiment. This makes the task of learning about economic parameters all the more difficult. Procedures for using such data to answer questions of economic importance are the subject matter of this book. k k k 6 CHAPTER 1 An Introduction to Econometrics 1.4.1 k Experimental Data One way to acquire information about the unknown parameters of economic relationships is to conduct or observe the outcome of an experiment. In the physical sciences and agriculture, it is easy to imagine controlled experiments. Scientists specify the values of key control variables and then observe the outcome. We might plant similar plots of land with a particular variety of wheat, and then vary the amounts of fertilizer and pesticide applied to each plot, observing at the end of the growing season the bushels of wheat produced on each plot. Repeating the experiment on N plots of land creates a sample of N observations. Such controlled experiments are rare in business and the social sciences. A key aspect of experimental data is that the values of the explanatory variables can be fixed at specific values in repeated trials of the experiment. One business example comes from marketing research. Suppose we are interested in the weekly sales of a particular item at a supermarket. As an item is sold it is passed over a scanning unit to record the price and the amount that will appear on your grocery bill. But at the same time, a data record is created, and at every point in time the price of the item and the prices of all its competitors are known, as well as current store displays and coupon usage. The prices and shopping environment are controlled by store management, so this “experiment” can be repeated a number of days or weeks using the same values of the “control” variables. There are some examples of planned experiments in the social sciences, but they are rare because of the difficulties in organizing and funding them. A notable example of a planned experiment is Tennessee’s Project Star.1 This experiment followed a single cohort of elementary school children from kindergarten through the third grade, beginning in 1985 and ending in 1989. In the experiment children and teachers were randomly assigned within schools into three types of classes: small classes with 13–17 students, regular-sized classes with 22–25 students, and regular-sized classes with a full-time teacher aide to assist the teacher. The objective was to determine the effect of small classes on student learning, as measured by student scores on achievement tests. We will analyze the data in Chapter 7 and show that small classes significantly increase performance. This finding will influence public policy toward education for years to come. 1.4.2 Quasi-Experimental Data It is useful to distinguish between “pure” experimental data and “quasi”-experimental data. A pure experiment is characterized by random assignment. In the example where varying amounts of fertilizer and pesticides are applied to plots of land for growing wheat, the different applications of fertilizer and pesticides are randomly assigned to different plots. In Tennessee’s Project Star, students and teachers are randomly assigned to different sized classes with and without a teacher’s aide. In general, if we have a control group and a treatment group, and we want to examine the effect of a policy intervention or treatment, pure experimental data are such that individuals are randomly assigned to the control and treatment groups. Random assignment is not always possible however, particularly when dealing with human subjects. With quasi-experimental data, allocation to the control and treatment groups is not random but based on another criterion. An example is a study by Card and Krueger that is studied in more detail in Chapter 7. They examined the effect of an increase in New Jersey’s minimum wage in 1992 on the number of people employed in fast-food restaurants. The treatment group was fast-food restaurants in New Jersey. The control group was fast-food restaurants in eastern Pennsylvania where there was no change in the minimum wage. Another example is the effect on spending habits of a change in the income tax rate for individuals above a threshold income. The treatment group is the group with incomes above the threshold. The control group is those with incomes below the threshold. When dealing with quasi-experimental data, one must be aware that the effect of the treatment may be confounded with the effect of the criterion for assignment. ............................................................................................................................................ 1 See https://dataverse.harvard.edu/dataset.xhtml?persistentId=hdl:1902.1/10766 for program description, public use data, and extensive literature. k k k 1.5 Economic Data Types 1.4.3 7 Nonexperimental Data An example of nonexperimental data is survey data. The Public Policy Research Lab at Louisiana State University (www.survey.lsu.edu) conducts telephone and mail surveys for clients. In a telephone survey, numbers are selected randomly and called. Responses to questions are recorded and analyzed. In such an environment, data on all variables are collected simultaneously, and the values are neither fixed nor repeatable. These are nonexperimental data. Such surveys are carried out on a massive scale by national governments. For example, the Current Population Survey (CPS)2 is a monthly survey of about 50,000 households conducted by the U.S. Bureau of the Census. The survey has been conducted for more than 50 years. The CPS website says “CPS data are used by government policymakers and legislators as important indicators of our nation’s economic situation and for planning and evaluating many government programs. They are also used by the press, students, academics, and the general public.” In Section 1.8 we describe some similar data sources. 1.5 Economic Data Types Economic data comes in a variety of “flavors.” In this section we describe and give an example of each. In each example, be aware of the different data characteristics, such as the following: k 1. Data may be collected at various levels of aggregation: ⚬ micro—data collected on individual economic decision-making units such as individuals, households, and firms. ⚬ macro—data resulting from a pooling or aggregating over individuals, households, or firms at the local, state, or national levels. 2. Data may also represent a flow or a stock: ⚬ flow—outcome measures over a period of time, such as the consumption of gasoline during the last quarter of 2018. ⚬ stock—outcome measured at a particular point in time, such as the quantity of crude oil held by ExxonMobil in its U.S. storage tanks on November 1, 2018, or the asset value of the Wells Fargo Bank on July 1, 2018. 3. Data may be quantitative or qualitative: ⚬ quantitative—outcomes such as prices or income that may be expressed as numbers or some transformation of them, such as real prices or per capita income. ⚬ qualitative—outcomes that are of an “either-or” situation. For example, a consumer either did or did not make a purchase of a particular good, or a person either is or is not married. 1.5.1 Time-Series Data A time-series is data collected over discrete intervals of time. Examples include the annual price of wheat in the United States and the daily price of General Electric stock shares. Macroeconomic data are usually reported in monthly, quarterly, or annual terms. Financial data, such as stock prices, can be recorded daily, or at even higher frequencies. The key feature of time-series data is that the same economic quantity is recorded at a regular time interval. For example, the annual real gross domestic product (GDP) for the United States is depicted in Figure 1.1. A few values are given in Table 1.1. For each year, we have the recorded value. The data are annual, or yearly, and have been “deflated” by the Bureau of Economic Analysis to billions of real 2009 dollars. ............................................................................................................................................ 2 www.census.gov/cps/ k k k 8 CHAPTER 1 An Introduction to Econometrics 17,000 16,000 15,000 GDP 14,000 13,000 12,000 11,000 10,000 9,000 1992 1996 2000 2004 Year 2008 2012 2016 FIGURE 1.1 Real U.S. GDP, 1994–2014.3 U.S. Annual GDP (Billions of Real 2009 Dollars) T A B L E 1.1 k 1.5.2 Year GDP 2006 2007 2008 2009 2010 2011 2012 2013 2014 14,613.8 14,873.7 14,830.4 14,418.7 14,783.8 15,020.6 15,354.6 15,583.3 15,961.7 Cross-Section Data A cross-section of data is collected across sample units in a particular time period. Examples are income by counties in California during 2016 or high school graduation rates by state in 2015. The “sample units” are individual entities and may be firms, persons, households, states, or countries. For example, the CPS reports results of personal interviews on a monthly basis, covering items such as employment, unemployment, earnings, educational attainment, and income. In Table 1.2, we report a few observations from the March 2013 survey on the variables RACE, EDUCATION, SEX, and WAGE (hourly wage rate).4 There are many detailed questions asked of the respondents. ............................................................................................................................................ 3 Source: www.bea.gov/national/index.htm 4 In the actual raw data, the variable descriptions are coded differently to the names in Table 1.2. We have used shortened versions for convenience. k k k 1.6 The Research Process T A B L E 1.2 9 Cross-Section Data: CPS, March 2013 Variables Individual RACE EDUCATION SEX WAGE 1 2 3 4 5 6 7 8 9 White White Black White White White White Black White Assoc Degree Master’s Degree Bachelor’s Degree High School Graduate Master’s Degree Master’s Degree Master’s Degree Assoc Degree Some College, No Degree Male Male Male Female Male Female Female Female Female 10.00 60.83 17.80 30.38 12.50 49.50 23.08 28.95 9.20 1.5.3 k Panel or Longitudinal Data A “panel” of data, also known as “longitudinal” data, has observations on individual micro-units that are followed over time. For example, the Panel Study of Income Dynamics (PSID)5 describes itself as “a nationally representative longitudinal study of nearly 9000 U.S. families. Following the same families and individuals since 1969, the PSID collects data on economic, health, and social behavior.” Other national panels exist, and many are described at “Resources for Economists,” at www.rfe.org. To illustrate, data from two rice farms6 are given in Table 1.3. The data are annual observations on rice farms (or firms) over the period 1990–1997. The key aspect of panel data is that we observe each micro-unit, here a farm, for a number of time periods. Here we have amount of rice produced, area planted, labor input, and fertilizer use. If we have the same number of time period observations for each micro-unit, which is the case here, we have a balanced panel. Usually the number of time-series observations is small relative to the number of micro-units, but not always. The Penn World Table7 provides purchasing power parity and national income accounts converted to international prices for 182 countries for some or all of the years 1950–2014. 1.6 The Research Process Econometrics is ultimately a research tool. Students of econometrics plan to do research or they plan to read and evaluate the research of others, or both. This section provides a frame of reference and guide for future work. In particular, we show you the role of econometrics in research. Research is a process, and like many such activities, it flows according to an orderly pattern. Research is an adventure, and can be fun! Searching for an answer to your question, seeking new knowledge, is very addictive—for the more you seek, the more new questions you will find. A research project is an opportunity to investigate a topic that is important to you. Choosing a good research topic is essential if you are to complete a project successfully. A starting point is the question “What are my interests?” Interest in a particular topic will add pleasure to the ............................................................................................................................................ 5 http://psidonline.isr.umich.edu 6 These data were used by O’Donnell, C.J. and W.E. Griffiths (2006), Estimating State-Contingent Production Frontiers, American Journal of Agricultural Economics, 88(1), 249–266. 7 www.rug.nl/ggdc/productivity/pwt k k k 10 CHAPTER 1 An Introduction to Econometrics T A B L E 1.3 k Panel Data from Two Rice Farms FARM YEAR PROD AREA LABOR 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 1990 1991 1992 1993 1994 1995 1996 1997 1990 1991 1992 1993 1994 1995 1996 1997 7.87 7.18 8.92 7.31 7.54 4.51 4.37 7.27 10.35 10.21 13.29 18.58 17.07 16.61 12.28 14.20 2.50 2.50 2.50 2.50 2.50 2.50 2.25 2.15 3.80 3.80 3.80 3.80 3.80 4.25 4.25 3.75 160 138 140 127 145 123 123 87 184 151 185 262 174 244 159 133 FERT 207.5 295.5 362.5 338.0 337.5 207.2 345.0 222.8 303.5 206.0 374.5 421.0 595.7 234.8 479.0 170.0 research effort. Also, if you begin working on a topic, other questions will usually occur to you. These new questions may put another light on the original topic or may represent new paths to follow that are even more interesting to you. The idea may come after lengthy study of all that has been written on a particular topic. You will find that “inspiration is 99% perspiration.” That means that after you dig at a topic long enough, a new and interesting question will occur to you. Alternatively, you may be led by your natural curiosity to an interesting question. Professor Hal Varian8 suggests that you look for ideas outside academic journals—in newspapers, magazines, etc. He relates a story about a research project that developed from his shopping for a new TV set. By the time you have completed several semesters of economics classes, you will find yourself enjoying some areas more than others. For each of us, specialized areas such as health economics, economic development, industrial organization, public finance, resource economics, monetary economics, environmental economics, and international trade hold a different appeal. If you find an area or topic in which you are interested, consult the Journal of Economic Literature (JEL) for a list of related journal articles. The JEL has a classification scheme that makes isolating particular areas of study an easy task. Alternatively, type a few descriptive words into your favorite search engine and see what pops up. Once you have focused on a particular idea, begin the research process, which generally follows steps like these: 1. Economic theory gives us a way of thinking about the problem. Which economic variables are involved, and what is the possible direction of the relationship(s)? Every research project, given the initial question, begins by building an economic model and listing the questions (hypotheses) of interest. More questions will arise during the research project, but it is good to list those that motivate you at the project’s beginning. 2. The working economic model leads to an econometric model. We must choose a functional form and make some assumptions about the nature of the error term. ............................................................................................................................................ 8 Varian, H. How to Build an Economic Model in Your Spare Time, The American Economist, 41(2), Fall 1997, pp. 3–10. k k k 1.7 Writing an Empirical Research Paper 11 3. Sample data are obtained and a desirable method of statistical analysis chosen, based on initial assumptions and an understanding of how the data were collected. 4. Estimates of the unknown parameters are obtained with the help of a statistical software package, predictions are made, and hypothesis tests are performed. 5. Model diagnostics are performed to check the validity of assumptions. For example, were all of the right-hand side explanatory variables relevant? Was an adequate functional form used? 6. The economic consequences and the implications of the empirical results are analyzed and evaluated. What economic resource allocation and distribution results are implied, and what are their policy-choice implications? What remaining questions might be answered with further study or with new and better data? These steps provide some direction for what must be done. However, research always includes some surprises that may send you back to an earlier point in your research plan or that may even cause you to revise it completely. Research requires a sense of urgency, which keeps the project moving forward, the patience not to rush beyond careful analysis, and the willingness to explore new ideas. 1.7 k Writing an Empirical Research Paper Research rewards you with new knowledge, but it is incomplete until a research paper or report is written. The process of writing forces the distillation of ideas. In no other way will your depth of understanding be so clearly revealed. When you have difficulty explaining a concept or thought, it may mean that your understanding is incomplete. Thus, writing is an integral part of research. We provide this section as a building block for future writing assignments. Consult it as needed. You will find other tips on writing economics papers on the book website, www.principlesofeconometrics.com. 1.7.1 Writing a Research Proposal After you have selected a specific topic, it is a good idea to write up a brief project summary, or proposal. Writing it will help to focus your thoughts about what you really want to do. Show it to your colleagues or instructor for preliminary comments. The summary should be short, usually no longer than 500 words, and should include the following: 1. A concise statement of the problem 2. Comments on the information that is available, with one or two key references 3. A description of the research design that includes a. the economic model b. the econometric estimation and inference methods c. data sources d. estimation, hypothesis testing, and prediction procedures, including the econometric software and version used 4. The potential contribution of the research 1.7.2 A Format for Writing a Research Report Economic research reports have a standard format in which the various steps of the research project are discussed and the results interpreted. The following outline is typical. k k k 12 k CHAPTER 1 An Introduction to Econometrics 1. Statement of the Problem The place to start your report is with a summary of the questions you wish to investigate as well as why they are important and who should be interested in the results. This introductory section should be nontechnical and should motivate the reader to continue reading the paper. It is also useful to map out the contents of the following sections of the report. This is the first section to work on and also the last. In today’s busy world, the reader’s attention must be garnered very quickly. A clear, concise, well-written introduction is a must and is arguably the most important part of the paper. 2. Review of the Literature Briefly summarize the relevant literature in the research area you have chosen and clarify how your work extends our knowledge. By all means, cite the works of others who have motivated your research, but keep it brief. You do not have to survey everything that has been written on the topic. 3. The Economic Model Specify the economic model that you used and define the economic variables. State the model’s assumptions and identify hypotheses that you wish to test. Economic models can get complicated. Your task is to explain the model clearly, but as briefly and simply as possible. Don’t use unnecessary technical jargon. Use simple terms instead of complicated ones when possible. Your objective is to display the quality of your thinking, not the extent of your vocabulary. 4. The Econometric Model Discuss the econometric model that corresponds to the economic model. Make sure you include a discussion of the variables in the model, the functional form, the error assumptions, and any other assumptions that you make. Use notation that is as simple as possible, and do not clutter the body of the paper with long proofs or derivations; these can go into a technical appendix. 5. The Data Describe the data you used, as well as the source of the data and any reservations you have about their appropriateness. 6. The Estimation and Inference Procedures Describe the estimation methods you used and why they were chosen. Explain hypothesis testing procedures and their usage. Indicate the software used and the version, such as Stata 15 or EViews 10. 7. The Empirical Results and Conclusions Report the parameter estimates, their interpretation, and the values of test statistics. Comment on their statistical significance, their relation to previous estimates, and their economic implications. 8. Possible Extensions and Limitations of the Study Your research will raise questions about the economic model, data, and estimation techniques. What future research is suggested by your findings, and how might you go about performing it? 9. Acknowledgments It is appropriate to recognize those who have commented on and contributed to your research. This may include your instructor, a librarian who helped you find data, or a fellow student who read and commented on your paper. 10. References An alphabetical list of the literature you cite in your study, as well as references to the data sources you used. Once you’ve written the first draft, use your computer’s spell-check software to check for spelling errors. Have a friend read the paper, make suggestions for clarifying the prose, and check your logic and conclusions. Before you submit the paper, you should eliminate as many errors as possible. Your work should look good. Use a word processor, and be consistent with font sizes, section headings, style of footnotes, references, and so on. Often software developers provide templates for term papers and theses. A little searching for a good paper layout before beginning is a good idea. Typos, missing references, and incorrect formulas can spell doom for an otherwise excellent paper. Some do’s and don’ts are summarized nicely, and with good humor, by Deidre N. McClosky in Economical Writing, 2nd edition (Prospect Heights, IL: Waveland Press, Inc., 1999). While it is not a pleasant topic to discuss, you should be aware of the rules of plagiarism. You must not use someone else’s words as if they were your own. If you are unclear about what you can and cannot use, check with the style manuals listed in the next paragraph, or consult k k k 1.8 Sources of Economic Data 13 your instructor. Your university may provide a plagiarism-checking software, such as Turnitin or iThenticate, that will compare your paper to millions of online sources and look for problem areas. There are some free online versions as well. The paper should have clearly defined sections and subsections. The pages, equations, tables, and figures should be numbered. References and footnotes should be formatted in an acceptable fashion. A style guide is a good investment. Two classics are the following: • The Chicago Manual of Style, 16th edition, is available online and in other formats. • A Manual for Writers of Research Papers, Theses, and Dissertations: Chicago Style for Students and Researchers, 8th edition, by Kate L. Turabian; revised by Wayne C. Booth, Gregory G. Colomb, and Joseph M Williams (2013, University of Chicago Press). 1.8 Sources of Economic Data Economic data are much easier to obtain since the development of the World Wide Web. In this section we direct you to some places on the Internet where economic data are accessible. During your study of econometrics, browse some of the sources listed to gain some familiarity with data availability. 1.8.1 Links to Economic Data on the Internet There are a number of fantastic sites on the World Wide Web for obtaining economic data. k Resources for Economists (RFE) www.rfe.org is a primary gateway to resources on the Internet for economists. This excellent site is the work of Bill Goffe. Here you will find links to sites for economic data and sites of general interest to economists. The Data link has these broad data categories: • U.S. Macro and Regional Data Here you will find links to various data sources such as the Bureau of Economic Analysis, Bureau of Labor Statistics, Economic Reports of the President, and the Federal Reserve Banks. • Other U.S. Data Here you will find links to the U.S. Census Bureau, as well as links to many panel and survey data sources. The gateway to U.S. government agencies is FedStats (fedstats.sites.usa.gov). Once there, click on Agencies to see a complete list of U.S. government agencies and links to their homepages. • World and Non-U.S. Data Here there are links to world data, such as at the CIA World Factbook and the Penn World Tables, as well as international organizations such as the Asian Development Bank, the International Monetary Fund, the World Bank, and so on. There are also links to sites with data on specific countries and sectors of the world. • Finance and Financial Markets Here are links to sources of U.S. and world financial data on variables such as exchange rates, interest rates, and share prices. • Journal Data and Program Archives Some economic journals post data used in articles. Links to these journals are provided here. (Many of the articles in these journals will be beyond the scope of undergraduate economics majors.) National Bureau of Economic Research (NBER) access to a great amount of data. There are headings for • Macro Data • Industry Productivity and Digitalization Data • International Trade Data k www.nber.org/data provides k k 14 CHAPTER 1 An Introduction to Econometrics • • • • • Individual Data Healthcare Data—Hospitals, Providers, Drugs, and Devices Demographic and Vital Statistics Patent and Scientific Papers Data Other Data Economagic Some websites make extracting data relatively easy. For example, Economagic (www.economagic.com) is an excellent and easy-to-use source of macro time series (some 100,000 series available). The data series are easily viewed in a copy and paste format, or graphed. 1.8.2 Interpreting Economic Data In many cases it is easier to obtain economic data than it is to understand the meaning of the data. It is essential when using macroeconomic or financial data that you understand the definitions of the variables. Just what is the index of leading economic indicators? What is included in personal consumption expenditures? You may find the answers to some questions like these in your textbooks. Another resource you might find useful is A Guide to Everyday Economic Statistics, 7th edition, by Gary E. Clayton and Martin Gerhard Giesbrecht, (Boston: Irwin/McGraw-Hill, 2009). This slender volume examines how economic statistics are constructed, and how they can be used. 1.8.3 k Obtaining the Data Finding a data source is not the same as obtaining the data. Although there are a great many easy-to-use websites, “easy-to-use” is a relative term. The data will come packaged in a variety of formats. It is also true that there are many, many variables at each of these websites. A primary challenge is identifying the specific variables that you want, and what exactly they measure. The following examples are illustrative. The Federal Reserve Bank of St. Louis 9 has a system called FRED (Federal Reserve Economic Data). Under “Categories” there are links to financial variables, population and labor variables, national accounts, and many others. Data on these variables can be downloaded in a number of formats. For reading the data, you may need specific knowledge of your statistical software. Accompanying Principles of Econometrics, 5e, are computer manuals for Excel, EViews, Stata, SAS, R, and Gretl to aid this process. See the publisher website www.wiley.com/college/hill, or the book website at www.principlesofeconometrics.com for a description of these aids. The CPS (www.census.gov/cps) has a tool called DataFerrett. This tool will help you find and download data series that are of particular interest to you. There are tutorials that guide you through the process. Variable descriptions, as well as the specific survey questions, are provided to aid in your selection. It is somewhat like an Internet shopping site. Desired series are “ticked” and added to a “Shopping Basket.” Once you have filled your basket, you download the data to use with specific software. Other Web-based data sources operate in this same manner. One example is the PSID.10 The Penn World Tables11 offer data downloads in both Excel and Stata formats. You can expect to find massive amounts of readily available data at the various sites we have mentioned, but there is a learning curve. You should not expect to find, download, and process the data without considerable work effort. Being skilled with Excel and statistical software is a must if you plan to regularly use these data sources. ............................................................................................................................................ 9 https://fred.stlouisfed.org 10 http://psidonline.isr.umich.edu 11 www.rug.nl/ggdc/productivity/pwt k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 15 Probability Primer LEARNING OBJECTIVES Remark Learning Objectives and Keywords sections will appear at the beginning of each chapter. We urge you to think about, and possibly write out answers to the questions, and make sure you recognize and can define the keywords. If you are unsure about the questions or answers, consult your instructor. When examples are requested in Learning Objectives sections, you should think of examples not in the book. k Based on the material in this primer, you should be able to 1. Explain the difference between a random variable and its values, and give an example. 9. Use the definition of expected value for a discrete random variable to compute expectations, given a pdf f (x) and a function g(X) of X. 2. Explain the difference between discrete and continuous random variables, and give examples of each. 10. Define the variance of a discrete random variable, and explain in what sense the values of a random variable are more spread out if the variance is larger. 3. State the characteristics of a probability density function (pdf ) for a discrete random variable, and give an example. 11. Use a joint pdf (table) for two discrete random variables to compute probabilities of joint events and to find the (marginal) pdf of each individual random variable. 4. Compute probabilities of events, given a discrete probability function. 5. Explain the meaning of the following statement: ‘‘The probability that the discrete random variable takes the value 2 is 0.3.’’ 12. Find the conditional pdf for one discrete random variable given the value of another and their joint pdf . 6. Explain how the pdf of a continuous random variable is different from the pdf of a discrete random variable. 13. Work with single and double summation notation. 14. Give an intuitive explanation of statistical independence of two random variables, and state the conditions that must hold to prove statistical independence. Give examples of two independent random variables and two dependent random variables. 7. Show, geometrically, how to compute probabilities given a pdf for a continuous random variable. 8. Explain, intuitively, the concept of the mean, or expected value, of a random variable. 15 k k Trim Size: 8in x 10in 16 k Page 16 Probability Primer 15. Define the covariance and correlation between two random variables, and compute these values given a joint probability function of two discrete random variables. 17. Use Statistical Table 1, Cumulative Probabilities for the Standard Normal Distribution, and your computer software to compute probabilities involving normal random variables. 16. Find the mean and variance of a sum of random variables. 18. Use the Law of Iterated Expectations to find the expected value of a random variable. KEYWORDS conditional expectation conditional pdf conditional probability continuous random variable correlation covariance cumulative distribution function discrete random variable expected value k Hill5e c01a.tex V1 - 12/21/2017 2:43pm experiment indicator variable iterated expectation joint probability density function marginal distribution mean normal distribution population probability probability density function random variable standard deviation standard normal distribution statistical independence summation operations variance We assume that you have had a basic probability and statistics course. In this primer, we review some essential probability concepts. Section P.1 defines discrete and continuous random variables. Probability distributions are discussed in Section P.2. Section P.3 introduces joint probability distributions, defines conditional probability and statistical independence. In Section P.4, we digress and discuss operations with summations. In Section P.5, we review the properties of probability distributions, paying particular attention to expected values and variances. In Section P.6, we discuss the important concept of conditioning, and how knowing the value of one variable might provide information about, or help us predict, another variable. Section P.7 summarizes important facts about the normal probability distribution. In Appendix B, “Probability Concepts,” are enhancements and additions to this material. P.1 Random Variables Benjamin Franklin is credited with the saying “The only things certain in life are death and taxes.” While not the original intent, this bit of wisdom points out that almost everything we encounter in life is uncertain. We do not know how many games our football team will win next season. You do not know what score you will make on the next exam. We don’t know what the stock market index will be tomorrow. These events, or outcomes, are uncertain, or random. Probability gives us a way to talk about possible outcomes. A random variable is a variable whose value is unknown until it is observed; in other words, it is a variable that is not perfectly predictable. Each random variable has a set of possible values it can take. If W is the number of games our football team wins next year, then W can take the values 0, 1, 2, …, 13, if there are a maximum of 13 games. This is a discrete random variable since it can take only a limited, or countable, number of values. Other examples of discrete random variables are the number of computers owned by a randomly selected household, and the number of times you will visit your physician next year. A special case occurs when a random variable can only be one of two possible values—for example, in a phone survey, if you are asked if you are a college graduate or not, your answer can only be “yes” or “no.” Outcomes like this can be characterized by an indicator variable taking the values one if yes or zero if no. Indicator variables are discrete and are used to represent qualitative characteristics such as sex (male or female) or race (white or nonwhite). k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.2 Probability Distributions Page 17 17 The U.S. GDP is yet another example of a random variable, because its value is unknown until it is observed. In the third quarter of 2014 it was calculated to be 16,164.1 billion dollars. What the value will be in the second quarter of 2025 is unknown, and it cannot be predicted perfectly. GDP is measured in dollars and it can be counted in whole dollars, but the value is so large that counting individual dollars serves no purpose. For practical purposes, GDP can take any value in the interval zero to infinity, and it is treated as a continuous random variable. Other common macroeconomic variables, such as interest rates, investment, and consumption, are also treated as continuous random variables. In finance, stock market indices, like the Dow Jones Industrial Index, are also treated as continuous. The key attribute of these variables that makes them continuous is that they can take any value in an interval. P.2 k Probability Distributions Probability is usually defined in terms of experiments. Let us illustrate this in the context of a simple experiment. Consider the objects in Table P.1 to be a population of interest. In statistics and econometrics, the term population is an important one. A population is a group of objects, such as people, farms, or business firms, having something in common. The population is a complete set and is the focus of an analysis. In this case the population is the set of ten objects shown in Table P.1. Using this population, we will discuss some probability concepts. In an empirical analysis, a sample of observations is collected from the population of interest, and using the sample observations we make inferences about the population. If we were to select one cell from the table at random (imagine cutting the table into 10 equally sized pieces of paper, stirring them up, and drawing one of the slips without looking), that would constitute a random experiment. Based on this random experiment, we can define several random variables. For example, let the random variable X be the numerical value showing on a slip that we draw. (We use uppercase letters like X to represent random variables in this primer). The term random variable is a bit odd, as it is actually a rule for assigning numerical values to experimental outcomes. In the context of Table P.1, the rule says, “Perform the experiment (stir the slips, and draw one) and for the slip that you obtain assign X to be the number showing.” The values that X can take are denoted by corresponding lowercase letters, x, and in this case the values of X are x = 1, 2, 3, or 4. For the experiment using the population in Table P.1,1 we can create a number of random variables. Let Y be a discrete random variable designating the color of the slip, with Y = 1 denoting T A B L E P.1 The Seussian Slips: A Population 1 2 3 4 4 2 3 3 4 4 ............................................................................................................................................ 1A table suitable for classroom experiments can be obtained at www.principlesofeconometrics.com/poe5/extras/ table_p1. We thank Veronica Deschner McGregor for the suggestion of “One slip, two slip, white slip, blue slip” for this experiment, inspired by Dr. Seuss’s “One Fish Two Fish Red Fish Blue Fish (I Can Read It All by Myself),” Random House Books for Young Readers (1960). k k Trim Size: 8in x 10in 18 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 18 Probability Primer a shaded slip and Y = 0 denoting a slip with no shading. The numerical values that Y can take are y = 0, 1. Consider X, the numerical value on the slip. If the slips are equally likely to be chosen after shuffling, then in a large number of experiments (i.e., shuffling and drawing one of the ten slips), 10% of the time we would observe X = 1, 20% of the time X = 2, 30% of the time X = 3, and 40% of the time X = 4. These are probabilities that the specific values will occur. We would say, for example, P(X = 3) = 0.3. This interpretation is tied to the relative frequency of a particular outcome’s occurring in a large number of experiment replications. We summarize the probabilities of possible outcomes using a probability density function (pdf ). The pdf for a discrete random variable indicates the probability of each possible value occurring. For a discrete random variable X the value of the pdf f (x) is the probability that the random variable X takes the value x, f (x) = P(X = x). Because f (x) is a probability, it must be true that 0 ≤ f (x) ≤ 1 and, if X takes n possible values x1 ,…, xn , then the sum of their probabilities must be one ( ) ( ) ( ) (P.1) 𝑓 x1 + 𝑓 x2 + · · · + 𝑓 xn = 1 For discrete random variables, the pdf might be presented as a table, such as in Table P.2. As shown in Figure P.1, the pdf may also be represented as a bar graph, with the height of the bar representing the probability with which the corresponding value occurs. The cumulative distribution function (cdf ) is an alternative way to represent probabilities. The cdf of the random variable X, denoted F(x), gives the probability that X is less than or equal to a specific value x. That is, F(x) = P(X ≤ x) (P.2) k T A B L E P.2 Probability Density Function of X x f (x) 1 2 3 4 0.1 0.2 0.3 0.4 k 0.45 0.40 Probability 0.35 0.30 0.25 0.20 0.15 0.10 X value 0.05 1 2 3 4 FIGURE P.1 Probability density function for X. k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.2 Probability Distributions E X A M P L E P.1 Page 19 19 Using a cdf Using the probabilities in Table P.2, we find that F(1) = P(X ≤ 1) = 0.1, F(2) = P(X ≤ 2) = 0.3, F(3) = P(X ≤ 3) = 0.6, and F(4) = P(X ≤ 4) = 1. For example, using the pdf f (x) we compute the probability that X is less than or equal to 2 as F(2) = P(X ≤ 2) = P(X = 1) + P(X = 2) = 0.1 + 0.2 = 0.3 Since the sum of the probabilities P(X = 1) + P(X = 2) + P(X = 3) + P(X = 4) = 1, we can compute the probability that X is greater than 2 as P(X > 2) = 1 − P(X ≤ 2) = 1 − F(2) = 1 − 0.3 = 0.7 An important difference between the pdf and cdf for X is revealed by the question “Using the probability distribution in Table P.2, what is the probability that X = 2.5?” This probability is zero because X cannot take this value. The question “What is the probability that X is less than or equal to 2.5?” does have an answer. F(2.5) = P(X ≤ 2.5) = P(X = 1) + P(X = 2) = 0.1 + 0.2 = 0.3 The cumulative probability can be calculated for any x between −∞ and +∞. Continuous random variables can take any value in an interval and have an uncountable number of values. Consequently, the probability of any specific value is zero. For continuous random variables, we talk about outcomes being in a certain range. Figure P.2 illustrates the pdf f (x) of a continuous random variable X that takes values of x from 0 to infinity. The shape is representative of the distribution for an economic variable such as an individual’s income or wage. Areas under the curve represent probabilities that X falls in an interval. The cdf F(x) is defined as in (P.2). For this distribution, P(10 < X < 20) = F(20) − F(10) = 0.52236 − 0.17512 k = 0.34724 (P.3) How are these areas obtained? The integral from calculus gives the area under a curve. We will not compute many integrals in this book.2 Instead, we will use the computer and compute cdf values and probabilities using software commands. 0 f(x) P(10 < X < 20) 0 10 20 x 30 40 50 FIGURE P.2 Probability density function for a continuous random variable. ............................................................................................................................................ 2 See Appendix A.4 for a brief explanation of integrals, and illustrations using integrals to compute probabilities in Appendix B.2.1. The calculations in (P.3) are explained in Appendix B.3.9. k k Trim Size: 8in x 10in 20 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 20 Probability Primer Joint, Marginal, and Conditional Probabilities P.3 Working with more than one random variable requires a joint probability density function. For the population in Table P.1 we defined two random variables, X the numeric value of a randomly drawn slip and the indicator variable Y that equals 1 if the selected slip is shaded, and 0 if it is not shaded. Using the joint pdf for X and Y we can say “The probability of selecting a shaded 2 is 0.10.” This is a joint probability because we are talking about the probability of two events occurring simultaneously; the selection takes the value X = 2 and the slip is shaded so that Y = 1. We can write this as P(X = 2 and Y = 1) = P(X = 2, Y = 1) = 𝑓 (x = 2, y = 1) = 0.1 The entries in Table P.3 are probabilities f (x, y) = P(X = x, Y = y) of joint outcomes. Like the pdf of a single random variable, the sum of the joint probabilities is 1. P.3.1 Marginal Distributions Given a joint pdf , we can obtain the probability distributions of individual random variables, which are also known as marginal distributions. In Table P.3, we see that a shaded slip, Y = 1, can be obtained with the values x = 1, 2, 3, and 4. The probability that we select a shaded slip is the sum of the probabilities that we obtain a shaded 1, a shaded 2, a shaded 3, and a shaded 4. The probability that Y = 1 is P(Y = 1) = 𝑓Y (1) = 0.1 + 0.1 + 0.1 + 0.1 = 0.4 k k This is the sum of the probabilities across the second row of the table. Similarly the probability of drawing a white slip is the sum of the probabilities across the first row of the table, and P(Y = 0) = fY (0) = 0 + 0.1 + 0.2 + 0.3 = 0.6, where fY (y) denotes the pdf of the random variable Y. The probabilities P(X = x) are computed similarly by summing down across the values of Y. The joint and marginal distributions are often reported as in Table P.4.3 Joint Probability Density Function for X and Y T A B L E P.3 x T A B L E P.4 y 1 2 3 4 0 1 0 0.1 0.1 0.1 0.2 0.1 0.3 0.1 Joint and Marginal Probabilities y/x 1 2 3 4 f (y) 0 1 f (x) 0 0.1 0.1 0.1 0.1 0.2 0.2 0.1 0.3 0.3 0.1 0.4 0.6 0.4 1.0 ............................................................................................................................................ 3 Similar calculations for continuous random variables use integration. See Appendix B.2.3 for an illustration. k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.3 Joint, Marginal, and Conditional Probabilities P.3.2 Page 21 21 Conditional Probability What is the probability that a randomly chosen slip will take the value 2 given that it is shaded? This question is about the conditional probability of the outcome X = 2 given that the outcome Y = 1 has occurred. The effect of the conditioning is to reduce the set of possible outcomes. Conditional on Y = 1 we only consider the four possible slips that are shaded. One of them is a 2, so the conditional probability of the outcome X = 2 given that Y = 1 is 0.25. There is a one in four chance of selecting a 2 given only the shaded slips. Conditioning reduces the size of the population under consideration, and conditional probabilities characterize the reduced population. For discrete random variables the probability that the random variable X takes the value x given that Y = y is written P(X = x|Y = y). This conditional probability is given by the conditional pdf f (x|y) P(X = x, Y = y) 𝑓 (x, y) = (P.4) 𝑓 (x|y) = P(X = x|Y = y) = P(Y = y) 𝑓Y (y) where fY (y) is the marginal pdf of Y. E X A M P L E P.2 Calculating a Conditional Probability 𝑓 (x = 2|y = 1) = P(X = 2|Y = 1) Using the marginal probability P(Y =1) = 0.4, the conditional pdf of X given Y = 1 is obtained by using (P.4) for each value of X. For example, 𝑓 (x = 2, y = 1) P(X = 2, Y = 1) = P(Y = 1) 𝑓Y (1) 0.1 = 0.25 = 0.4 = k A key point to remember is that by conditioning we are considering only the subset of a population for which the condition holds. Probability calculations are then based on the “new” population. We can repeat this process for each value of X to obtain the complete conditional pdf given in Table P.5. P.3.3 Statistical Independence When selecting a shaded slip from Table P.1, the probability of selecting each possible outcome, x = 1, 2, 3, and 4 is 0.25. In the population of shaded slips the numeric values are equally likely. The probability of randomly selecting X = 2 from the entire population, from the marginal pdf , is P(X = 2) = fX (2) = 0.2. This is different from the conditional probability. Knowing that the slip is shaded tells us something about the probability of obtaining X = 2. Such random variables are dependent in a statistical sense. Two random variables are statistically independent, or simply independent, if the conditional probability that X = x given that Y = y is the same as the unconditional probability that X = x. This means, if X and Y are independent random variables, then P(X = x|Y = y) = P(X = x) T A B L E P.5 x 𝑓 (x|y = 1) (P.5) Conditional Probability of X Given Y = 1 1 2 3 4 0.25 0.25 0.25 0.25 k k Trim Size: 8in x 10in 22 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 22 Probability Primer Equivalently, if X and Y are independent, then the conditional pdf of X given Y = y is the same as the unconditional, or marginal, pdf of X alone. 𝑓 (x|y) = 𝑓 (x, y) = 𝑓X (x) 𝑓Y (y) (P.6) Solving (P.6) for the joint pdf , we can also say that X and Y are statistically independent if their joint pdf factors into the product of their marginal pdf s P(X = x, Y = y) = 𝑓 (x, y) = 𝑓X (x) 𝑓Y (y) = P(X = x) × P(Y = y) (P.7) If (P.5) or (P.7) is true for each and every pair of values x and y, then X and Y are statistically independent. This result extends to more than two random variables. The rule allows us to check the independence of random variables X and Y in Table P.4. If (P.7) is violated for any pair of values, then X and Y are not statistically independent. Consider the pair of values X = 1 and Y = 1. P(X = 1, Y = 1) = 𝑓 (1, 1) = 0.1 ≠ 𝑓X (1) 𝑓Y (1) = P(X = 1) × P(Y = 1) = 0.1 × 0.4 = 0.04 The joint probability is 0.1 and the product of the individual probabilities is 0.04. Since these are not equal, we can conclude that X and Y are not statistically independent. P.4 k A Digression: Summation Notation ∑ Throughout this book we will use a summation sign, denoted by the symbol , to shorten algebraic expressions. Suppose the random variable X takes the values x1 , x2 , …, x15 . The sum of these values is x1 + x2 + · · · + x15 . Rather than write this sum out each time we will represent ∑15 ∑15 it as i=1 xi , so that i=1 xi = x1 + x2 + · · · + x15 . If we sum n terms, a general number, then the ∑n summation will be i=1 xi = x1 + x2 + · · · + xn . In this notation ∑ • The symbol is the capital Greek letter sigma and means “the sum of.” • The letter i is called the index of summation. This letter is arbitrary and may also appear as t, j, or k. ∑n • The expression i=1 xi is read “the sum of the terms xi , from i equals 1 to n.” • The numbers 1 and n are the lower limit and upper limit of summation. The following rules apply to the summation operation. Sum 1. The sum of n values x1 , …, xn is n ∑ i=1 xi = x1 + x2 + · · · + xn Sum 2. If a is a constant, then n ∑ i=1 axi = a n ∑ xi i=1 Sum 3. If a is a constant, then n ∑ a = a + a + · · · + a = na i=1 k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.5 Properties of Probability Distributions Page 23 23 Sum 4. If X and Y are two variables, then n ( ∑ n n ) ∑ ∑ xi + yi = xi + yi i=1 i=1 i=1 Sum 5. If X and Y are two variables, then n ( ∑ n n ) ∑ ∑ axi + byi = a xi + b yi i=1 i=1 i=1 Sum 6. The arithmetic mean (average) of n values of X is n ∑ x= i=1 n xi = x1 + x2 + · · · + xn n Sum 7. A property of the arithmetic mean (average) is that n ( ∑ n n n n n ) ∑ ∑ ∑ ∑ ∑ xi − x = xi − x = xi − nx = xi − xi = 0 i=1 i=1 i=1 i=1 i=1 i=1 Sum 8. We often use an abbreviated form of the summation notation. For example, if f (x) is a function of the values of X, n ( ) ( ) ( ) ( ) ∑ 𝑓 xi = 𝑓 x1 + 𝑓 x2 + · · · + 𝑓 xn i=1 ∑ ( ) = 𝑓 xi (“Sum over all values of the index i ”) i = k ∑ 𝑓 (x) (“Sum over all possible values of X”) x Sum 9. Several summation signs can be used in one expression. Suppose the variable Y takes n values and X takes m values, and let f (x, y) = x + y. Then the double summation of this function is m n m n ( ) ∑ ) ∑ ∑ ( ∑ xi + yj 𝑓 xi , yj = i=1 j=1 i=1 j=1 To evaluate such expressions work from the innermost sum outward. First set i = 1 and sum over all values of j, and so on. That is, m n m[ ( ) ∑ ) ( ) ( )] ∑ ∑ ( 𝑓 xi , y1 + 𝑓 xi , y2 + · · · + 𝑓 xi , yn 𝑓 xi , yj = i=1 j=1 i=1 The order of summation does not matter, so m n n m ) ∑ ) ∑ ∑ ( ∑ ( 𝑓 xi , yj = 𝑓 xi , yj i=1 j=1 P.5 j=1 i=1 Properties of Probability Distributions Figures P.1 and P.2 give us a picture of how frequently values of the random variables will occur. Two key features of a probability distribution are its center (location) and width (dispersion). A key measure of the center is the mean, or expected value. Measures of dispersion are variance, and its square root, the standard deviation. k k Trim Size: 8in x 10in 24 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 24 Probability Primer P.5.1 Expected Value of a Random Variable The mean of a random variable is given by its mathematical expectation. If X is a discrete random variable taking the values x1 , … , xn , then the mathematical expectation, or expected value, of X is ) ) ) ( ( ( (P.8) E(X) = x1 P X = x1 + x2 P X = x2 + · · · + xn P X = xn The expected value, or mean, of X is a weighted average of its values, the weights being the probabilities that the values occur. The uppercase letter “E” represents the expected value operation. E(X) is read as “the expected value of X.” The expected value of X is also called the mean of X. The mean is often symbolized by μ or μX . It is the average value of the random variable in an infinite number of repetitions of the underlying experiment. The mean of a random variable is the population mean. We use Greek letters for population parameters because later on we will use data to estimate these real world unknowns. In particular, keep separate the population mean μ and the arithmetic (or sample) mean x that we introduced in Section P.4 as Sum 6. This can be particularly confusing when a conversation includes the term “mean” without the qualifying term “population” or “arithmetic.” Pay attention to the usage context. E X A M P L E P.3 Calculating an Expected Value For the population in Table P.1, the expected value of X is E(X) = 1 × P(X = 1) + 2 × P(X = 2) + 3 × P(X = 3) + 4 × P(X = 4) = (1 × 0.1) + (2 × 0.2) + (3 × 0.3) + (4 × 0.4) = 3 k k For a discrete random variable the probability that X takes the value x is given by its pdf f (x), P(X = x) = f (x). The expected value in (P.8) can be written equivalently as ( ) ( ) ( ) μX = E(X) = x1 𝑓 x1 + x2 𝑓 x2 + · · · + xn 𝑓 xn n ( ) ∑ ∑ (P.9) = xi 𝑓 xi = x𝑓 (x) x i=1 Using (P.9), the expected value of X, the numeric value on a randomly drawn slip from Table P.1 is μX = E(X) = 4 ∑ x𝑓 (x) = (1 × 0.1) + (2 × 0.2) + (3 × 0.3) + (4 × 0.4) = 3 x=1 What does this mean? Draw one “slip” at random from Table P.1, and observe its numerical value X. This constitutes an experiment. If we repeat this experiment many times, the values x = 1, 2, 3, and 4 will appear 10%, 20%, 30%, and 40% of the time, respectively. The arithmetic average of all the numerical values will approach μX = 3, as the number of experiments becomes large. The key point is that the expected value of the random variable is the average value that occurs in many repeated trials of an experiment. For continuous random variables, the interpretation of the expected value of X is unchanged—it is the average value of X if many values are obtained by repeatedly performing the underlying random experiment.4 ............................................................................................................................................ 4 Since there are now an uncountable number of values to sum, mathematically we must replace the “summation over all possible values” in (P.9) by the “integral over all possible values.” See Appendix B.2.2 for a brief discussion. k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.5 Properties of Probability Distributions P.5.2 Page 25 25 Conditional Expectation Many economic questions are formulated in terms of conditional expectation, or the conditional mean. One example is “What is the mean (expected value) wage of a person who has 16 years of education?” In expected value notation, what is E(WAGE|EDUCATION = 16)? For a discrete random variable, the calculation of conditional expected value uses (P.9) with the conditional pdf 𝑓 (x|y) replacing f (x), so that ∑ μX|Y = E(X|Y = y) = x𝑓 (x|y) x E X A M P L E P.4 Calculating a Conditional Expectation Using the population in Table P.1, what is the expected numerical value of X given that Y = 1, the slip is shaded? The conditional probabilities 𝑓 (x|y = 1) are given in Table P.5. The conditional expectation of X is E(X|Y = 1) = 4 ∑ x𝑓 (x|1) = 1 × 𝑓 (1|1) + 2 × 𝑓 (2|1) The average value of X in many repeated trials of the experiment of drawing from the shaded slips is 2.5. This example makes a good point about expected values in general, namely that the expected value of X does not have to be a value that X can take. The expected value of X is not the value that you expect to occur in any single experiment. x=1 + 3 × 𝑓 (3|1) + 4 × 𝑓 (4|1) = 1(0.25) + 2(0.25) + 3(0.25) + 4(0.25) = 2.5 k What is the conditional expectation of X given that Y = y if the random variables are statistically independent? If X and Y are statistically independent the conditional pdf f (x|y) equals the pdf of X alone, f (x), as shown in (P.6). The conditional expectation is then ∑ ∑ E(X|Y = y) = x𝑓 (x|y) = x𝑓 (x) = E(X) x x If X and Y are statistically independent, conditioning does not affect the expected value. P.5.3 Rules for Expected Values Functions of random variables are also random. If g(X) is a function of the random variable X, such as g(X) = X 2 , then g(X) is also random. If X is a discrete random variable, then the expected value of g(X) is obtained using calculations similar to those in (P.9). ] ∑ [ (P.10) E g(X) = g(x) 𝑓 (x) x For example, if a is a constant, then g(X) = aX is a function of X, and ] ∑ [ E(aX) = E g(X) = g(x) 𝑓 (x) x ∑ ∑ = ax𝑓 (x) = a x𝑓 (x) x x = aE(X) Similarly, if a and b are constants, then we can show that E(aX + b) = aE(X) + b If g1 (X) and g2 (X) are functions of X, then ] [ ] [ ] [ E g1 (X) + g2 (X) = E g1 (X) + E g2 (X) (P.11) (P.12) This rule extends to any number of functions. Remember the phrase “the expected value of a sum is the sum of the expected values.” k k Trim Size: 8in x 10in 26 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 26 Probability Primer P.5.4 Variance of a Random Variable The variance of a discrete or continuous random variable X is the expected value of ]2 [ g(X) = X − E(X) The variance of a random variable is important in characterizing the scale of measurement and the spread of the probability distribution. We give it the symbol σ2 , or σ2X , read “sigma squared.” The variance σ2 has a Greek symbol because it is a population parameter. Algebraically, letting E(X) = μ, using the rules of expected values and the fact that E(X) = μ is not random, we have var(X) = σ2X = E(X − μ)2 ) ( ) ( = E X 2 − 2μX + μ2 = E X 2 − 2μE(X) + μ2 ( ) = E X 2 − μ2 (P.13) We use the letters “var” to represent variance, and var(X) ( is) read as “the variance of X,” where X is a random variable. The calculation var(X) = E X2 − μ2 is usually simpler than var(X) = E(X − μ)2 , but the solution is the same. E X A M P L E P.5 Calculating a Variance For the population in Table P.1, we have shown that E(X) = μ = 3. Using (P.10), the expectation of the random variable g(X) = X 2 is k Then, the variance of the random variable X is ( ) var(X) = σ2X = E X 2 − μ2 = 10 − 32 = 1 4 4 ( ) ∑ ∑ x2 𝑓 (x) E X 2 = g(x) 𝑓 (x) = x=1 k x=1 ] [ ] [ ] [ ] [ = 12 × 0.1 + 22 × 0.2 + 32 × 0.3 + 42 × 0.4 = 10 The square root of the variance is called the standard deviation; it is denoted by σ or sometimes as σX if more than one random variable is being discussed. It also measures the spread or dispersion of a probability distribution and has the advantage of being in the same units of measure as the random variable. A useful property of variances is the following. Let a and b be constants, then var(aX + b) = a2 var(X) (P.14) An additive constant like b changes the mean (expected value) of a random variable, but it does not affect its dispersion (variance). A multiplicative constant like a affects the mean, and it affects the variance by the square of the constant. To see this, let Y = aX + b. Using (P.11) E(Y) = μY = aE(X) + b = aμX + b Then [( )2 ] [( ( ] ))2 = E aX + b − aμX + b var(aX + b) = var(Y) = E Y − μY [ ( [( )2 ] )2 ] = E a2 X − μX = E aX − aμX [( )2 ] = a2 var(X) = a2 E X − μX The variance of a random variable is the average squared difference between the random variable X and its mean value μX . The larger the variance of a random variable, the more “spread out” the k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.5 Properties of Probability Distributions Page 27 27 0.8 0.7 0.6 f (x) 0.5 0.4 Smaller variance Larger variance 0.3 0.2 0.1 0.0 –1 0 1 2 3 x 4 5 6 7 FIGURE P.3 Distributions with different variances. values of the random variable are. Figure P.3 shows two pdf s for a continuous random variable, both with mean μ = 3. The distribution with the smaller variance (the solid curve) is less spread out about its mean. P.5.5 Expected Values of Several Random Variables Let X and Y be random variables. The rule “the expected value of the sum is the sum of the expected values” applies. Then5 k E(X + Y) = E(X) + E(Y) (P.15) E(aX + bY + c) = aE(X) + bE(Y) + c (P.16) Similarly The product of random variables is not as easy. E(XY) = E(X)E(Y) if X and Y are independent. These rules can be extended to more random variables. P.5.6 Covariance Between Two Random Variables The covariance between X and Y is a measure of linear association between them. Think about two continuous variables, such as height and weight of children. We expect that there is an association between height and weight, with taller than average children tending to weigh more than the average. The product of X minus its mean times Y minus its mean is ( )( ) X − μX Y − μY (P.17) In Figure P.4, we plot values (x and y) of X and Y that have been constructed so that E(X) = E(Y) = 0. The x and y values of X and Y fall predominately in quadrants I and III, so that the arithmetic average of the values (x − μX )(y − μY ) is positive. We define the covariance between two random variables as the expected (population average) value of the product in (P.17). [( )( )] cov(X, Y) = σXY = E X − μX Y − μY = E(XY) − μX μY (P.18) We use the letters “cov” to represent covariance, and cov(X, Y) is read as “the covariance between X and Y,” where X and Y are random variables. The covariance σXY of the random variables underlying Figure P.4 is positive, which tells us that when the values x are greater ............................................................................................................................................ 5 These results are proven in Appendix B.1.4. k k Trim Size: 8in x 10in Page 28 Probability Primer II I III IV 0 –4 y 4 28 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k –4 0 x 4 FIGURE P.4 Correlated data. than μX , then the values y also tend to be greater than μY ; and when the values x are below μX , then the values y also tend to (be less )( than μY . )If the random variables values tend primarily to fall in quadrants II and IV, then x − μX y − μY will tend to be negative and σXY will be negative. If the random variables values are spread evenly across the four quadrants, and show neither positive nor negative association, then the covariance is zero. The sign of σXY tells us whether the two random variables X and Y are positively associated or negatively associated. Interpreting the actual value of σXY is difficult because X and Y may have different units of measurement. Scaling the covariance by the standard deviations of the variables eliminates the units of measurement, and defines the correlation between X and Y k σ cov(X, Y) = XY ρ= √ √ var(X) var(Y) σX σY (P.19) As with the covariance, the correlation ρ between two random variables measures the degree of linear association between them. However, unlike the covariance, the correlation must lie between –1 and 1. Thus, the correlation between X and Y is 1 or –1 if X is a perfect positive or negative linear function of Y. If there is no linear association between X and Y, then cov(X, Y) = 0 and ρ = 0. For other values of correlation the magnitude of the absolute value |ρ| indicates the “strength” of the linear association between the values of the random variables. In Figure P.4, the correlation between X and Y is ρ = 0.5. E X A M P L E P.6 Calculating a Correlation To illustrate the calculation, reconsider the population in Table P.1 with joint pdf given in Table P.4. The expected value of XY is E(XY) = 1 4 ∑ ∑ xy𝑓 (x, y) The random variable X has expected value E(X) = μX = 3 and the random variable Y has expected value E(Y) = μY = 0.4. Then the covariance between X and Y is cov(X, Y) = σXY = E(XY) − μX μY = 1 − 3 × (0.4) = −0.2 y=0 x=1 = (1 × 0 × 0) + (2 × 0 × 0.1) + (3 × 0 × 0.2) + (4 × 0 × 0.3) + (1 × 1 × 0.1) + (2 × 1 × 0.1) + (3 × 1 × 0.1) + (4 × 1 × 0.1) The correlation between X and Y is cov(X, Y) −0.2 = √ = −0.4082 ρ= √ √ √ var(X) var(Y) 1 × 0.24 = 0.1 + 0.2 + 0.3 + 0.4 =1 k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.6 Conditioning Page 29 29 If X and Y are independent random variables, then their covariance and correlation are zero. The converse of this relationship is not true. Independent random variables X and Y have zero covariance, indicating that there is no linear association between them. However, just because the covariance or correlation between two random variables is zero does not mean that they are necessarily independent. There may be more complicated nonlinear associations such as X 2 + Y 2 = 1. In (P.15) we obtain the expected value of a sum of random variables. There are similar rules for variances. If a and b are constants, then var(aX + bY) = a2 var(X) + b2 var(Y) + 2ab cov(X, Y) (P.20) A significant point to note is that the variance of a sum is not just the sum of the variances. There is a covariance term present. Two special cases of (P.20) are var(X + Y) = var(X) + var(Y) + 2cov(X, Y) (P.21) var(X − Y) = var(X) + var(Y) − 2cov(X, Y) (P.22) To show that (P.22) is true, let Z = X − Y. Using the rules of expected value E(Z) = μZ = E(X) − E(Y) = μX − μY The variance of Z = X − Y is obtained using the basic definition of variance, with some substitution, [( ] [( ( )2 ] ))2 = E X − Y − μX − μY var(X − Y) = var(Z) = E Z − μZ } {[ ( ) ( )]2 X − μX − Y − μY =E )2 )( )} ( ( + Y − μY − 2 X − μX Y − μY [( [( [( )2 ] )2 ] )( )] + E Y − μY − 2E X − μX Y − μY = E X − μX k =E {( X − μX )2 k = var(X) + var(Y) − 2cov(X, Y) If X and Y are independent, or if cov(X, Y) = 0, then var(aX + bY) = a2 var(X) + b2 var(Y) (P.23) var(X ± Y) = var(X) + var(Y) (P.24) These rules extend to more random variables. P.6 Conditioning In Table P.4, we summarized the joint and marginal probability functions for the random variables X and Y defined on the population in Table P.1. In Table P.6 we make two modifications. First, the probabilities are expressed as fractions. The many calculations below are simpler using arithmetic Joint, Marginal, and Conditional Probabilities T A B L E P.6 y/x 0 1 f (x) f (x|y = 0) f (x|y = 1) 1 2 3 4 0 1/10 1/10 0 1/4 1/10 1/10 2/10 1/6 1/4 2/10 1/10 3/10 2/6 1/4 3/10 1/10 4/10 3/6 1/4 f ( y) f ( y|x = 1) f ( y|x = 2) f ( y|x = 3) f ( y|x = 4) 6/10 4/10 0 1 1/2 1/2 k 2/3 1/3 3/4 1/4 Trim Size: 8in x 10in 30 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 30 Probability Primer with fractions. Second, we added the conditional probability functions (P.4) for Y given each of the values that X can take and the conditional probability functions for X given each of the values that Y can take. Now would be a good time for you to review Section P.3.2 on conditional probability. For example, what is the probability that Y = 0 given that X = 2? That is, if we only consider population members with X = 2, what is the probability that Y = 0? There are only two population elements with X = 2, one with Y = 0 and one with Y = 1. The probability of randomly selecting Y = 0 is one-half. For discrete random variables, the conditional probability is calculated as the joint probability divided by the probability of the conditioning event. 𝑓 (y = 0|x = 2) = P(Y = 0|X = 2) = P(X = 2, Y = 0) 1∕10 1 = = P(X = 2) 2∕10 2 In the following sections, we discuss the concepts of conditional expectation and conditional variance. P.6.1 Conditional Expectation Many economic questions are formulated in terms of a conditional expectation, or the conditional mean. One example is “What is the mean wage of a person who has 16 years of education?” In expected value notation, what is E(WAGE|EDUC = 16)? The effect of conditioning on the value of EDUC is to reduce the population of interest to only individuals with 16 years of education. The mean, or expected value, of wage for these individuals may be quite different than the mean wage for all individuals regardless of years of education, E(WAGE), which is the unconditional expectation or unconditional mean. For discrete random variables,6 the calculation of a conditional expected value uses equation (P.9) with the conditional pdf replacing the usual pdf , so that ∑ ∑ E(X|Y = y) = x x𝑓 (x|y) = x xP(X = x|Y = y) (P.25) ∑ ∑ E(Y|X = x) = y y𝑓 (y|x) = y yP(Y = y|X = x) k E X A M P L E P.7 Conditional Expectation Using the population in Table P.1, what is the expected numerical value of X given that Y = 1? The conditional probabilities P(X = x|Y = 1) = 𝑓 (x|y = 1) = 𝑓 (x|1) are given in Table P.6. The conditional expectation of X is ∑4 E(X|Y = 1) = x=1 x𝑓 (x|1) [ ] [ ] [ ] = 1 × 𝑓 (1|1) + 2 × 𝑓 (2|1) + 3 × 𝑓 (3|1) [ ] + 4 × 𝑓 (4|1) = 1(1∕4) + 2(1∕4) + 3(1∕4) + 4(1∕4) = 10∕4 = 5∕2 The average value of X in many repeated trials of the experiment of drawing from the shaded slips (Y = 1) is 2.5. This example makes a good point about expected values in general, namely that the expected value of X does not have to be a value that X can take. The expected value of X is not the value that you expect to occur in any single experiment. It is the average value of X after many repetitions of the experiment. What is the expected value of X given that we only consider values where Y = 0? Confirm that E(X|Y = 0) = 10∕3. For comparison purposes recall from Section P.5.1 that the unconditional expectation of X is E(X) = 3. Similarly, if we condition on the X values, the conditional expectations of Y are ∑ E(Y|X = 1) = y y𝑓 (y|1) = 0(0) + 1(1) = 1 ∑ E(Y|X = 2) = y y𝑓 (y|2) = 0(1∕2) + 1(1∕2) = 1∕2 ∑ E(Y|X = 3) = y y𝑓 (y|3) = 0(2∕3) + 1(1∕3) = 1∕3 ∑ E(Y|X = 4) = y y𝑓 (y|4) = 0(3∕4) + 1(1∕4) = 1∕4 Note that E(Y|X) varies as X varies; it is a function of X. For comparison, the unconditional expectation of Y, E(Y), is E(Y) = ∑ y y𝑓 (y) = 0(6∕10) + 1(4∕10) = 2∕5 ............................................................................................................................................ 6 For continuous random variables the sums are replaced by integrals. See Appendix B.2. k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.6 Conditioning P.6.2 Page 31 31 Conditional Variance The unconditional variance of a discrete random variable X is [( )2 ] ∑( )2 x − μX 𝑓 (x) = var(X) = σ2X = E X − μX (P.26) x It measures how much variation there is in X around the unconditional mean of X, μX . For example, the unconditional variance var(WAGE) measures the variation in WAGE around the unconditional mean E(WAGE). In (P.13) we show that equivalently ( ) ∑ var(X) = σ2X = E X 2 − μ2X = x2 𝑓 (x) − μ2X (P.27) x In Section P.6.1 we discussed how to answer the question “What is the mean wage of a person who has 16 years of education?” Now we ask “How much variation is there in wages for a person who has 16 years of education?” The answer to this question is given by the conditional variance, var (WAGE|EDUC = 16). The conditional variance measures the variation in WAGE around the conditional mean E(WAGE|EDUC = 16) for individuals with 16 years of education. The conditional variance of WAGE for individuals with 16 years of education is the average squared difference in the population between WAGE and the conditional mean of WAGE, } {[ ]2 WAGE − E(WAGE | EDUC = 16) || var(WAGE | EDUC = 16) = E ⏟⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏟⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏟ | EDUC = 16 | ⏟⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏟⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏞⏟ conditional mean | conditional variance k To obtain the conditional variance we modify the definitions of variance in equations (P.26) and (P.27); replace the unconditional mean E(X) = μX with the conditional mean E(X|Y = y), and the unconditional pdf f (x) with the conditional pdf f (x|y). Then } ∑( {[ )2 ]2 | x − E(X|Y = y) 𝑓 (x|y) var(X|Y = y) = E X − E(X|Y = y) |Y = y = (P.28) | x or ]2 ∑ ]2 ( ) [ [ var(X|Y = y) = E X 2 |Y = y − E(X|Y = y) = x2 𝑓 (x|y) − E(X|Y = y) (P.29) x E X A M P L E P.8 Conditional Variance For the population in Table P.1, the unconditional variance of X is var(X) = 1. What is the variance of X given that Y = 1? To use (P.29) first compute ) ( E X 2 |Y = 1 ∑ = x2 𝑓 (x|Y = 1) x = 12 (1∕4) + 22 (1∕4) + 32 (1∕4) + 42 (1∕4) = 15∕2 Then ) ∑ ( E X 2 |Y = 0 = x2 𝑓 (x|Y = 0) x = 12 (0) + 22 (1∕6) + 32 (2∕6) + 42 (3∕6) = 35∕3 Then ) [ ]2 ( var(X|Y = 0) = E X 2 |Y = 0 − E(X|Y = 0) = 35∕3 −(10∕3)2 = 5∕9 ( ) [ ]2 var(X|Y = 1) = E X 2 |Y = 1 − E(X|Y = 1) = 15∕2 −(5∕2)2 = 5∕4 In this case, the conditional variance of X, given that Y = 1, is larger than the unconditional variance of X, var(X) = 1. To calculate the conditional variance of X given that Y = 0, we first obtain k In this case, the conditional variance of X, given that Y = 0, is smaller than the unconditional variance of X, var(X) = 1. These examples have illustrated that in general the conditional variance can be larger or smaller than the unconditional variance. Try working out var(Y|X = 1), var(Y|X = 2), var(Y|X = 3), and var(Y|X = 4). k Trim Size: 8in x 10in 32 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 32 Probability Primer P.6.3 Iterated Expectations The Law of Iterated Expectations says that we can find the expected value of Y in two steps. First, find the conditional expectation E(Y|X). Second, find the expected value E(Y|X) treating X as random. ] ∑ [ (P.30) Law of Iterated Expectations∶ E(Y) = EX E(Y|X) = x E(Y|X = x) 𝑓X (x) In this expression we put an “X” subscript in the expectation EX [E(Y|X)] and the probability function fX (x) to emphasize that we are treating X as random. The Law of Iterated Expectations is true for both discrete and continuous random variables. E X A M P L E P.9 k Iterated Expectation [ ] [ ] ∑ ∑ EX E(Y|X) = EX g(X) = x g(x)𝑓X (x) = x E(Y|X = x) 𝑓X (x) [ ] [ ] = E(Y|X = 1) 𝑓X (1) + E(Y|X = 2) 𝑓X (2) ] [ ] [ + E(Y|X = 3) 𝑓X (3) + E(Y|X = 4) 𝑓X (4) Consider the conditional expectation E(X|Y = y) = ∑ x xf (x|y). As we computed in Section P.6.1, E(X|Y = 0) = 10/3 and E(X|Y = 1) = 5/2. Similarly, the conditional ∑ expectation E(Y|X = x) = y yf (y|x). For the population in Table P.1, these conditional expectations were calculated in Section P.6.1 to be E(Y|X = 1) = 1, E(Y|X = 2) = 1/2, E(Y|X = 3) = 1/3 and E(Y|X = 4) = 1/4. Note that E(Y|X = x) changes when x changes. If X is allowed to vary randomly7 then the conditional expectation varies randomly. The conditional expectation is a function of X, or E(Y|X) = g(X), and is random when viewed this way. Using (P.10) we can find the expected value of g(X). = 1(1∕10) + (1∕2)(2∕10) + (1∕3)(3∕10) + (1∕4)(4∕10) = 2∕5 If we draw many values x from the population in Table P.1, the average of E(Y|X) is 2/5. For comparison the “unconditional” expectation of Y is E(Y) = 2/5. EX [E(Y|X)] and E(Y) are the same. Proof of the Law of Iterated Expectations To prove the Law of Iterated Expectations we make use of relationships between joint, marginal, and conditional pdf s that we introduced in Section P.3. In Section P.3.1 we discussed marginal distributions. Given a joint pdf f (x, y) we can obtain the marginal pdf of y alone fY (y) by summing, for each y, the joint pdf f (x, y) across all values of the variable we wish to eliminate, in this case x. That is, for Y and X, 𝑓 (y) = 𝑓Y (y) = 𝑓 (x) = 𝑓X (x) = ∑ x 𝑓 (x, y) ∑ y 𝑓 (x, y) (P.31) Because f ( ) is used to represent pdf s in general, sometimes we will put a subscript, X or Y, to be very clear about which variable is random. Using equation (P.4) we can define the conditional pdf of y given X = x as 𝑓 (y|x) = 𝑓 (x, y) 𝑓X (x) Rearrange this expression to obtain 𝑓 (x, y) = 𝑓 (y|x) 𝑓X (x) (P.32) A joint pdf is the product of the conditional pdf and the pdf of the conditioning variable. ............................................................................................................................................ 7 Imagine shuffling the population elements and randomly choosing one. This is an experiment and the resulting number showing is a value of X. By doing this repeatedly X varies randomly. k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.6 Conditioning Page 33 33 To show that the Law of Iterated Expectations is true8 we begin with the definition of the expected value of Y, and operate with the summation. [ ] ] [ ∑ ∑ ∑ substitute for 𝑓 (y) E(Y) = y𝑓 (y) = y 𝑓 (x, y) y = ∑ [ y y ∑ y = x [ ∑ ∑ x ∑ y x ] 𝑓 (y|x) 𝑓X (x) ] y𝑓 (y|x) 𝑓X (x) E(Y|x)𝑓X (x) x ] [ = EX E(Y|X) = [ substitute for 𝑓 (x, y) ] [ ] change order of summation [ recognize the conditional expectation ] While this result may seem an esoteric oddity it is very important and widely used in modern econometrics. P.6.4 Variance Decomposition Just as we can break up the expected value using the Law of Iterated Expectations we can decompose the variance of a random variable into two parts. ] ] [ [ (P.33) Variance Decomposition∶ var(Y) = varX E(Y|X) + EX var(Y|X) k This “beautiful” result9 says that the variance of the random variable Y equals the sum of the variance of the conditional mean of Y given X and the mean of the conditional variance of Y given X. In this section we will discuss this result.10 Suppose that we are interested in the wages of the population consisting of working adults. How much variation do wages display in the population? If WAGE is the wage of a randomly drawn population member, then we are asking about the variance of WAGE, that is, var(WAGE). The variance decomposition says ] ] [ [ var(WAGE) = varEDUC E(WAGE|EDUC) + EEDUC var(WAGE|EDUC) E(WAGE|EDUC) is the expected value of WAGE given a specific value of education, such as EDUC = 12 or EDUC = 16. E(WAGE|EDUC = 12) is the average WAGE in the population, given that we only consider workers who have 12 years of education. If EDUC changes then the conditional mean E(WAGE|EDUC) changes, so that E(WAGE|EDUC = 16) is not the same as E(WAGE|EDUC = 12), and in fact we expect E(WAGE|EDUC = 16) > E(WAGE|EDUC = 12); more education means more “human capital” and thus [ the average wage] should be higher. The first component in the variance decomposition varEDUC E(WAGE|EDUC) measures the variation in E(WAGE|EDUC) due to variation in education. [ ] The second part of the variance decomposition is EEDUC var(WAGE|EDUC) . If we restrict our attention to population members who have 12 years of education, the mean wage is E(WAGE|EDUC = 12). Within the group of workers who have 12 years of education we will observe wide ranges of wages. For example, using one sample of CPS data from 2013,11 wages for those with 12 years of education varied from $3.11/hour to $100.00/hour; for those with 16 years of education wages varied from $2.75/hour to $221.10/hour. For workers with 12 and 16 years of education that variation is measured by var(WAGE|EDUC = 12) and ............................................................................................................................................ 8 The proof for continuous variables is in Appendix B.2.4. 9 Tony O’Hagan, “A Thing of Beauty,” Significance Magazine, Volume 9 Issue 3 (June 2012), 26–28. 10 The proof of the variance decomposition is given in Appendix B.1.8 and Example B.1. 11 The data file cps5. k k Trim Size: 8in x 10in 34 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 34 Probability Primer [ ] var(WAGE|EDUC = 16). The term EEDUC var(WAGE|EDUC) measures the average of var(WAGE|EDUC) as education changes. To summarize, the variation of WAGE in the population can be attributed to two sources: variation in the conditional mean E(WAGE|EDUC) and variation due to changes in education in the conditional variance of WAGE given education. P.6.5 Covariance Decomposition Recall that the covariance between two random variables Y and X is cov(X, Y) = [( )( )] E X − μX Y − μY . For discrete random variables this is cov(X, Y) = ∑ ∑( x y x − μX )( ) y − μY 𝑓 (x, y) By using the relationships between marginal, conditional and joint pdf s we can show ) ∑( x − μX E(Y|X = x) 𝑓 (x) cov(X, Y) = (P.34) x Recall that E(Y|X) = g(X) so this result says that the covariance between X and Y can be calculated as the expected value of X, minus its mean, times a function of X, cov(X, Y) = EX [(X − μX )E(Y|X)]. An important special case is important in later chapters. When the conditional expectation of Y given X is a constant, E(Y|X = x) = c, then ) ) ∑( ∑( x − μX E(Y|X = x) 𝑓 (x) = c x − μX 𝑓 (x) = 0 cov(X, Y) = x k x A special case is E(Y|X = x) = 0, which by direct substitution implies cov(X, Y) = 0. E X A M P L E P.10 Covariance Decomposition To illustrate we compute cov(X, Y) for the population in Table P.1 using the covariance decomposition. We have computed that cov(X, Y) = −0.2 in Section P.5.6. The ingredients are the values of the random variable X, its mean μX = 3, the probabilities P(X = x) = f (x) and conditional expectations Using the covariance decomposition we have ) ∑( x − μX E(Y|X = x) 𝑓 (x) cov(X, Y) = x = (1 − 3)(1)(1∕10) + (2 − 3)(1∕2)(2∕10) + (3 − 3)(1∕3)(3∕10) + (4 − 3)(1∕4)(4∕10) = −2∕10 − 1∕10 + 1∕10 = −2∕10 = −0.2 E(Y|X = 1) = 1, E(Y|X = 2) = 1∕2, E(Y|X = 3) = 1∕3 and E(Y|X = 4) = 1∕4 P.7 We see that the covariance decomposition yields the correct result, and it is convenient in this example. The Normal Distribution In the previous sections we discussed random variables and their pdf s in a general way. In real economic contexts, some specific pdf s have been found to be very useful. The most important is the normal distribution. If X(is a normally distributed random variable with mean μ and variance ) σ2 , it is symbolized as X ∼ N μ, σ2 . The pdf of X is given by the impressive formula [ ] −(x − μ)2 1 𝑓 (x) = √ exp , −∞ < x < ∞ (P.35) 2σ2 2πσ2 k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.7 The Normal Distribution Page 35 35 0.4 f (x) 0.3 N(0, 1) N(2, 1) 0.2 N(0, 4) 0.1 0 –4 –2 0 2 4 6 x ( ) FIGURE P.5 Normal probability density functions N 𝛍, 𝛔2 . k where exp(a) denotes the exponential12 function ea . The mean μ and variance σ2 are the parameters of this distribution and determine its center and dispersion. The range of the continuous normal random variable is from minus infinity to plus infinity. Pictures of the normal pdf s are given in Figure P.5 for several values of the mean and variance. Note that the distribution is symmetric and centered at μ. Like all continuous random variables, probabilities involving normal random variables are found as areas under the pdf . For calculating probabilities both computer software and statistical tables values make use of the relation between a normal random variable and its “standardized” equivalent. A standard random variable is one that has a normal pdf with mean 0 and ) ( normal variance 1. If X ∼ N μ, σ2 , then Z= X−μ ∼ N(0, 1) σ (P.36) The standard normal random variable Z is so widely used that its pdf and cdf are given their own special notation. The cdf is denoted Φ(z) = P(Z ≤ z). Computer programs, and Statistical Table 1 in Appendix D give values of Φ(z). The pdf for the standard normal random variable is ( ) 1 ϕ(z) = √ exp −z2 ∕2 , −∞ < z < ∞ 2π Values of the density function are given in Statistical Table 6 in Appendix D. To calculate normal probabilities, remember that the distribution is symmetric, so that P(Z > a) = P(Z < −a), and P(Z > a) = P(Z ≥ a), since the probability of any one point is zero for a continuous random variable. If X ∼ N(μ, σ2 ) and a and b are constants, then ( ) (a − μ) ( X−μ a−μ a − μ) =Φ (P.37) P(X ≤ a) = P ≤ =P Z≤ σ σ σ σ ( ) (a − μ) ( X−μ a−μ a − μ) =1−Φ (P.38) P(X > a) = P > =P Z> σ σ σ σ ( ) ( ) (a − μ) a−μ b−μ b−μ (P.39) P(a ≤ X ≤ b) = P ≤Z≤ =Φ −Φ σ σ σ σ ............................................................................................................................................ 12 See Appendix A.1.2 for a review of exponents. k k Trim Size: 8in x 10in 36 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 36 Probability Primer E X A M P L E P.11 Normal Distribution Probability Calculation For example, if X ∼ N(3, 9), then P(4 ≤ X ≤ 6) = P(0.33 ≤ Z ≤ 1) = Φ(1) − Φ(0.33) = 0.8413 − 0.6293 = 0.2120 k In addition to finding normal probabilities ) we sometimes must find a value zα of a standard nor( zα is called the 100𝛂-percentile. For mal random variable such that P Z ≤ zα (= α. The value ) example, z0.975 is the value of Z such that P Z ≤ z0.975 = 0.975. This particular percentile can be found using Statistical Table 1, Cumulative Probabilities for the Standard Normal Distribution. The cumulative probability associated with the value z =1.96 is P(Z ≤ 1.96) = 0.975, so that the 97.5 percentile is z0.975 = 1.96. Using Statistical Table 1 we can only roughly obtain other percentiles. Using the cumulative probabilities P(Z ≤ 1.64) = 0.9495 and P(Z ≤ 1.65) = 0.9505 we can say that the 95th percentile of the standard normal distribution is between 1.64 and 1.65, and is about 1.645. Luckily computer software makes these approximations )unnecessary. ( ) The inverse normal ( function finds percentiles zα given α. Formally, if P Z ≤ zα = Φ zα = α then zα = Φ−1 (α). Econometric software, even spreadsheets, have the inverse normal function built in. Some commonly used percentiles are shown in Table P.7. In the last column are the percentiles rounded to fewer decimals. It would be useful for you to remember the numbers 2.58, 1.96, and 1.645. An interesting and useful fact about the normal distribution( is that) a weighted sum ) ( of normal random variables has a normal distribution. That is, if X1 ∼ N μ1 , σ21 and X2 ∼ N μ2 , σ22 then ) ( (P.40) Y = a1 X1 + a2 X2 ∼ N μY = a1 μ1 + a2 μ2 , σ2Y = a21 σ21 + a22 σ22 + 2a1 a2 σ12 ) ( where σ12 = cov X1 , X2 . A number of important probability distributions are related to the normal distribution. The t-distribution, the chi-square distribution, and the F-distribution are discussed in Appendix B. T A B L E P.7 Standard Normal Percentiles 𝛂 z𝛂 = 𝚽−1 (𝛂) Rounded 2.57583 2.32635 1.95996 1.64485 1.28155 −1.28155 −1.64485 −1.95996 −2.32635 −2.57583 2.58 2.33 1.96 1.645 1.28 −1.28 −1.645 −1.96 −2.33 −2.58 0.995 0.990 0.975 0.950 0.900 0.100 0.050 0.025 0.010 0.005 k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.7 The Normal Distribution Page 37 37 The Bivariate Normal Distribution P.7.1 Two continuous random variables, X and Y, have a joint normal, or bivariate normal, distribution if their joint pdf takes the form { [( ( ) )( ) x − μX 2 x − μX y − μY 1 exp − − 2ρ 𝑓 (x, y) = √ σX σX σY 2πσX σY 1 − ρ2 } ( ) ]/ ) ( y − μY 2 2 + 2 1−ρ σY where −∞ < x < ∞, −∞ < y < ∞. The parameters μX and μY are the means of X and Y, σ2X and σ2Y are the variances of X and Y, so that σX and σY are the standard deviations. The parameter ρ is the correlation between X and Y. If cov(X, Y) = σXY then σ cov(X, Y) = XY ρ= √ √ var(X) var(Y) σX σY 4 2 –4 y k The complex equation for f (x, y) defines a surface in three-dimensional space. In Figure P.6a13 we depict the surface if μX = μY = 0, σX = σY = 1, and ρ = 0.7. The positive correlation means there is a positive linear association between the values of X and Y, as described in Figure P.4. Figure P.6b depicts the contours of the density, the result of slicing the density horizontally, at a given height. The contours are more “cigar-shaped” the larger the absolute value of the correlation ρ. In Figure P.7a the correlation is ρ = 0. In this case the joint density is symmetrical and the contours in Figure P.7b are circles. If X and Y are jointly normal then they are statistically independent if, and only if, ρ = 0. –2 0 x 0 –2 2 4 –4 –2 0 y 2 4 –4 –4 –2 (a) 0 x 2 4 (b) FIGURE P.6 The bivariate normal distribution: 𝛍X = 𝛍Y = 0, 𝛔X = 𝛔Y = 1, and 𝛒 = 0.7. ............................................................................................................................................ 13 “The Bivariate Normal Distribution” from the Wolfram Demonstrations Project http://demonstrations.wolfram.com/. Figures P.6, P.7, and P.8 represent the interactive graphics on the site as static graphics for the primer. The site permits easy manipulation of distribution parameters. The joint density function figure can be rotated and viewed from different angles. k k Trim Size: 8in x 10in 38 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 38 Probability Primer 4 y 2 0 –4 –2 4 x 0 –2 2 0 2 –2 4 –4 y –4 –4 –2 2 0 x (a) 4 (b) FIGURE P.7 The bivariate normal distribution: 𝛍X = 𝛍Y = 0, 𝛔X = 𝛔Y = 1, and 𝛒 = 0. k There are several relations between the normal, bivariate normal, and the conditional distributions that are used in statistics and econometrics. First, if X and Y have a bivariate normal dis-) ( 2 , σ tribution then the marginal distributions of X and Y are normal distributions too, X ∼ N μ X X ) ( 2 and Y ∼ N μY , σY . Second, the conditional distribution for Y given X is normal, with conditional mean E(Y|X) ( )= 2 2 2 1 − ρ . − βμ and β = σ ∕σ , and conditional variance var(Y|X) = σ α + βX, where α = μ Y ( X XY X Y [ )] Or Y|X ∼ N α + βX, σ2Y 1 − ρ2 . Three noteworthy points about these results are (i) that the conditional mean is a linear function of X, and is called a linear regression function; (ii) the conditional variance is constant and does not vary with X; and (iii) the conditional variance is smaller than the unconditional variance if ρ ≠ 0. In Figure P.814 we display a joint normal density with Conditional distribution of Y when X = x Bivariate normal density 20 y 15 0 5 x 10 10 5 15 0 x 20 10 20 15 5 0 (a) 10 y (b) FIGURE P.8 (a) Bivariate normal distribution with 𝛍X = 𝛍Y = 5, 𝛔X = 𝛔Y = 3, and 𝛒 = 0.7; (b) conditional distribution of Y given X = 10. ............................................................................................................................................ 14 “The Bivariate Normal and Conditional Distributions” from the Wolfram Demonstrations Project http://demonstrations.wolfram.com/TheBivariateNormalAndConditionalDistributions/. Both the bivariate distribution and conditional distributions can be rotated and viewed from different perspectives. k k Trim Size: 8in x 10in k Hill5e c01a.tex V1 - 12/21/2017 2:43pm P.8 Exercises Page 39 39 μX = μY = 5, σX = σY = 3, and ρ = 0.7. The covariance between X and Y is σXY = ρσX σY = 0.7 × 3 × 3 = 6.3 so that β = σXY ∕σ2X = 6.3∕9 = 0.7 and α = μY – βμX = 5 – 0.7 × 5 = 1.5. The conditional mean of Y given X = 10 is E(Y|X = 10) = ( α + βX ) = 1.5 ( + 0.7X )= 1.5 + 0.7 × 10 = 8.5. The conditional variance is var(Y|X = 10) = σ2Y 1 − ρ2 = 32 1 − 0.72 = 9(0.51) = 4.59. That is, the conditional distribution is (Y|X = 10) ∼ N(8.5, 4.59). P.8 Exercises Answers to odd-numbered exercises are on the book website www.principlesofeconometrics.com/poe5. P.1 Let x1 = 17, x2 = 1, x3 = 0; y1 = 5, y2 = 2, y3 = 8. Calculate the following: ∑2 a. xi ∑i=1 3 b. xy t=1( t t ) ∑3 ∕3 [Note: x is called the arithmetic average or arithmetic mean.] c. x = x i i=1 ) ∑3 ( d. x −x i=1 i )2 ∑3 ( e. x −x i=1 i ) (∑ 2 3 f. x2 − 3x i=1 i ) (∑ )( ) ∑3 ( 3 g. y ∕3 x − x yi − y where y = i=1 i i=1 i (∑ ) 3 h. − 3x y x y j j j=1 k P.2 Express each of the following sums in summation notation. ) ( ) ( ) ( ) ( a. x1 ∕y1 + x2 ∕y2 + x3 ∕y3 + x4 ∕y4 b. y2 + y3 + y4 c. x1 y1 + x2 y2 + x3 y3 + x4 y4 d. x3 y5 + x4 y6 + x5 y7 ( ) ( ) e. x3 ∕y23 + x4 ∕y24 ) ( ) ( ) ( ) ( f. x1 − y1 + x2 − y2 + x3 − y3 + x4 − y4 P.3 Write out each of the following sums and compute where possible. ) ∑3 ( a − bxi a. i=1 ∑4 2 t b. ) ∑t=1 2 ( 2 c. 2x + 3x + 1 x=0 ∑4 𝑓 (x + 3) d. ∑x=2 3 𝑓 (x, y) e. ∑x=1 4 ∑2 f. (x + 2y) x=3 y=1 P.4 Show algebraically that )2 (∑n 2 ) ∑n ( 2 a. x − nx x −x = i=1 ( i )( )i=1 (i ∑n ) ∑n b. x y − nx y xi − x yi − y = i=1 i i ) ∑i=1 n ( c. x −x =0 j=1 j P.5 Let SALES denote the monthly sales at a bookstore. Assume SALES are normally distributed with a mean of $50,000 and a standard deviation of $6000. a. Compute the probability that the firm has a month with SALES greater than $60,000. Show a sketch. b. Compute the probability that the firm has a month with SALES between $40,000 and $55,000. Show a sketch. c. Find the value of SALES that represents the 97th percentile of the distribution. That is, find the value SALES0.97 such that P(SALES > SALES0.97 ) = 0.03. d. The bookstore knows their PROFITS are 30% of SALES minus fixed costs of $12,000. Find the probability of having a month in which PROFITS were zero or negative. Show a sketch. [Hint: What is the distribution of PROFITS?] k k Trim Size: 8in x 10in 40 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 40 Probability Primer P.6 A venture capital company feels that the rate of return (X) on a proposed investment is approximately normally distributed with a mean of 40% and a standard deviation of 10%. a. Find the probability that the return X will exceed 55%. b. The banking firm who will fund the venture sees the rate of return differently, claiming that venture capitalists are always too optimistic. They perceive that the distribution of returns is V = 0.8X − 5%, where X is the rate of return expected by the venture capital company. If this is correct, find the probability that the return V will exceed 55%. P.7 At supermarkets sales of “Chicken of the Sea” canned tuna vary from week to week. Marketing researchers have determined that there is a relationship between sales of canned tuna and the price of canned tuna. Specifically, SALES = 50000 − 100 PRICE. SALES is measured as the number of cans per week and PRICE is measured in cents per can. Suppose PRICE over the year can be considered (approximately) a normal random variable with mean μ = 248 cents and standard deviation σ = 10 cents. a. Find the expected value of SALES. b. Find the variance of SALES. c. Find the probability that more than 24,000 cans are sold in a week. Draw a sketch illustrating the calculation. d. Find the PRICE such that SALES is at its 95th percentile value. That is, let SALES0.95 be the 95th percentile of SALES. Find the value PRICE0.95 such that P (SALES > SALES0.95 ) = 0.05. P.8 The Shoulder and Knee Clinic knows that their expected monthly revenue from patients depends on their level of advertising. They hire an econometric consultant who reports that their expected monthly revenue, measured in $1000 units, is given by the following equation E(REVENUE|ADVERT) = 100 + 20 ADVERT, where ADVERT is advertising expenditure in $1000 units. The econometric consultant also claims that REVENUE is normally distributed with variance var(REVENUE|ADVERT) = 900. k a. Draw a sketch of the relationship between expected REVENUE and ADVERT as ADVERT varies from 0 to 5. b. Compute the probability that REVENUE is greater than 110 if ADVERT = 2. Draw a sketch to illustrate your calculation. c. Compute the probability that REVENUE is greater than 110 if ADVERT = 3. d. Find the 2.5 and 97.5 percentiles of the distribution of REVENUE when ADVERT = 2. What is the probability that REVENUE will fall in this range if ADVERT = 2? e. Compute the level of ADVERT required to ensure that the probability of REVENUE being larger than 110 is 0.95. P.9 Consider the U.S. population of registered voters, who may be Democrats, Republicans or independents. When surveyed about the war with ISIS, they were asked if they strongly supported war efforts, strongly opposed the war, or were neutral. Suppose that the proportion of voters in each category is given in Table P.8: T A B L E P.8 Table for Exercise P.9 Against Political Party Republican Independent Democrat 0.05 0.05 0.35 War Attitude Neutral In Favor 0.15 0.05 0.05 0.25 0.05 0 a. Find the “marginal” probability distributions for war attitudes and political party affiliation. b. What is the probability that a randomly selected person is a political independent given that they are in favor of the war? c. Are the attitudes about war with ISIS and political party affiliation statistically independent or not? Why? k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.8 Exercises Page 41 41 d. For the attitudes about the war assign the numerical values AGAINST = 1, NEUTRAL = 2, and IN FAVOR = 3. Call this variable WAR. Find the expected value and variance of WAR. e. The Republican party has determined that monthly fundraising depends on the value of WAR from month to month. In particular the monthly contributions to the party are given by the relation (in millions of dollars) CONTRIBUTIONS = 10 + 2 × WAR. Find the mean and standard deviation of CONTRIBUTIONS using the rules of expectations and variance. P.10 A firm wants to bid on a contract worth $80,000. If it spends $5000 on the proposal it has a 50–50 chance of getting the contract. If it spends $10,000 on the proposal it has a 60% chance of winning the contract. Let X denote the net revenue from the contract when the $5000 proposal is used and let Y denote the net revenue from the contract when the $10,000 proposal is used. X −5,000 75,000 a. b. c. d. f (x) 0.5 0.5 y −10,000 70,000 f (y) 0.4 0.6 If the firm bases its choice solely on expected value, how much should it spend on the proposal? Compute the variance of X. [Hint: Using scientific notation simplifies calculations.] Compute the variance of Y. How might the variance of the net revenue affect which proposal the firm chooses? P.11 Prior to presidential elections citizens of voting age are surveyed. In the population, two characteristics of voters are their registered party affiliation (republican, democrat, or independent) and for whom they voted in the previous presidential election (republican or democrat). Let us draw a citizen at random, defining these two variables. k ⎧−1 ⎪ PARTY = ⎨ 0 ⎪ 1 ⎩ { −1 VOTE = 1 registered republican independent or unregistered registered democrat voted republican in previous election voted democratic in previous election a. Suppose that the probability of drawing a person who voted republication in the last election is 0.466, and the probability of drawing a person who is registered republican is 0.32, and the probability that a randomly selected person votes republican given that they are a registered republican is 0.97. Compute the joint probability Prob[PARTY = −1, VOTE = −1]. Show your work. b. Are these random variables statistically independent? Explain. P.12 Based on years of experience, an economics professor knows that on the first principles of economics exam of the semester 13% of students will receive an A, 22% will receive a B, 35% will receive a C, 20% will receive a D, and the remainder will earn an F. Assume a 4 point grading scale (A = 4, B = 3, C = 2, D = 1, and F = 0). Define the random variable GRADE = 4, 3, 2, 1, 0 to be the grade of a randomly chosen student. a. What is the probability distribution f (GRADE) for this random variable? b. What is the expected value of GRADE? What is the variance of GRADE? Show your work. c. The professor has 300 students in each class. Suppose that the grade of the ith student is GRADEi and that the probability distribution of grades f (GRADEi ) is the same for all students. Define ∑300 CLASS_ AVG = i=1 GRADEi ∕300. Find the expected value and variance of CLASS_AVG. d. The professor has estimated that the number of economics majors coming from the class is related to the grade on the first exam. He believes the relationship to be MAJORS = 50 + 10CLASS_AVG. Find the expected value and variance of MAJORS. Show your work. P.13 The LSU Tigers baseball team will play the Alabama baseball team in a weekend series of two games. Let W = 0, 1, or 2 equal the number of games LSU wins. Let the weekend’s weather be designated as Cold or Not Cold. Let C = 1 if the weather is cold and C = 0 if the weather is not cold. The joint probability function of these two random variables is given in Table P.9, along with space for the marginal distributions. k k Trim Size: 8in x 10in 42 Hill5e c01a.tex V1 - 12/21/2017 2:43pm k Page 42 Probability Primer T A B L E P.9 C=1 C=0 f (w) Table for Exercise P.13 W=0 W=1 W=2 f (c) (i) 0.07 (v) 0.12 0.14 (vi) 0.12 (iii) 0.61 (ii) (iv) a. Fill in the blanks, (i)–(vi). b. Using the results of (a), find the conditional probability distribution of the number of wins, W, conditional on the weather being warm, C = 0. Based on a comparison of the conditional probability distribution f (w|C = 0) and the marginal distribution f (w), can you conclude that the number of games LSU wins W is statistically independent of the weather conditions C, or not? Explain. c. Find the expected value of the number of LSU wins, W. Also find the conditional expectation E(W|C = 0). Show your work. What kind of weather is more favorable for the LSU Tigers baseball team? d. The revenue of vendors at the LSU Alex Box baseball stadium depends on the crowds, which in turn depends on the weather. Suppose that food sales FOOD = $10,000 − 3000C. Use the rules for expected value and variance to find the expected value and standard deviation of food sales. k P.14 A clinic specializes in shoulder injuries. A patient is randomly selected from the population of all clinic clients. Let S be the number of doctor visits for shoulder problems in the past six months. Assume the values of S are s = 1, 2, 3, or 4. Patients at the shoulder clinic are also asked about knee injuries. Let K = the number of doctor visits for knee injuries during the past six months. Assume the values of K are k = 0, 1 or 2. The joint probability distribution of the numbers of shoulder and knee injuries is shown in Table P.10. Use the information in the joint probability distribution to answer the following questions. Show brief calculations for each T A B L E P.10 Table for Exercise P.14 Knee = K Shoulder = S 1 2 3 4 f (k) 0 1 2 0.15 0.06 0.02 0.02 0.09 0.06 0.10 0.08 f (s) 0.2 0.10 0.33 a. What is the probability that a randomly chosen patient will have two doctor visits for shoulder problems during the past six months? b. What is the probability that a randomly chosen patient will have two doctor visits for shoulder problems during the past six months given that they have had one doctor visit for a knee injury in the past six months? c. What is the probability that a randomly chosen patient will have had three doctor visits for shoulder problems and two doctor visits for knee problems in the past six months? d. Are the number of doctor visits for knee and shoulder injuries statistically independent? Explain. e. What is the expected value of the number of doctor visits for shoulder injuries from this population? f. What is the variance of the number of doctor visits for shoulder injuries from this population? P.15 As you walk into your econometrics exam, a friend bets you $20 that she will outscore you on the exam. Let X be a random variable denoting your winnings. X can take the values 20, 0 [if there is a tie], or −20. You know that the probability distribution for X, f (x), depends on whether she studied for k k Trim Size: 8in x 10in Hill5e c01a.tex V1 - 12/21/2017 2:43pm k P.8 Exercises Page 43 43 the exam or not. Let Y = 0 if she studied and Y = 1 if she did not study. Consider the following joint distribution Table P.11. T A B L E P.11 Joint pdf for Exercise P.15 Y X a. b. c. d. e. 0 1 f (x) −20 0 20 (i) (iii) 0.10 0 0.15 (iv) (ii) 0.25 (v) f (y) (vi) 0.60 Fill in the missing elements (i)–(vi) in the table. Compute E(X). Should you take the bet? What is the probability distribution of your winnings if you know that she did not study? Find your expected winnings given that she did not study. Use the Law of Iterated Expectations to find E(X). P.16 Breast cancer prevalence in the United Kingdom can be summarized for the population (data are in 1000s) as in Table P.12. T A B L E P.12 Table for Exercise P.16 k k Sex Suffers from Breast Cancer Not Suffering from Breast Cancer Total Female Male Total 550 30,868 31,418 3 30,371 30,374 553 61,239 61,792 Compute the probability that a randomly drawn person has breast cancer. Compute the probability that a randomly drawn female has breast cancer. Compute the probability that a person is female given that the person has breast cancer. What is the conditional probability function for the prevalence of breast cancer given that the person is female? e. What is the conditional probability function for the prevalence of breast cancer given that the person is male? a. b. c. d. P.17 A continuous random variable Y has pdf 𝑓 (y) = a. b. c. d. e. f. { 2y 0 < y < 1 0 otherwise Sketch the pdf . Find the cdf , F(y) = P(Y ≤ y) and sketch it. [Hint: Requires calculus.] Use the pdf and a geometric argument to find the probability P(Y ≤ 1/2). Use the cdf from part (b) to compute P(Y ≤ 1/2). Using the pdf and a geometric argument find the probability P(1/4 ≤ Y ≤ 3/4). Use the cdf from part (b) to compute P(1/4 ≤ Y ≤ 3/4). P.18 Answer each of the following: a. An internal revenue service auditor knows that 3% of all income tax forms contain errors. Returns are assigned randomly to auditors for review. What is the probability that an auditor will have to k Trim Size: 8in x 10in 44 k Hill5e c01a.tex V1 - 12/21/2017 2:43pm Page 44 Probability Primer b. c. d. e. f. view four tax returns until the first error is observed? That is, what is the probability of observing three returns with no errors, and then observing an error in the fourth return? Let Y be the number of independent trials of an experiment before a success is observed. That is, it is the number of failures before the first success. Assume each trial has a probability of success of p and a probability of failure of 1 − p. Is this a discrete or continuous random variable? What is the set of possible values that Y can take? Can Y take the value zero? Can Y take the value 500? Consider the pdf f (y) = P(Y = y) = p(1 − p)y . Using this pdf compute the probability in (a). Argue that this probability function generally holds for the experiment described in (b). Using the value p = 0.5, plot the pdf in (c) for y = 0, 1, 2, 3, 4. ∑∞ ∑∞ Show that y=0 𝑓 (y) = y=0 p(1 − p)y = 1. [Hint: If |r| < 1 then 1 + r + r2 + r3 + · · · = 1/(1−r).] Verify for y = 0, 1, 2, 3, 4 that the cdf P(Y ≤ y) = 1 − (1 − p)y+1 yields the correct values. P.19 Let X and Y be random variables with expected values μ = μX = μY and variances σ2 = σ2X = σ2Y . Let Z = (2X + Y)/2. a. b. c. d. Find the expected value of Z. Find the variance of Z assuming X and Y are statistically independent. Find the variance of Z assuming that the correlation between X and Y is −0.5. Let the correlation between X and Y be −0.5. Find the correlation between aX and bY, where a and b are any nonzero constants. P.20 Suppose the pdf of the continuous random variable X is f (x) = 1, for 0 < x < 1 and f (x) = 0 otherwise. k a. Draw a sketch of the pdf . Verify that the area under the pdf for 0 < x < 1 is 1. b. Find the cdf of X. [Hint: Requires the use of calculus.] c. Compute the probability that X falls in each of the intervals [0, 0.1], [0.5, 0.6], and [0.79, 0.89]. Indicate the probabilities on the sketch drawn in (a). d. Find the expected value of X. e. Show that the variance of X is 1/12. f. Let Y be a discrete random variable taking the values 1 and 0 with conditional probabilities P(Y = 1|X = x) = x and P(Y = 0|X = x) = 1 − x. Use the Law of Iterated Expectations to find E(Y). g. Use the variance decomposition to find var(Y). P.21 A fair die is rolled. Let Y be the face value showing, 1, 2, 3, 4, 5, or 6 with each having the probability 1/6 of occurring. Let X be another random variable that is given by { Y if Y is even X= 0 if Y is odd a. b. c. d. e. f. ( ) Find E(Y), E Y2 , and var(Y). ( ) What is the probability distribution for X? Find E(X), E X2 , and var(X). Find the conditional probability distribution of Y given each X. Find the conditional expected value of Y given each value of X, E(Y|X). ( ) Find the probability distribution of Z = XY. Show that E(Z) = E(XY) = E X2 . Find cov(X, Y). P.22 A large survey of married women asked “How many extramarital affairs did you have last year?” 77% said they had none, 5% said they had one, 2% said two, 3% said three, and the rest said more than three. Assume these women are representative of the entire population. a. What is the probability that a randomly selected married woman will have had one affair in the past year? b. What is the probability that a randomly selected married woman will have had more than one affair in the past year? c. What is the probability that a randomly chosen married woman will have had less than three affairs in the past year? d. What is the probability that a randomly chosen married woman will have had one or two affairs in the past year? e. What is the probability that a randomly chosen married woman will have had one or two affairs in the past year, given that they had at least one? k k Trim Size: 8in x 10in k Hill5e c01a.tex V1 - 12/21/2017 2:43pm P.8 Exercises Page 45 45 P.23 Let NKIDS represent the number of children ever born to a woman. / The possible values of NKIDS are nkids = 0, 1, 2, 3, 4, . . . . Suppose the pdf is f (nkids) = 2nkids (7.389nkids!), where ! denotes the factorial operation. a. Is NKIDS a discrete or continuous random variable? b. Calculate the pdf for nkids = 0, 1, 2, 3, 4. Sketch it. [Note: It may be convenient to use a spreadsheet or other software to carry out tedious calculations.] c. Calculate the probabilities P[NKIDS ≤ nkids] for nkids = 0, 1, 2, 3, 4. Sketch the cumulative distribuiton function. d. What is the probability that a woman will have more than one child. e. What is the probability that a woman will have two or fewer children? P.24 Five baseballs are thrown to a batter who attempts to hit the ball 350 feet or more. Let H denote the number of successes, with the pd for having h successes being f (h) = 120 × 0.4h × 0.65−h ∕[h!(5−h)!], where ! denotes the factorial operation. a. Is H a discrete or continuous random variable? What values can it take? b. Calculate the probabilities that the number of successes h = 0, 1, 2, 3, 4, and 5. [Note: It may be convenient to use a spreadsheet or other software to carry out tedious calculations.] Sketch the pdf . c. What is the probability of two or fewer successes? d. Find the expected value of the random variable H. Show your work. e. The prizes are $1000 for the first success, $2000 for the second success, $3000 for the third success, and so on. What is the pdf for the random variable PRIZE, which is the total prize winnings? f. Find the expected value of total prize winnings, PRIZE. P.25 An author knows that a certain number of typographical errors (0, 1, 2, 3, …) are on each book page. Define the random variable T equaling the number of errors per page. Suppose that T has a Poisson distribution [Appendix B.3.3], with pdf , f (t) = μt exp(−μ)/t!, where ! denotes the factorial operation, and μ = E(T) is the mean number of typographical errors per page. k a. If μ = 3, what is the probability that a page has one error? What is the probability that a page has four errors? b. An editor independently checks each word of every page and catches 90% of the errors, but misses 10%. Let Y denote the number of errors caught on a page. The values of y must be less than or equal to the actual number t of errors on the page. Suppose that the number of errors caught on a page with t errors has a binomial distribution [Appendix B.3.2]. g(y|t, p = 0.9) = t! 0.9y 0.1t−y , y = 0, 1, … , t y!(t − y)! Compute the probability that the editor finds one error on a page given that the page actually has four errors. c. Find the joint probability P[Y = 3, T = 4]. d. It can be shown that the probability the editor will find Y errors on a page follows a Poisson distribution with mean E(Y) = 0.9μ. Use this information to find the conditional probability that there are T = 4 errors on a page given that Y = 3 are found. k k k CHAPTER 2 The Simple Linear Regression Model LEARNING OBJECTIVES Remark Learning Objectives and Keywords sections will appear at the beginning of each chapter. We urge you to think about and possibly write out answers to the questions, and make sure you recognize and can define the keywords. If you are unsure about the questions or answers, consult your instructor. When examples are requested in Learning Objectives sections, you should think of examples not in the book. k Based on the material in this chapter you should be able to 1. Explain the difference between an estimator and an estimate, and why the least squares estimators are random variables, and why least squares estimates are not. 2. Discuss the interpretation of the slope and intercept parameters of the simple regression model, and sketch the graph of an estimated equation. 3. Explain the theoretical decomposition of an observable variable y into its systematic and random components, and show this decomposition graphically. 4. Discuss and explain each of the assumptions of the simple linear regression model. 5. Explain how the least squares principle is used to fit a line through a scatter plot of data. Be able to define the least squares residual and the least squares fitted value of the dependent variable and show them on a graph. 6. Define the elasticity of y with respect to x and explain its computation in the simple linear regression model when y and x are not transformed in any way, and when y and/or x have been transformed to model a nonlinear relationship. 7. Explain the meaning of the statement ‘‘If regression model assumptions SR1–SR5 hold, then the least squares estimator b2 is unbiased.’’ In particular, what exactly does ‘‘unbiased’’ mean? Why is b2 biased if an important variable has been omitted from the model? 8. Explain the meaning of the phrase ‘‘sampling variability.’’ )2 ∑( 9. Explain how the factors σ2 , xi − x , and N affect the precision with which we can estimate the unknown parameter β2 . 46 k k k 2.1 An Economic Model 12. Explain the difference between an explanatory variable that is fixed in repeated samples and an explanatory variable that is random. 10. State and explain the Gauss–Markov theorem. 11. Use the least squares estimator to estimate nonlinear relationships and interpret the results. KEYWORDS assumptions asymptotic biased estimator BLUE degrees of freedom dependent variable deviation from the mean form econometric model economic model elasticity exogenous variable Gauss–Markov theorem heteroskedastic k 13. Explain the term ‘‘random sampling.’’ homoskedastic independent variable indicator variable least squares estimates least squares estimators least squares principle linear estimator log-linear model nonlinear relationship prediction quadratic model random error term random-x regression model regression parameters repeated sampling sampling precision sampling properties scatter diagram simple linear regression analysis simple linear regression function specification error strictly exogenous unbiased estimator Economic theory suggests many relationships between economic variables. In microeconomics, you considered demand and supply models in which the quantities demanded and supplied of a good depend on its price. You considered “production functions” and “total product curves” that explained the amount of a good produced as a function of the amount of an input, such as labor, that is used. In macroeconomics, you specified “investment functions” to explain that the amount of aggregate investment in the economy depends on the interest rate and “consumption functions” that related aggregate consumption to the level of disposable income. Each of these models involves a relationship between economic variables. In this chapter, we consider how to use a sample of economic data to quantify such relationships. As economists, we are interested in questions such as the following: If one variable (e.g., the price of a good) changes in a certain way, by how much will another variable (the quantity demanded or supplied) change? Also, given that we know the value of one variable, can we forecast or predict the corresponding value of another? We will answer these questions by using a regression model. Like all models, the regression model is based on assumptions. In this chapter, we hope to be very clear about these assumptions, as they are the conditions under which the analysis in subsequent chapters is appropriate. 2.1 47 An Economic Model In order to develop the ideas of regression models, we are going to use a simple, but important, economic example. Suppose that we are interested in studying the relationship between household income and expenditure on food. Consider the “experiment” of randomly selecting households from a particular population. The population might consist of households within a particular city, state, province, or country. For the present, suppose that we are interested only in households with an income of $1000 per week. In this experiment, we randomly select a number of households from this population and interview them. We ask the question “How much did you spend per person on food last week?” Weekly food expenditure, which we denote as y, is a random variable since the value is unknown to us until a household is selected and the question is asked and answered. k k k 48 CHAPTER 2 The Simple Linear Regression Model Remark In the Probability Primer and Appendices B and C, we distinguished random variables from their values by using uppercase (Y) letters for random variables and lowercase (y) letters for their values. We will not make this distinction any longer because it leads to complicated notation. We will use lowercase letters, like “y,” to denote random variables as well as their values, and we will make the interpretation clear in the surrounding text. The continuous random variable y has a probability density function (which we will abbreviate as pdf ) that describes the probabilities of obtaining various food expenditure values. If you are rusty or uncertain about probability concepts, see the Probability Primer and Appendix B at the end of this book for a comprehensive review. The amount spent on food per person will vary from one household to another for a variety of reasons: some households will be devoted to gourmet food, some will contain teenagers, some will contain senior citizens, some will be vegetarian, and some will eat at restaurants more frequently. All of these factors and many others, including random, impulsive buying, will cause weekly expenditures on food to vary from one household to another, even if they all have the same income. The pdf f(y) describes how expenditures are “distributed” over the population and might look like Figure 2.1. The pdf in Figure 2.1a is actually a conditional pdf since it is “conditional” upon household income. If x = weekly household income = $1000, then the conditional pdf is f(y|x = $1000). The conditional mean, or expected value, of y is E(y|x = $1000) = μy|x and is our population’s mean weekly food expenditure per person. k k Remark The expected value of a random variable is called its “mean” value, which is really a contraction of population mean, the center of the probability distribution of the random variable. This is not the same as the sample mean, which is the arithmetic average of numerical values. Keep the distinction between these two usages of the term “mean” in mind. The conditional variance of y is var(y|x = $1000) = σ2 , which measures the dispersion of household expenditures y about their mean μy|x . The parameters μy|x and σ2 , if they were known, would give us some valuable information about the population we are considering. If we knew these parameters, and if we knew that the conditional distribution f (y|x = $1000) was normal, f(y|x = 1000) f(y|x = 1000) f(y|x) μy|x y f(y|x = 1000) f(y|x = 2000) μy|2000 μy|1000 (b) (a) FIGURE 2.1 (a) Probability distribution f ( y|x = 1000) of food expenditure y given income x = $1000. (b) Probability distributions of food expenditure y given incomes x = $1000 and x = $2000. k y k 2.2 An Econometric Model 49 ) ( N μy|x , σ2 , then we could calculate probabilities that y falls in specific intervals using properties of the normal distribution. That is, we could compute the proportion of the household population that spends between $50 and $75 per person on food, given $1000 per week income. As economists, we are usually interested in studying relationships between variables, in this case the relationship between y = weekly food expenditure per person and x = weekly household income. Economic theory tells us that expenditure on economic goods depends on income. Consequently, we call y the “dependent variable” and x the “independent” or “explanatory” variable. In econometrics, we recognize that real-world expenditures are random variables, and we want to use data to learn about the relationship. An econometric analysis of the expenditure relationship can provide answers to some important questions, such as: If weekly income goes up by $100, how much will average weekly food expenditures rise? Or, could weekly food expenditures fall as income rises? How much would we predict the weekly per person expenditure on food to be for a household with an income of $2000 per week? The answers to such questions provide valuable information for decision makers. k Using … per person food spending information … one can determine the similarities and disparities in the spending habits of households of differing sizes, races, incomes, geographic areas, and other socioeconomic and demographic features. This information is valuable for assessing existing market conditions, product distribution patterns, consumer buying habits, and consumer living conditions. Combined with demographic and income projections, this information may be used to anticipate consumption trends. The information may also be used to develop typical market baskets of food for special population groups, such as the elderly. These market baskets may, in turn, be used to develop price indices tailored to the consumption patterns of these population groups. [Blisard, Noel, Food Spending in American Households, 1997–1998, Electronic Report from the Economic Research Service, U.S. Department of Agriculture, Statistical Bulletin Number 972, June 2001] From a business perspective, if we are managers of a supermarket chain (or restaurant, or health food store, etc.), we must consider long-range plans. If economic forecasters are predicting that local income will increase over the next few years, then we must decide whether, and how much, to expand our facilities to serve our customers. Or, if we plan to open franchises in high-income and low-income neighborhoods, then forecasts of expenditures on food per person, along with neighborhood demographic information, give an indication of how large the stores in those areas should be. In order to investigate the relationship between expenditure and income, we must build an economic model and then a corresponding econometric model that forms the basis for a quantitative or empirical economic analysis. In our food expenditure example, economic theory suggests that average weekly per person household expenditure on food, represented mathematically by the conditional mean E(y|x) = μy|x , depends on household income x. If we consider households with different levels of income, we expect the average expenditure on food to change. In Figure 2.1b, we show the pdfs of food expenditure for two different levels of weekly income, $1000 and $2000. Each conditional pdf f (y|x) shows that expenditures will be distributed about a mean value μy|x , but the mean expenditure by households with higher income is larger than the mean expenditure by lower income households. In order to use data, we must now specify an econometric model that describes how the data on household income and food expenditure are obtained and that guides the econometric analysis. 2.2 An Econometric Model Given the economic reasoning in the previous section, and to quantify the relationship between food expenditure and income, we must progress from the ideas in Figure 2.1, to an econometric model. First, suppose a three-person household has an unwavering rule that each week they k k k 50 k CHAPTER 2 The Simple Linear Regression Model spend $80 and then also spend 10 cents of each dollar of income received on food. Let y = weekly household food expenditure ($) and let x = weekly household income ($). Algebraically their rule is y = 80 + 0.10x. Knowing this relationship we calculate that in a week in which the household income is $1000, the household will spend $180 on food. If weekly income increases by $100 to $1100, then food expenditure increases to $190. These are predictions of food expenditure given income. Predicting the value of one variable given the value of another, or others, is one of the primary uses of regression analysis. A second primary use of regression analysis is to attribute, or relate, changes in one variable to changes in another variable. To that end, let “Δ” denote “change in” in the usual algebraic way. A change in income of $100 means that Δx = 100. Because of the spending rule y = 80 + 0.10x the change in food expenditure is Δy = 0.10Δx = 0.10(100) = 10. An increase in income of $100 leads to, or causes, a $10 increase in food expenditure. Geometrically, the rule is a line with “y-intercept” 80 and slope Δy∕Δx = 0.10. An economist might say that the household “marginal propensity to spend on food is 0.10,” which means that from each additional dollar of income 10 cents is spent on food. Alternatively, in a kind of economist shorthand, the “marginal effect of income on food expenditure is 0.10.” Much of economic and econometric analysis is an attempt to measure a causal relationship between two economic variables. Claiming causality here, that is, changing income leads to a change in food expenditure, is quite clear given the household’s expenditure rule. It is not always so straightforward. In reality, many other factors may affect household expenditure on food; the ages and sexes of the household members, their physical size, whether they do physical labor or have desk jobs, whether there is a party following the big game, whether it is an urban or rural household, whether household members are vegetarians or into a paleo-diet, as well as other taste and preference factors (“I really like truffles”) and impulse shopping (“Wow those peaches look good!”). Lots of factors. Let e = everything else affecting food expenditure other than income. Furthermore, even if a household has a rule, strict or otherwise, we do not know it. To account for these realities, we suppose that the household’s food expenditure decision is based on the equation y = β1 + β2 x + e (2.1) In addition to y and x, equation (2.1) contains two unknown parameters, β1 and β2 , instead of “80” and “0.10,” and an error term e, which represents all those other factors (everything else) affecting weekly household food expenditure. Imagine that we can perform an experiment on the household. Let’s increase the household’s income by $100 per week and hold other things constant. Holding other things constant, or holding all else (everything else) equal, is the ceteris paribus assumption discussed extensively in economic principles courses. Let Δx = 100 denote the change in household income. Assuming everything else affecting household food expenditure, e, is held constant means that Δe = 0. The effect of the change in income is Δy = β2 Δx + Δe = β2 Δx = β2 × 100. The change in weekly food expenditure Δy = β2 × 100 is explained by, or caused by, the change in income. The unknown parameter β2 , the marginal propensity to spend on food from income, tells us the proportion of the increase in income used for food purchases; it answers the “how much” question “How much will food expenditure change given a change in income, holding all else constant?” The experiment in the previous paragraph is not feasible. We can give a household an extra $100 income, but we cannot hold all else constant. The simple calculation of the marginal effect of an increase in income on food expenditure Δy = β2 × 100 is not possible. However, we can shed light on this “how much” question by using regression analysis to estimate β2 . Regression analysis is a statistical method that uses data to explore relationships between variables. A simple linear regression analysis examines the relationship between a y-variable and one x-variable. It is said to be “simple” not because it is easy, but because there is only one x-variable. The y-variable is called the dependent variable, the outcome variable, the explained variable, the left-hand-side variable, or the regressand. In our example, the dependent variable is k k k 2.2 An Econometric Model 51 y = weekly household expenditure on food. The variable x = weekly household income is called the independent variable, the explanatory variable, the right-hand-side variable, or the regressor. Equation (2.1) is the simple linear regression model. All models are abstractions from reality and working with models requires assumptions. The same is true for the regression model. The first assumption of the simple linear regression model is that relationship (2.1) holds for the members of the population under consideration. For example, define the population to be three-person households in a given geographic region, say southern Australia. The unknowns β1 and β2 are called population parameters. We assert the behavioral rule y = β1 + β2 x + e holds for all households in the population. Each week food expenditure equals β1 , plus a proportion β2 of income, plus other factors, e. The field of statistics was developed because, in general, populations are large, and it is impossible (or impossibly costly) to examine every population member. The population of three-person households in a given geographic region, even if it is only a medium-sized city, is too large to survey individually. Statistical and econometric methodology examines and analyzes a sample of data from the population. After analyzing the data, we make statistical inferences. These are conclusions or judgments about a population based on the data analysis. Great care must be taken when drawing inferences. The inferences are conclusions about the particular population from which the data were collected. Data on households from southern Australia may, or may not, be useful for making inferences, drawing conclusions, about households from the southern United States. Do Melbourne, Australia, households have the same food spending patterns as households in New Orleans, Louisiana? That might be an interesting research topic. If not, then we may not be able to draw valid conclusions about New Orleans household behavior from the sample of Australian data. k 2.2.1 Data Generating Process k The sample of data, and how the data are actually obtained, is crucially important for subsequent inferences. The exact mechanisms for collecting a sample of data are very discipline specific (e.g., agronomy is different from economics) and beyond the scope of this book.1 For the household food expenditure example, let us assume that we can obtain a sample at a point in time [these are cross-sectional data] consisting of N data pairs that are randomly selected from the population. ) ( Let yi , xi denote the ith data pair, i = 1, … , N. The variables yi and xi are random variables, because their values are not ) until they are observed. Randomly selecting households makes ( known the first observation pair y , x 1 1 statistically independent of all other data ( ) ( pairs, ) and each observation pair yi , xi is statistically independent of every other data pair, yj , xj , where i ≠ j. We ) ( further assume that the random variables yi and xi have a joint pdf f yi , xi that describes their distribution of values. We often do not know the exact nature of the joint distribution (such as bivariate normal; see Probability Primer, Section P.7.1), but all pairs drawn from the same population are assumed to follow the same joint pdf , and, thus, the data pairs are not only statistically independent but are also identically distributed (abbreviated i.i.d. or iid). Data pairs that are iid are said to be a random sample. If our first assumption is true, that the behavioral ( rule)y = β1 + β2 x + e holds for all households in the population, then restating (2.1) for each yi , xi data pair yi = β1 + β2 xi + ei , i = 1, … , N (2.1) This is sometimes called the data generating process (DGP) because we assume that the observable data follow this relationship. ............................................................................................................................................ 1 See, for example, Paul S. Levy and Stanley Lemeshow (2008) Sampling of Populations: Methods and Applications, 4th Edition, Hoboken, NJ: John Wiley and Sons, Inc. k k 52 CHAPTER 2 The Simple Linear Regression Model 2.2.2 k The Random Error and Strict Exogeneity The second assumption ) simple regression model (2.1) concerns the “everything else” ( of the term e. The variables yi , xi are random variables because we do not know what values they take until a particular household is chosen and they are observed. The error term ei is also a random variable. All the other factors affecting food expenditure except income will be different for each population household if for no other reason that everyone’s tastes and preferences are different. Unlike food expenditure and income, the random error term ei is not observable; it is unobservable. We cannot measure tastes and preferences in any direct way, just as we cannot directly measure the economic “utility” derived from eating a slice of cake. The second regression assumption is that the x-variable, income, cannot be used to predict the value of ei , the effect of the collection of all other factors affecting the food expenditure by the ith household. Given 2 an income value xi for the ith household, the best (optimal) ) predictor of the random error ei is ( the conditional expectation, or conditional E ei |xi . The assumption that xi cannot be used ) ( mean, to predict ei is equivalent to saying E ei |xi = 0. That is, given a household’s income we cannot do any better than predicting that the random error is zero; the effects of all other factors on food expenditure average out, in a very specific way, to zero. We will discuss other situations in which this might or might )not be true in Section 2.10. For now, recall(from)the Probability ( )Primer, ( Section P.6.5, that E ei |xi = 0 has two implications. The first is E ei |xi = 0 =⇒ E ei = 0; if the conditional expected value of the random error is zero, then the unconditional expectation of the random error is also zero. In the population, the average effect of all the omitted factors summarized by the random error( term) is zero. ( ) The second implication is E ei |xi = 0 =⇒ cov ei , xi = 0. If the conditional expected value of the random error is zero, then ei , the random error for the ith observation, has covariance zero and correlation zero, with the corresponding observation xi . In our example, the random component ei , representing all factors affecting food expenditure except income for the ith household, is uncorrelated with income for that household. You might wonder how that could possibly be shown to be true. After all, ei is unobservable. The answer is that it is very hard work. You must convince yourself and your audience that anything that might have been omitted from the model is not correlated with xi . The primary tool is economic reasoning: your own intellectual experiments (i.e., thinking), reading literature( on the ) topic and discussions with colleagues or classmates. And we really can’t prove that E ei |xi = 0 is true with absolute certainty in most economic models. ( ) ( We) noted that E ei |xi = 0 has two implications. If either of the implications is not true, then E ei |xi = 0 is not true, that is, ( ) ( ) ( ) E ei |xi ≠ 0 if (i) E ei ≠ 0 or if (ii) cov ei , xi ≠ 0 ) ( In the first case, if the population average of the random errors ei is not zero, then E e |x i i ( ) ( ) ≠ 0. In a certain sense, we will be able to work around the ) when E ei ≠(0, say) if E ei = 3, ( case as you will see below. The second implication of E ei |xi = 0 is that cov ei , xi = 0; the random error for the ith observation has ) covariance and correlation with the ith observation ( zero variable x is said to be exogeon the explanatory variable. If cov ei , xi = 0, the (explanatory ) nous, providing our first assumption that the pairs yi , xi are iid holds. When x is exogenous, regression analysis )can be used successfully to estimate β1 and β2 . To differentiate ( ) the weaker ( from the stronger condition E e = 0, we say |x condition cov ei , xi = 0, simple( exogeneity, i i ) ( ) that x is strictly exogenous if E ei |xi = 0. If cov ei , xi ≠ 0, then x is said to be endogenous. When x is endogenous, it is more difficult, sometimes much more difficult, to carry out statistical inference. A great deal will be said about exogeneity and strict exogeneity in the remainder of this book. ............................................................................................................................................ 2 You will learn about optimal prediction in Appendix 4C. k k k 2.2 An Econometric Model E X A M P L E 2.1 A Failure of the Exogeneity Assumption Consider a regression model exploring the relationship between a working person’s wage and their years of education, using a random sample of data. The simple regression model is WAGEi = β1 + β2 EDUCi + ei , where WAGEi is the hourly wage rate of the ith randomly selected person and EDUC ) i is their years of education. The pairs ( WAGEi , EDUCi from the random sample are assumed to be iid. In this model, the random error ei accounts for all those factors other than EDUCi that affect a person’s wage rate. What might some of those factors be? Ability, intelligence, 2.2.3 perseverance, and industriousness are all important characteristics of an employee and likely to influence their wage rate. Are any of these factors which are bundled into ei likely to be correlated with EDUCi ? A few moments reflection will lead you to say “yes.” It is very plausible that those with higher education have higher ability, intelligence, perseverance, and industriousness. Thus, there is a strong argument that EDUCi is an endogenous regressor in this regression and that the strict exogeneity assumption fails. The Regression Function The importance ( of) the strict exogeneity assumption is the following. If the strict exogeneity assumption E ei |xi = 0 is true, then the conditional expectation of yi given xi is ) ) ( ( E yi |xi = β1 + β2 xi + E ei |xi = β1 + β2 xi , i = 1, … , N (2.2) ) ( The conditional expectation E yi |xi = β1 + β2 xi in (2.2) is called the regression function, or population regression function. It says that in the population the average value of the dependent variable for the ith observation, conditional ( on )xi , is given by β1 + β2 xi . It also says that given a change in x, Δx, the resulting change in E yi |xi is β2 Δx holding all else constant, in the sense that given xi the average of the random errors is zero, and any change in x is not correlated with any corresponding change in the random error e. In this case, we can say that a change in ) x leads ( to, or causes, a change in the expected (population average) value of y given xi , E yi |x(i . ) The regression function in (2.2) is shown in Figure 2.2, with y-intercept β1 = E yi |xi = 0 and slope ) ) ( ( ΔE yi |xi dE yi |xi β2 = = (2.3) Δxi dxi where Δ denotes “change in” and dE(y|x)∕dx denotes the “derivative” of E(y|x) with respect to x. We will not use derivatives to any great extent in this book, and if you are not too familiar with the concept you can think of “d” as a stylized version of Δ and go on. See Appendix A.3 for a discussion of derivatives. E( y| x) Average expenditure k 53 E(y| x) = β1 + β2 x ΔE( y| x) β2 = Δx ΔE(y| x) dE(y | x) = Δx dx β1 x Income FIGURE 2.2 The economic model: a linear relationship between average per person food expenditure and income. k k k 54 CHAPTER 2 The Simple Linear Regression Model E X A M P L E 2.2 Strict Exogeneity in the Household Food Expenditure Model The strict exogeneity assumption is that the average of everything else affecting the food expenditure of the ith household, given the income of the ith household, is zero. Could this be true? One test of this possibility is the question “Using the income of the ith household, can we predict the value of ei , the combined influence of all factors affecting food expenditure other than income?” If( the )answer is yes, then strict exogeneity fails. If not, then E ei |xi = 0 may be a plausible assumption. And if it is, then equation (2.1) can be interpreted as a causal model, and β2 can be thought of as the marginal effect of income on expected (average) household food expenditure, (holding ) all else constant, as shown in equation (2.3). If E ei |xi ≠ 0 then xi can be used to predict a nonzero value for ei , which in turn will affect the value of yi . In this case, β2 will not capture all the effects of an income change, and the model cannot be interpreted as causal. Another important consequence of the assumption of strict exogeneity is that it allows us to think of the econometric model as decomposing the dependent variable into two components: one that varies systematically as the values of the independent variable change and another that is random That is, the econometric model yi = β1 + β2 xi + ei can be broken into two parts: ) ( “noise.” E yi |xi = β1 + β2 xi and the random error, ei . Thus ) ( yi = β1 + β2 xi + ei = E yi |xi + ei k The values ( of) the dependent variable yi vary systematically due to variation in the conditional mean E yi |xi = β1 + β2 xi , as the value of the explanatory variable changes, and the values of the dependent variable yi vary randomly due to ei . The conditional pdf s of e and y are identical except for their location, as shown in Figure 2.3. Two values of food expenditure y1 and y2 for households with x = $1000 of weekly income are shown in Figure 2.4 relative to their conditional mean. There will be variation in household expenditures on food from one household to another because of variations in tastes and preferences, and everything else. Some will spend more than the average value for households with the same income, and some will ( spend less.)If we knew β1 and β2 , then we could compute the conditional mean expenditure E yi |x = 1000 = β1 + β2 (1000) and also the value of the random errors e1 and e2 . We never know β1 and β2 so we can never compute e1 and e2 . What we are assuming, however, is that at each level of income x the average value of all that is represented by the random error is zero. 2.2.4 Random Error Variation We have ) the assumption that the conditional expectation of the random error term is ( made zero, E ei |xi = 0. For the random error term we are interested in both its conditional mean, f(∙∣x = 1000) f (e∣x = 1000) E(e∣x = 1000) = 0 f (y∣x = 1000) E(y∣x = 1000) = β1 + β2(1000) FIGURE 2.3 Conditional probability densities for e and y. k k k 2.2 An Econometric Model 55 f(y∣x = 1000) f(y∣x = 1000) e1 e2 y1 y2 E(y∣x = 1000) = β1 + β2(1000) y FIGURE 2.4 The random error. or expected value, and its variance. Ideally, the conditional variance of the random error is constant. ) ( (2.4) var ei |xi = σ2 This is the homoskedasticity (also spelled homoscedasticity) assumption. At each xi the variation of the random error component is the same. Assuming the population relationship yi = β1 + β2 xi + ei the conditional variance of the dependent variable is ) ( ) ( ) ( var yi |xi = var β1 + β2 xi + ei |xi = var ei |xi = σ2 k The simplification works because by conditioning on xi we are treating it as if it is known and therefore not random. Given xi the component β1 + β2 xi is not random, so the variance rule (P.14) applies. This was an explicit assumption in Figure 2.1(b) where the pdf s f (y|x = 1000) and f (y|x( = 2000) have the same variance, σ2 . If strict exogeneity holds, then the regression function ) is E yi |xi = β1 + β2 xi , as shown in Figure 2.2. The conditional distributions f (y|x = 1000) and f (y|x = 2000) are placed along the conditional mean function in Figure 2.5. In the household expenditure example, the idea is that for a particular level of household income x, the values of household food expenditure will vary randomly about the conditional mean due to the assumption that at each x the average value of the random error e is zero. Consequently, at each level of income, household food expenditures are centered on the regression function. The conditional homoskedasticity assumption implies that at each level of income the variation in food f(y∣x) y re ditu d Foo en exp 0) 0) x= f(y∣ 100 x= f(y∣ 200 β1 + β2x = E(y∣x) μy∣1000 x= 100 0 x= 200 0 Hou μy∣2000 seho ld in com e x FIGURE 2.5 The conditional probability density functions for y, food expenditure, at two levels of income. k k k 56 CHAPTER 2 The Simple Linear Regression Model expenditure about its mean is the same. That means that at each and every level of income ( we)are equally uncertain about how far food expenditures might fall from their mean value, E yi |xi = β1 + β2 xi . Furthermore, this uncertainty does not depend on income or anything else. If this ) ( assumption is violated, and var ei |xi ≠ σ2 , then the random errors are said to be heteroskedastic. 2.2.5 Variation in x 2.2.6 Error Normality ) ( In a regression analysis, one of the objectives is to estimate β2 = ΔE yi |xi ∕Δxi . If we are to hope that a sample of data can be used to estimate the effects of changes in x, then we must observe some different values of the explanatory variable x in the sample. Intuitively, if we collect data only on households with income $1000, we will not be able to measure the effect of changing income on the average value of food expenditure. Recall from elementary geometry that “it takes two points to determine a line.” The minimum number of x-values in a sample of data that will allow us to proceed is two. You will find out in Section 2.4.4 that in fact the more different values of x, and the more variation they exhibit, the better our regression analysis will be. k In the discussion surrounding Figure 2.1, we explicitly made the assumption that food expenditures, given income, were normally distributed. In Figures 2.3–2.5, we implicitly made the assumption of conditionally normally distributed errors and dependent variable by drawing classically bell-shaped curves. It is not at all necessary for the random errors to be conditionally normal in order for regression analysis to “work.” However, as you will discover in Chapter 3, when samples are small, it is advantageous for statistical inferences that the random errors, and dependent variable y, given each x-value, are normally distributed. The normal distribution has a long and interesting history,3 as a little Internet searching will reveal. One argument for assuming regression errors are normally distributed is that they represent a collection of many different factors. The Central Limit Theorem, see Appendix C.3.4, says roughly that collections of many random factors tend toward having a normal distribution. In the context of the food expenditure model, if we consider that the random errors reflect tastes and preferences, it is entirely plausible that the random errors at each income level are normally When the assumption of con) ) ( distributed. ( ditionally normal errors is made, we write ei |xi ∼ N 0, σ2 and also then yi |xi ∼ N β1 + β2 xi , σ2 . It is a very strong assumption when it is made, and as mentioned it is not strictly speaking necessary, so we call it an optional assumption. 2.2.7 Generalizing the Exogeneity Assumption ) ( So far we have assumed that the data pairs yi , xi have been drawn from a random sample and are iid. What happens if the sample values of the explanatory variable are correlated? And how might that happen? A lack of independence occurs naturally when using financial or macroeconomic time-series data. Suppose we observe the monthly report on new housing starts, yt , and the current )30-year ( fixed mortgage rate, xt , and we postulate the model yt = β1 + β2 xt + et . The data yt , xt can be described as macroeconomic time-series data. In contrast to cross-section data where we have observations on a number of units (say households or firms or persons or countries) at a given point in time, with time-series data we have observations over time on a number of variables. It is customary to use(a “t”)subscript to denote time-series data and to use T to denote the sample size. In the data pairs yt , xt , t = 1, … , T, both yt and xt are random because we do not know the values ............................................................................................................................................ 3 For example, Stephen M. Stigler (1990) The History of Statistics: The Measurement of Uncertainty, Reprint Edition, Belknap Press, 73–76. k k k 2.2 An Econometric Model 57 until they are observed. Furthermore, each of the data series is likely to be correlated across time. For example, the monthly fixed mortgage rate is likely to change slowly over time making the ( ) rate at time t correlated with the rate at time t − 1. The assumption that the pairs yt , xt represent random iid draws from a probability distribution is not realistic. When considering the exogeneity assumption for this case, we need to be concerned not just with possible correlation between xt and et , but also with possible correlation between et and every other value of the explanatory variable, namely, xs , s = 1, 2, … , T. If xs is correlated with xt , then it is possible that xs (say, the mortgage rate in one month) may have an impact on yt (say, housing starts in the next month). Since it is xt , not x(s that)appears in the equation yt = β1 + β2 xt + et , the effect of xs will be included in et , implying E et |xs (≠ 0. We ) could use xs to help predict the value of et . This possibility is That is, independence of the pairs ruled out when the pairs y , x t t( are )assumed to be(independent. ( ) ) yt , xt and the assumption E et |xt = 0 imply E et |xs = 0 for all s = 1, 2, … , T. To extend the strict assumption to models where the values of x are correlated, ( exogeneity ) we need to assume E et |xs = 0 for all (t, s) = 1, 2, … , T. This means that we cannot predict the random error at time t,( et , using any of the values of the explanatory variable. Or, in terms of ) our earlier notation, E ei |xj = 0 for all (i, j) = 1, 2, … , N. To write this assumption in a more ) ( convenient form, we introduce the notation 𝐱 = x1 , x2 , … , xN . That is, we are using x to denote all sample observations on the explanatory variable. Then, a more general way of writing the ( ) strict exogeneity assumption is E e |𝐱 = 0, i = 1, 2, … , N. From this assumption, we can also i ( ) write E yi |𝐱 = β1 + β2 xi for i = 1, 2, … , N. This assumption is discussed further in ( the) context of alternative types of data in Section 2.10 and in Chapter 9. The assumption E ) (ei |𝐱 )= 0, ( i = 1, 2, … , N, is a weaker assumption than assuming E ei |xi = 0 and that the pairs yi , xi are independent, and it enables us to derive a number of results for cases where different observations on x may be correlated as well as for the case where they are independent. k k 2.2.8 Error Correlation ( ) In addition ( ) ( ) to possible correlations between a random error for one household ei or one time period et being correlated with the value of an explanatory variable for another household xj ( ) or time period xs , it is possible that there are correlations between the random error terms. With cross-sectional data, data on households, individuals, or firms collected at one point in time, there may be a lack of statistical independence between random errors for individuals who are spatially connected. That is, suppose that we collect observations on two (or more) individuals who live in the same neighborhood. It is very plausible that there are similarities among people who live in a particular neighborhood. Neighbors can be expected to have similar incomes if the homes in a neighborhood are homogenous. Some suburban neighborhoods are popular because of green space and schools for young children, meaning households may have members similar in ages and interests. We might add a spatial component s to the error and say that the random errors ei (s) and ej (s) for the ith and jth households are possibly correlated because of their common location. Within a larger sample of data, there may be clusters of observations with correlated errors because of the spatial component. In a time-series context, your author is writing these pages on the tenth anniversary of Hurricane Katrina, which devastated the U.S. Gulf Coast and the city of New Orleans, Louisiana, in particular. The impact of that shock did not just happen and then go away. The effect of that huge random event had an effect on housing and financial markets during August 2005, and also in September, October, and so on, to this day. Consequently, the random errors ) in the population ( ) re( lationship yt = β1 + β2 xt + et are correlated over time, so that cov et , et+1 ≠ 0, cov et , et+2 ≠ 0, and so on. This is called serial correlation, or autocorrelation, in econometrics. The starting point in regression analysis(is to assume that there is no error correlation. In ) time-series models, we start by assuming cov e , e |𝐱 = 0 for t ≠ s, and for cross-sectional data t s ) ( we start by assuming cov ei , ej |𝐱 = 0 for i ≠ j. We will cope with failure of these assumptions in Chapter 9. k k 58 CHAPTER 2 The Simple Linear Regression Model 2.2.9 Summarizing the Assumptions We summarize the starting assumptions of the simple regression model in a very general way. In our summary we use subscripts i and j but the assumptions are general, and apply equally to time-series data. If these assumptions hold, then regression analysis can successfully )estimate ( the(unknown population parameters β and β and we can claim that β = ΔE y |x 1 2 2 i i ∕Δxi = ) dE yi |xi ∕dxi measures a causal effect. We begin our study of regression analysis and econometrics making these strong assumptions about the DGP. For future reference, the assumptions are named SR1–SR6, “SR” denoting “simple regression.” Econometrics is in large part devoted to handling data and models for which these assumptions may not hold, leading to modifications of usual methods for estimating β1 and β2 , testing hypotheses, and predicting outcomes. In Chapters 2 and 3, we study the simple regression model under these, or similar, strong assumptions. In Chapter 4, we introduce modeling issues and diagnostic testing. In Chapter 5, we extend our model to multiple regression analysis with more than one explanatory variable. In Chapter 6, we treat modeling issues concerning the multiple regression model, and starting in Chapter 8 we address situations in which SR1–SR6 are violated in one way or another. Assumptions of the Simple Linear Regression Model ) ( SR1: Econometric Model All data pairs yi , xi collected from a population satisfy the relationship yi = β1 + β2 xi + ei , i = 1, … , N k SR2:( Strict Exogeneity The conditional expected value of the random error ei is zero. If ) 𝐱 = x1 , x2 , … , xN , then ( ) E ei |x = 0 If strict exogeneity holds, then the population regression function is ( ) E yi |x = β1 + β2 xi , i = 1, … , N and ( ) yi = E yi |x + ei , i = 1, … , N SR3: Conditional Homoskedasticity The conditional variance of the random error is constant. ) ( var ei |x = σ2 SR4: Conditionally Uncorrelated Errors The conditional covariance of random errors ei and ej is zero. ( ) cov ei , ej |x = 0 for i ≠ j SR5: Explanatory Variable Must Vary In a sample of data, xi must take at least two different values. SR6: Error Normality (optional) The conditional distribution of the random errors is normal. ) ( ei |x ∼ N 0, σ2 The random error e and the dependent variable y are both random variables, and as we have shown, the properties of one variable can be determined from the properties of the other. There is, however, one interesting difference between them: y is “observable” and e is “unobservable.” k k k 2.3 Estimating the Regression Parameters 59 If the regression parameters ) β1 and β2 were known, then for any values yi and xi we could ( = y − β + β x calculate e i 1 2 i . This is illustrated in Figure 2.4. Knowing the regression function ( ) i E yi |𝐱 = β1 + β2 xi we could separate yi into its systematic and random parts. However, β1 and β2 are never known, and it is impossible to calculate ei . What comprises the error term e? The random error e represents all factors affecting y other than x, or what we have called everything (else. )These factors cause individual observations yi to differ from the conditional mean value E yi |𝐱 = β1 + β2 xi . In the food expenditure example, what factors can(result ) in a difference between household expenditure per person yi and its conditional mean E yi |𝐱 = β1 + β2 xi ? 1. We have included income as the only explanatory variable in this model. Any other economic factors that affect expenditures on food are “collected” in the error term. Naturally, in any economic model, we want to include all the important and relevant explanatory variables in the model, so the error term e is a “storage bin” for unobservable and/or unimportant factors affecting household expenditures on food. As such, it adds noise that masks the relationship between x and y. 2. The error term e captures any approximation error that arises because the linear functional form we have assumed may be only an approximation to reality. 3. The error term captures any elements of random behavior that may be present in each individual. Knowing all the variables that influence a household’s food expenditure might not be enough to perfectly predict expenditure. Unpredictable human behavior is also contained in e. k If we have omitted( some ) important factor, or made any other serious specification error, then assumption SR2 E ei |𝐱 = 0 will be violated, which will have serious consequences. 2.3 k Estimating the Regression Parameters E X A M P L E 2.3 Food Expenditure Model Data The economic and econometric models we developed in the previous section are the basis for using a sample of data to estimate the intercept and slope parameters, β1 and β2 . For illustration we examine typical data on household food expenditure and weekly income from a random sample of 40 households. Representative observations and summary statistics are given in Table 2.1. We control for household size by considering only three-person households. The values of y are weekly food expenditures for a three-person household, in dollars. Instead of measuring income in dollars, we measure it in units of $100, because a $1 increase in income has a numerically small effect on food expenditure. Consequently, for the first household, the reported income is $369 per week with weekly food expenditure of $115.22. For the 40th household, weekly income is $3340 and weekly food expenditure is $375.73. The complete data set of observations is in the data file food. k T A B L E 2.1 Food Expenditure and Income Data Observation (household) Food Expenditure ($) i 1 2 39 40 Sample mean Median Maximum Minimum Std. dev. yi 115.22 135.98 ⋮ 257.95 375.73 Summary Statistics 283.5735 264.4800 587.6600 109.7100 112.6752 Weekly Income ($100) xi 3.69 4.39 29.40 33.40 19.6048 20.0300 33.4000 3.6900 6.8478 k 60 CHAPTER 2 The Simple Linear Regression Model Remark In this book, data files are referenced with a descriptive name in italics such as food. The actual files which are located at the book websites www.wiley.com/college/hill and www.principlesofeconometrics.com come in various formats and have an extension that denotes the format, for example, food.dat, food.wf1, food.dta, and so on. The corresponding data definition file is food.def . We assume that the expenditure data in Table 2.1 satisfy the assumptions SR1–SR5. That is, we assume that the regression model yi = β1 + β2 xi + ei describes a population relationship and that the random error has conditional expected value zero. This implies that the conditional expected value of household food expenditure is a linear function of income. The conditional variance of y, which is the same as that of the random error e, is assumed constant, implying that we are equally uncertain about the relationship between y and x for all observations. Given x the values of y for different households are assumed uncorrelated with each other. Given this theoretical model for explaining the sample observations on household food expenditure, the problem now is how to use the sample information in Table 2.1, specific values of yi and xi , to estimate the unknown regression parameters β1 and β2 . These parameters represent the unknown intercept and slope(coefficients for the food expenditure–income relationship. If we ) represent the 40 data points as yi , xi , i = 1, … , N = 40, and plot them, we obtain the scatter diagram in Figure 2.6. Remark 0 100 200 300 400 500 600 700 It will be our notational convention to use i subscripts for cross-sectional data observations, with the number of sample observations being N. For time-series data observations, we use the subscript t and label the total number of observations T. In purely algebraic or generic situations, we may use one or the other. y = weekly food expenditure in $ k 0 10 20 x = weekly income in $100 FIGURE 2.6 Data for the food expenditure example. k 30 40 k k 2.3 Estimating the Regression Parameters 61 ( ) Our problem is to estimate the location of the mean expenditure line E yi |𝐱 = β1 + β2 xi . We would expect this line to be somewhere in the middle of all the data points since it represents population mean, or average, behavior. To estimate β1 and β2 , we could simply draw a freehand line through the middle of the data and then measure the slope and intercept with a ruler. The problem with this method is that different people would draw different lines, and the lack of a formal criterion makes it difficult to assess the accuracy of the method. Another method is to draw a line from the expenditure at the smallest income level, observation i = 1, to the expenditure at the largest income level, i = 40. This approach does provide a formal rule. However, it may not be a very good rule because it ignores information on the exact position of the remaining 38 observations. It would be better if we could devise a rule that uses all the information from all the data points. 2.3.1 The Least Squares Principle To estimate β1 and β2 we want a rule, or formula, that tells us how to make use of the sample observations. Many rules are possible, but the one that we will use is based on the least squares principle. This principle asserts that to fit a line to the data values we should make the sum of the squares of the vertical distances from each point to the line as small as possible. The distances are squared to prevent large positive distances from being canceled by large negative distances. This rule is arbitrary, but very effective, and is simply one way to describe a line that runs through the middle of the data. The intercept and slope of this line, the line that best fits the data using the least squares principle, are b1 and b2 , the least squares estimates of β1 and β2 . The fitted line itself is then (2.5) ŷ i = b1 + b2 xi k The vertical distances from each point to the fitted line are the least squares residuals. They are given by (2.6) ê i = yi − ŷ i = yi − b1 − b2 xi These residuals are depicted in Figure 2.7a. Now suppose we fit another line, any other line, to the data. Denote the new line as ŷ ∗i = b∗1 + b∗2 xi where b∗1 and b∗2 are any other intercept and slope values. The residuals for this line, ê ∗i = yi − ŷ ∗i , are shown in Figure 2.7b. The least squares estimates b1 and b2 have the property that the sum of their squared residuals is less than the sum of squared residuals for any other line. That is, if SSE = N ∑ i=1 ê 2i is the sum of squared least squares residuals from (2.6) and SSE∗ = N ∑ i=1 2 ê ∗i = N ( ∑ i=1 yi − ŷ ∗i )2 is the sum of squared residuals based on any other estimates, then SSE < SSE∗ no matter how the other line might be drawn through the data. The least squares principle says that the estimates b1 and b2 of β1 and β2 are the ones to use, since the line using them as intercept and slope fits the data best. The problem is to find b1 and b2 in a convenient way. Given the sample observations on y and x, we want to find values for the unknown parameters β1 and β2 that minimize the “sum of squares” function N ( ) ∑ )2 ( yi − β1 − β2 xi S β1 , β2 = i=1 k k k 62 CHAPTER 2 The Simple Linear Regression Model y y = b1 + b2 x y4 e4 y3 y4 e3 y3 y2 e2 y1 e1 y2 y1 x1 x2 x3 x4 x (a) y e3* y*2 y*1 y* = b*1 + b2* x e4* y3 y4* y* 3 e2* e*1 k y = b1 + b2 x y4 y2 k y1 x1 x2 x3 x4 x (b) FIGURE 2.7 (a) The relationship among y, ê , and the fitted regression line. (b) The residuals from another fitted line. This is a straightforward calculus problem, the details of which are given in Appendix 2A. The formulas for the least squares estimates of β1 and β2 that give the minimum of the sum of squared residuals are The Ordinary Least Squares (OLS) Estimators ∑( b2 = where y = ∑ yi ∕N and x = ∑ )( ) xi − x yi − y )2 ∑( xi − x b1 = y − b2 x (2.7) (2.8) xi ∕N are the sample means of the observations on y and x. We will call the estimators b1 and b2 , given in equations (2.7) and (2.8), the ordinary least squares estimators. “Ordinary least squares” is abbreviated as OLS. These least squares estimators are called “ordinary,” despite the fact that they are extraordinary, because these estimators are used day in and day out in many fields of research in a routine way, and to distinguish them k k 2.3 Estimating the Regression Parameters 63 from other methods called generalized least squares, and weighted least squares, and two-stage least squares, all of which are introduced later in this book. The formula for b2 reveals why we had to assume [SR5] that in the sample xi must take at least two different values. If xi = 5, for example, for all observations, then b2 in (2.7) is mathematically undefined and does not exist since its numerator and denominator are zero! If we plug the sample values yi and xi into (2.7) and (2.8), then we obtain the least squares estimates of the intercept and slope parameters β1 and β2 . It is interesting, however, and very important, that the formulas for b1 and b2 are perfectly general and can be used no matter what the sample values turn out to be. This should ring a bell. When the formulas for b1 and b2 are taken to be rules that are used whatever the sample data turn out to be, then b1 and b2 are random variables. When actual sample values are substituted into the formulas, we obtain numbers that are the observed values of random variables. To distinguish these two cases, we call the rules or general formulas for b1 and b2 the least squares estimators. We call the numbers obtained when the formulas are used with a particular sample least squares estimates. • Least squares estimators are general formulas and are random variables. • Least squares estimates are numbers that we obtain by applying the general formulas to the observed data. The distinction between estimators and estimates is a fundamental concept that is essential to understand everything in the rest of this book. E X A M P L E 2.4a k Estimates for the Food Expenditure Function Using the least squares estimators (2.7) and (2.8), we can obtain the least squares estimates for the intercept and slope parameters β1 and β2 in the food expenditure example using the data in Table 2.1. From (2.7), we have )( ) ∑( xi − x yi − y 18671.2684 = 10.2096 = b2 = )2 ∑( 1828.7876 x −x i and from (2.8) b1 = y − b2 x = 283.5735 − (10.2096)(19.6048) = 83.4160 A convenient way to report the values for b1 and b2 is to write out the estimated or fitted regression line, with the estimates rounded appropriately: ŷ i = 83.42 + 10.21xi This line is graphed in Figure 2.8. The line’s slope is 10.21, and its intercept, where it crosses the vertical axis, is 83.42. The least squares fitted line passes through the middle of the data in a very precise way, since one of the characteristics of the fitted line based on the least squares parameter estimates is that it( passes through the point defined by the sample ) means, x, y = (19.6048, 283.5735). This follows directly from rewriting (2.8) as y = b1 + b2 x. Thus, the “point of the means” is a useful reference value in regression analysis. Interpreting the Estimates Once obtained, the least squares estimates are interpreted in the context of the economic model under consideration. k The value b2 = 10.21 is an estimate of β2 . Recall that x, weekly household income, is measured in $100 units. The regression slope β2 is the amount by which expected weekly expenditure on food per household increases when household weekly income increases by $100. Thus, we estimate that if weekly household income goes up by $100, expected weekly expenditure on food will increase by approximately $10.21, holding all else constant. A supermarket executive with information on likely changes in the income and the number of households in an area could estimate that it will sell $10.21 more per typical household per week for every $100 increase in income. This is a very valuable piece of information for long-run planning. Strictly speaking, the intercept estimate b1 = 83.42 is an estimate of the expected weekly food expenditure for a household with zero income. In most economic models we must be very careful when interpreting the estimated intercept. The problem is that we often do not have any data points near x = 0, something that is true for the food expenditure data shown in Figure 2.8. If we have no observations in the region where income is zero, then our estimated relationship may not be a good approximation to reality in that region. So, although our estimated model suggests that a household with zero income is expected to spend $83.42 per week on food, it might be risky to take this estimate literally. This is an issue that you should consider in each economic model that you estimate. k k CHAPTER 2 The Simple Linear Regression Model 500 y 300 400 y = 83.42 + 10.21 x 200 – y = 283.57 100 y = weekly food expenditure in $ 600 700 64 0 – x = 19.60 0 10 20 x = weekly income in $100 30 40 FIGURE 2.8 The fitted regression. k k Elasticities Income elasticity is a useful way to characterize the responsiveness of consumer expenditure to changes in income. See Appendix A.2.2 for a discussion of elasticity calculations in a linear relationship. The elasticity of a variable y with respect to another variable x is ε= percentage change in y 100(Δy∕y) Δy x • = = percentage change in x 100(Δx∕x) Δx y In the linear economic model given by (2.1), we have shown that β2 = ΔE(y|x) Δx so the elasticity of mean expenditure with respect to income is ε= E X A M P L E 2.4b ΔE(y|x) x x • = β2 • Δx E(y|x) E(y|x) (2.9) Using the Estimates To estimate this elasticity we replace β2 by b2 = 10.21. We must also replace “x” and “E(y|x)” by something, since in a linear model the elasticity is different on each point on the regression line. Most commonly, ( ) the elasticity is calculated at the “point of the means” x, y = (19.60, 283.57) because it is a representative point on the regression line. If we calculate the income elasticity at the point of the means, we obtain 19.60 x = 0.71 ε̂ = b2 = 10.21 × 283.57 y k k 2.3 Estimating the Regression Parameters This estimated income elasticity takes its usual interpretation. We estimate that a 1% increase in weekly household income will lead to a 0.71% increase in expected weekly household expenditure ( )on food, when x and y take their sample mean values, x, y = (19.60, 283.57). Since the estimated income elasticity is less than one, we would classify food as a “necessity” rather than a “luxury,” which is consistent with what we would expect for an average household. Prediction The estimated equation can also be used for prediction or forecasting purposes. Suppose that we wanted to predict average weekly food expenditure for a household with a weekly income of $2000. This prediction is carried out by substituting x = 20 into our estimated equation to obtain ŷ i = 83.42 + 10.21xi = 83.42 + 10.21(20) = 287.61 We predict that a household with a weekly income of $2000 will on average spend $287.61 per week on food. Computer Output 65 different and uses different terminology to describe the output. Despite these differences, the various outputs provide the same basic information, which you should be able to locate and interpret. The matter is complicated somewhat by the fact that the packages also report various numbers whose meaning you may not know. For example, using the food expenditure data, the output from the software package EViews is shown in Figure 2.9. In the EViews output, the parameter estimates are in the “Coefficient” column, with names “C,” for constant term (the estimate b1 ) and INCOME (the estimate b2 ). Software programs typically name the estimates with the name of the variable as assigned in the computer program (we named our variable INCOME) and an abbreviation for “constant.” The estimates that we report in the text are rounded to two significant digits. The other numbers that you can ∑ 2 recognize at this time are SSE = ê i = 304505.2, which is called “Sum squared resid,” and the sample mean of y, ∑ y = yi ∕N = 283.5735, which is called “Mean dependent var.” We leave discussion of the rest of the output until later. Many different software packages can compute least squares estimates. Every software package’s regression output looks k k Dependent Variable: FOOD_EXP Method: Least Squares Sample: 1 40 Included observations: 40 C INCOME R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood F-statistic Prob(F-statistic) Coefficient Std. Error t-Statistic Prob. 83.41600 10.20964 43.41016 2.093264 1.921578 4.877381 0.0622 0.0000 0.385002 0.368818 89.51700 304505.2 −235.5088 23.78884 0.000019 Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Hannan-Quinn criter Durbin-Watson stat 283.5735 112.6752 11.87544 11.95988 11.90597 1.893880 FIGURE 2.9 EViews regression output. 2.3.2 Other Economic Models We have used the household expenditure on food versus income relationship as an example to introduce the ideas of simple regression. The simple regression model can be applied to estimate k k 66 CHAPTER 2 The Simple Linear Regression Model the parameters of many relationships in economics, business, and the social sciences. The applications of regression analysis are fascinating and useful. For example, • If the hourly wage rate of electricians rises by 5%, how much will new house prices increase? • If the cigarette tax increases by $1, how much additional revenue will be generated in the state of Louisiana? • If the central banking authority raises interest rates by one-half a percentage point, how much will consumer borrowing fall within six months? How much will it fall within one year? What will happen to the unemployment rate in the months following the increase? • If we increase funding on preschool education programs in 2018, what will be the effect on high school graduation rates in 2033? What will be the effect on the crime rate by juveniles in 2028 and subsequent years? The range of applications spans economics and finance, as well as most disciplines in the social and physical sciences. Any time you ask how much a change in one variable will affect another variable, regression analysis is a potential tool. Similarly, any time you wish to predict the value of one variable given the value of another then least squares regression is a tool to consider. 2.4 k Assessing the Least Squares Estimators Using the food expenditure data, we have estimated the parameters of the regression model yi = β1 + β2 xi + ei using the least squares formulas in (2.7) and (2.8). We obtained the least squares estimates b1 = 83.42 and b2 = 10.21. It is natural, but, as we shall argue, misguided, to ask the question “How good are these estimates?” This question is not answerable. We will never know the true values of the population parameters β1 or β2 , so we cannot say how close b1 = 83.42 and b2 = 10.21 are to the true values. The least squares estimates are numbers that may or may not be close to the true parameter values, and we will never know. Rather than asking about the quality of the estimates we will take a step back and examine the quality of the least squares estimation procedure. The motivation for this approach is this: if we were to collect another sample of data, by choosing another set of 40 households to survey, we would have obtained different estimates b1 and b2 , even if we had carefully selected households with the same incomes as in the initial sample. This sampling variation is unavoidable. Different samples will yield different estimates because household food expenditures, yi , i = 1, … , 40, are random variables. Their values are not known until the sample is collected. Consequently, when viewed as an estimation procedure, b1 and b2 are also random variables, because their values depend on the random variable y. In this context, we call b1 and b2 the least squares estimators. We can investigate the properties of the estimators b1 and b2 , which are called their sampling properties, and deal with the following important questions: 1. If the least squares estimators b1 and b2 are random variables, then what are their expected values, variances, covariances, and probability distributions? 2. The least squares principle is only one way of using the data to obtain estimates of β1 and β2 . How do the least squares estimators compare with other procedures that might be used, and how can we compare alternative estimators? For example, is there another estimator that has a higher probability of producing an estimate that is close to β2 ? We examine these questions in two steps to make things easier. In the first step, we investigate the properties of the least squares estimators conditional on the values of the explanatory variable in the sample. That is, conditional on x. Making the analysis conditional on x is equivalent to saying that, when we consider all possible samples, the household income values in the sample stay the k k k 2.4 Assessing the Least Squares Estimators 67 same from one sample to the next; only the random errors and food expenditure values change. This assumption is clearly not realistic but it simplifies the analysis. By conditioning on x, we are holding it constant, or fixed, meaning that we can treat the x-values as “not random.” In the second step, in Section 2.10, we return to the random sampling assumption ) ( considered and recognize that yi , xi data pairs are random, and randomly selecting households from a population leads to food expenditures and incomes that are random. However, even in this case and treating x as random, we will discover that most of our conclusions that treated x as nonrandom remain the same. In either case, whether we make the analysis conditional on x or make the analysis general by treating x as random, the answers to the questions above depend critically on whether the assumptions SR1–SR5 are satisfied. In later chapters, we will discuss how to check whether the assumptions we make hold in a specific application, and what we might do if one or more assumptions are shown not to hold. Remark We will summarize the properties of the least squares estimators in the next several sections. “Proofs” of important results appear in the appendices to this chapter. In many ways, it is good to see these concepts in the context of a simpler problem before tackling them in the regression model. Appendix C covers the topics in this chapter, and the next, in the familiar and algebraically easier problem of estimating the mean of a population. k The Estimator b2 Formulas (2.7) and (2.8) are used to compute the least squares estimates b1 and b2 . However, they are not well suited for examining theoretical properties of the estimators. In this section, we rewrite the formula for b2 to facilitate its analysis. In (2.7), b2 is given by )( ) ∑( xi − x yi − y b2 = )2 ∑( xi − x 2.4.1 This is called the deviation from the mean form of the estimator because the data have their sample means subtracted. Using assumption SR1 and a bit of algebra (Appendix 2C), we can write b2 as a linear estimator, N ∑ (2.10) b2 = wi yi i=1 where x −x wi = ∑( i )2 xi − x (2.11) The term wi depends only on x. Because we are conditioning our analysis on x, the term wi is treated as if it is a constant. We remind you that conditioning on x is equivalent to treating x as given, as in a controlled, repeatable experiment. Any estimator that is a weighted average of yi ’s, as in (2.10), is called a linear estimator. This is an important classification that we will speak more of later. Then, with yet more algebra (Appendix 2D), we can express b2 in a theoretically convenient way, ∑ b2 = β2 + wi ei (2.12) where ei is the random error in the linear regression model yi = β1 + β2 xi + ei . This formula is not useful for computations, because it depends on β2 , which we do not know, and on the ei ’s, k k k 68 CHAPTER 2 The Simple Linear Regression Model which are unobservable. However, for understanding the sampling properties of the least squares estimator, (2.12) is very useful. 2.4.2 The Expected Values of b1 and b2 The OLS estimator b2 is a random variable since its value is unknown until ) a sample is collected. ( What we will show is that if our model assumptions hold, then E b2 |𝐱 = β2 ; that is, given x the of expected value of b2 is equal to the true parameter β2 . When the expected value of any(estimator ) a parameter equals the true parameter value, then that estimator is unbiased. Since E b2 |𝐱 = β2 , the least squares estimator b2 given x is an unbiased estimator of β2 . In Section ( ) 2.10, we will show that the least squares estimator b2 is unconditionally unbiased also, E b2 = β2 . The intuitive meaning of unbiasedness comes from the sampling interpretation of mathematical expectation. Recognize that one sample of size N is just one of many samples that we could have been selected. If the formula for b2 is used to estimate β2 in each of those possible samples, then, if our assumptions are valid, the average value of the estimates b2 obtained from all possible samples will be β2 . We will show that this result is true so that we can illustrate the part played by the assumptions of the linear regression model. In (2.12), what parts are random? The parameter β2 is not random. It is a population parameter we are trying to estimate. Conditional on x we can treat xi as if it is not random. Then, conditional on x, wi is not random either, as it depends only on the values of xi . The only random factors in (2.12) are the random error terms ei . We can find the conditional expected value of b2 using the fact that the expected value of a sum is the sum of the expected values: k ) ( ) ( ) ( ∑ E b2 |x = E β2 + wi ei |x = E β2 + w1 e1 + w2 e2 + · · · + wN eN |x ( ) ( ) ( ) ( ) = E β2 + E w1 e1 |x + E w2 e2 |x + · · · + E wN eN |x ) ∑ ( = β2 + E wi ei |x ( ) ∑ = β2 + wi E ei |x = β2 k (2.13) The rules of expected values are fully discussed in the Probability Primer, P.5, ) ( and) ( Section Appendix B.1.1. In the last line of (2.13), we use two assumptions. First, E wi ei |𝐱 = wi E ei |𝐱 because conditional on x the terms wi are not random, and constants can ( be)factored out of expected values. Second, we have relied on the assumption that E ) ( ) ei |𝐱 = 0. Actually, if ( E ei |𝐱 = c, where c is any constant value, such as 3, then E b2 |𝐱 = β2 . Given x, the OLS estimator ) b2 is an unbiased estimator of the regression parameter β2 . On the other hand, if ( E ei |𝐱 ≠ 0 and it depends on x in ) way, then b2 is a biased estimator of β2 . One leading ( some case in which the assumption E ei |𝐱 = 0 fails is due to omitted variables. Recall that ei contains everything else affecting yi other than xi . If we have omitted ) anything that ( is )important ( and that is correlated with x then we would expect that E ei |𝐱 ≠ 0 and E b2 |𝐱 ≠ β2 . In Chapter 6 we discuss this omitted variables bias. Here we have shown that conditional on x, and under ( )SR1–SR5, the least squares estimator is linear and unbiased. In Section 2.10, we show that E b2 = β2 without conditioning on x. The unbiasedness of the estimator b2 is an important sampling property. On average, over all possible samples from the population, the least squares estimator is “correct,” on average, and this is one desirable property of an estimator. This statistical property by itself does not mean that b2 is a good estimator of β2 , but it is part of the story. The unbiasedness property is related to what happens in all possible samples of data from the same population. The fact that b2 is unbiased does not imply anything about what might happen in just one sample. An individual estimate (a number) b2 may be near to, or far from, β2 . Since β2 is never known we will never know, given k k 2.4 Assessing the Least Squares Estimators 69 one sample, whether our estimate is “close” to β2 or not. Thus, the estimate b2 = 10.21 may be close to β2 or not. ( ) The least squares estimator b1 of β1 is also an unbiased estimator, and E b1 |𝐱 = β1 if the model assumptions hold. 2.4.3 Sampling Variation To illustrate how the concept of unbiased estimation relates to sampling variation, we present in Table 2.2 least squares estimates of the food expenditure model from 10 hypothetical random samples (data file table2_2) of size N = 40 from the same population with the same incomes as the households given in Table 2.1. Note the variability of the least squares parameter estimates from sample to sample. This sampling variation is due to the fact that we obtain 40 different households in each sample, and their weekly food expenditure varies randomly. The property of unbiasedness is about the average values of b1 and b2 if used in all possible samples of the same size drawn from the same population. The average value of b1 in these 10 samples is b1 = 96.11. The average value of b2 is b2 = 8.70. If we took the averages of estimates from more samples, these averages would approach the true parameter values β1 and β2 . Unbiasedness does not say that an estimate from any one sample is close to the true parameter value, and thus we cannot say that an estimate is unbiased. We can say that the least squares estimation procedure (or the least squares estimator) is unbiased. 2.4.4 k The Variances and Covariance of b1 and b2 Table 2.2 shows that the least squares estimates of β1 and β2 vary from sample to sample. Understanding this variability is a key to assessing the reliability and sampling precision of an estimator. We now obtain the variances and covariance of the estimators b1 and b2 . Before presenting the expressions for the variances and covariance, let us consider why they are important to know. The variance of the random variable b2 is the average of the squared distances ) between the possible ( values of the random variable and its mean, which we now know is E b2 |𝐱 = β2 . The conditional variance of b2 is defined as {[ ( ) ( )]2 | } var b2 |x = E b2 − E b2 |x |x | T A B L E 2.2 Estimates from 10 Hypothetical Samples Sample b1 b2 1 2 3 4 5 6 7 8 9 10 93.64 91.62 126.76 55.98 87.26 122.55 91.95 72.48 90.34 128.55 8.24 8.90 6.59 11.23 9.14 6.80 9.84 10.50 8.75 6.99 k k k 70 CHAPTER 2 The Simple Linear Regression Model f2(b2∣x) f1(b2∣x) β2 FIGURE 2.10 Two possible probability density functions for b2 . k It measures the spread of the probability( distribution ( of b)2 . In Figure 2.10 are graphs of two ) possible probability distributions of b2 , f1 b2 |𝐱 and f2 b2 |𝐱 , that have the same mean value but different variances. ( ( ) ) The pdf f2 b2 |𝐱 has a smaller variance than f1 b2 |𝐱 . (Given) a choice, we (are interested in ) b b have the pdf f |𝐱 , rather than f |𝐱 . With the estimator precision and would prefer that b 2 2 2 1 2 ( ) probability is more concentrated around the true parameter value β2 givdistribution f2 b2(|𝐱 , the ) ing, relative to f1 b2 |𝐱 , a higher probability of getting an estimate that is close to β2 . Remember, getting an estimate close to β2 is a primary objective of regression analysis. The variance of an estimator measures the precision of the estimator in the sense that it tells us how much the estimates can vary from sample to sample. Consequently, we often refer to the sampling variance or sampling precision of an estimator. The smaller the variance of an estimator is, the greater the sampling precision of that estimator. One estimator is more precise than another estimator if its sampling variance is less than that of the other estimator. We will now present and discuss the conditional variances and covariance of b1 and b2 . Appendix 2E contains the derivation of the variance of the least squares estimator b2 . If the regression model assumptions SR1–SR5 are correct (assumption SR6 is not required), then the variances and covariance of b1 and b2 are [ ] ∑ 2 xi ( ) 2 var b1 |x = σ (2.14) )2 ∑( N xi − x ( ) σ2 var b2 |x = ∑( )2 xi − x [ ] ( ) −x 2 cov b1 , b2 |x = σ ∑( )2 xi − x (2.15) (2.16) At the beginning of this section we said that for unbiased estimators, smaller variances are better than larger variances. Let us consider the factors that affect the variances and covariance in (2.14)–(2.16). 1. The variance of the random error term, σ2 , appears in each of the expressions. It reflects the dispersion of the values y about their expected value E(y|x). The greater the variance σ2 , the greater is that dispersion, and the greater is the uncertainty about where the values of y fall relative to their conditional mean E(y|x). When σ2 is larger, the information we have about β1 and β2 is less precise. In Figure 2.5, the variance is reflected in the spread of the probability distributions f (y|x). The larger the variance term σ2 , the greater is the uncertainty in the statistical model, and the larger the variances and covariance of the least squares estimators. k k k 2.4 Assessing the Least Squares Estimators yi yi y y x (a) xi 71 yi = b1 + b2 xi x xi (b) FIGURE 2.11 The influence of variation in the explanatory variable x on precision of estimation: (a) low x variation, low precision: (b) high x variation, high precision. k )2 ∑( 2. The sum of squares of the values of x about their sample mean, xi − x , appears in each of the variances and in the covariance. This expression measures how spread out about their mean are the sample values of the independent or explanatory variable x. The more they are spread out, the larger the sum of squares. The less they are spread out, the smaller the sum of squares. You may recognize this sum of squares as the numerator of the sample variance of )2 ∑( the x-values. See Appendix C.4. The larger the sum of squares, xi − x , the smaller the conditional variances of the least squares estimators and the more precisely we can estimate the unknown parameters. The intuition behind this is demonstrated in Figure 2.11. Panel (b) is a data scatter in which the values of x are widely spread out along the x-axis. In panel (a), the data are “bunched.” Which data scatter would you prefer given the task of fitting a line by hand? Pretty clearly, the data in panel (b) do a better job of determining where the least squares line must fall, because they are more spread out along the x-axis. 3. The larger the sample size N, the smaller the variances and covariance of the least squares estimators; it is better to have more sample data than less. The sample size N appears in each of the variances and covariance because each of the sums consists ( ) )2of N terms. Also, ∑( N appears explicitly in var b1 |𝐱 . The sum of squares term xi − x gets larger as N increases because each of the terms in the sum is positive or zero (being zero if x happens to equal for an observation). Consequently, as N gets larger, both ( )its sample ( mean value ) var b2 |𝐱 and cov b1 , b2 |𝐱 get smaller, since the sum (of squares appears in their denomina) tor. The sums in the numerator and denominator of var b1 |𝐱 both get larger as N gets larger and(offset) one another, leaving the N in the denominator as the dominant term, ensuring that var b1 |𝐱 also gets smaller as N gets larger. ( ) ∑ 4. The term xi2 appears in var b1 |𝐱 . The larger this term is, the larger the variance of the least squares estimator b1 . Why is this so? Recall that the intercept parameter β1 is the expected value of y given that x = 0. The farther our data are from x = 0, the more difficult it is to interpret β1 , as in the food expenditure example, and the more difficult it is to accurately ∑ estimate β1 . The term xi2 measures the squared distance of the data from the origin, x) = 0. ( ∑ If the values of x are near zero, then xi2 will be small, and this will reduce var b1 |𝐱 . But ∑ if the values (of x) are large in magnitude, either positive or negative, the term xi2 will be large and var b1 will be larger, other things being equal. ( ) 5. The sample mean of the x-values appears in cov b1 , b2 |𝐱 . The absolute magnitude of the covariance increases with an increase in magnitude of the sample mean x, and the covariance has a sign opposite to that of x. The reasoning here can be seen from Figure 2.11. In panel (b) k k k 72 CHAPTER 2 The Simple Linear Regression Model the least squares fitted line must pass through the point of the means. Given a fitted line through the data, imagine the effect of increasing the estimated slope b2 . Since the line must pass through the point of the means, the effect must be to lower the point where the line hits the vertical axis, implying a reduced intercept estimate b1 . Thus, when the sample mean is positive, as shown in Figure 2.11, there is a negative covariance between the least squares estimators of the slope and intercept. 2.5 The Gauss–Markov Theorem What can we say about the least squares estimators b1 and b2 so far? • The estimators are perfectly general. Formulas (2.7) and (2.8) can be used to estimate the unknown parameters β1 and β2 in the simple linear regression model, no matter what the data turn out to be. Consequently, viewed in this way, the least squares estimators b1 and b2 are random variables. • The least squares estimators are linear estimators, as defined in (2.10). Both b1 and b2 can be written as weighted averages of the yi values. • If assumptions SR1–SR5 ( ) hold, then the ( least ) squares estimators are conditionally unbiased. This means that E b1 |𝐱 = β1 and E b2 |𝐱 = β2 . • Given x we have expressions for the variances of b1 and b2 and their covariance. Furthermore, we have argued that for any unbiased estimator, having a smaller variance is better, as this implies we have a higher chance of obtaining an estimate close to the true parameter value. k Now we will state and discuss the famous Gauss–Markov theorem, which is proven in Appendix 2F. Gauss–Markov Theorem: Given x and under the assumptions SR1–SR5 of the linear regression model, the estimators b1 and b2 have the smallest variance of all linear and unbiased estimators of β1 and β2 . They are the best linear unbiased estimators (BLUE) of β1 and β2 . Let us clarify what the Gauss–Markov theorem does, and does not, say. 1. The estimators b1 and b2 are “best” when compared to similar estimators, those that are linear and unbiased. The theorem does not say that b1 and b2 are the best of all possible estimators. 2. The estimators b1 and b2 are best within their class because they have the minimum variance. When comparing two linear and unbiased estimators, we always want to use the one with the smaller variance, since that estimation rule gives us the higher probability of obtaining an estimate that is close to the true parameter value. 3. In order for the Gauss–Markov theorem to hold, assumptions SR1–SR5 must be true. If any of these assumptions are not true, then b1 and b2 are not the best linear unbiased estimators of β1 and β2 . 4. The Gauss–Markov theorem does not depend on the assumption of normality (assumption SR6). 5. In the simple linear regression model, if we want to use a linear and unbiased estimator, then we have to do no more searching. The estimators b1 and b2 are the ones to use. This explains k k