Estimating fully observed recursive mixed-process models with cmp David Roodman Probit model: Link function (g) induces likelihoods for each possible outcome fε(y*–x'β) Area: 1–Fε(0–x'β) xi'β y=g(y*)=1{y*>0} y=1 0 y=0 Area: Fε(0–x'β) y* (latent) y (observed) Relabeling left graph for ε scale: “error link” function (h) induces likelihoods for each possible outcome fε(ε) Area: 0 –xi'β y=h(ε)=1{ε>–xi'β} y=1 y=0 (h(ε)=g(x'β +ε)) Area: ε (latent) y (observed) Just change g() to get new models Tobit (censored) Ordered probit With generalization, embraces multinomial and rank-ordered probit, truncated regression… Compute likelihood same way Given yi, determine feasible value(s) for ε – If just one, Li = normal density at that point – If a range, Li = cumulative density over range For models that censor some observations (Tobit), L=Π Li combines cumulative and point densities. Amemiya (1973): maximizing L is consistent Multiple equations (SUR) For each obs, likelihood reached as before Given y, determine feasible set for ε and integrate normal density over it Feasible set can be point, ray, square, half plane… Cartesian product of points, line segments, rays, lines. Bivariate probit Suppose for obs i, yi1= yi2=0 Feasible range for ε is: Integral of fε(ε)=φ(ε;Σ) over this: Can use built-in binormal(). Similar for y=(0,1)′, (1,0)′, (1,1)′. Mixed uncensored-probit Suppose for obs i, we observe some y=(yi1, 0)′ Feasible range for ε is a ray: Integral of fε(ε)=φ(ε;Σ) over this: Integral of 2-D normal distribution over a ray. Hard with built-in functions Requires additional math Conditional modeling—“c” in cmp Model can vary by observation—depend on data – Worker retraining evaluation • Model employment for all subjects • Model program uptake only for those in cities where offered – Classical Heckman selection modeling • • • • Model selection (probit) for every observation Model outcome (linear) for complete observations Likelihood for incomplete obs is one-equation probit Likelihood for complete obs is that on previous slide – Myriad possibilities Recursive systems y’s can appear on RHS in each other’s equations Matrix of y coefficients must be upper triangular I.e.: System must have clearly defined stages. E.g.: – SUR (several equations, one stage) – 2SLS If system is fully modeled and truly recursive, then estimation is FIML If system has simultaneity and the early equation stages instrument, then LIML If system is Fact Recursive Fully observed (y’s appear in RHS but never y*’s) then likelihoods developed for SUR still work Can treat y’s in RHS just like x’s sureg and biprobit can be IV estimators! Rarely understood, not proved in general in literature Greene (1998): “surprisingly”…“seem not to be widely known” Wooldridge (e-mail 2009): “I came to this realization somewhat late, although I’ve known it for a couple of years now.” I prove, perhaps not rigorously Maybe too simple for great econometricians to bother publishing General recursive, fully observed system cmp can fit: conditional recursive mixed-process systems Processes: Linear, probit, tobit, ordered probit, multinomial probit, interval regression, truncated regression Can emulate: Built-in: probit, ivprobit , treatreg , biprobit, oprobit, mprobit, asmprobit, tobit, ivtobit, cnreg, intreg, truncreg, heckman, heckprob User-written: triprobit, mvprobit, bitobit, mvtobit, oheckman, (partly) bioprobit Required. One exp for each equation. Tell cmp model type for each eq and can vary by observation Emulation examples 200 Heteroskedasticity can make censored models not just inefficient but inconsistent -50 0 50 y 100 150 Tobit example: error variance rises with x -20 0 20 40 60 x y* True model Censored values Fitted model 80 Implementation innovation: ghk2() Mata implementation of Geweke-Hajivassiliou-Keane algorithm for estimating cumulative normal densities above dimension 2. Differs from built-in ghkfast(): Accepts lower as well as upper bounds E.g., integrate over cube [a1,b1]× [a2,b2]× [a3,b3] (otherwise requires 23 calls instead of 1) Optimized for many observations & few simulation draws/observation Does not “pivot” coordinates. Pivoting can improve precision, but creates discontinuities when draws are few. (ghkfast() now lets you turn off pivoting.) Implementation innovation: “lfd1” In Stata ML, using an lf likelihood evaluator assumes that (A1) for each eq, ml computes numerically with 2 calls per eq, then analytically. And for Hessian, # of calls is quadratic in # of eq Using a d1 evaluator, ml does not assume A1. But does (A2) require evaluator to provide scores For Hessian, # of calls in linear in # of parameters Two unrelated changes create unnecessary trade-off ml is missing an “lfd1” type that assumes A1 and A2— would make Hessian with # of calls linear in # of eq. Solution: pseudo-d2. d2 routine efficiently takes over (numerical) computation of Hessian Good for score-computing evaluators for which Possible extensions Marginal effects that reflect interactions between equations (Multi-level) random effects Dropping full observability—y*’s on right Rank-ordered multinomial probit References Roodman, David. 2009. Estimating fully observed recursive mixed-process models with cmp. Working Paper 168. Washington, DC: Center for Global Development. Roodman, David, and Jonathan Morduch. 2009. The Impact of Microcredit on the Poor in Bangladesh: Revisiting the Evidence. Working Paper 174. Washington, DC: Center for Global Development.