Uploaded by rimbaud2019

J Public Adm Res Theory-2013-Zhu-395-428

JPART 23:395–428
Panel Data Analysis in Public
Administration: Substantive and
Statistical Considerations
Ling Zhu
University of Houston
Panel data analysis has become a popular tool for researchers in public policy and public
administration. Combining information from both spatial and temporal dimensions, panel
data allow researchers to use repeated observations of the same units (e.g., government
agencies, public organizations, public managers, etc.), and could increase both quantity
and quality of the empirical information. Nonetheless, practices of choosing different
panel model specifications are not always guided by substantive considerations. Using a
state-level panel data set related to public health administration as an example, I compare
four categories of panel model specifications: (1) the fixed effects model, (2) the random effects model, (3) the random coefficients (heterogeneous parameter) model, and
(4) ­linear dynamic models. I provide an overview of the substantive consideration relevant
to different statistical specifications. Furthermore, I compare estimation results and discuss
how these different model choices may lead to different substantive interpretations. Based
on model comparisons, I demonstrate several potential problems of different panel models. I conclude with a discussion on how to choose among different models based on
substantive and theoretical considerations.
Introduction
Empirical research in public administration has been enriched by the availability of
panel data (Eom, Ho, Lee, and Xu 2008).1 Longitudinal data, tracking a particular
sample of spatial units, increase both quantity and quality of the empirical information and have several advantages over cross-section or time-series data (Hsiao 2003).
An earlier version of this article was presented at the 2012 Fall Research Conference of the Association for
Public Policy Analysis and Management. I thank the panel participants for their feedback on the earlier draft.
I thank Solé Prillaman and Andrea Eckelman for research assistance. I am indebted to Kenneth J. Meier,
Andrew B. Whitford, and the anonymous reviewers for their comments and suggestions. All errors remain
mine. Address correspondence to the author at lzhu4@central.uh.edu.
1 Panel
data refer to data combining information of both spatial and temporal dimensions, which allow
researchers to use repeated observations of the same units over time (Baltagi 2008; Hsiao 2003). It is also
referred as cross-section-time-series (CSTS) data (Beck and Katz 1995), time-series-cross-section (TSCS) data,
or “pooled” data.
doi:10.1093/jopart/mus064
Advance Access publication December 10, 2012
© The Author 2012. Published by Oxford University Press on behalf of the Journal of Public Administration Research
and Theory, Inc. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Abstract
396
Journal of Public Administration Research and Theory
2 For
example, Cornwell and Kellough (1994) use panel data on federal civil service agencies to examine
what account for workforce diversity. Boyne, James, John, and Petrovsky (2010) use a panel data design to
study organizational performance and turnover in 148 English local governments in 4 years. Soss, Fording, and
Schram (2011) use a panel data set of 24 Florida Workforce Board regions across 30 months to test the effect
of the Florida Welfare Transition program on decisions to sanction clients. Meier and O’Toole have conducted
a series studies on public management and organizational performance using panel designs pooling data of
more than 1,000 Texas school districts across multiple years (Meier and O’Toole 2002, 2003, 2008).
3 A new development in the panel data literature is to estimate panel data models in a Bayesian context
(Gelman and Hill 2007; Pang 2010). The overview on panel data models in this article, however, primarily
focuses on the non-Bayesian context.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Stimson (1985, 916) asserts “many of the possible threats to valid inferences are specific to either cross-sectional or time-series design, and many of them can be jointly
controlled by incorporating both space and time into the analysis.”
Students of public policy and public administration particularly benefit from
panel data analysis.2 Policy analysts often encounter data limitations when assessing
the impact of a policy, and only limited annual policy data are available (Stimson
1985). Panel data can increase the number of observations by pooling different timeseries together. With greater degrees of freedom, researchers are able to investigate
complex empirical associations among variables such as the interaction between
space and time. Public administration scholars often study governments, large organizations, social and institutional contexts, which do not vary much within a single
time-series. Panel data could outperform single-dimension data by providing the ability to deal with rarely changing variables and unobserved heterogeneity across units
(Allison 2009; Plümper and Troeger 2007).
The methodological sophistication of panel data analysis, moreover, has evolved
over the past two decades. The early work relies on ordinary least square (OLS) techniques (i.e., completely pooling) (Mundlak 1978; Stimson 1985; Taylor 1980). The
next generation of methodology focuses on the analysis of covariates associated
with spatial and temporal dimensions and statistical theories that are employed to
justify error assumptions (Arellano and Honore 1999; Wooldridge 2002). Various
dynamic and spatial econometric models are developed to handle complex panel data
structures (Baltagi 2008). A recent development of this methodology enterprise is
the effort of linking statistical models to substantive and theoretical considerations
(Granato and Scioli 2004; Plümper, Troeger, and Manow 2005; Primo, Jacobsmeier,
and Milyo 2007; Whitten and Williams 2011).3
Despite the increasing availability of data and statistical models, panel data analysis still remains to be an “art” because of several enduring issues. Firstly, there has
been a long-standing debate on the utility of the fixed effects (FE) model (Baltagi,
Bresson, and Pirotte 2003; Greene 2011a; Nickell 1981). Secondly, although pooling could increase the statistical power of an empirical data set, it could also make a
statistical model vulnerable to problems caused by both cross-sectional and temporal dimensions. Adjusting for correlated errors along both the spatial and temporal
dimensions could be quite challenging. Thirdly, when theory is weak (i.e., researchers
do not have a strong and explicit theory to guide the choice of explanatory variables), estimating a consistent and efficient panel model could be difficult. Beck (2011)
points out that panel data always suffer from potential omitted variable bias, thus
leaving a room for competing model choices and different specification errors. As a
Zhu
Panel Data Analysis in Public Administration
Panel Data Models: An Overview
In this and the subsequent section, panel models are illustrated with a data set of
public health administration. In this empirical example, the cross-sectional units are
50 American states. Annual data are available for all 50 states from 1990 to 2006.
The data set is used to investigate the characteristics of state health care systems
and their effects on citizens’ access to health insurance. Table 1 reports the descriptive statistics of all variables in this panel data set. Three institutional variables are
included in the data set: the public ownership of state health care systems, the public
source of financing state Medicaid programs, and the generosity of state Medicaid
eligibility rules. Although this empirical example is based on data related to public
health administration, it resembles a few common characteristics of a panel data set
in public administration.
Firstly, it is quite common to see a panel application with more cross-sectional
units than time units in public administration and public management.4 Secondly,
although data variation exists across spatial and temporal dimensions, some key theoretical variables (in the empirical example, institutional variables) do not vary substantially across time. For example, from 1990 to 2006, Delaware, New Hampshire,
Rhode Island, and Vermont did not experience any institutional change in the
ownership of their health care systems (within variation equals 0). States such as
Arizona, Maine, and New Jersey experienced incremental changes in the public ownership (within-state mean <10% public hospitals, standard deviation <2). Overall,
states do not change their Medicaid eligibility rules dramatically. Only a few states
experienced substantial policy changes in their Medicaid income eligibility rules,
such as Arizona (from 107% in 2003 to 200% after 2003) and New Jersey (from 100%
in 2004 to 115% in 2005). Thirdly, different theories and empirical research show
that many factors, combined, affect the dependent variable (state-level uninsured
rates). Markowitze, Gold, and Rice (1991) contend that people’s employment status
plays a direct role in determining their access to health insurance. They also find that
4 There
are many other panel data applications that have similar data structures; for example, a panel data
set of large-N organizations across multiple years, longitudinal data for a large number of public managers,
and so on.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
result, the “art” of panel data analysis hinges on conjectures about the theoretical
relationship, specific data structures, and particular assumptions about the source of
bias. Knowing the pros and cons of different types of panel analytic models, therefore, becomes essential.
This paper provides a synopsis of commonly used panel data models, reviews
advantages and disadvantages of each panel method, and illustrates the analytic
applications with a panel data set of public health administration across 50 states
and 17 years. After comparing how different substantive conclusions could be drawn
from different model applications, a brief discussion on extensions pertaining discrete
variables is presented. Overall, this article makes the argument that statistical models
are essentially tools for theory building. Researchers need to link their model choices
to substantive considerations, not simply to statistical assumptions.
397
398
Journal of Public Administration Research and Theory
Table 1
Descriptive Statistics of the Panel Data Set: 50 States from 1990 to 2006
Variable
Mean
SD
15.56
4.46
24.21
18.64
Public finance
15.09
3.88
Eligibility
87.70
56.36
5.15
12.76
1.39
3.57
47.99
25.38
10.28
7.25
12.61
9.46
8.67
1.94
% Black population
% Latino population
% Population older than 65 years
18.53
4.25
3.99
1.37
% Population whose body mass index scores are
above 30
% Population who reported themselves as being
in poor health
Dependent variable
Uninsured
Institutional variables
Public ownership
Demographic variables
Black population
Latino population
Aged population
Health risk factors
Obesity rate
Perceived poor health
% Population without health insurance
% Community hospitals owned by
government
State Medicaid spending as the percentage of
total personal health care spending
State Medicaid income eligibility limits for
working adults, as a percentage of the federal
poverty level (FPL)
State unemployment rate
State poverty rate
Berry, Ringquist, Fording, and Hanson (1998)’s
measure of state liberalism
Notes: Data for the dependent variable are aggregated from the Current Population Survey (CPS)’s Annual Social and Economic
Supplement, the American Community Survey, and the Survey of Income and Program Participation. Data are archived in the
Census Bureau historical tables, Health Insurance Coverage Status and Type of Coverage by State, Persons Under 65. Data for
Public Ownership are drawn from the American Hospital Association (AHA)’s annual data on the number of community hospitals by type. Data for Medicaid spending are drawn from the US Department of Health and Human Services, the Center for
Medicaid and Medicare Services Expenditure Reports. Data for state Medicaid eligibility rules are drawn from the Kaiser Family
Foundation’s policy report on state Medicaid eligibility rules (Heberlein, Brooks, Guyer, Artiga, and Stephens 2011). Data for
the economic variables are drawn from the Bureau of Labor Statistics (BLS) annual estimates on state unemployment, and US
Census Bureau, Income, Poverty, and Health Insurance Coverage in the United States, Current Population Reports. Data for the
state government ideology are based on Berry, Ringquist, Fording, and Hanson (1998) and accessed from http://www.bama.
ua.edu/~rcfording/stateideology.html. Data for the demographic variables and health risk factors are drawn from the Centers for
Diseases Control and Prevention (CDC) WONDER Current Population Information, and Behavior Risk Factor Surveillance
System (BRFSS).
individual characteristics, such as race, ethnicity, and health risks, affect access to
health insurance. Nelson, Bolen, Wells, Smith, and Bland (2004) find evidence that
states’ fiscal responsibilities in health care affect citizens’ insurance take-up. Hacker
(2004) contends that the structure of welfare policies and the characteristics of the
American hybrid health care system are important determinants of access to care.
In sum, various factors need to be considered as substantively important variables
when analyzing state-level uninsured rates, including economic variables, the institutional characteristics of a state health care system, demographic variables, and
health risks.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Economic variables
Unemployment
Poverty
Political variable
State liberalism
Measurement
Zhu
Panel Data Analysis in Public Administration
399
The generic panel model equation can be written as:
Yi,t = α + Xi,t β + ei,t(1)
The Unobserved Heterogeneity
One important assumption of the population average model is that all units are
homogenous. After all, researchers pool different spatial units across time in seeking
a commonality of these units. As for the empirical example, equation (1) implies unit
homogeneity across all state–year observations. This assumption is often applied to
work done in scientific laboratories, but rarely holds in nonrandomized observational
studies (Holland 1986; Rosenbaum 2005; Wilson and Bulter 2007). Unobserved unit
heterogeneity may bias statistical estimation and lead to invalid causal inferences.
Rosenbaum (2005, 148) contends “observational studies vary markedly in sensitivity
to unobserved biases.”
The Fixed Effects Model
The FE model deals with unobserved heterogeneity by using unit-specific intercepts
(Greene 2011b). Equation (2) is used to denote the unit-specific fixed effects, and fi is
defined based on a vector of spatial units, zi.
Yi,t = α + Xi,t β + fi + ei,t
f i = z iγ
(2)
The FE model can be estimated based on a full set of unit dummy variables or by
mean-centering the dependent variable and all the explanatory variables to “clear”
5 Essentially,
consistency and efficiency are two desirable properties of a good estimator. Here, I mainly
focus on threats to consistency. Various panel data techniques related to improving estimation efficiency will be
discussed in the subsequent section.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
In equation (1), i denotes the spatial dimension (i.e., states), t denotes time (i.e., year),
α is the intercept, β refers to the vector of slope coefficients, and Xi,t refers to the vector of regressors. This generic model equation refers to the application of complete
pooling. In other words, we do not speculate that state-specific effects or time-specific
effects exist with regard to how institutional characteristics of state health care systems, economic factors, and demographics affect state-level uninsured rates (Gelman
and Hill 2007). This is also referred as the population average model (Baltagi 2008).
In essence, the parameter of interest in a population average model is the mean
effect across time (within) and spatial (between) units. Although complete pooling
produces a simple and generic model, it could lead to biased estimation due to two
major threats: unobserved heterogeneity across spatial units (i.e., states) and temporal dependency across time units.5 There are various types of panel models developed
to handle one or both issues. Each panel model has its pros and cons and is linked to
different substantive relationships.
400
Journal of Public Administration Research and Theory
The Random Effects Model
Different from the FE model, which conceptualizes each state as having its own
baseline, the random effects (RE) model assumes the intercept to be some random
deviation from the underlying mean intercept. Greene (2011b, 347) points out that if
“the unobserved individual heterogeneity, however formulated, can be assumed to be
uncorrelated with included variables, then the model may be formulated as”:
Yi,t = α +Xi,t β + ui + ei,t(3)
According to equation (3), the RE model specifies that ui is a state-specific random
element, such that “there is but a single draw that enters the regression identically in
each period” (Greene 2011b, 347). The RE model has several desirable properties.
One is that time-invariant variables can be included in the panel model (Wilson and
Bulter 2007). Furthermore, if we have a large-N panel data set and the random effect
is uncorrelated with the regressors, the RE effects model is more efficient than the
FE model. The estimation efficiency, in addition, increases as the number of crosssection units (N) increases. As for the state-level panel data example, the model based
on equation (3) will have more degrees of freedom than the model based on equation
(2), because equation (3) does not estimate state-specific intercepts. However, if the
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
cross-unit heterogeneity (Allison 2009). Both the least square dummy variable
(LSDV) approach and the mean-centering approach account for within-variance and
eliminate between-variance. In other words, the FE model controls for cross-unit heterogeneity that is not captured by the conditional mean, Xi,t β.
When cross-unit heterogeneity (fi ) is correlated with the regressors (Xi,t), failure to account for the heterogeneity could lead to biased estimation of β. In such a
situation, the FE model is better than the OLS specification because it improves the
estimation consistency. The FE model, however, has a few limitations. The LSDV
specification includes a large number of dummy variables to account for each unit
as a specific source of unobserved heterogeneity. It may become problematic when
T is relatively small. Both LSDV approach and mean-centering approach have been
accused of absorbing a lot of cross-sectional variance. This can be a serious problem
if the panel data set contains much greater cross-sectional variation than cross-time
variation (Kennedy 1998). The LSDV approach is also deemed to be a crude method
to model unmeasured heterogeneity. In other words, unit dummy variables can only
tell researchers that there is cross-unit heterogeneity but cannot show what factors
explicitly cause the heterogeneity.
Considering the aforementioned empirical example, if the substantive interest is
to investigate how the institutional arrangements of state health care systems affect
state-level uninsured rates, it is reasonable to expect that these institutional variables
may capture some cross-state heterogeneity. Estimating the panel model by including a full set of state dummy variables may throw away cross-state variations that
are captured by the institutional variables and lead to artificial null findings on the
institutional variables.
Zhu
Panel Data Analysis in Public Administration
The Random Coefficients Model
Comparing equations (2) and (3), we can see that the key difference between the FE
model and the RE model is substantive. Equations (2) and (3) conceptualize different sources of cross-unit heterogeneity. Equation (2) implies part of the state-specific
effects is not captured by the vector of regressors and need to be controlled (the
covariates are completely correlated with unobserved state-specific effects). Equation
(3) implies that the substantive variables (i.e., institutional, economic, demographic,
and health risk variables) may well capture most state-specific effects, and thus including state dummy variables is unnecessary (the covariates are not correlated with the
unobserved state-specific effects). In equation (3), ui is theorized as a unit-specific random intercept. Greene (2011b, 347) contends that with a sufficiently rich data set, one
can also estimate models including both random intercepts and random coefficients,
expressed as equation (4):
Yi,t = α + Xi,t (β + hi) + ui + ei,t(4)
Based on equation (4), both the intercept and the slopes vary by states, i. This type of
random coefficients model (RCM) can also be estimated based on time, t. Beck and
Katz (2007) argue that panel data analysis essentially comes down to the question
about how much to pool observations across time and space. With a series of Monte
Carlo experiments, they show that RCM estimators obtained via the maximum-likelihood method outperform pooled OLS, unit-by-unit OLS (FE) with finite sample
panel data.6
6 The
aforementioned FE, RE, and RCM models all recognize the multi-level structure of the dataset.
Although I discuss these model considerations based on the nonnested state-level CSTS data example, similar
considerations can apply to hierarchical linear models that capture more complex multi-level structures. For
example, if a panel dataset contains both individual-level, organization-level, and state-level units (e.g., public
managers from different state agencies in different years), one may consider a three-level nested parameter
structure for fixed effects, random effects, or random coefficients [see Gelman and Hill (2007) for more detailed
discussion on complex/nested multi-level specifications].
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
random intercept ui is correlated with Xi,t, the RE model could produce inconsistent
estimation. Hence, it is always critical to consider whether the random effects (ui) are
correlated with substantively important variables in the model.
In addition, the Hausman model specification test can be used to evaluate the
consistency and efficiency of the fixed effects and the random effects specification
(Hausman 1978). Large χ2 statistics produced by the Hausman test suggest that the
RE model is inconsistent and one may not consider the RE model as an appropriate
model choice. When both fixed effects and random effects specifications are consistent, model-fit statistics can be used to choose between the two different specifications.
For example, both mean squared error (MSE) and root mean squared error (RMSE)
reflect estimation bias, and the model associated with smaller MSE/RMSE statistics
would be preferred (James and Stein 1961).
401
402
Journal of Public Administration Research and Theory
Time Dependence and Dynamic Panel Models
The second major threat to consistent panel estimation is the issue of temporal
dependence. Most public administration data feature first-order autoregressive processes, whereby the current status of organizations, governments, or policy networks
is a function of their own past (De Boef 2001). In other words, administrative processes are inertial and organizational memory persists. Formally, a first-order autoregressive process can be written as:7
Yi,t = ρYi,t−1 + ei,t(5)
7 Statistically,
we can also define a higher-order autoregressive process by changing the notation of Yi,t
into Yi,t−2,3,4 … k. A higher-order autoregressive process may occur if an administrative process is affected by
cyclical factors, such as budgetary and election cycles. The first-order autoregressive process, however, is most
commonly seen in public administration. In addition, if a panel data set only has very short T (for instance,
pooling hundreds of local government units in 4 or 5 years), estimating a higher-order autoregressive process
would be unfeasible. Box and Jenkins (1970) lay out the theoretical foundation of modeling an autoregressive
process. De Boef (2001) and De Boef and Granato (1997) provide nice discussion on how various political/
policy processes can be conceptualized by an autoregressive model.
8 With a CSTS setup, one can examine the common (mean) dependence across all included series (ρ) or
panel-specific dependence (ρi).
9 The Levin–Lin–Chu panel unit root test assumes that all units share the same AR(1) process but allows
specifying the different unit effects, time dynamics, and time trends. An alternative test to the aforementioned
tests is a multivariate augmented Dickey–Fuller test using the seemingly unrelated regression (SUR) estimator.
This test, however, requires that T must be larger than N (Taylor and Sarno 1998).
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
In equation (5), ρ characterizes the feature of organizational memory.8 An administrative process carries permanent memory if ρ equals to 1 or −1 (i.e., the presence of
panel unit root).
If ρ ϵ (−1, 1), a process is stationary and the organizational memory is not permanent (De Boef 2001). According to De Boef (2001, 81), a stationary process is
strongly autoregressive when | ρ | is near 1 (formally ρ = 1 + ϵ, and ϵ is a small
negative fraction), and a process mainly carries short-term memory when | ρ | is near
0. When pooling different states, we assume that all these administrative processes are
bounded by a common set of relationships, which are constant over time (Pesaran
and Smith 1995). Dynamic data could violate this time-consistent assumption and
thus can lead to spurious regressions and biased coefficient estimation (Beck 2001;
Dickey and Fuller 1981; Im, Pesaran, and Shin 2003; Maddala and Wu 1999).
The first consideration of estimating panel dynamics is to examine the dynamic
nature of the dependent variable. Various statistical tests can be used to evaluate the
stationarity of the dependent variable. For example, both the augmented Dickey–
Fuller unit root approach and the Phillips–Perron approach can be applied to a panel
data set. According to Maddala and Wu (1999), both approaches assume that all
series are not stationary against the alternative that at least one series in the panel is
stationary. If a panel data set is balanced, one can also use the Im–Pesaran–Shin (Im,
Pesaran, and Shin 2003) test or the Levin–Lin–Chu panel unit root test (Levin, Lin,
and Chu 2002).9
Once it has been determined if the dependent variable is panel stationary, two
substantive questions need to be further considered. The first substantive question
Zhu
Panel Data Analysis in Public Administration
403
is how one could conceptualize temporal dynamics and model persistence (Wawro
2002). In the state health care example, the substantive question is whether the impact
of health care institutions persists through time. If so, how can the long-term and
short-term impacts be reflected by statistical models? Permanent memory, long-term
equilibrium relationships, and short-term dynamics represent substantively different
administrative/organizational processes and thus require different statistical specifications. The second question is whether to model data dynamics as a nuisance (ei,t),
or to incorporate the temporal dynamics in the conditional mean (Xi,t β). Dynamic
specifications differ based on different substantive considerations and different data
structures of temporal dependence.
When the dependent variable is panel stationary (| ρ | ≠ 1), one can specify an autoregressive distributed lag (ADL) model either by including a lagged dependent variable
(LDV) or by estimating a static model with serially correlated error terms (e.g., an
AR(1) error specification).10 In public administration, one common practice of modeling stationary panel data is to include a LDV as a control for organizational memory.
This dynamic specification is usually deemed partial adjustment of behavior over
time (Wawro 2002). The inclusion of a LDV is also justified by the substantive consideration that large organizations, institutional contexts, and government agencies
are normally inertial. The LDV approach is a parsimonious way to capture both the
long-term and short-term effects of regressors (Beck and Katz 1996; Wawro 2002).
An additional attraction of including lagged terms of the dependent variable
is that they are usually easy to estimate and interpret. In the state health care example, if the dynamic model is specified as equation (6), the autoregressive parameter,
ρ, captures how the long-term effects of health care institution variables (Xi,t) are
distributed over time. βk, the vector of coefficients corresponding to the health care
institutional variables (Xk), captures the short-term effects. One can also calculate the
total long-term effect based on the approximation, [βk / (1 − ρ)]. The size of βk and
ρ determines the magnitude of the short-term effect and the persistence of the longterm effect. When ρ is large, the short-term effect, βk, is slowly discounted over time
and the total long-term effect cumulates over a long time period. Conversely, if ρ is
small, the short-term effect will be substantially discounted in the future and the total
long-term effect will be relatively small.
Yi,t = ρYi,t−1 + Xi,t β + ei,t(6)
The inclusion of a LDV, nevertheless, has been accused of absorbing covariates associated with other regressors and could potentially wash out the explanatory power
of other exogenous regressors (Achen 2000; Keele and Kelly 2006). Substantively,
the LDV approach also makes a strict underlying assumption that the effects of all
state health care variables dissipate exponentially over the long run. Combining a
10 Beck
(1991) argues that when data are stationary, the LDV specification is not fundamentally different
from a static model with an AR(1) specification. Both specifications belong to the family of ADL model.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Stationary Panel Data: Autoregressive Distributed Lag Models
404
Journal of Public Administration Research and Theory
Dealing with Nonstationary Panel Data
If panel unit root presents, estimating a LDV model can lead to biased results.
There are alternative dynamic panel methods to the LDV approach, based on different substantive considerations of short-term and long-term relationships. The
class of error correction models (ECM) was firstly proposed to deal with integrated
and near-integrated dynamic data (De Boef and Granato 1997; Engle and Granger
1987; Westerlund 2007) and later applied to stationary autoregressive processes (De
Boef 2001; De Boef and Keele 2008). Despite the varying forms, the error correction
method deals with integrated or near-integrated dynamic data by differencing the
dependent variable (i.e., Δyi,t) and including changes and lagged values of explanatory
variables and lagged values of the dependent variable as the regressors.11 Following
De Boef (2001, 84), a generalized ECM model can be written as:
Δyi,t = α + β1Δxi,t − γ (yi,t−1 − xi,t−1) + β2 xi,t−1 + ei,t(7)
Using the state health care panel example, Δyi,t (the annual change in state uninsured
rates) is affected by Δxi,t (the annual change in government Medicaid spending) and
the long-run relationship between yi,t (uninsured rates) and xi,t (Medicaid spending).12
11 See
De Boef (2001) for the comparison between the Engle and Granger two-step ECM and the
generalized one-step ECM. De Boef and Keele (2008) provide a thorough overview of different variants of the
generalized one-step ECM. De Boef and Granato (1997) contend that integrated and near-integrated processes
have very similar statistical properties in finite samples. Hence, ECM is deemed a common statistical method to
characterize long-run relationships represented in both integrated and near-integrated processes.
12 For simplicity, I only included one explanatory variable in equation (7). To illustrate the substantive
relationships, I use xi,t to denote the Public Finance variable (i.e., state Medicaid spending). Similar lag and
difference terms may apply to other explanatory variables if the dynamic relationships are the same.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
LDV with fixed effects in panel data analysis, futhermore, could potentially lead to
biased estimation in some empirical contexts. Wawro (2002) points out that combining a LDV with fixed effects may produce finite sample autoregressive bias to the
coefficient of the LDV (i.e., the “Nickell bias” that attenuates the coefficient of the
LDV) when T is very small (Nickell 1981; Phillips and Sul 2007; Wawro 2002). Based
on their recent Monte Carlo experiments, Keele and Kelly (2006) report that the performance of a LDV model is still desirable if the dependent variable is very dynamic
and if serial autocorrelation in the error term is very small. If these two conditions
are met, the bias caused by LVD diminishes as T increases.
In the state health care example, the panel data set includes data for 17 years.
It grants some degree of freedom for estimating a simple dynamic relationship with
a LDV. The data may also permit a more complex dynamic specification than the
LDV specification. If T is very short, estimating complex dynamic relationships would
become unfeasible. Substantive relationships between state health care variables and
the uninsured rates, furthermore, will be different in a static model with a serially correlated error term. Such a model specification only captures the mean association
between the level of uninsured rates and the level of a particular state health care
variable.
Zhu
Panel Data Analysis in Public Administration
405
In particular, the long-run relationship between the level of uninsured rates and state
Medicaid spending is reflected by −γ (yi,t−1 − xi,t−1) + β2 xi,t−1, whereby γ represents
how disequilibrium between uninsured rates and spending is “corrected” over long
run. Approximately, the long-run effects of spending can be calculated based on
1 − (β2 / γ) (Alogoskoufis and Smith 1990; Campbell and Shiller 1988).
Alternatively, one may deal with panel unit root by differencing both the dependent variable and the key explanatory variables. Such a first-difference model is often
seen as a particular case in the ECM family, whereby the long-run relationship is
restricted to be 0 (Beck 1991; De Boef and Keele 2008). The first-difference model is
easy to estimate. Empirically, it may not perform well because it does not account for
the long-run relationship in the process (Beck 1991).
Most aforementioned dynamic specifications add a lagged term of the dependent
variable as a regressor. It violates the strict exogenous assumption of OLS due to
the inclusion of an endogenous lagged term. The issue is further complicated if the
dynamic specification includes error terms (such as panel-specific fixed effects), which
are correlated with the endogenous lagged term (Anderson and Hsiao 1981; Wawro
2002). Various instrument variable (IV) approaches have been proposed to deal with
this additional complication. Anderson and Hsiao (1981) proposed using a secondorder change measure of the dependent variable (Δyi,t − 2) or a second-order LDV
(yi,t − 2) as instruments to remove the correlation between panel-specific effects and the
endogenous lagged term. Arellano and Bond (1991) modified the Anderson–Hsiao
estimator and proposed the generalized method of movements (GMM). The GMM
dynamic panel estimators have become increasingly popular because of a few nice
properties. It is designed to be suitable for panel data with small T and large N. It
is flexible to deal with regressors that are not strictly exogenous by instrumenting
endogenous regressors with their own lags. The Arrellano–Bond method, moreover,
is flexible to model dynamics associated with both levels and changes. The major
drawbacks of the GMM method include: (1) researchers need to predetermine which
regressors are not strictly exogenous; and (2) it is usually an arbitrary decision on how
many instrument variables need to be included. Another caveat is that although the
GMM approach can handle endogenous variables in a flexible way, the IV procedure
is not a panacea to all the endogeneity issues.
Estimation Efficiency: Corrections for Nonspherical Errors
Aside from the aforementioned considerations on consistency, researchers also need
to consider the efficiency of a particular estimator that they choose. Several issues may
rise and affect estimation efficiency. Public administration scholars often model panel
data with large N and small T, whereby some key theoretical variables do not change
in a short time period. In the aforementioned empirical example, two institutional
variables tend to be very stable over time within each state: state Medicaid eligibility
rules and the public ownership of state health care systems. In this situation, if the
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Additional Considerations: Endogenous Lags and the Instrument Variable Approach
406
Journal of Public Administration Research and Theory
13 There
has been a debate, however, on the quality of standard errors produced by the FEVD method (Beck
2011). Scholars who are skeptical to the FEVD method contend that it often produces small standard errors,
leads to overconfidence in inferences, and thus caution needs to be warranted when using this method (Breusch,
Ward, Nguyen, and Kompas 2011; Greene 2011a).
14 In this section, I only provide a brief overview on available FGLS (feasible generalized least squares)
methods for improving efficiency. The relevant statistical tests for detecting different nonspherical errors will be
illustrated along various empirical models in the next section.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
within variation is substantively important, estimating a FE model could lose a substantial amount of efficiency, because of the rarely changing variables. Plümper and
Troeger (2007) modified the conventional FE model and developed the fixed effects
vector decomposition (FEVD) method to deal with time-invariant (or rarely varying)
variables. Despite the longish debate on the quality of standard errors produced by
the FEVD method (Beck 2011), Plümper and Troeger (2007)’s method could be useful to pooled data analysis if key theoretical variables are time-invariant and estimating within effects is substantively critical.13
Another potential challenge to efficiency is the heterogeneity in the error term, ei,t.
In theory, there are three sources of nonspherical errors: (1) heteroskedasticity across
units (or across panels), (2) contemporaneous correlation across panels, and (3) serial
autocorrelation. Various methods are available to detect and correct nonspherical errors.
If unit-specific heteroskedasticity occurs, one can estimate the White standard errors to
improve estimation efficiency. If heteroskedasticity only occurs across states, not within
states, clustered standard errors can be estimated to improve estimation efficiency. Beck
and Katz (1995, 1996) proposed to use the panel-corrected standard errors (PCSE) to
deal with the situation whereby both panel heteroskedasticity and spatial correlation
across panels occur. Serial autocorrelation can be corrected by implementing the Prais–
Winsten procedure, such as a generalized AR(1) correction or a panel-specific correction,
PSAR(1) (Beck 2001). The general AR(1) correction only estimates one common serial
autocorrelation parameter by averaging across all included series. PSAR(1), however, estimates a unique autocorrelation parameter for each series. PSAR(1) could yield a better
model fit than the general AR(1) if great heterogeneity is detected in the error structure.
Estimating PSAR(1) is less desirable, however, when a panel data set contains very short
T. Researchers also need to consider whether they have substantive reasons to justify a
heterogeneous structure of the serially correlated error terms (Beck and Katz 1995).14
Newer approaches for handling spatial dependence have also gained scholarly
attention. A few scholars have proposed spatial lag models to handle spatial dependence across panels. The spatial lag approach and the PCSE approach handle cross-unit
interdependence differently. The spatial lag approach treats spatial interdependence
as a substantively relevant term, whereas the PCSE approach models spatial interdependence as nuisance (Franzese and Hayes 2007).
Beside the aforementioned models, there are many mixed applications in the
practice of panel data analysis. Researchers often need to make the specification
decisions that are related to issues of modeling heterogeneity and time dynamics.
Because the error term (ei,t) is always an estimated quantity, any misspecification in
the conditional mean (Xi,t β) would lead to nonspherical errors. Hence, it is usually a
sound practice to use substantive considerations to guide decisions in specifying the
conditional mean first. In the next section, I use the aforementioned state-level panel
example to illustrate different panel model specifications.
Zhu
Panel Data Analysis in Public Administration
407
Panel Data Analysis: Empirical Illustrations
Modeling Heterogeneity: OLS, Fixed Effects, or Random Effects?
The nature of the dependent variable (state-level uninsured rates) demonstrates that
substantively interesting variation exists across 50 states. The three institutional variables also demonstrate that the characteristics of the health care system in a state may
not vary substantially across time. The important implication of these two substantive
Figure 1 State-Level Uninsured Rates, 1990–2006
The Percentage of State Residents Who Live without Health Insurance (Persons Under 65)
30
20
10
0
30
20
10
0
30
20
10
0
30
20
10
0
30
20
10
0
30
20
10
0
30
20
10
0
2005
2000
1995
1990
2005
2000
1995
1990
Year
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
The first step of analyzing a panel data set is to examine the nature of the dependent variable along the two dimensions, T and N. Beck (2001) and Beck and Katz
(1995) emphasize that it is critical to evaluate the relative statistical powers along the
two dimensions, because TSCS data and CSTS data are suitable for different models. The aforementioned state health care panel data set contains more N (50 states)
than T (17 years). It is also necessary to examine the substantive characteristics of
the dependent variable. Figure 1 presents data variation by state and year. Figure 1
demonstrates that the cross-state difference in health insurance coverage exhibits a
persistent pattern across 17 years. Within most states, the uninsured rates vary incrementally over the 17-year period.
408
Journal of Public Administration Research and Theory
15 The
variable measuring public finance in health insurance is lagged by 2 years because budgetary decisions
are made based on fiscal cycles, and in some states, they are made biannually.
16 In table 2, the first FE model is estimated using the STATA 12 command “xtreg..., fe”. The same
estimation can be obtained using “areg..., absorb(state id)”. The second FE model is estimated based on the least
square specification with fixed effects dummies, “xi:xtreg...i.state”. These two procedures produce the same slope
coefficients as the ones obtained based on the mean-centering approach (i.e., removing group means from both
the dependent variable and the explanatory variables). The mean-centering approach, however, produces smaller
standard errors than the ones computed using “xtreg..., fe” or “xi:xtreg...i.state,” because the two STATA fixed
effects procedures do not adjust for changes in the degree of freedom when computing the standard errors.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
considerations is that the fitted panel model needs to capture the cross-state data
variation. As such, estimating an OLS model and a FE model including state dummies could be problematic. Table 2 reports the comparison of four models: OLS,
FE using the mean-centering approach, FE including a full set of state dummies,
and RE.15 Table 2 shows that estimation results are very sensitive to different model
specifications. The slope coefficients change substantially from the OLS specification
to the two FE specifications, indicating that the complete pooling strategy may not
be appropriate. The two FE models clearly demonstrate the problem of estimating
within effects when a large proportion of variance exists across space. The within and
between R2 statistics suggest that the fixed effects specification absorbs a lot of data
variation across states, and thus the two models capture very little data variation in
the panel data set.
The second FE model is particularly problematic because the between R2 becomes
1 after including a full set of state dummies. The between R2 indicates that the second
FE model is a saturated model. This is because state dummies fully control away any
contextual difference in states. This can be deemed to be double-counting data variation driven by the three state health care variables. If the theoretical expectation is
that the levels of all three state health care institutional variables determine the mean
level of state uninsured rates, one cannot estimate a FE model that absorbs all the
institutional differences across states.16
The fourth model in table 2 is a one-way RE model, which assumes that the
group effect (ui) is a random draw from the underlying population effect and is not
correlated with any time-invariant unit (i.e., state dummies). Is the RE model a more
appropriate specification than the OLS model and the two FE models? If ui is uncorrelated with any state-specific effect, the RE model is consistent and more efficient
than the OLS and the two FE models. Running the Hausman test based on comparing the first FE model and the RE model in table 2, the test produces a χ2 of 39.25
(df = 11, p = .000). The test result suggests that the RE model is not consistent and
produces biased estimations.
What does the comparison in table 2 mean? Both the OLS and the RE model are
not consistent, suggesting that one may take the risk of using predetermined factors
to explain the state uninsured rates without a good control for cross-unit heterogeneity. The two FE models, however, are also problematic. The two FE models purely rely
on within-state changes and exclude data variations across states, which are substantively relevant. Plümper, Troeger, and Manow (2005) contend that if the substantive
consideration is to predict level effects, one could choose not to include unit dummies.
They argue that the substantive tradeoff one needs to make is to choose between mild
Zhu
Panel Data Analysis in Public Administration
409
Table 2
Estimation Results: Comparing OLS, FE, and RE Models
(1) OLS
Variable
Coeff.
(T-score)
SE
0.002
0.031
0.006
0.004
0.085
0.041
0.012
0.012
0.060
0.026
0.082
1.058
Coeff.
(Z-score)
−0.005
(−1.04)
0.109*
(1.98)
0.0185
(0.61)
−0.007
(−1.65)
0.204*
(2.46)
0.279**
(5.22)
−0.366
(−1.60)
0.280**
(3.07)
0.702**
(3.02)
−0.062
(−1.70)
−0.118
(−1.66)
4.308
(1.02)
750
—
0.116
0.060
0.058
1.855
(3) FE 2
SE
0.005
0.055
0.031
0.004
0.083
0.053
0.229
0.091
0.233
0.036
0.070
4.234
Coeff.
(Z-score)
−0.005
(−1.04)
0.109*
(1.98)
0.0185
(0.61)
−0.007
(−1.65)
0.204*
(2.46)
0.279**
(5.22)
−0.366
(−1.60)
0.280**
(3.07)
0.702**
(3.02)
−0.062
(−1.70)
−0.118
(−1.66)
11.12
(1.58)
750
—
0.116
1.000
0.842
1.855
(4) RE
SE
0.005
0.055
0.031
0.004
0.083
0.053
0.229
0.091
0.233
0.036
0.070
7.098
Coeff.
(Z-score)
0.011**
(−2.93)
0.076
(1.66)
0.037*
(2.57)
−0.007
(−1.71)
0.186*
(2.33)
0.342**
(6.94)
0.080**
(2.59)
0.259**
(8.46)
0.170
(1.27)
−0.078**
(−3.07)
−0.051
(−0.74)
6.502**
(3.09)
750
—
0.097
0.744
0.627
1.888
SE
0.004
0.046
0.014
0.004
0.080
0.049
0.031
0.031
0.134
0.025
0.069
2.107
Notes: FE1 is the FE model estimated by using the mean-centering approach. FE2 is the FE model estimated including a full set
of state-specific dummies. T-statistics and Z-statistics are reported in parentheses.
Significance level: *p < .05, **p < .01, two-tailed test.
omitted variable bias caused by excluding unit dummies and the saturated model
including all unit dummies. This recommendation, however, is an arbitrary decision
on choosing between two biased models.
Considering Panel Dynamics
Despite that the state-level uninsured rates vary incrementally over time in most states,
data dynamics need to be considered when modeling such a state-level panel data set.
As aforementioned, it is essential to examine whether the pooled 50 time-series are
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
−0.011**
(−5.11)
−0.064*
(−2.09)
Public ownership
0.0107
(1.76)
State liberalism
−0.010*
(−2.56)
Unemployment
0.220**
(2.60)
Poverty
0.539**
(13.13)
Black population 0.0494**
(4.17)
Latino population 0.228**
(18.92)
Aged population −0.210**
(−3.51)
Obesity rate
−0.052*
(−1.98)
Perceived poor
0.127
health
(1.55)
Intercept
10.82**
(10.22)
N
750
Adjusted R2
0.663
Within R2
—
Between R2
—
Overall R2
—
RMSE
2.590
Medicaid
eligibility
Public financet−2
(2) FE 1
410
Journal of Public Administration Research and Theory
17 If
the dependent variable is panel stationary and T is very small, dynamic specifications normally are
unnecessary. If panel unit root is detected, regardless the size of T, one should consider modeling the dynamics,
and the ECM approach would be an appropriate solution. If T is large and the dependent variable is highly
autoregressive (i.e., ρ takes a large or near 1 value), the LDV approach could be a parsimonious way to specify
the dynamics.
18 The RE model in table 3 and the LDV model in table 4 are estimated based on the same specification
except the inclusion of a LDV.
19 The only unaffected regressor is Unemployment.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
panel stationary. If so, how much uninsured rates in the current year are dependent
on those in the last year?
Table 3 reports the results of different panel unit root tests, with different number
of lags, a linear trend term, or a drift. Table 3 includes consistent evidence that the
dependent variable is panel stationary when considering a linear trend term, a drift, a
first-order lag, or a combination of these test specifications. If one lags the dependent
variable by 2 years and adds a linear trend term, the Phillip–Perron test still rejects
the null hypothesis of panel unit root. Therefore, one can conclude that the dependent
variable is panel stationary.
Figure 2 further examines the first-order correlation between Uninsured Ratest
and Uninsured Ratest−1 for each state. This descriptive figure is produced by running
50 separate bivariate models, regressing Uninsured Ratest on Uninsured Ratest−1 for
each state. We are interested in examining: (1) if the correlation for each state is near
1 (an integration or near-integration), and (2) if not near 1, if the correlation is statistically significant (an autoregressive process that captures some long-term memory).
Figure 2 shows that most states have mean correlations around 0.5. Statistically significant correlations are observed in 23 states. Insignificant correlations are observed
in 27 states. Table 3 and figure 2 together show that the dependent variable, state-level
uninsured rates, is panel stationary. The specific form of the autoregressive process,
moreover, slightly varies by states.
What would happen if one applied a LDV specification to stationary panel data
with some heterogeneity in the autoregressive parameter? Given that nearly half of
the states do not see a significant correlation between Uninsured Ratest and Uninsured
Ratest−1, including a LDV may wash out some of the level effects by controlling away
level variance without explaining if the long-term relationship between the dependent
variable and the key explanatory variables is common across all states. Similarly, running a first-differenced model or a GMM model becomes unnecessary.17
Table 4 reports the estimation results based on the five dynamic model specifications: the LDV approach, the first-difference specification, the Anderson–Hsiao specification, and two Arellano–Bond GMM specifications. First, the estimation results
of the LDV model show that the average correlation between the DV and the LDV is
0.678. Comparing the coefficients produced by the LDV model and the RE model in
table 2, one may find that including the LDV inflates the overall model fit and within
R2.18 The LDV model, however, does not increase the explanatory power of most
regressors. Instead, including the LDV attenuates the slope coefficients of most exogenous regressors in the model.19 For example, the coefficient of Medicaid Eligibility is
−0.011 in the RE model and becomes −0.004 after including the LDV. The coefficient
of Poverty is shrunk by about 50%, changing from 0.342 to 0.179.
Zhu
Panel Data Analysis in Public Administration
411
Table 3
Different Panel Unit Root Tests, State Uninsured Rates from 1990 to 2006
Test
df
252.269
225.483
189.549
179.152
152.546
117.507
440.559
255.269
254.386
249.366
179.152
179.228
171.487
—
—
—
—
—
—
—
—
100
100
100
100
100
100
100
100
100
100
100
100
100
—
—
—
—
—
—
—
—
T-statistics
—
—
—
—
—
—
—
—
—
—
—
—
—
−2.366
−1.986
−2.668
−2.140
−17.142
−15.264
−20.506
−17.937
p
.000
.000
.000
.000
.000
.111
.000
.000
.000
.000
.000
.000
.000
.000
.000
.000
.242
.000
.000
.000
.457
Notes: ADF, augmented Dickey–Fuller test; PP, Phillips–Perron test; IPShin, Im–Pearson–Shin test; LLC, Levin–Lin–Chu test.
All the four tests have the null hypothesis of panel nonstationarity.
Table 4 also demonstrates a similar issue of applying a linear dynamic specification based on the GMM approach. Both Arellano–Bond’s models produce smaller
coefficients of the LDV than the LDV model. Using the Arellano–Bond method, the
average correlation between the DV and the LDV is less than 0.40, indicating that
the long-term effects are not very persistent. The dynamic specification, moreover,
reduces the coefficient size of many exogenous regressors (e.g., Medicaid Eligibility,
Unemployment, Poverty, Black Population, and Latino Population).
The first-difference approach and the Anderson–Hsiao specification, furthermore,
are also not good model choices given that the DV is panel stationary. Differencing
both the DV and the regressors removes a large proportion of variance and thus makes
both within R2 and between R2 near zero. Substantively, the two FE models in table 2
and the AH specification in table 4 can be viewed as two typical examples of misspecification: the saturated model and the null model. The unit-specific FE model specification produces a saturated model because it overfits the data. The AH specification
produces a null model that barely explains any variation of state uninsured rates.
Adjusting for Coefficient Stability and the RCM
Because a large proportion of data variation exists across states and the change of
uninsured rates is not very dynamic, one needs to consider a panel model that captures
cross-state variance. The RE model in table 2 produces biased coefficients because it
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
ADF, no trend, lag = 0
ADF, no trend, lag = 1
ADF, no trend, lag = 2
ADF, with trend, lag = 0
ADF, with trend, lag = 1
ADF, with trend, lag = 2
ADF, with drift, lag = 0
PP, no trend, lag = 0
PP, no trend, lag = 1
PP, no trend, lag = 2
PP, with trend, lag = 0
PP, with trend, lag = 1
PP, with trend, lag = 2
IPShin, no trend, lag = 1
IPShin, no trend, lag = 2
IPShin, with trend, lag = 1
IPShin, with trend, lag = 2
LLC, no trend, lag = 1
LLC, no trend, lag = 2
LLC, with trend, lag = 1
LLC, with trend, lag = 2
χ2
412
Journal of Public Administration Research and Theory
Figure 2 Illustration of First-order Correlations between Uninsuredt and Uninsuredt−1
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
does not capture the possible state-level heterogeneity caused by state-specific effects.
However, running a FE model including all state dummies is also problematic. Thus,
the substantive consideration becomes evaluating how unobserved heterogeneity is
caused by specific states and is not fully explained by the regressors in the model. In
other words, one needs to detect which states are “average states” explained by the
model, and which states are outlier states that require additional model considerations.
Outlier states may be detected by analyzing residuals after fitting a baseline
model (e.g., an OLS model). Evaluating the standardized residuals or the Cook’s
distance normally helps to detect outlier states. Another way to evaluate outlier cases
is to run a weighted (robust) regression analysis weighting observations based on the
absolute value of residuals.
Obesity rate
Aged population
Latino population
Black population
Poverty
Unemployment
State liberalism
Public ownership
Public financet−2
Medicaid eligibility
Uninsuredt−1
0.678**
(26.29)
−0.004*
(−2.33)
−0.0263
(−1.20)
0.002
(0.41)
−0.004
(−1.46)
0.182**
(2.98)
0.179**
(5.52)
0.011
(1.27)
−0.072**
(6.88)
−0.084
(−1.95)
−0.004
(−0.21)
Coeff.
(Z-score)
−0.003
(−0.40)
0.198*
(2.18)
−0.027
(−0.59)
0.002
(0.033)
0.216
(1.78)
0.129**
(2.62)
−1.435
(−1.85)
−0.139
(−0.34)
−0.199
(−0.43)
−0.105
(−1.58)
0.002
0.019
0.043
0.010
0.009
0.033
0.061
0.003
0.004
0.022
—
Coeff.
(Z-score)
0.026
SE
(2) FD
0.066
0.458
0.403
0.775
0.049
0.121
0.006
0.046
0.091
0.007
—
SE
0.143**
(2.99)
−0.003
(−0.47)
0.200*
(2.19)
−0.069
(−1.45)
0.004
(0.68)
0.190
(1.42)
0.112*
(2.15)
−1.446
(−1.79)
0.0867
(0.21)
−0.754
(−1.40)
−0.076
(−1.14)
Coeff.
(Z-score)
(3) AH
0.066
0.538
0.419
0.806
0.052
0.134
0.006
0.048
0.091
0.007
0.047
SE
0.384**
(8.21)
−0.008
(−1.10)
0.298**
(3.89)
−0.045
(−1.01)
0.005
(0.81)
0.150
(1.55)
0.201**
(3.35)
−0.892*
(−2.04)
0.0257
(0.18)
−0.425
(−0.94)
−0.026
(−0.52)
Coeff.
(Z-score)
(4) AB-1
0.050
0.453
0.145
0.436
0.060
0.097
0.006
0.045
0.077
0.007
0.046
SE
0.356**
(10.18)
−0.007
(−1.51)
0.277**
(6.96)
−0.044**
(2.71)
0.004*
(2.03)
0.213**
(5.25)
0.181**
(3.89)
−0.257
(−0.31)
−0.052
(−0.24)
0.305
(0.46)
−0.039
(−1.36)
Coeff.
(Z-score)
0.029
0.671
0.218
0.830
0.047
0.040
0.002
0.016
0.040
0.005
0.035
SE
Continued
(5) AB-2
Panel Data Analysis in Public Administration
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Variable
(1) LDV
Table 4
Estimation Results: Comparing the LDV, the First-difference, the Anderson−Hsiao, and the Arellano–Bond GMM Specifications
Zhu
413
3.233**
(3.97)
750
0.214
0.981
0.829
1.862
Intercept
0.814
0.059
SE
0.284
(1.84)
700
0.036
0.020
0.034
2.005
−0.076
(−1.20)
Coeff.
(Z-score)
(2) FD
0.152
0.064
SE
0.130
(0.82)
650
0.036
0.340
0.022
—
−0.078
(−1.23)
Coeff.
(Z-score)
(3) AH
0.158
0.064
SE
18.81**
(2.71)
700
—
—
—
—
−0.097
(−1.15)
Coeff.
(Z-score)
(4) AB-1
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
−0.095*
(−2.05)
4.030
(0.36)
700
—
—
—
—
6.931
Coeff.
(Z-score)
0.084
SE
(5) AB-2
SE
11.256
0.046
Notes: LV refers to the RE model with a lagged dependent variable. FD refers to the first-difference specification by differencing both the dependent variable and all the explanatory variables. AH
refers to the Anderson–Hsiao specification that transforms variables by taking the first-order difference and using ΔYi,t−2 as the instrument. The lagged DV term in Model (3) is Uninsured Ratesi, t−2.
AB-1 refers to the Arellano–Bond GMM specification, using the one-step estimation approach. AB-2 refers to the Arellano–Bond GMM specification, using the two-step estimation approach. In
both AB models, the number of instruments equals to 131. Z-statistics are reported in parentheses.
Significance level: *p < .05, **p < .01, two-tailed test.
N
Within R2
Between R2
Overall R2
RMSE
0.018
(0.31)
Coeff.
(Z-score)
(1) LDV
Perceived poor
health
Variable
Table 4 (continued)
414
Journal of Public Administration Research and Theory
Zhu
Panel Data Analysis in Public Administration
The Beck–Katz Procedure and Beyond
Despite the considerations of modeling heterogeneity and panel dynamics, one also
needs to consider if heterogeneity is detected in the error term, ei,t. After estimating the
OLS model, it is necessary to test for heteroskedasticity, group-based heterogeneity,
cross-section dependence, and serial autocorrelation (usually testing for the first-order
auto-correlation).
After estimating an OLS model including dummy variables for outlier states, the
Breusch–Pagan/Cook–Weisberg test for heteroskedasticity reports a χ2 of 20.72 and a
p-value of .0000. This suggests that heteroskedasticity is detected. The Arellano–Bond
test for AR(1) reports a significant Z-score, suggesting that serial autocorrelation is
detected. After estimating a FE model including the same set of regressors and the full
set of year dummy variables, the modified Wald test reports a χ2 of 299.81(df = 50,
p = .0000), suggesting that panel-wise heteroskedasticity is detected. Lastly, the Frees
method (Frees 1995) is used to test spatial dependence for the panel data with small
T and large N. The spatial dependence test reports a Z-score of 1.057, which is larger
than the corresponding critical values and suggests spatial dependence. The LSDV
model reported in table 5, therefore, needs to be adjusted for correlated errors.
Beck and Katz (1995) develop a robust standard error approach adjusting for
cross-unit dependence. Table 6 reports five models estimated based on the Beck
and Katz approach, with different error term specifications. Model (1) is estimated
adjusting for spatial dependence, panel-wise heteroskedasticity, and panel-specific
serial autocorrelations. Model (2) only corrects for panel-specific serial autocorrelation and panel-wise heteroskedasticity without correcting for spatial dependence.
Model (3) specifies PSAR(1) but assumes independent error structure across panels.
20 Cases
with large residuals are always downweighted when running the robust regression analysis. The
weights are closely corresponding to the Cook’s distance.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Approximately, as residuals increase, the estimated weights decrease. Hence,
states with relatively small weights can be considered as outlier states.20 Figure 3
illustrates the outlier states detected based on residual analysis, which need to be
controlled. Another possible source of unobserved heterogeneity might be the timespecific effects (i.e., heterogeneous parameters across time). A set of year dummies
can be included in the model to control for time-specific effects. In other words, I use
each year as its own baseline to adjust for unstable coefficient estimations across years.
An alternative way to think about heterogeneity is that parameters may vary
across years, but the variation is random based on a general mean baseline. The
RCM is estimated based on this alternative consideration and uses second-level
(year-level) parameters to capture heterogeneity across time. Table 5 reports the
comparison between the RCM model and the least square model with a set of year
dummies. Both models are estimated by controlling for the outlier states detected
based on residual analysis. The two models produce comparable results for Medicaid
Eligibility, Unemployment, Latino Population, and Obesity Rate; the intercepts differ
substantially across two models and so do the coefficients of Public Finance, Public
Ownership, Poverty, Aged Population, and Perceived Poor Health.
415
Journal of Public Administration Research and Theory
Figure 3 Illustration of Outlier States
25
30
Arkansas
Oklahoma
Florida Arkansas
Utah
20
15
10
Uninsured Rates
Uninsured Rates
30
25
Arkansas
Arkansas
20
15
10
Hawaii
Hawaii
5
100
80
60
Medicaid Eligibility
30
30
Arkansas
15
10
Hawaii
Hawaii
Uninsured Rates
Arkansas
Oklahoma
Florida
Utah
Nevada
20
5
40
20
50
40
30
20
10
Public Ownership
25
Oklahoma
Utah Florida
25
20
15
10
Hawaii
Hawaii
5
7
6
5
4
3
16
14
12
10
8
Public Finance
Arkansas
Oklahoma
Arkansas Nevada
Florida
Utah
Unemployment
Model (1) produces larger standard errors than those in Models (2) and (3) for
most variables, except for Medicaid Eligibility, Black Population and Perceived Poor
Health. In this empirical case, ignoring spatial dependence may lead to overconfidence in the statistical inference. Models (4) and (1) take the same error specification, but Model (4) only specifies a common AR(1). Hence, Model (1) and Model
(4) show the statistical difference between using AR(1) and PSAR(1). Given the
heterogeneous correlation pattern shown in figure 2, it is not surprising to see that
Model (1) produces a larger R2 than Model (4). The coefficients corresponding to
Public Finance, Public Ownership, and Obesity Rate become insignificant in Model
(4). The coefficient size of Aged Population is substantially attenuated. Model (5)
is estimated by including a LDV, without an AR(1) or PSAR(1) specification for
the error term. Model (5) also corrects for panel-wise heteroskedasticity and spatial autocorrelation. Again, including a LDV attenuates most coefficients. This is
because the LDV picks up some long-term effects and the other β coefficients only
carry short-term effects.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Hawaii
Hawaii
5
Uninsured Rates
416
Zhu
Panel Data Analysis in Public Administration
417
Table 5
Estimation Results: Comparing the LSDV and the RCM Specifications
(1) LSDV
Variable
Medicaid eligibility
Public financet−2
Public ownership
State liberalism
Poverty
Black population
Latino population
Aged population
Obesity rate
Perceived poor health
Intercept
Within R2
Between R2
Overall R2
Random effects
SD (cons)
SD (residuals)
N
Log likelihood
Coeff. (Z-score)
SE
Coeff. (Z-score)
SE
−0.010**
(−3.05)
−0.0346
(−0.75)
0.0302*
(2.48)
−0.004
(−1.08)
0.303**
(3.10)
0.327**
(7.03)
0.090**
(3.34)
0.243**
(8.42)
0.110
(0.76)
−0.152**
(−2.59)
0.0319
(0.49)
10.898**
(4.36)
0.245
0.850
0.740
0.003
−0.012**
(−6.43)
−0.062*
(−2.05)
0.008
(1.45)
−0.001
(−0.30)
0.258**
(2.92)
0.564**
(15.47)
0.044**
(3.95)
0.196**
(16.14)
−0.330**
(−4.62)
−0.155**
(−3.97)
0.147*
(2.12)
13.433**
(12.00)
—
—
—
0.002
—
—
750
—
0.046
0.012
0.004
0.098
0.046
0.027
0.029
0.146
0.059
0.066
2.501
0.900
2.163
750
−1659.927
0.029
0.005
0.004
0.090
0.036
0.011
0.012
0.073
0.039
0.069
1.112
0.213
0.057
Notes: LSDV refers to the least square dummy variable specification. RCM refers to the random coefficients model. Z-statistics
are reported in parentheses.
Significance level: *p < .05, **p < .01, two-tailed test.
Comparing estimation results between Model (1) (the LSDV–PCSE model) in
table 6 and the RCM in table 5, RCM produces smaller population-averaged coefficients
than the LSDV–PCSE model for all three institutional variables and State Liberalism.
The coefficients for the two economic variables (Poverty and Unemployment), Latino
Population, and the two health risk variables are larger in the RCM model than those
in the LSDV–PCSE model. The RCM model reported in table 5 only captures random
intercepts across years and assumes that the regression lines for a regressor are parallel across different years. As for Perceived Poor Health, the big shift in coefficient size
indicates that there might be additional variance related to this variable, which is not
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Unemployment
(2) RCM
0.002
−0.017**
(−9.74)
0.136**
(3.11)
0.019*
(2.40)
−0.003
(−0.62)
0.160
(1.21)
0.403**
(8.08)
0.050**
(3.38)
0.171**
(10.98)
−0.435**
(−2.92)
−0.059
(−1.21)
−0.051
(−0.90)
13.47**
(6.67)
750
0.880
2.019
0.056
0.048
0.149
0.016
0.015
0.050
0.132
0.005
0.008
0.044
—
—
−0.017**
(−8.62)
0.136**
(3.51)
0.019**
(2.76)
−0.003
(−0.70)
0.160
(1.94)
0.403**
(10.34)
0.050**
(3.27)
0.171**
(9.42)
−0.435**
(−3.91)
−0.059*
(−2.23)
−0.051
(−0.86)
13.47**
(8.70)
750
0.880
—
1.549
0.059
0.026
0.111
0.018
0.015
0.039
0.083
0.004
0.007
0.039
0.002
—
(2) LSDV PSAR(1)
Coeff.
Robust
(Z-score)
SE
−0.016**
(−7.59)
0.136**
(3.45)
0.0189**
(2.70)
−0.003
(−0.71)
0.160
(1.89)
0.403**
(10.91)
0.050**
(3.31)
0.171**
(9.15)
−0.435**
(−3.63)
−0.059*
(−2.19)
−0.051
(−0.83)
13.47**
(8.02)
750
0.880
—
1.679
0.061
0.027
0.120
0.019
0.015
0.037
0.085
0.004
0.007
0.039
0.002
—
(3) LSDV PSAR(1)
Coeff.
(Z-score)
SE
−0.016**
(−7.34)
0.059
(1.40)
0.016
(1.69)
−0.002
(−0.56)
0.155
(1.07)
0.407**
(7.45)
0.057**
(3.21)
0.222**
(14.77)
−0.284*
(−2.36)
−0.055
(−0.95)
0.016
(0.24)
12.04**
(6.75)
750
0.633
—
1.783
0.067
0.058
0.120
0.015
0.018
0.055
0.146
0.005
0.009
0.042
0.002
—
(4) LSDV AR(1)
Coeff.
(Z-score)
PCSE
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
0.603**
(10.05)
−0.005**
(−3.64)
−0.001
(−0.04)
0.001
(0.20)
−0.002
(−0.60)
0.153
(1.63)
0.225**
(5.52)
0.009
(0.96)
0.077**
(4.95)
−0.191**
(−2.88)
−0.016
(−0.42)
0.011
(0.19)
5.092**
(4.37)
750
0.837
1.165
0.059
0.037
0.067
0.016
0.009
0.041
0.094
0.003
0.005
0.005
0.001
0.060
(5) LDV
Coeff.
(Z-score)
PCSE
Notes: LSDV refers to the least square dummy variable specification. LDV refers to the lagged dependent variable specification. AR(1) and PSAR(1) refer to first-order serial autocorrelation and
panel-specific serial autocorrelation, respectively. Z-statistics are reported in parentheses.
Significance level: *p < .05, **p < .01, two-tailed test.
N
R2
Perceived poor
health
Intercept
Obesity rate
Aged population
Latino population
Black population
Poverty
Unemployment
State liberalism
Public ownership
Public financet−2
Medicaid eligibility
Uninsuredt−1
Variable
(1) LSDV PSAR(1)
Coeff.
(Z-score) PCSE
Table 6
Estimation Results: Comparing LSDV and LDV with Different Standard Error Specifications
418
Journal of Public Administration Research and Theory
Zhu
Panel Data Analysis in Public Administration
21 Figure
5 is generated based on a robust analysis incrementally dropping year panels. Both models are not
robust when the number of years is very small. When including all the years from 1998 to 2006, results based
on the RCM specification are quite robust.
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
incorporated in the RCM. It is likely that citizens’ perceptions about their own health
status may vary across years. To adjust for possible parameter heterogeneity, one can
add a random slope term into the RCM for the variable Perceived Poor Health.
Model (1) in table 7 reports the estimation results of the RCM model with varying intercepts by year and varying slopes for Perceived Poor Health. Examining the
RCM model in table 7 and Model (1) (the LSDV–PCSE model) in table 6, substantive
findings are comparable based on these two models. Controlling for the institutional
characteristics of state health care systems and the macroeconomy (Unemployment
and Poverty), state liberalism does not predict the level of state uninsured rates. Both
institutional and economic factors matter substantively for explaining state uninsured
rates. The uninsured rates are slightly lower in states with generous Medicaid eligibility rules and more public finance in health care than in states with tight Medicaid
eligibility rules and less public finance in health care. Both unemployment and poverty
strongly predict the level of uninsured rates.
The RCM model in table 7 is preferred to the LSDV–PCSE model in table 6
because of two reasons. First, the RCM specification better captures the nature of
heterogeneity across state-year observations. Although the substantive findings for
many variables based on population-averaged coefficients in two models are similar,
the RCM specification explicitly models heterogeneity and is more flexible to capture
varying parameters across time. Figure 4 illustrates homogeneity and heterogeneity
in parameters. In this panel data set, the coefficients of the poverty measure do not
vary across different years. The coefficients of the variable for perceived poor health,
however, exhibit a heterogeneous pattern across different years. Although the LSDV–
PCSE specification also deals with heterogeneity across time by including a set of year
dummies, it does not accurately reflect the varying slopes for this particular variable.
When parameter heterogeneity is detected, it is better to model the heterogeneous
coefficients instead of simply assuming that all varying parameters are nested in the
year dummy variables. The RCM model reports findings on public ownership and
public finance that make more substantive sense. The LSDV–PCSE model reports
positive associations between public ownership, which is opposite to the substantive
expectation. The RCM model, however, reports negative associations between the two
institutional variables and the uninsured rates.
Second, the RCM specification performs better than the LSDV–PCSE model
based on coefficient stability at the population-averaged level. Figure 5 compares the
performances of the two models based on the population-averaged coefficients of
six variables: Medicaid Eligibility, Public Finance, Public Ownership, Unemployment,
Poverty, and Latino Population.21 All the six subfigures demonstrate that the LSDV–
PCSE model produces the slope coefficients that are more sensitive to changes in
year panels than those produced by the RCM model. Figure 5(a–c) shows that the
RCM model yields a more stable mean estimation for the three institutional variables:
Medicaid Eligibility, Public Finance, and Public Ownership when T is relatively large.
The coefficients of Poverty and Unemployment are both insensitive in two models,
419
420
Journal of Public Administration Research and Theory
Table 7
RCM with Random Intercept and Random Slope
(1) RCM
Variable
(2) RCM–EC
Coeff. (Z-score)
SE
Coeff. (Z-score)
SE
0.002
−0.005**
(−3.16)
−0.020
(-0.88)
0.002
(0.45)
−0.001
(−0.43)
—
0.002
Unemploymentt-1
−0.012**
(6.29)
−0.060*
(−2.03)
0.008
(1.45)
−0.001
(−0.36)
0.238**
(2.73)
—
∆Unemploymentt
—
—
Uninsuredt−1
− Unempt−1
Poverty
—
—
0.563**
(15.48)
0.044**
(3.94)
0.197**
(16.37)
−0.330**
(−4.63)
−0.149**
(−3.84)
0.136
(1.74)
13.453**
(12.10)
0.036
Medicaid eligibility
Public financet−2
Public ownership
State liberalism
Black population
Latino population
Aged population
Obesity rate
Perceived poor health
Intercept
Random effects
SD (poor health)
SD (cons)
SD (residuals)
N
Log likelihood
0.134
0.675
2.158
750
−1659.16
0.005
0.003
0.087
—
0.011
0.012
0.071
0.039
0.078
1.111
0.072
0.313
0.057
−0.276**
(−3.80)
0.010
(0.08)
−0.391**
(−13.71)
0.226**
(6.98)
0.013
(1.47)
0.075**
(6.78)
−0.171**
(−3.03)
−0.054
(1.83)
0.037
(0.63)
5.747**
(6.16)
0.089
0.484
1.696
750
−1476.76
0.023
0.004
0.003
—
0.073
0.126
0.029
0.032
0.009
0.011
0.057
0.030
0.060
0.934
0.050
0.187
0.044
Notes: Model (1) is estimated including random intercepts and random slopes for Perceived Poor Health. Model (2) is estimated
including random intercepts, random slopes for Perceived Poor Health and an EC specification for Unemployment. Z-statistics
are reported in parentheses.
Significance level: *p < .05, **p < .01, two-tailed test.
especially Unemployment. The RCM model produces slightly different populationaveraged coefficients than the LSDV–PCSE because the RCM model does not assume
that every year generates its unique baseline value.
In addition, figure 5(d) shows that the effects of Unemployment, estimated in both
the RCM and LSDV–PCSE model, may not be time-consistent, especially in years
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Unemployment
0.029
Zhu
Panel Data Analysis in Public Administration
421
Figure 4 Homogeneity and Heterogeneity in Parameter
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
30
20
10
0
30
20
10
20
10
0
30
20
10
0
25
20
15
10
5
25
20
15
10
5
(a) State Poverty Rate: Parameter Homogeneity
Fitted Values
Uninsured Rate
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
30
20
10
0
30
20
10
0
30
20
10
0
30
20
10
0
10
8
6
4
2
10
8
6
4
2
(b) Self−Reported Poor Health Status (% population): Parameter Heterogeneity
Fitted values
Uninsured Rate
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
0
30
.1
.15
.2
.8
1
.25
1.2
−1.5 −1 −.5
−.04 −.02
0
0
.5
.02
1
.04
Year<1994
.6
Year<1995
.4
Year<1994
95% CIs
prior to 1998. This may point to the possibility that there is some long-term equilibrium
relationship between unemployment and uninsured rates. This might be a reasonable
substantive consideration given that most state health care systems are labor market–
driven. Model (2) in table 7 incorporates this substantive consideration and adds an
error correction (EC) specification for Unemployment. As Model (2) shows, with a
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Year<1994
Year<1996
Full Sample
Year<2006
Year<2005
Year<2004
Year<2003
Year<2002
Year<2001
Year<2000
Year<1999
Year<1998
Year<1997
Year<1996
Year<1995
Year<1994
Full Sample
Year<2006
Year<2005
Year<2004
Year<2003
Year<2002
Year<2001
Year<2000
Year<1999
Year<1998
Year<1997
Year<1995
Year<1999
Year<1998
Year<1997
Year<1996
Year<1995
Year<1994
Full Sample
Year<2006
Year<2005
Year<2004
Year<2003
Year<2002
Year<2001
Year<2000
Year<1999
Year<1998
Year<1997
Year<1995
Year<1997
Year<1998
Year<1999
Year<2000
Year<2001
Year<2002
Year<2003
Year<2004
Year<2005
Year<2006
Full Sample
Year<1994
Year<1995
Year<1996
Year<1997
Year<1998
Year<1999
Year<2000
Year<2001
Year<2002
Year<2003
Year<2004
Year<2005
Year<2006
Full Sample
(f) Sample: Latino Population
(e) Sample:Poverty
Year<1996
(d) Sample: Unemployment
(c) Sample: Ownership
Year<1996
(b) Sample: Public Finance
(a) Sample: Medicaid Eligibility
RCM
95% CIs
LSDV−PCSE
Year<2000
Year<2001
Year<2002
Year<2003
Year<2004
Year<2005
Year<2006
Full Sample
.2
−.3 −.2 −.1
0
−.03 −.02 −.01
.1
0
.2
.01
Journal of Public Administration Research and Theory
422
Figure 5 Comparing Coefficients’ Stability: LSDV–PCSE and RCM
Zhu
Panel Data Analysis in Public Administration
Discussion: Implications for Observational Studies Using Panel Data
In this article, I have discussed various panel models suitable for different substantive
considerations. Panel data can become powerful research designs for answering questions of interest in public administration. Panel data analysis can be useful in analyzing incremental changes in policies, comparing policy institutions, or generalizing
substantive relationships across organizational contexts. The empirical example discussed in this article, however, does not represent all types of panel data sets. Beside the
CSTS example discussed in this article (i.e., continuous dependent variable with large
N and small T), a panel data set may also have larger T than N, or contain hierarchical data structures (e.g., individual-level observations nested in social groups, organizations nested in different levels of government, and so on). Public administration
scholars may also be concerned with different substantive relationships that require
alternative panel data approaches. For example, the generalized estimating equation
(GEE) method would be appropriate if the substantive question concerns inferring
average groups effects instead of unit-specific effects. Statistically, the GEE method is
flexible in that it allows “for a range of substantively-motivated correlation patterns
within [data] clusters and offer the potential for valuable substantive insights into the
dynamics of that correlation” (Zorn 2001, 470).22 More complex multi-level specifications should be considered when the substantive questions focus on micro–macro connections, such as how institutional contexts are linked to micro-level administrative
decisions (Gelman and Hill 2007). Public administration scholars, furthermore, often
explore substantive policy outcomes pertaining to small-domain estimations, such as
public health outcomes for racial/ethnic minorities, unemployment in local communities, failing organizations, etc. Advanced statistical methods such as Bayesian analysis
22 See
Whitford and Yates (2009) for a recent example of applying GEE models in public policy and public
administration research. The GEE models estimate the population-averaged expectation of the dependent
variable instead of unit-specific or time-specific conditional means (Zorn 2001).
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
dynamic specification for the Unemployment variable, the dependent variable has now
taken a differencing transformation. Added in the model are a change term, a lag
term of Unemployment, and the disequilibrium term (Uninsuredt−1 – Unemploymentt−1).
In Model (2), significant coefficients are observed for Unemploymentt−1 and (Uninsur
edt−1 –Unemploymentt−1), indicating that state uninsured rates and unemployment, on
average, are tracking each other from a long run. The two models in table 7 clearly
show some statistical tradeoffs. Adding EC as a dynamic specification shrinks coefficients for other regressors because Model (2) reflects how these variables affect changes
in uninsured rates. Without specifying the long-run dynamic relationship, Model (1)
slightly underestimates the effect of Unemployment on uninsured rates. Nevertheless,
coefficient signs remain the same in Models (1) and (2), pointing toward the same
direction of relationships. If the substantive focus is to examine how economic factors
(e.g., unemployment) affect people’s access to health care, one certainly needs to put
more efforts into thinking about the nuanced difference between long-term and shortterm effects.
423
424
Journal of Public Administration Research and Theory
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
(Gill and Meier 2000; Wagner and Gill 2005) would be called for making substantive
inferences based on small-domain statistics.
Beyond the aforementioned different substantive considerations, many public
administration data sets contain discrete policy variables. For example, public managers’ rating of their organizational performance can be recorded by ordered grade
scales. The frequency of networking activities can be coded as count data or ordinal scales. Statistical specifications suitable for linear probability models are needed
to model panel data with discrete dependent variables (Baltagi 2008; Honore and
Kryiazidou 2000). Although statistical considerations may vary in accordance with
different data structures, the preceding theoretical discussion on panel methods and
the analytic example led to several general implications for observational studies in
using panel data.
First, and most importantly, different models make different underlying
assumptions about the nature and the source of heterogeneity. Pooling data across
time and space, as Stimson (1985) warns, may raise challenges to valid inferences
because unobserved heterogeneity could be time-dependent and spatially dependent.
Although panel data have advantages over pure cross-sectional and time-series data,
they require scholars to consider how to model heterogeneity substantively. If heterogeneity mainly exists across the 50 states, the underlying data-generating process
may be applied to some states, but not others. If heterogeneity is time-dependent, one
needs to consider how to model heterogeneous dynamics. The state-level panel data
example shows that some conventional techniques may produce problematic results.
For example, the unit-specific FE model is widely used in observational studies to
control for the potential omitted variable bias. Running a FE model by states, however, produces artificial null findings for substantively interesting variables. The oneway RE model, furthermore, is inconsistent because it assumes unit homogeneity.
The preferred model choice, in this example, is the RCM specification that fits the
data with partial pooling.
Second, dummy variables are effective ways to control for unobserved heterogeneity, but they are not substantively helpful to explain why heterogeneity is state-specific
or year-specific. In the analytic example, a few states are identified to be outlier states
that do not fit the partial pooling model. Using dummy variables to control for these
outlier states improves estimation reliability, but tells nothing about why in Hawaii a
relatively large increase in unemployment did not lead to substantial changes in the
uninsured rates. Nor can the state dummy variables explain why in Arkansas, with
the same Medicaid eligibility rules and the same level of public finance, the uninsured
rate was around 28% in one year and around 25% in another year (see figure 3).
Further theorization on these substantively interested states is warranted. Could it
be possible that the macroeconomic forces interact with health care institutions to
affect citizens’ access to health care? If this conjecture were true, an interaction term
between unemployment and the institutional variables could be added into the model
to evaluate how institutional arrangements buffer the economic shocks. Could it also
be possible that Hawaii is a pioneering state mandating comprehensive health insurance, and thus the uninsured rate in Hawaii is less responsive to the macroeconomy
than that in a different state? To evaluate this conjecture, one would need to include
a new variable measuring state mandates on employment-based health insurance. In
Zhu
Panel Data Analysis in Public Administration
References
Achen, Christopher H. 2000. Why lagged dependent variables can suppress the explanatory power
of other independent variables. Presented at the Annual Meeting of the Political Methodology
Section of the American Political Science Association, Los Angeles, July 2000.
Allison, Paul. 2009. Fixed-effects regression models. Thousand Oaks, CA: Sage Publications.
Alogoskoufis, George, and Ron Smith. 1990. On error correction models: Specification, interpretation, estimation. Journal of Economic Surveys 5:97–125.
23 In
the analytic example, spatial dependence is modeled away in the error term. One may also consider
estimating the spatial dependence using spatial lags (Baltagi et al. 2003; Hayes, Kachi, and Franzese 2010).
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
the RCM reported in table 7, these substantive effects are nested in the state dummy
variables.
Third, it is important to consider “whether the inclusion of a LDV and period
dummies absorb most of the theoretically interesting time-series variance in the data”
(Plümper, Troeger, and Manow 2005, 328). As for the empirical example, period dummies are chosen in the one-way RE model because of two reasons: (1) the dependent
variable is panel stationary, and (2) some of the coefficients are sensitive to data from
year to year. Alternatively, the RCM explicitly accounts for the parameter heterogeneity across years. This model specification may not be adequate, however, if the panel
data reflect some long-term equilibrium relationships. Scholars need to consider different methods based on the link between their substantive and statistical considerations.
Fourth, there are various FGLS methods that one can use to adjust for nonconstant error terms. These FGLS methods for estimating errors, however, should not
always be considered as the preferred techniques for panel data analysis. Because any
misspecification of the conditional mean can lead to correlated errors, one should
consider specifying the conditional mean first and then applying FGLS techniques for
the error term. Modeling heterogeneity as an important substance should be preferred
to modeling heterogeneity as a nuisance.
Although this article reviews many different panel applications, it does not
exhaust the long list of panel models.23 Nor does this article make the argument that
scholars only need to worry about methodological issues related to different panel
specifications.
Empirical models using observational panel data are usually vulnerable to omitted variable bias and measurement errors. Hence, scholars need to worry more about
how to build a panel data set at the first place (Wawro 2002). Rigorous theories and
measurement techniques can help to preempt many statistical issues for modeling
a multidimension data set. King (1998, 3) points out “sophisticated methods are
required only when data are problematic; extremely reliable data with sensational relationships require little if any statistical analysis.” In fact, data in public administration
do not always guarantee the statistical assumptions for fitting very simple models. It
is also difficult, if not impossible, to develop a single panel data approach suitable for
all data problems. In practice, researchers usually need to consider mixed applications
for conducting panel data analysis. All of these constraints should force researchers to
link substantive considerations to model specifications.
425
426
Journal of Public Administration Research and Theory
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Anderson, T. W., and Cheng Hsiao. 1981. Estimation of dynamic models with error components.
Journal of American Statistical Association 76:598–606.
Arellano, Manuel, and Bo Honore. 1999. Panel data models: Some recent development. In Handbook
of econometrics, eds. James J. Heckman and Leamer E. Edward, 3229–96. Elsevier.
Arellano, Manuel, and Stephen Bond. 1991. Some tests of specification for panel data: Monte Carlo
evidence and an application to employment equations. Review of Economic Studies 58:277–97.
Baltagi, Badi H. 2008. Econometric analysis of panel data, 4th ed. West Sussex, UK: John Wiley and
Sons.
Baltagi, Badi H., Georges Bresson, and Alain Pirotte. 2003. Fixed effects, random effects for
Hausman-Taylor? A pretest estimator. Economics Letters 79:361–9.
Beck, Nathaniel. 1991. Comparing dynamic specifications: The case of presidential approval.
Political Analysis 3:51–87.
———. 2001. Time-series-cross-section data: What have we learned in the past few years? Annual
Review of Political Science 4:271–93.
———. 2011. Of fixed-effects and time-invariant variables. Political Analysis 19:119–22.
Beck, Nathaniel, and Jonathan N. Katz. 1995. What to do (and not to do) with time-series crosssection data. American Political Science Review 89:634–47.
———. 1996. Nuisance versus substance: Specifying and estimating time-series cross-section models. Political Analysis 6:1–36.
———. 2007. Random coefficients models for time-series- cross-section data: Monte Carlo experiments. Political Analysis 15:182–95.
Berry, William D., Evan J. Ringquist, Richard C. Fording, and Russell L. Hanson. 1998. Measuring
citizen and government ideology in the American states: 1960–93. American Journal of Political
Science 42:327–48.
Box, George, and Gwilym Jenkins. 1970. Time series analysis: Forecasting and control. San Francisco,
CA: Holden-Day.
Boyne, George A., Oliver James, Peter John, and Nicolai Petrovsky. 2010. Does public service performance affect top management turnover? Journal of Public Administration Research and
Theory 20 (Suppl 2):i261–79.
Breusch, Trevor, Michael B. Ward, Hoa Thi Minh Nguyen, and Tom Kompas. 2011. FEVD: Just IV
or just mistaken? Political Analysis 19:165–9.
Campbell, John Y., and Robert J. Shiller. 1988. Interpreting cointegrated models. Journal of Dynamics
and Control 12:505–22.
Cornwell, Christopher, and Edward J. Kellough. 1994. Women and minorities in federal government
agencies: Examining new evidence from panel data. Public Administration Review 54:265–70.
De Boef, Suzanna. 2001. Modeling equilibrium relationships: Error correction models with strongly
autoregressive data. Political Analysis 9:78–94.
De Boef, Suzanna, and Jim Granato. 1997. Near-integrated data and the analysis of political relationships. American Journal of Political Science 41:619–40.
De Boef, Suzanna, and Luke Keele. 2008. Taking time seriously. American Journal of Political
Science 52:184–200.
Dickey, David A., and Wayne A. Fuller. 1981. Likelihood ratio statistics for autoregressive timeseries with a unit root. Econometrica 49:1057–72.
Engle, Robert F., and C. W. J. Granger. 1987. Co-integration and error correction: Representation,
estimation, and testing. Econometrica 55:251–76.
Eom, Tae Ho, Sock Hwan Lee, and Hua Xu. 2008. Introduction to panel data analysis. In Handbook
of research methods in public administration, 2nd ed, eds. Kaifeng Yang and Gerald J. Miller,
575–94. Boca Raton, FL: Auerbach Publications.
Franzese, Robert J., Jr., and Judy Hayes. 2007. Spatial-econometric models of cross-sectional interdependence in political science panel and times-series-cross-section data. Political Analysis
15:140–64.
Frees, Edward W. 1995. Assessing cross-sectional correlations in panel data. Journal of Econometrics
69:393–414.
Zhu
Panel Data Analysis in Public Administration
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Gelman, Andew, and Jennifer Hill. 2007. Data analysis using regression and multilevel/hierarchical
models. New York, NY: Cambridge Univ. Press.
Gill, Jeff, and Kenneth J. Meier. 2000. Public administration research and practice: A methodological manifesto. Journal of Public Administration Research and Theory 10:157–99.
Granato, Jim, and Frank Scioli. 2004. Puzzles, proverbs, and omega matrices: The scientific and
social significance of empirical implications of theoretical models (EITM). Perspectives on
Politics 2:313–23.
Greene, William. 2011a. Fixed effects vector decomposition: A magical solution to the problem of
time-invariant variables in fixed effects models? Political Analysis 19:135–46.
———. 2011b. Econometric analysis, 7th ed. Upper Saddle River, NJ: Prentice Hall.
Hacker, Jacob S. 2004. Privatizing risk without privatizing the welfare state: The hidden politics of
social policy retrenchment in the United States. American Political Science Review 98:243–60.
Hausman, Jerry A. 1978. Specification tests in econometrics. Econometrica 46:1251–71.
Hayes, Jude, Aya Kachi, and Robert J. Franzese, Jr. 2010. A spatial model incorporating dynamic,
endogenous network interdependence: A political science application. Statistical Methodology
7:406–28.
Heberlein, Martha, Tricia Brooks, Jocelyn Guyer, Samatha Artiga, and Jessica Stephens. 2011.
Holding steady, looking ahead: Annual findings of a 50-state survey of eligibility rules, enrollment and renewal procedures, and cost sharing practices in Medicaid and CHIP, 2010–2011. The
Henry J. Kaiser Family Foundation, Commission on Medicaid and Uninsured. http://www.kff.
org/medicaid/8130.cfm (accessed May 2011).
Holland, Paul W. 1986. Statistics and causal inference. Journal of American Statistical Association
81:945–60.
Honore, Bo E., and Ekaterini Kryiazidou. 2000. Panel data discrete choice models with lagged
dependent variables. Econometrica 68:839–74.
Hsiao, Cheng. 2003. Analysis of panel data, 2nd ed. New York, NY: Cambridge Univ. Press.
Im, So Kyung, Hashem M. Pesaran, and Yongcheol Shin. 2003. Testing for unit roots in heterogeneous panels. Journal of Econometrics 115:53–74.
James, W., and Charles Stein. 1961. Estimation with quadratic loss. In Proceedings of the Fourth
Berkeley Symposium on mathematical statistics and probability, Volume 1: Contributions to the
theory of statistics, ed. Jerzy Neyman, 361–79. Berkeley, CA: Univ. of California Press.
Keele, Luke, and Nathan J. Kelly. 2006. Dynamic models for dynamic theories: The ins and outs of
lagged dependent variables. Political Analysis 14:186–205.
Kennedy, Peter. 1998. A guide to econometrics. Cambridge, MA: MIT Press.
King, Gary. 1998. Unifying political methodology: The likelihood theory of statistical inference. Ann
Arbor, MI: Michigan Univ. Press.
Levin, Andrew, Chien-Fu Lin, and Chia-Shang James Chu. 2002. Unit root tests in panel data:
Asymptotic and finite sample properties. Journal of Econometrics 108:1–24.
Maddala, Gangadharrao S., and Shaowen Wu. 1999. A comparative study of unit root tests with
panel data and a new simple test. Oxford Bulletin of Economics and Statistics 61:631–52.
Markowitze, Michael A., Marsha Gold, and Thomas Rice. 1991. Determinants of health insurance
status among young adults. Medical Care 29:6–19.
Meier, Kenneth J., and Laurence J. O’Toole, Jr. 2002. Public management and organizational performance: The effect of managerial quality. Journal of Policy Analysis and Management 21:629–43.
———. 2003. Public management and educational performance: The impact of managerial networking. Public Administration Review 63:689–99.
———. 2008. Management theory and Occam’s Razor: How public organizations buffer the environment. Administration and Society 39:931–58.
Mundlak, Yair. 1978. On the pooling of time series and cross section data. Econometrica 46:69–85.
Nelson, David E., Julie Bolen, Henry R. Wells, Suzanne M. Smith, and Shayne Bland. 2004.
State trend in uninsurance among individuals aged 18 to 64 years: United States, 1992–2001.
American Journal of Public Health 94:1992–7.
Nickell, Stephen. 1981. Biases in dynamic models with fixed effects. Econometrica 49:1417–26.
427
428
Journal of Public Administration Research and Theory
Downloaded from http://jpart.oxfordjournals.org/ at Rutgers University on December 1, 2015
Pang, Xun. 2010. Modeling heterogeneity and serial correlation in binary time-series cross-sectional
data: A Bayesian multilevel model with AR(p) Errors. Political Analysis 18:470–98.
Pesaran, Hashem M., and Ron Smith. 1995. Estimating long-run relationships from dynamic heterogeneous panels. Journal of Econometrics 68:79–113.
Phillips, Peter C. B., and Songguy Sul. 2007. Bias in dynamic panel estimation with fixed effects,
incidental trends and cross section dependence. Journal of Econometrics 137:162–88.
Plümper, Thomas, and Vera E. Troeger. 2007. Efficient estimation of time-invariant and rarely changing variables in finite sample panel analyses with unit fixed-effects. Political Analysis 15:124–39.
Plümper, Thomas, Vera E. Troeger, and Philip Manow. 2005. Panel data analysis in comparative
politics: Linking method to theory. European Journal of Political Research 44:327–54.
Primo, David M., Matthew L. Jacobsmeier, and Jeffrey Milyo. 2007. Estimating the impact of state
policies and institutions with mixed-level data. State Politics and Policy Quarterly 4:446–59.
Rosenbaum, Paul R. 2005. Heterogeneity and causality: Unit heterogeneity and design sensitivity in
observational studies. American Statistical Association 59:147–52.
Soss, Joe, Richard Fording, and Sanford F. Schram. 2011. The organization of discipline: From
performance management to perversity and punishment. Journal of Public Administration
Research and Theory 21 (Suppl 2):i203–32.
Stimson, James. 1985. Regression in space and time: A statistical essay. American Journal of Political
Science 29:914–47.
Taylor, William E. 1980. Small sample considerations in estimations from panel data. Journal of
Econometrics 13:2032–223.
Taylor, Mark P., and Lucio Sarno. 1998. The behavior of real exchange rates during the post Bretton
Woods period. Journal of International Economics 46:281–312.
Wagner, Kevin, and Jeff Gill. 2005. Bayesian inference in public administration research: Substantive
differences from somewhat different assumptions. Journal of Public Administration 28:5–35.
Wawro, Gregory. 2002. Estimating dynamic panel data models in political science. Political Analysis
10:24–48.
Westerlund, Joakim. 2007. Testing for error correction in panel data. Oxford Bulletin of Economics
and Statistics 69:709–48.
Whitford, Andrew B., and Jeff Yates. 2009. Presidential rhetoric and the public agenda: Constructing
the war on drugs. Baltimore, MD: Johns Hopkins Univ. Press.
Whitten, Guy, and Laron K. Williams. 2011. Buttery guns and welfare hawks: The politics of defense
spending in advanced industrial democracies. American Journal of Political Science 55:117–34.
Wilson, Sven E., and Daniel M. Bulter. 2007. A lot more to do: The sensitivity of time-series crosssection analysis to simple alternative specifications. Political Analysis 15:101–23.
Wooldridge, Jeffrery M. 2002. Econometric analysis of cross-section and panel data. Cambridge, MA:
MIT Univ. Press.
Zorn, Christopher J. W. 2001. Generalized estimating equation models for correlated data: A review
with applications. American Journal of Political Science 45:470–90.