THE ARTS CHILD POLICY CIVIL JUSTICE EDUCATION ENERGY AND ENVIRONMENT This PDF document was made available from www.rand.org as a public service of the RAND Corporation. Jump down to document6 HEALTH AND HEALTH CARE INTERNATIONAL AFFAIRS NATIONAL SECURITY POPULATION AND AGING PUBLIC SAFETY SCIENCE AND TECHNOLOGY SUBSTANCE ABUSE TERRORISM AND HOMELAND SECURITY TRANSPORTATION AND INFRASTRUCTURE WORKFORCE AND WORKPLACE The RAND Corporation is a nonprofit research organization providing objective analysis and effective solutions that address the challenges facing the public and private sectors around the world. Support RAND Browse Books & Publications Make a charitable contribution For More Information Visit RAND at www.rand.org Explore RAND Health View document details This product is part of the RAND Corporation reprint series. RAND reprints present previously published journal articles, book chapters, and reports with the permission of the publisher. RAND reprints have been formally reviewed in accordance with the publisher’s editorial policy, and are compliant with RAND’s rigorous quality assurance standards for quality and objectivity. M. Susan Marquis Stephen H. Long Effects of ‘‘Second Generation’’ Small Group Health Insurance Market Reforms, 1993 to 1997 In the mid-1990s, several state legislatures enacted a ‘‘second generation’’ of small group health insurance reforms that required guaranteed issue of all products and prohibited the use of health as a rating factor. We use data from two large employer surveys to compare the behavior of small business in nine states that adopted these reforms between 1993 and 1997 to the behavior of small business in 11 states and the District of Columbia, where neither of these small group health insurance market reforms existed prior to 1997 (N ⫽ 8,465 in 1993; N ⫽ 12,219 in 1997). Our analyses focus on several outcomes: health insurance offer and enrollment rates in any employer plan, and in an HMO plan; turnover in offer decisions; and premiums, variability in premiums, and the rate of change in premiums. Overall, we Ànd no effect of small group reform on any of the outcomes; the sign of the effect is not consistent across reform states, the estimates rarely attain statistical signiÀcance, and they show no consistent pattern across the outcomes within each state. Therefore, predictions of the harm these regulations might cause to the market have not come to pass. On the other hand, proponents’ hopes for a solution to low coverage rates among small businesses have not materialized either. Many states enacted insurance market reforms during the 1990s to try to improve access to and affordability of health insurance coverage for employees in small Àrms. These reforms governed rules of issue and how insurers could set premiums for coverage. Reforms related to rules of issue were intended to eliminate insurer underwriting practices that prevented groups or some individuals in a group from purchasing coverage. Such reforms included: requirements that insurers issue a plan to all employers or individuals who wish to purchase it (guaranteed issue); requirements that coverage be renewed (guaranteed renewal); limits on the period during which pre-existing conditions can be excluded from coverage; and exemptions from pre-existing condition ex- clusions for people who change group insurers (portability). By 1997, 47 states had enacted some combination of access reforms (Curtis et al. 1999). Following these independent state efforts, the federal government mandated comprehensive access reforms for all states by enacting the Health Insurance Portability and Accountability Act of 1996 (HIPAA). Rating reforms restrict the extent to which premiums for a given set of beneÀts can vary across groups with different characteristics. Rating reforms are intended to enlarge the risk pool on which premiums for small employers are based, thus making insurance more affordable for high-risk groups and encouraging them to purchase insurance. On the other hand, small employers that have a good risk proÀle may pay high- M. Susan Marquis, Ph.D., and Stephen H. Long, Ph.D., are senior economists at RAND. This research was supported by the Robert Wood Johnson Foundation (grants no. 026935 and 028651) and the National Center for Health Statistics (contract no. 0009930281). Address correspondence to the authors at RAND, 1200 S. Hayes St., Arlington, VA 22202-5050. Inquiry 38: 365–380 (Winter 2001/2002). 䉷 2001 Blue Cross and Blue Shield Association and Blue Cross and Blue Shield of the Rochester Area. 0046-9580/01/3804–0365$1.25 365 Inquiry/Volume 38, Winter 2001/2002 er rates after reforms are enacted because insurers will be limited in the extent to which they can take into account this favorable proÀle. The federal government did not address rate regulation under HIPAA; hence, this type of reform is solely a state option. The effect that insurance reforms have on the small group market is theoretically ambiguous. If underwriting restrictions and rating reforms attract high-risk groups into the market, then average premiums may rise and some businesses may drop coverage. Some opponents of these reforms fear that they will drive the small group market into a ‘‘death spiral,’’ thereby contributing to rising rates of uninsured (see for example Curtis et al. 1999; Blumberg and Nichols 1996; Thorpe 1992). However, proponents point out that reforms may encourage greater price competition by limiting the extent to which insurers can compete on the basis of risk selection (Buchmueller and Jensen 1997). This may lead to lower prices and more coverage. In this paper, we examine the effect of guaranteed issue provisions and rating reforms enacted in the mid1990s on the extent and nature of employer-sponsored insurance for small groups. We extend the previous literature by focusing on a time during which a number of states moved to tighten rating reforms and to require guaranteed issue of all insurance products. We also examine a wider range of potential outcomes of reforms than most of the earlier literature. The continuing problem of the uninsured has led to new proposals to try to combat it. Some call for greater regulation of insurance markets (Herzlinger 2000). Others call for subsidies to private coverage, but observe that the success of these efforts will depend on having well-functioning insurance markets (Bilheimer and Colby 2001). Thus, understanding the role that market reforms have played is important to the design of future policy intended to increase the proportion of insured workers in small Àrms. Literature Review The empirical evidence on the effect of small group insurance market reforms is somewhat ambiguous. Early case studies of reforms in speciÀc states found little change in coverage overall, though some states experienced improvements in coverage while others experienced a loss (Chollet and Paul 1995; Wessner 1994; Institute for Health Policy Solutions 1995). More recent analyses have used primarily two types of data. One set of studies has used data from the Current Population Survey (CPS) to investigate how the number of uninsured and the number cov366 ered by private insurance of all types varies as market reforms vary across states or time or, in most cases, both. Most of these studies investigated changes in the uninsured rate or private coverage rate either for the entire nonelderly population or for all adults. This potentially understates the effect of reforms on the smaller target population of employees in small business. Custer (1998), using data from the 1998 CPS, concluded that both guaranteed issue and rating reforms led to increases in the number of uninsured. However, this conclusion was based on differences in uninsured rates in a cross-section of states and may reÁect geographic differences other than market regulation. Several analysts have pooled data from multiple years of the CPS to control for state-speciÀc effects in the uninsured rate in analyzing the role of reforms. Two of these studies, one for 1989–94 and a second for 1989–95, found that neither access nor rating reforms in the small group market affected the uninsured rate (Sloan and Conover 1998; Zuckerman and Rajan 1999).1 In contrast, a third study using data from 1989–95 concluded that guaranteed issue regulations and rating reforms had statistically signiÀcant but countervailing effects on the uninsured rate (Marsteller et al. 1998).2 Guaranteed issue was associated with a small decrease of about 1 percentage point in the uninsured rate, which was offset when rating reforms were also in place. However, Buchmueller and DiNardo (1998) concluded that strict community rating reforms do not affect uninsured rates among all adults or among workers in small Àrms; their conclusion was based on a comparison of trends in CPS uninsured rates in a state enacting community rating (New York), with trends in a state with no reforms (Pennsylvania) and a state with only modest reforms (Connecticut). Another strand of literature has analyzed employer data to investigate the effect of insurance market reforms on employer decisions to offer insurance. In contrast to the population analyses, the employer-level analyses focus on the target population of small business. These studies, too, have produced mixed results. Hing and Jensen (1999) used variation among states in the insurance market reforms that had been enacted prior to 1993 to study access and rating reforms. They concluded that the probability of a small Àrm offering insurance was about two to three percentage points higher in states that had reforms in place for several years than in states that had not enacted reforms by 1993. As the authors observe, however, states that had better functioning markets may have been states to enact reforms early, and this ‘‘Second Generation’’ Reforms self-selection may account for their cross-section results. Jensen and Morrisey (1999) pooled time-series and cross-sectional data on small employers to analyze reforms enacted over the period 1989 to 1995. Overall, they did not Ànd statistically signiÀcant effects of access reforms. However, sample size—and hence power—was small and point estimates suggested that reforms increased small Àrm offer rates by about two to seven percentage points. Similarly, Simon (1999) reports small and generally insigniÀcant effects of reforms enacted between 1993 and 1997 on employer offer rates.3 Buchmueller and Jensen (1997) also report that more small Àrms in California offered health insurance following comprehensive reforms that went into effect in 1993; this result was especially pronounced among Àrms with fewer than 10 employees. However, similar patterns were not found in other states enacting reforms during the study period. Moreover, an increase in offer rates of similar magnitude was seen among very small Àrms in states that had enacted reforms prior to the study period, suggesting temporal trends might be at work. In sum, the preponderance of evidence suggests that small group market reforms have had little effect on the decisions of small employers to offer insurance. However, many of these previous studies looked at reforms enacted in the early ’90s. Subsequently, a number of states moved to tighter issue and rating reforms, which might have a greater effect than the ‘‘Àrst generation’’ of reforms. In addition, most of the earlier studies did not differentiate the extent of reform—for example, guaranteed issue for all products vs. only some standard products, or tight vs. loose rating restrictions. As a consequence, they might underestimate the effect of stringent reform. A contribution of our research is that we used a detailed database describing small group insurance market regulations in each state during the period 1990 to 1997 to differentiate the extent of reform.4 Also, we studied reforms adopted between 1994 and 1996, which was a period when states were moving toward tighter rating regulations—especially prohibition of the use of health as a rating factor. In contrast, rating reforms covered in many of the earlier studies permitted the use of health ratings, though often limiting the permissible range of variation associated with health differences among groups (Curtis et al. 1999). Finally, we go beyond an exploration of the effect of the reforms on employer offer rates and employee enrollments to examine a number of other hypotheses about reforms, as described later. Data and Methods Data Our data are from the National Employer Health Insurance Survey (NEHIS) and the 1997 Robert Wood Johnson Foundation (RWJF) Employer Health Insurance Survey. The NEHIS, sponsored by the National Center for Health Statistics, completed interviews with 34,604 private employers with one or more employees in 1994. The nationwide sample was designed to provide reliable state-level estimates of employers and employees.5 The 1997 RWJF survey, conducted by RAND and the Research Triangle Institute, interviewed 21,545 private employers in the continental United States. The sample was concentrated in the 60 communities followed by the RWJF Community Tracking Study and in states having signiÀcant small group rating reforms. These cases were supplemented by a sample from the remainder of the continental United States to better represent the nation’s business establishments.6 The sampling frame for both surveys was Dun’s Market IdentiÀers national census of employment establishments. The NEHIS sample was stratiÀed by state, and within state by Àrm and establishment size. (Firm size refers to the number of employees at all locations of a business nationwide; establishment size refers to employment at a single location of a business.) The RWJF survey was stratiÀed by geographic area (as deÀned previously) and by size of the establishment. Sampled establishments in each study are weighted to account for different sampling probabilities and for nonresponse. The weighted samples in both surveys represent all private employment establishments that have at least one employee.7 The NEHIS survey included establishments in all 50 states and the District of Columbia, whereas the RWJF covered establishments in the continental United States. We restricted our analysis of the NEHIS to the geography covered by the RWJF survey. The distribution of weighted establishments by establishment size, Àrm size, and industry was similar across the surveys. In addition, as we describe next, our analysis controls for employer characteristics and, hence, for any small differences between the realized samples. The surveys used similar questions and deÀnitions. Both collected information from employers using computer-assisted telephone interviews (CATI). We subjected the databases to the same algorithms to edit data for consistency and outlier values, and to impute missing data. Both surveys were conducted with the person or persons in each establishment most knowl367 Inquiry/Volume 38, Winter 2001/2002 edgeable about health beneÀts and business and worker characteristics. The NEHIS survey was administered during 1994 and asked respondents to report characteristics of coverage as of the end of 1993. The RWJF survey was administered during 1997 and employers were asked about coverage as of the date of the interview. The response rate to the NEHIS was 71%; for the RWJF survey, it was 60%. These are comparable to response rates for two other recent large-scale surveys of employer health insurance (Gabel, Ginsburg, and Hunt 1997; Branscome et al. 2000). The surveys collected information about the employer, the number of workers, and the distribution of key characteristics of these workers. They also collected detailed information about every health insurance plan offered by the business, including the number of employees enrolled in the plan, premiums, and the type of plan. In addition, the RWJF survey asked about the percentage change in premiums across all offered plans that each employer experienced between 1996 and 1997. The establishment is the unit of observation for most analyses in this paper. However, the numbers of employees and health plan enrollees at each establishment also allow us to develop sampling weights to make estimates for employees and enrollees for some analyses. Estimation Strategy Our analysis of the effects of reforms focuses on guaranteed issue requirements and rating reforms that prohibit the use of health status as a factor. We do this because these represent the ‘‘second generation’’ of reforms—few states required guaranteed issue of all products nor prohibited rating based on health status prior to 1993.8 We examine nine states that enacted these reforms between 1993 and 1996. Thus, the NEHIS data from 1993 provide outcomes prior to reforms, and the RWJF survey data from 1997 provide outcomes subsequent to reforms.9 Six of the reform study states had neither guaranteed issue nor health rating prohibitions prior to 1993 but added one of these provisions to law between 1993 and 1996. Three of the states required that some products be guaranteed issue prior to 1993, but enacted tighter issue regulations or health rating prohibitions between 1993 and 1996.10 We distinguish between states that prohibited health as a rating factor with little or no restriction on the use of age (high and low premiums based on age were allowed to differ by more than 100%, or a two-to-one band), and those that prohibited health and imposed a prohibition or tight restriction on the use of age as well. 368 To control for temporal effects, we compare outcomes for small employers in the nine reform states with outcomes in a group of 11 states and the District of Columbia where the reforms were not enacted.11 These states are Alabama, Arkansas, Georgia, Illinois, Indiana, Louisiana, Michigan, Pennsylvania, New Mexico, Nevada, and West Virginia.12 These 12 areas did not provide for guaranteed issue, nor did they prohibit health rating prior to 1997.13 Table 1 summarizes the provisions in the nine reform states, indicates the dates of reform, and presents the sample sizes for our study in each state and the comparison group. We deÀne small employers to be those with 50 or fewer workers, because this is the deÀnition adopted in eight of the nine states. We measure the effect of the regulations in three ways. First, we compare outcomes in the small group market in 1997 in each of the reform states with outcomes in the 12 nonreform areas taken as a group. However, these cross-section comparisons are biased by self-selection if there are market characteristics that affect states’ decisions to enact regulations and affect market outcomes. To control for such statespeciÀc effects, we examine changes in outcomes between the pre-reform period (1993) and the post-reform period (1997), and test whether changes over the period in each of the reform states differ from those in the control group. The latter change represents a control for temporal factors. This is typically referred to as a ‘‘difference-in-differences’’ estimate. The difference-in-differences estimate accounts for time-invariant, state-speciÀc effects; however, there may be state-speciÀc changes over time unrelated to regulations that could bias this comparison. To control for state-speciÀc temporal effects, we assume that these factors affect medium as well as small employers in the state. (We deÀne medium employers to be those with 51 to 150 employees.) We then can compare the change in outcomes over time between small employers and medium employers in reform states with the change between small and medium employers in the control states. This is referred to as a ‘‘difference-indifference-in-differences’’ estimate. Formally, we Àt the following model: Y ⫽ b0 ⫹ b1 ·YR97 ⫹ b2 ·SMALL ⫹ b3 ·STATE ⫹ b4 ·(STATE·YR97) ⫹ b5 ·(SMALL·STATE) ⫹ b6 ·(SMALL·YR97) ⫹ b 7 ·(SMALL·STATE·YR97) ⫹ b8 ·X ⫹ ⑀. (1) Table 1. Small group reform provisions and sample sizes, 1993 and 1997 12 Nonreform statesa Reform provisions 1993 provisions Guarantee issue Rating: health variation Rating: age variation restricted to 100% spread 1997 provisions Guarantee issue Rating: health variation Rating: age variation restricted to 100% spread Date of guarantee issue legislation Date of rating legislation Largest size group affected MD None Allowed None Allowed None Allowed All Prohibited Restricted NJ NY WA CA MN FL OR CT None None None None Some Some Some None Allowed Allowed Allowed Allowed Allowed Allowed Allowed Allowed Not restricted Not restricted Not restricted Not restricted Not restricted Not restricted Not restricted Not restricted Not restricted Not restricted All Prohibited Restricted All Prohibited Prohibited Not restricted NA Jul-94 Jan-94 Apr-93 Not restricted Not restricted Not restricted Not restricted Jul-94 Jul-93 Jul-93 Apr-94 Oct-96 Not restricted NA NA NA Jul-95b 50 Jan-96 49c Apr-93 50 Jan-96 50 1993 sample sizes Groups size 1–50 Groups size 51–150 4,436 1,028 375 85 434 100 571 108 426 97 624 115 1997 sample sizes Groups size 1–50 Groups size 51–150 2,188 440 295 568 97 1,336 193 828 110 1,433 170 d All Prohibited All Allowed NA 50 All Allowed NA 50 All Prohibited All Prohibited Restricted Some Prohibited Jan-94 50 Mar-93 25 Jul-95 50 412 114 473 101 368 89 346 76 1,435 134 1,422 141 1,461 166 1,253 112 369 ‘‘Second Generation’’ Reforms Source: National Employer Health Insurance Survey, 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey, and Institute for Health Policy Solutions database of small group reform provisions. Note: NA ⫽ not applicable. a Alabama, Arkansas, District of Columbia, Georgia, Illinois, Indiana, Louisiana, Michigan, Pennsylvania, New Mexico, Nevada, West Virginia. b Maryland enacted prohibitions on health as a rating factor, but with limited restrictions on age effective July 1994. Tighter restrictions on age variation became effective on the date listed. c Increased to 50 in October 1997. d Fewer than 50 observations. Inquiry/Volume 38, Winter 2001/2002 The variable YR97 is an indicator that takes the value 1 for observations from the 1997 RWJF survey and 0 for those from the NEHIS. SMALL is an indicator that takes the value 1 for businesses that are parts of Àrms with 50 or fewer employees and 0 for businesses of Àrms that have 51 to 150 employees. STATE represents nine indicator variables included in the regression for the reform states. For example, there is one variable that takes the value 1 for employers in Maryland and 0 otherwise; similar variables are included for employers in each of the other reform states. The vector X represents other characteristics of the employer included to control for differences over time and across states in the composition of employers. We deÀne these control variables subsequently. The cross-section estimate of the effect of reform is given by b2 ⫹ b4 ⫹ b6 ⫹ b7; the difference-in-differences estimate is b6 ⫹ b7; and the difference-in-difference-in-differences estimate is b7. We Àt a model of this form for each of the outcomes described later. We use the Àtted model to predict outcomes for each reform state in each year and for the grouped control states in each year assuming that the employer composition is the same in all states and over time.14 We do this by predicting outcomes for each employer in our RWJF study sample as if it were observed in the state and year in question. That is, our RWJF sample is used to deÀne the distribution of characteristics, X, for the small- and medium-size employers in our prediction sample. For each of these employers, we use the Àtted equation given earlier to predict what the outcome would be in 1993 if the employer were in one of the nonreform states, by setting the value for all the nonreform state indicators to 0 and for YR97 to 0. We then set the YR97 indicator to 1 and predict the 1997 outcome for employers in the nonreform states. Similarly, by varying the values of the indicators for the reform states we predict the 1993 and 1997 outcomes for a constant mix of employers in each of these states. We report the mean of the predicted values for small employers for each year and state in the tables to follow. The difference-in-differences and difference-in-difference-in-differences measures based on the predicted values are also given. Predicted values for medium size employers, used in the difference-in-differencein-differences, are reported in the Appendix Table. Control Measures In addition to the state, year, and size indicators, our regression models include other employer character370 istics to control for differences among states in the composition of employers and employees. These characteristics include: industry; detailed indicators for the Àrm and establishment size; the percentage of low-income workers at the establishment (workers earning less than $7 per hour); the percentage of union employees; the percentage of employees working full time (35 hours or more per week); and the age of the business. We also include indicators for region in our analysis, so that our comparison between a reform state and the pool of contrast states is not confounded by regional differences. Outcome Measures The outcomes that we examine include insurance offer and enrollment rates in any employer-sponsored health plan and in a health maintenance organization (HMO) plan; turnover in offer decisions; and premiums, variability in premiums, and the change in premiums. We examine both the percentage of employers offering health insurance and the percentage of employees in Àrms offering insurance. The variable in each case is the dichotomous response to the question posed to employers as to whether they offer health insurance to employees. When we use establishment weights in the analysis, the result is an estimate of the percentage of employers offering insurance. Using employee weights, we produce an estimate of the percentage of employees in businesses that offer insurance. Because the outcome is dichotomous, we Àt logistic regression models using the variables shown in equation 1. We also investigate the percentage of all employees who are enrolled in their employer’s health insurance plan. This is a continuous variable and we Àt a linear regression model as shown in equation 1 for analysis of this outcome. Some analysts believe that small group insurance reforms may increase the role of HMOs in the small group market. Many HMOs use community rating and do not limit enrollment on the basis of health factors. Reforms that prohibit these practices by commercial insurers may, therefore, make the small group market more attractive to HMOs. Some support for the hypothesis has been demonstrated in other studies (Buchmueller and Jensen 1997; Morrisey and Jensen 1997). To explore this hypothesis, we Àt a logistic regression model to examine the percentage of insurance-offering employers that offer an HMO as one of the plans and the percentage of enrollees who are in an HMO plan. Each health insurance plan offered was classiÀed as an HMO or other type of plan based on the survey ‘‘Second Generation’’ Reforms respondent’s self-assessment of the plan type, aided by complete deÀnitions as needed. Rating reform is intended to lower the price of insurance for high-risk groups, but, as we noted earlier, the effect on average premiums depends on whether the reform results in cross-subsidies that drive low-risk groups from the market (Curtis et al. 1999; Thorpe 1992). The two employer surveys measured premiums for single coverage and for family coverage for each health insurance plan offered by employers in the sample. To account for differences in the scope and depth of beneÀts of different insurance plans, the premiums for each plan were adjusted by an estimate of the actuarial value of the plan beneÀts. The actuarial value for a plan ranges between 0 and 1 and represents the share of expenditures on health care by a standardized group of enrollees that the plan would reimburse. Using the 1987 National Medical Expenditure Survey, we generated the distribution of expenditures for a representative crosssection of the population under age 65. The 1987 expenditures were adjusted to total spending for 1993 and 1997 using the aggregate spending in each year from the National Health Accounts, developed by the Health Care Financing Administration. We then used information about plan beneÀts from the employer surveys to determine the amount that each plan would reimburse.15 We adjusted premiums by multiplying them by the ratio of the average actuarial value for all plans to the plan-speciÀc actuarial value. We regressed adjusted premiums on the state, year, and size indicators and the control characteristics described earlier to investigate the effect of the market reforms on premium levels. We added type of plan (HMO, preferred provider organization or point-ofservice, indemnity) as a control variable for the premium regressions. The unit of analysis in the premium regressions is the insurance plan. In Àtting the regression, we used establishment sampling weights multiplied by the share of the employer’s total enrollment in the plan. For businesses offering more than one plan, therefore, this effectively examines the average single premium and the average family premium for policies purchased by the employer.16 Rating reforms restrict the range of variation in premiums that an insurer is permitted to charge in the small group market. Thus, they were designed to reduce variability in premiums among employers and to promote greater stability over time in premiums faced by an employer. To examine the effects of reforms on variability in premiums over time, our outcome measure is the residual variance in plan pre- miums after accounting for the variance attributable to the control factors (regression variance). That is, it is the residual variance from the premium regression, measured as the square of the regression residual, for each establishment’s insurance plan. We divided the residual variance in premiums by the squared value of the mean premium for employers in the state, year, and size group. That is, our measure is essentially the coefÀcient of variation squared (called the relvariance) (Kish 1967). It is a measure of variance relative to the level of prices to account for differences in prices over time and geography that may affect the absolute value of the variance.17 To assess the effects on stability in premiums over time, we study the percentage of employers reporting premium changes of greater than 10% (either increase or decrease) between 1996 and 1997 in the RWJF survey. Absent restrictions on insurer pricing, small employers can experience extreme rate variation which, in turn, could result in high turnover of coverage (IHPS 1995). If reforms are successful in limiting the year-to-year variation in premiums that small employers face, then there may be greater stability in employers’ decisions to offer insurance (Buchanan and Marquis 1999). Employers interviewed in the 1997 RWJF survey were asked if they offered health insurance two years previously as well as currently. We used these questions to measure whether the employer made a change in whether to offer coverage. This variable is used to assess whether rating reform affects stability in the market. Because this measure comes from one point in time, our model does not include year indicators and the interaction of year and other characteristics. Limitations Our data come from two very large surveys and so provide an ample number of observations to detect effects of reform, if they exist. Nonetheless, there are some limitations of our data that may hamper our ability to detect effects. First, our study period may be too short to detect the full effects of small group reform, because it may take some time for insurers and employers to adjust to the regulations once they are implemented. Second, we study a period during which there was considerable activity in states to introduce reforms in the small group market. As a result, our control group was limited to a dozen states, omitting a broad middle group from the analysis. Finally, there is the potential for selection bias in our comparison. Many of the states that we did not study adopted some regulations on rating and guar371 12 Nonreform states MD NJ NY WA CA MN FL OR CT 33.5 38.7 ⫺2.2 6.6* 7.6 37.9 32.6 ⫺8.3* ⫺3.9 ⫺.8 48.2 41.5 .6 ⫺5.3 ⫺5.9 39.7 36.9 ⫺4.0 ⫺1.4 ⫺.5 39.6 44.2 3.3 6.0 10.1 40.5 44.1 3.2 5.0 3.4 45.3 48.4 7.5* 4.5 8.4 42.3 39.2 ⫺1.7 ⫺1.7 5.7 43.4 46.6 ⫺11.1* 3.8 23.7* 50.3 43.8 ⫺13.9* ⫺5.9* ⫺18.1 54.6 60.0 2.3 6.0 7.1 57.1 55.4 ⫺2.3 ⫺1.1 ⫺2.5 57.2 59.4 1.7 2.8 6.0 53.9 59.5 1.8 6.2* 8.9 62.7 66.3 8.6* 4.2 4.8 54.8 50.9 ⫺6.8* ⫺3.3 7.7 34.7 38.7 ⫺.3 3.0 2.1 37.1 33.6 ⫺5.4* ⫺4.5* ⫺20.8* 45.4 43.3 4.3 ⫺3.1 ⫺8.0 38.4 37.3 ⫺1.7 ⫺2.1 ⫺1.4 34.9 39.0 0.0 3.1 1.0 33.5 40.2 1.2 5.7* 1.0 42.5 47.1 8.1* 3.6 ⫺3.5 39.4 38.9 ⫺.1 ⫺1.5 ⫺4.1 Percentage of employers offering insurance 1993 1997 1997 difference Difference in differences Difference in difference in differencesa 42.3 40.9 NA NA NA 43.6 45.9 5.0 3.7 Percentage of employees in Àrms offering insurance 1993 1997 1997 difference Difference in differences Difference in difference in differencesa 58.3 57.7 NA NA NA 60.8 60.1 2.4 ⫺.1 38.0 39.0 NA NA NA 36.5 36.2 ⫺2.8 ⫺1.3 b b Percentage of employees enrolled 1993 1997 1997 difference Difference in differences Difference in difference in differencesa b Source: National Employer Health Insurance Survey, 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey, and Institute for Health Policy Solutions database of small group reform provisions. Notes: ‘‘1997 difference’’ compares outcomes in each reform state to those in the 12 nonreform states. ‘‘Difference in differences’’ compares the 1993 to 1997 change in outcomes in each reform state to the change in the 12 nonreform states. ‘‘Difference in difference in differences’’ compares difference-in-differences measures for employers with 50 or fewer workers to those with 51 to 150 workers. These measures are not applicable, ‘‘NA,’’ to the nonreform state control group. a Predicted values for 1993 and 1997 for employers with 51 to 150 workers are given in the Appendix. b Fewer than 50 observations. * SigniÀcantly different at .05 level. Inquiry/Volume 38, Winter 2001/2002 372 Table 2. Insurance offer and enrollment rates for small business, 1993 and 1997 Source: 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey and Institute for Health Policy Solutions database of small group reform provisions. Notes: ‘‘Difference’’ compares turnover in each reform state to that in the 12 nonreform states. ‘‘Difference vs. difference for medium size groups’’ compares the difference for employers with 50 or fewer workers to that for employers with 51 to 150 workers. a Predicted values for 1993 and 1997 for employers with 51 to 150 workers are given in the Appendix. b Fewer than 50 observations. * SigniÀcantly different at .05 level. 18.0 3.3 5.0 8.5 ⫺6.2* ⫺.8 16.6 1.9 ⫺3.7 b Percentage of employers changing offer decision 1995 to 1997 Difference Difference vs. difference for medium size groups 14.7 NA NA 16.8 2.1 15.7 1.0 3.1 17.0 2.3 3.5 10.5 ⫺4.2 ⫺.1 9.3 ⫺5.4 ⫺2.0 16.5 1.8 ⫺2.4 OR WA NY NJ MD 12 Nonreform states Table 3. Turnover in offer rates for small business by state reform provisions, 1995 and 1997 CA MN FL CT ‘‘Second Generation’’ Reforms anteed issue during the study period, though not as stringent as the reforms we investigated. Some of our study states adopted less stringent reforms as interim steps toward the reforms we studied. States that introduced the ‘‘second generation’’ reforms may be those that found earlier reforms did no harm; in contrast, states that did not move forward may have had less positive experience with interim reforms. That is, if reforms were adopted by all states, the effects might have differed from those we observed. Results Our evidence from the employer surveys suggests that the reforms did not affect coverage (Table 2). There are few statistically signiÀcant Àndings. Although the cross-section estimate, the difference-indifferences estimate, and the difference-in-differencein-differences estimate often suggest the same direction of effect for a state, there is not a consistent direction of effect across the states. Four of the reform states introduced guaranteed issue for all products and rating reforms (Maryland, New Jersey, New York, and Washington). The crosssection estimates of the effect of reform are mostly statistically insigniÀcant; they suggest positive effects in two states, but negative effects in the other two states. Similarly, the difference-in-differences estimate and the difference-in-difference-in-differences estimates are generally statistically insigniÀcant and yield mixed signs for these four states.18 Two of the states introduced access reforms but did not prohibit health as a rating factor (California and Minnesota). There are no statistically signiÀcant differences between these states and the control states in our three estimates for the percentage of employers offering insurance. Furthermore, the direction of effect is positive in one state and negative in the other. Finally, the other three states (Florida, Oregon, and Connecticut) introduced health rating prohibitions and, in two of the states, strengthened access reforms as well. Again, comparison with the control states suggests statistically insigniÀcant differences, and the direction of effects is of mixed sign. Similar patterns are found when we examine the percentage of employees offered insurance and the percentage of employees enrolled in insurance.19 Previous research has suggested that the effects of small group reform might differ for the smallest of small businesses.20 To test for this, we added to our model of offer rates interaction terms for Àrms with fewer than 10 workers. For each state, the direction of effect was similar for very small (⬍10 workers) 373 12 Nonreform states Percentage of employers offering HMO (if offer insurance) 1993 1997 1997 difference Difference in differences Difference in difference in differencesa Percentage of enrollees in HMOs 1993 1997 1997 difference Difference in differences Difference in difference in differencesa MD NJ NY WA CA MN FL OR CT 14.0 30.0 NA NA NA 37.0 58.5 28.5* 5.5 5.6 35.8 5.8 14.2* 29.8* 21.0 43.7 13.7* 6.7 5.1 21.1 33.9 3.9 ⫺3.2 7.2 44.7 57.2 27.2* ⫺3.5 ⫺8.5 23.5 46.1 16.1* 6.6 18.1 21.4 53.0 23.0* 15.6* 20.8 40.0 57.6 27.6* 1.6 4.2 13.9 37.2 7.2 7.3 24.2 15.2 28.9 NA NA NA 32.9 56.2 27.3* 9.6 8.6 33.8 4.9 11.5* 15.5 18.6 43.2 14.3* 10.9 5.5 18.6 22.5 ⫺6.4 ⫺4.8 ⫺13.5 43.5 53.6 24.7* ⫺3.6 13.0 31.1 52.9 24.0* 8.1 ⫺.6 25.6 56.2 27.3* 16.9* 29.6* 37.1 55.5 26.6* 4.7 16.1 13.7 36.4 7.5 9.0 33.9* b b Source: National Employer Health Insurance Survey, 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey, and Institute for Health Policy Solutions database of small group reform provisions. Notes: ‘‘1997 difference’’ compares outcomes in each reform state to those in the 12 nonreform states. ‘‘Difference in differences’’ compares the 1993 to 1997 change in outcomes in each reform state to the change in the 12 nonreform states. ‘‘Difference in difference in differences’’ compares difference-in-differences measures for employers with 50 or fewer workers to those with 51 to 150 workers. These measures are not applicable, ‘‘NA,’’ to the nonreform state control group. a Predicted values for 1993 and 1997 for employers with 51 to 150 workers are given in the Appendix. b Fewer than 50 observations. * SigniÀcantly different at .05 level. Inquiry/Volume 38, Winter 2001/2002 374 Table 4. Offer and enrollment rates for HMOs for small business, 1993 and 1997 Table 5. Monthly premiums and variation in premiums for small business, adjusted for plan beneÀts, 1993 and 1997 12 Nonreform states MD Premiums for employee-only coverage ($) 1993 1997 1997 difference Difference in differences Difference in difference in differencesa 159 172 NA NA NA 170 166 ⫺6 ⫺17 Premiums for family coverage ($) 1993 1997 1997 difference Difference in differences Difference in difference in differencesa 392 404 NA NA NA 430 428 24* ⫺14 c c Relvariance in premiums for employee-only coverage 1993 1997 1997 difference Difference in differences Difference in difference in differencesa 20.7 17.4 NA NA NA 15.7 8.7 ⫺8.7* ⫺3.7 Relvariance in premiums for family coverage 1993 1997 1997 difference Difference in differences Difference in difference in differencesa 11.2 12.9 NA NA NA 11.0 6.5 ⫺6.4* 6.2* NJ NY WA CA MN FL OR CT 192 205 33* 0 27 196 201 29 ⫺8 5 144 158 ⫺14 1 8 169 170 ⫺2 ⫺12 17 138 150 ⫺22* ⫺1 14 161 170 ⫺2 ⫺4 ⫺3 130 156 ⫺16* 13 21 209 214 42* ⫺8 12 452 474 70* 10 27 446 432 28* ⫺26* ⫺28 387 384 ⫺20 ⫺15 ⫺13 416 413 9 ⫺15 ⫺5 376 380 ⫺24 ⫺8 ⫺14 406 436 32* ⫺18 ⫺14 341 384 ⫺20 31* 11 485 503 99* 6 ⫺2 30.0 22.0 4.6 ⫺4.7 ⫺3.5 19.3 16.1 ⫺1.3 .1 2.0 16.8 15.0 ⫺2.4 1.5 4.1 18.7 13.5 ⫺3.9 ⫺1.9 ⫺5.1 17.9 13.5 ⫺3.9 ⫺1.1 7.3 16.9 18.8 1.4 5.2 9.7 20.9 22.5 5.1 4.9 4.8 14.6 16.8 3.9 .5 ⫺1.3 17.9 16.8 3.9 ⫺2.8 ⫺3.5 8.1 9.2 ⫺3.7 ⫺.6 ⫺3.2 11.7 11.7 ⫺1.2 ⫺1.7 ⫺5.8 9.6 8.4 ⫺4.5* ⫺2.9 ⫺1.7 7.3 9.5 ⫺3.4 .5 ⫺2.4 15.6 14.0 1.1 3.3 ⫺5.2 b c 19.1 25.8 8.4* 10.0* 16.6 b c 9.8 13.3 .4 1.8 ⫺6.0 375 ‘‘Second Generation’’ Reforms Source: National Employer Health Insurance Survey, 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey, and Institute for Health Policy Solutions database of small group reform provisions. Notes: Observed premiums have been adjusted to account for differences in plan beneÀts (see ‘‘Outcome Measures’’ in text). ‘‘1997 difference’’ compares outcomes in each reform state to those in the 12 nonreform states. ‘‘Difference in differences’’ compares the 1993 to 1997 change in outcomes in each reform state to the change in the 12 nonreform states. ‘‘Difference in difference in differences’’ compares difference-in-differences measures for employers with 50 or fewer workers to those with 51 to 150 workers. These measures are not applicable, ‘‘NA,’’ to the nonreform state control group. a Predicted values for 1993 and 1997 for employers with 51 to 150 workers are given in the Appendix. b Residual variance from premium level regression divided by the square of the mean premium, times 100. c Fewer than 50 observations. * SigniÀcantly different at .05 level. 376 Source: 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey and Institute for Health Policy Solutions database of small group reform provisions. Notes: ‘‘Difference’’ compares the rate of large premium change in each reform state to that in the 12 nonreform states. ‘‘Difference in differences’’ compares the difference for Àrms with 50 or fewer workers to that for Àrms with 51 to 150 workers. a Fewer than 50 observations. * SigniÀcantly different at .05 level. 16.4 .4 ⫺1.7 18.9 2.4 7.0 28.7 12.2* 7.2 19.8 3.3 1.8 13.0 ⫺3.5 .8 22.4 5.9 ⫺9.1 8.8 ⫺7.7 ⫺12.3* 18.2 1.7 3.5 a OR FL MN CA WA NY NJ MD 15.3 ⫺1.2 16.5 NA NA Percentage of employers Difference Difference in differences 12 Nonreform states Table 6. Small Àrms with premium changes of more than 10% between 1996 and 1997 or other small (10–50 workers) employers (not shown). However, again there are few statistically signiÀcant Àndings and no consistent direction of effect across the states. We also do not Ànd evidence that reforms have reduced turnover in offers among employers in the small group insurance market (Table 3). In cross-section, eight of the nine differences are not statistically signiÀcant; moreover, the point estimates would suggest turnover is greater (not less) in six of the reform states. Using trends among medium employers as a control, we also Ànd insigniÀcant differences between the reform states and the control states, and the results are of mixed sign.21 HMO penetration in the small group market was generally greater in the reform states than in the nonreform states, both in 1993 and in 1997 (Table 4). The 1997 difference estimates for HMO offers are all positive and many are signiÀcant. Overall, the likelihood that an employer offered an HMO in 1997 was about 18 percentage points higher in the reform states than the nonreform states; enrollments in the HMOs in reform states were also about 18 percentage points higher. These differences are statistically signiÀcant, but this could be attributed as easily to the higher initial offer rates as it could to reform. The difference-in-differences estimate and the difference-indifference-in-differences estimate for HMO market penetration are positive in most reform states, although few are statistically signiÀcant. Averaged over all the reform states these are also positive, but not statistically signiÀcant. At best, this offers modest support for the hypothesis that reforms advanced opportunities for HMOs in the small group market. The reforms do not appear to have had a substantial effect either on the level or on the variability of premiums adjusted for plan beneÀts in the small group market (Table 5).22 In cross-section, we Ànd higher premiums in some reform states than in the control states, but lower premiums in others. This is a weak test, however, because the 1997 differences primarily reÁect geographic price differences that are also observed in 1993. The stronger tests, the difference-in-differences estimate and the difference-indifference-in-differences measure, are generally insigniÀcant and of mixed signs.23 Moreover, there is not evidence of less variability in premiums among small employers in reform states. In fact, in most measures, we obtain almost as many positive point estimates as negative ones and very few of the estimates are statistically signiÀcant. Similarly, we do not Ànd fewer small employers who faced large pre- CT Inquiry/Volume 38, Winter 2001/2002 ‘‘Second Generation’’ Reforms mium changes in either direction in reform states (Table 6). Very few of the differences between reform and control states are statistically signiÀcant, and point estimates are of mixed sign. Buchmueller and Jensen (1997) found that the upper tail of the premium distribution changed more than the lower tail after reform in California. We tested for this effect by measuring change in the percentage of premiums that exceeded 150% of the mean premium, both for employee-only and family coverage. There was no support in our data for this hypothesis. The 1997 difference measure for the density of the upper tail in the reform states in 1997 was positive 10 times and negative eight times across both types of premiums. Both the difference-in-differences measure and the difference-in-difference-in-differences measure were of mixed signs, all were insigniÀcant, and the measures were often inconsistent within a state for the two premium types (not shown). Discussion Our study adds to the recent literature on the effects of small group health insurance reform in four ways. First, it addresses the second generation of reforms, which were implemented more recently and were more stringent than the original reforms that several previous studies have addressed. Second, in choosing our study states we relied on measures of rate regulation that are more precise than those used by other researchers, who have used general secondary sources.24 Third, by limiting our comparisons to states with the most stringent reform provisions and those that had no reform, our tests were designed to Ànd an effect, if there is one.25 Finally, our employer survey data provided large samples of small businesses and their health insurance offerings, and they permit us to focus our analysis on the population to which the reforms were targeted.26 This provided us with a more convincing test of causation than if we had studied all employers or employees, or included other forms of private insurance in addition to employersponsored insurance. Overall, looking across the several outcomes and alternative estimators for each of them, we found little evidence that even stringent small group reform legislation led to a pattern of substantial, signiÀcant quantitative effects on the small group market across the nine study states by 1997. Despite the notable differences in approach and time period, these Àndings are consistent with those of three other recently published studies (Jensen and Morrisey 1999; Sloan and Conover 1998; Zuckerman and Rajan 1999). Both proponents and opponents of small group market reform will probably greet this news with mixed emotions. For proponents, we do not Ànd support for the hope that reform would lead to a significant narrowing of the variance in premiums or a sizable expansion in employment-based coverage among small groups. On the other hand, there might have been important improvements in access for some high-risk groups, which we could not detect with our analysis. For opponents of reforms of small group insurance markets, there should be some relief in hearing that stringent reform did not lead to marketwide price increases or sizable reductions in employment-based coverage. On the other hand, our analysis would not detect whether or not some lowrisk groups might have dropped coverage in response to speciÀc rate increases they faced. The reforms that we studied are in large part intended to eliminate some of the most serious problems in the small group market. They sought to guarantee that no small business is denied access to insurance, to protect groups that experience unexpected high medical expenditures from extreme premium volatility, and to shield high-risk groups from prohibitively high costs of insurance. However, our analysis is targeted toward Ànding overall effects across the whole population of small employers. This does not mean that there were not many individual gainers and losers under reform in the various states we studied. However, identiÀcation of these potential effects will require other designs and data sources. Notes Any views expressed in this paper are solely those of the authors, and no endorsement by the Robert Wood Johnson Foundation, the National Center for Health Statistics, or RAND is intended or should be inferred. The authors are grateful to Rick Curtis, Rafe Forland, Kevin Haugh, and Stephanie Lewis for coding and interpreting the small group reform data used to classify states in this research. The authors thank Linda Andrews, Roald Euller, and Ellen Harrison for their efforts in preparing the survey data Àles on which this paper is based. 1 Sloan and Conover (1998) studied nonelderly adults only during the period 1989–94. Zuckerman and Rajan (1999) studied all nonelderly during the period 1989–95. 2 This study used an augmented version of Zuckerman and Rajan’s data, for the same population and period, but it coded small group reforms differently. 377 Inquiry/Volume 38, Winter 2001/2002 3 Simon reports a signiÀcant but small (2 percentage point) fall in offer rates among small employers in states implementing reform relative to the trend in nonreform states. After adjusting for temporal change among large employers, however, the effect is not signiÀcant. 4 This database was developed by the Institute for Health Policy Solutions, see Curtis et al. (1999). 5 For details on the NEHIS, see National Center for Health Statistics (1997). 6 See Research Triangle Institute (1998); Kemper et al. (1996). 7 Excluded are self-employed individuals with no employees. 8 Prior to 1993, only two states, Vermont and Massachusetts, had legislation which guaranteed issue of all products in the small group market and prohibited health as a rating factor. This and other results on the history of state reform legislation are from an analysis of the IHPS database documenting the legislative history of reforms in all states (Curtis et al. 1999). 9 The reforms in one of our states—New York—were effective in April 1993. However, some policies had until their Àrst anniversary following the effective date to comply, and so we assume reforms had little effect in 1993. 10 California and Minnesota also introduced limits on the use of health status as a rating factor, but did not prohibit it. In California, the high and low premium based on health cannot differ by more than 20%; in Minnesota, the difference is limited to 67%. 11 Two of these 12 states, Michigan and Pennsylvania, had Blue Cross–Blue Shield plans that had signiÀcant market share and practiced guaranteed issue and community rating. We tested the sensitivity of the results shown in this paper by excluding these two populous states from the control. Our Àndings were the same and so are not shown here. 12 The states not included in our analysis enacted some new regulations concerning the issue of policies or rating reforms but did not adopt regulations as stringent as the reform states. However, because these states did undergo regulatory change that might affect outcomes, their inclusion could dampen our estimate of the effect of the reforms we study. Our objective was to maximize our ability to detect effects. 13 By mid-1997, most states had enacted legislation to guarantee issue of all products in the small group market to conform to the HIPAA legislation. We assume that this does not affect outcomes for 1997, because most contracts are written for the calendar year. None of our nonreform states adopted guarantee issue provisions for even a limited range of products prior to this. Some of our comparison states had rating reforms, however, but they permitted considerable variation in premiums based on age and health status (greater than 100% spread). 14 We report results for each of the reform states separately because we want to investigate whether there is a consistent pattern of effects across the states. Moreover, the states differed in the nature of reforms enacted (see Table 1). In contrast, we pool the 12 nonreform states. While the magnitude of a reform effect might be sensitive to the choice of a contrast, our conclusions stem in large part from a lack of consistency across the reform states—and different choices of the nonreform states to include in the contrast would not alter this Ànding. Furthermore, the contrast states are alike in that they did not enact reforms on the dimen- 378 15 16 17 18 19 20 21 22 23 24 sions we study. Moreover, as we noted earlier, the exclusion of two nonreform states in the contrast led to similar results. The expenditure distributions and actuarial values were designed and calculated for us by the Actuarial Research Corporation. We also produced estimates using enrollee weights and obtained essentially the same results, so they are not reported. Using the variance rather than relative variance as our outcome yielded similar results. The predicted values on all outcomes for medium-size employers, which enter into the difference-in-difference-in-differences calculation, are given in the Appendix Table. To conÀrm our Àndings from the employer survey, we conducted a similar analysis for the share of workers enrolled in their own employer group plan using data from the March 1993 and March 1997 Current Population Survey (CPS). With the CPS, we analyzed the difference-in-differences between each reform state and the 12 nonreform states in the percentage of employees in Àrms with fewer than 100 workers who held coverage from their own employer; the CPS does not permit us to produce estimates for Àrms of 50 or fewer workers. The results from the CPS also did not produce statistically signiÀcant estimates of the effect of reform, nor did the results suggest a consistent pattern of effects across states. See, for example, Buchmueller and Jensen (1997). Because we are comparing turnover in cross-section, we do not have controls for temporal effects. Moreover, the Àrst test which contrasts stability in each of the reform states with the nonreform states does not control for state-speciÀc effects. The second test uses stability of medium-size employers as control for state-speciÀc effects. These are weaker tests than employed for the other outcome measures. However, we report these results because improved stability in the small group market was an important goal of reforms. Because premiums are highly skewed, we also Àt models using the natural logarithm of premiums. These models produced few signiÀcant difference-in-differences or difference-in-difference-in-differences estimates of the effect of reform. As with the estimates of the effects on the mean, the effects on the median were not of consistent sign. We present the predictions from the model Àt to premiums on the dollar scale rather than the logarithm scale, because the former produces estimates of mean differences, whereas we would have to make assumptions about how variances vary across states to transform the predictions from the log scale to the dollar scale. Simon (1999), in contrast, concludes that states that have some form of guaranteed issue and some rating regulations have higher premiums than states without these regulations, based on an analysis of the NEHIS and the Medical Expenditure Panel Survey Insurance Component completed in 1996. However, the trend in premiums among small employers (those with 25 or fewer employees) did not differ signiÀcantly between reform and nonreform states and was very small in magnitude. The conclusion stems primarily from a negative trend in premiums for large employers (those with 100 or more employees) in the reform states relative to the nonreform states (that is, from a difference-in-difference-in-differences estimate of the effect). This is a step for further research urged by Marsteller et al. (1998). In addition, our choices of states were ‘‘Second Generation’’ Reforms based on implementation dates for both rate regulation and guaranteed issue that are more precise than the assumptions that have been used before. 25 By considering each reform state separately, in effect we follow Zuckerman and Rajan (1999), who advised examining reforms as packages. 26 Jensen and Morrisey (1999) acknowledged the desirability of large samples of employers for this type of research. References Bilheimer, L.T., and D.C. Colby. 2001. Expanding Coverage: ReÁections on Recent Efforts. Health Affairs 20(1): 83–95. Blumberg, L.J., and L.M. Nichols. 1996. First Do Not Harm: Developing Health Insurance Market Reform Packages. Health Affairs 15(3):35–53. Branscome, J.M., P.F. Cooper, J. Sommers, and J.P. Vistnes. 2000. Private Employer-Sponsored Health Insurance: New Estimates by State. Health Affairs 19(1):139–147. Buchanan, J. L., and M. S. Marquis. 1999. Who Gains and Who Loses with Community Rating for Small Business? Inquiry 36(1):30–43. Buchmueller, T.C., and G. A. Jensen. 1997. Small Group Reform in a Competitive Managed Care Market: The Case of California, 1993 to 1995. Inquiry 34(3):249– 263. Buchmueller, T., and J. DiNardo. 1998. Did Community Rating Induce an Adverse Selection Death Spiral? Evidence from New York, Pennsylvania, and Connecticut. Unpublished. University of California, Irvine. Chollet, D. J., and R.R. Paul. 1995. Community Rating: Issues and Experience. Washington, D.C.: Alpha Center. Curtis, R., S. Lewis, K. Haugh, and R. Forland. 1999. Health Insurance Reform in the Small-Group Market. Health Affairs 18(3):151–160. Custer, W.S. 1998. Health Insurance Coverage and the Uninsured. Washington, D.C.: Health Insurance Association of America. Gabel, J.R., P.B. Ginsburg, and K.A. Hunt. 1997. Small Employers and Their Health BeneÀts, 1988–1996: An Awkward Adolescence. Health Affairs 16(5):103–110. Herzlinger, R.E. 2000. Leveling the Playing Field. The New Democrat Blueprint 6:22. Hing, E., and G.A. Jensen. 1999. Health Insurance Portability and Accountability Act of 1996, Lessons from the States. Medical Care 37(7):692–705. Institute for Health Policy Solutions (IHPS). 1995. State Experiences with Community Rating and Related Reforms. Washington, D.C.: IHPS. Jensen, G.A., and M.A. Morrisey. 1999. Small Group Reform and Insurance Provision by Small Firms, 1989– 1995. Inquiry 36(2):176–187. Kemper, P., D. Blumenthal, J. M. Corrigan et al. 1996. The Design of the Community Tracking Study: A Longitudinal Study of Health System Change and Its Effects on People. Inquiry 33(2):195–208. Kish, L. 1967. Survey Sampling. New York: John Wiley and Sons, Inc. Marsteller, J.A., L. M. Nichols, A. Badawi, et al. 1998. Variations in the Uninsured: State and County Level Analyses. Washington, D.C.: The Urban Institute. Morrisey, M.A., and G.A. Jensen. 1997. Switching to Managed Care in the Small Employer Market. Inquiry 34(3): 237–248. National Center for Health Statistics (NCHS). 1997. Employer-Sponsored Health Insurance: State and National Estimates. Hyattsville, Md.: NCHS. Research Triangle Institute (RTI). 1998. 1997 Employer Health Insurance Survey: Final Methodology Report. Research Triangle Park, N.C.: RTI. Simon, K.J. 1999. The Impact of Small-Group Health Insurance Reform on the Price and Availability of Health BeneÀts. Working Paper. University of Maryland. Sloan, F.A., and C. J. Conover. 1998. Effects of State Reforms on Health Insurance Coverage of Adults. Inquiry 35(3):280–293. Thorpe, K.E. 1992. Expanding Employment-Based Health Insurance: Is Small Group Reform the Answer? Inquiry 29(2):128–136. Wessner, C. 1994 State Community Rating Reforms: Early Tests Produce Mixed Result. State Health Notes, Intergovernmental Health Policy Project 15(193):1–8. Zuckerman, S., and S. Rajan. 1999. An Alternative Approach to Measuring the Effects of Insurance Market Reforms. Inquiry 36(1):44–56. 379 Outcome 12 Nonreform states MD NJ NY WA CA MN FL OR CT 91.2 90.2 92.7 86.5 84.5 86.6 82.5 94.5 94.1 95.3 93.4 88.2 83.1 77.2 77.8 91.9 87.0 91.6 83.2 Percentage of employees offered insurance 1993 1997 90.5 93.9 93.2 89.7 73.2 77.7 93.3 95.0 97.3 94.1 93.9 88.2 88.4 81.7 82.4 92.2 95.0 85.7 78.1 Percentage of employees enrolled 1993 1997 59.5 56.6 55.9 55.9 53.9 54.5 67.9 67.7 69.7 63.7 60.1 58.2 57.4 51.0 52.8 68.8 73.0 56.1 55.8 5.6 6.5 3.6 4.3 11.9 13.3 2.3 6.0 26.3 25.8 28.1 44.8 30.5 35.2 50.2 70.3 36.7 40.3 40.1 50.0 43.2 55.7 25.6 23.8 18.4 30.8 27.6 49.4 15.8 40.9 57.6 57.4 39.2 64.3 52.8 56.5 37.0 42.0 31.8 23.3 Percentage of employers offering insurance 1993 1997 Percentage of employers changing offer decision 1995 to 1997 7.7 a a a a Percentage of employers offering HMO (if offer insurance) 1993 1997 24.6 39.7 44.5 Percentage of enrollees in HMOs 1993 1997 23.6 40.0 31.1 a a Monthly premiums for employee-only coverage ($) 1993 1997 152 168 155 Monthly premiums for family coverage ($) 1993 1997 388 392 412 Relvariance in premiums for employee-only coverage 1993 1997 Relvariance in premiums for family coverage 1993 1997 Percentage of Àrms with premium change of more 10%, 1996 to 1997 a a 184 186 184 200 145 167 160 160 133 147 136 164 129 150 182 191 445 432 410 416 404 406 409 403 370 380 370 406 350 374 463 475 11.7 16.6 6.1 .8 15.3 18.7 20.1 23.7 19.0 21.7 22.1 19.0 9.0 15.1 12.8 7.4 15.0 16.5 16.7 21.6 7.4 7.5 5.3 10.9 12.8 12.5 13.3 6.9 9.6 4.1 12.0 4.0 8.2 5.3 4.2 4.4 7.4 11.3 13.3 11.4 17.7 28.2 8.9 14.6 18.2 8.6 13.2 15.3 a a Source: National Employer Health Insurance Survey, 1997 Robert Wood Johnson Foundation Employer Health Insurance Survey, and Institute for Health Policy Solutions database of small group reform provisions. a Fewer than 50 observations. Inquiry/Volume 38, Winter 2001/2002 380 Appendix Table. Predicted 1993 and 1997 outcomes for medium-size employers.