[This Transcript is Unedited]
DEPARTMENT OF HEALTH AND HUMAN SERVICES
NATIONAL COMMITTEE ON VITAL AND HEALTH STATISTICS
SUBCOMMITTEE ON POPULATION HEALTH
February 27, 2009
Hubert Humphrey Building
200 Independence Avenue SW
CASET Associates, Ltd.
Fairfax, Virginia 22030
- Donald Steinwachs, Ph.D., Co-Chair
- William J. Scanlon, Ph.D, Co-chair
- Larry A. Green, M.D.
- Mark C. Hornbrook, Ph.D
- Garland Land, MPH
- Blackford Middleton, M.D.
- Walter G. Suarez, M.D.
TABLE OF CONTENTS
- Call to Order and Welcome – Donald Steinwachs, William Scanlon
- Panel 1: Modeling Health Insurance Data: Practice and Perspectives
- Beth McGlynn
- Jim Baumgardner
- Jonathan Gruber
- Bo Garrett
- Panel 2: Modeling Health Insurance Data: Practice and Perspectives, Continued
- Jessica Banthin
- Tom Selden
- John Sheils
- Joan Turek
- Linda Giannarelli
- Wrap Up, Final Comments
P R O C E E D I N G S (8:35 a.m.)
Agenda Item: Call to Order and Welcome
DR. STEINWACHS: I am Don Steinwachs, a member of the National Committee and
co-chair of the Populations Subcommittee. I’d like to welcome you to this
hearing on modeling health insurance data. This is the second in a series of
hearings that we have had. The first one looked at the adequacy of data and
information in being able to assess access and patterns of insurance coverage
and uninsurance in the nation, as we look forward to health reform initiatives.
This hearing is being broadcast live on the Web, so I ask you to speak into
microphones. Apparently they will become accustomed to you as you become
accustomed to them; it is a process.
Before we start the hearing and introduce our invited guests, and we are
very happy to have you with us today, I would like to ask that members of the
committee introduce themselves, and then also we will go around the room and
ask everyone here to introduce themselves, and then we will start the hearing.
I am Don Steinwachs, Johns Hopkins University, a member of the committee
and of the subcommittee.
DR. SCANLON: Bill Scanlon, Health Policy R&D, member of the committee
and member of the subcommittee.
DR. HORNBROOK: Mark Hornbrook, Kaiser Permanente, member of the committee
and member of the subcommittee.
DR. BREEN: Nancy Breen, National Cancer Institute, member of the
MR. HITCHCOCK: I am Dale Hitchcock from ASPE here at HHS, in the data
policy shop. I am a staff person to the subcommittee.
MS. BULLHEIMER: Linda Bullheimer at the National Center for Health
Statistics. I am an advisor to ASPE on this project.
MR. GARRETT: Bowen Garrett. I am a senior research associate at the Urban
Institute and Health Policy Center.
MS. MC GLYNN: Beth McGlynn. I am the Associate Director of RAND Health.
MR. BAUMGARDNER: Jim Baumgardner, Deputy Assistant Director of Health and
Human Resources from the Congressional Budget Office.
DR. SUAREZ: I am Walter Suarez. I am a member of the committee and a member
of this subcommittee.
DR. GREEN: Larry Green, University of Colorado, member of the committee and
member of the subcommittee, no conflicts.
MS. JACKSON: Debbie Jackson, National Center for Health Statistics,
DR. GIROSI: Federico Girosi, RAND. I am leading the modeling team of the
DR. BANTHIN: Jessica Banthin, Director of the Division of Modeling and
Simulation in the Center for Financing, Access and Cost Trends at AHRQ.
MS. NIVASSAN: Sherovasi Nivassan, Health Disparities Research Coordinator,
Division of Cancer Control and Population Sciences, NCI.
DR. SHORTRIDGE: Emily Shortridge, NORC at the University of Chicago. I am a
MR. SHEILS: I’m John Sheils with the Lewin Group.
MR. HOTT: Randy Hott, the Lewin Group.
MR. MUSCO: John Musco with the Office of Health Policy at ASPE.
MR. ARNETT: Ross Arnett with Econometrica.
MR. GOREITCHEV: Dimitri Goreitchev with Econometrica.
MR. SACARADE: Alex Sacarade, Econometrica.
MS. CARROLL: Monique Carroll, GE Health Care.
MR. O’HARA: Brad O’Hara from Census, small area estimates of health
MS. BICKFORD: Carol Bickford, American Nurses Association.
MR. ROSHOWSKI: Jim Roshowski, Center for Studying Health System Change.
MR. MERMAN: Gordon Merman, Office of Management and Budget.
MS. JAMISON: Missy Jamison, National Center for Health Statistics, CDC,
MS. KANAAN: Susan Kanaan, writer for the committee.
MS. SASHMONY: Nina Sashmony, ASPE.
MR. POISELL: John Poisell, CMS.
DR. STEINWACHS: Thank you very much, and welcome again. Before turning this
over to Bill Scanlon to introduce the first panel, I want to express a special
thanks to Dale Hitchcock and to Roshita Dorsey, who made this possible by
organizing and inviting all of you, and encouraging you to come, who are on the
panel and very much appreciated. Also, Debbie Jackson and Cynthia Sydney, who
made this possible too. So thank you very much.
DR. SCANLON: Let me extend my thanks to all those who made this possible,
as well as to all of you who agreed to come and talk with us.
As Don indicated, this is the second hearing that we are having with
respect to data for health reform. These two hearings have been done at the
request of ASPE, but they are continuing a focus of this subcommittee on the
issue of data availability for analysis that can help inform through policy
choices with respect to health and health care.
It goes back at least for the entire time that I have been on the full
committee. There have been two themes for those hearings. One is the issue of
access to data that exists. That is a real question not only for researchers
and analysts outside of government, but as it turns out, an issue within
government. So we have done a letter to the Secretary saying that we at least
need to think about how we can streamline this process within government.
The IOM has raised in its report earlier this month other sets of issues
which may apply more outside of government, but those are things that we need
to be thinking about too, how do we rationalize the access to data that exists.
The other theme is the issue of data that don’t exist but should, or we
would hope that they would. That is something that we need to also think about
some more seriously. Yesterday the full committee approved a letter to the
Secretary indicating that there is a real need for investing more resources in
data collection. There is an opportunity in the newest bill, the Stimulus Act,
to make some of these kinds of investments, but they only should be perceived
as a first step. Data collection is not a one time activity; it is something
that needs to be established on an ongoing basis, and there needs to be enough
adequacy for adequate investments so that we have the richness of data to be
able to understand the issues that we are facing.
In particular, it is totally inadequate to know what is happening on a
national level. We know too much about the variation in the country in terms of
that it exists, and we know too little about what explains that variation and
how we might change it in a positive way. So thinking about how to improve data
on a sub-national basis is something that is also important to us.
The last hearing was a focus of people who are data producers and some of
the data users from an analytic perspective. Today we are moving to what I
would think of as the heavy lifters, people who are trying to model various
aspects of the health care sector, which is the challenging activity in terms
of drawing upon multiple data sources, and trying to both make them work better
or harder, to learn more from the interactions, but it also increases the
demands or the requirements for the data. So it is something that we need to
know about in terms of making recommendations about future data activities
within the Department.
We recognize that while we are doing these hearings in the context of
anticipation for health reform, that a lot of things may happen with health
reform before additional data can be collected.
DR. GRUBER: Hello?
DR. SCANLON: Hello?
DR. GRUBER: Hello? John Gruber calling in. I can’t really hear someone on
DR. SCANLON: John, this is Bill Scanlon. We were just starting.
DR. GRUBER: I can’t really hear you.
(Remarks off the record.)
DR. SCANLON: John, thank you for joining us. We were just getting started.
We were just introducing the hearing and emphasizing that what we are
interested in here is the issue of data, access to data as well as production
of data that is adequate to be able to do the kinds of analyses that we need to
underpin health reform.
As we know, the health reform process is in some respects already underway
with respect to proposals. We are anticipating that this is going to be a
process, not an event, that there will be opportunities to weigh in with
additional information, additional analyses over time. So there is a question
in terms of what it is that can be done to try and facilitate and enhance this
At the last hearing, we heard strongly about this whole issue, since most
of our users of data were from government, about the problems and the need to
expedite access to data within government. This is a key issue.
Today we are interested in expanding this in terms of the area of modeling,
to know about some of the hurdles that you face as well as how you have
overcome them, the capacities that you have developed, and then things that you
could suggest for us to think about in terms of recommendations that the
Department may do in terms of making the capacity to do modeling stronger for
We are going to turn now to the first panel. We are very happy that you
agreed to participate. Beth McGlynn from RAND is going to be our first
presenter. If you would like to start, Beth, you have the floor.
Agenda Item: Panel 1: Modeling Health Insurance Data,
Practice Versus Perspective
MS. MC GLYNN: I want to start by thanking the committee for the opportunity
to talk to you about what we have learned about in particular the federal data
sets that are available for use in microsimulation modeling. I want to in
particular acknowledge Federico Girosi, who is the real brains behind the
operation. He will answer all your hard questions as we go along.
This is the outline of what we are going to talk about . I am going to give
you a brief overview of the model that we have developed. That is primarily to
set in context comments we make about the data sets that we have tried to use
and some of the challenges that we have encountered.
We are going to talk about our experiences with some of those data sets,
and then talk about dreams we have in our off hours about data sets, if they
only existed, what could we possibly do, and then end with some priorities for
where we hope you might recommend that the government go.
COMPARE is an initiative. It is an acronym that stands for Comprehensive
Assessment of Reform Efforts, that was begun at RAND about three years ago. It
has two goals, to provide a factual foundation for a national dialogue on
health reform options, and then to develop a set of tools that would help with
the development of legislation and policy options, both at the federal and
state level, as well as those things that might be under consideration by large
We look at the effects of policy changes on multiple dimensions of system
performance, including cost, quality, access. Then we have another category
which is operational feasibility, how hard is it to get from where we are today
to the future state of the world that is envisioned.
The dimensions are not so well highlighted, so I will read them to you. The
microsimulation model itself produces results in the areas of health, national
health expenditures, that is both in the aggregate and distributionally,
consumer financial risk, which looks at the impact on different kinds of
households, and the area of coverage. All of these are both aggregate and
COMPARE is an agent based microsimulation model. The agents are shown in
green here. Let me in particular highlight what is in the box. We have
developed a way of looking at not only the person level, but some other units
that are important for policy making purposes, so families, health insurance
eligibility units and tax units. As you know, those come together in different
ways as you are working. The arrows are some of the messages that go back and
forth between the different agents.
One of the things that is a little bit different about our model is that we
estimate premiums and insurance status endogenously. So that is one of the
features of this model. It gives you an overview.
We use a variety of data sources. The first three that are listed, the
SIPP, the MEPS-HC and the HRET/Kaiser are the core data sets. So the base
population is built out of the SIPP. That is linked to the MEPS-HC for the
purposes of coming up with health care expenditures, health status and
utilization patterns, and then the HRET/Kaiser data are used to describe
employer characteristics and employer health benefits.
You can see that we use a number of other data sets. Those are used in a
variety of ways, including to make sure that distributions look like national
distributions. In many cases we have to take the MEPS for instance and inflate
it to look like national health expenditure accounts. These are issues that are
familiar to everybody, but we do use quite a variety of data sources.
The agents that are in the microsimulation model, their behavior is modeled
in a number of different ways. Our starting point was a series of linear
progressions that describe the behavior of individuals and HIE use.
We have more recently moved into modeling behavior using utility
maximization methods and by the end of next month pretty much everything will
be converted to the utility maximization framework. Then the behavior affirms
is principally modeled using cost-benefit analysis.
The way the model works is, we are looking at a fairly short term
before-after. The results of the model describe the new state of the world at
equilibrium following the policy change. We estimate that that is about a two
to three year process. We use 2007 right now as the base year on which changes
We don’t do much in terms of macro facts. The employment variables are
static. We have been talking about how we can reconfigure the current state of
the world to reflect the economic downturn, and I think we have a method in
mind for doing that.
Any time you have a big shop like that, you have to have some kind of
timely data, and you have to have a base data set that you are using, and
ability to have access to the variables that help you understand things like,
are the unemployment effects located in a particular sector, and do we have
something in the data set that lets us know that there are differential effects
by sector, for example.
We don’t in our model have people changing jobs or being laid off in the
course of the simulation.
The analysis right now is performed at a national level. We can and have a
method in place for doing state level analysis. That would be done by
reweighting the data to make the synthetic universe look like the state that we
were looking at. But to do really good state level analysis requires some
additional state specific data, and we will talk in some of the limitations
about some of those issues.
To date we have modeled the following policy options, an individual mandate
with the National Health Insurance Exchange and subsidies, an employer mandate,
Medicaid SCHIP expansion, refundable tax credit, more like the Bush tax credit
than what McCain was proposing in the campaign, and then a Medicare buy-in
modeled after the recent Baucus proposal.
I should say, each of these were originally modeled as independents or
stand-alone policy options. Most recently what we have implemented is the combo
package proposed through Baucus. So that was the first time we put a number of
these options together. It gives you a sense of the kinds of things we have
We are also in the process of developing a provider module. The focus of
that will be on the behavioral responses of physicians and hospitals to changes
in payment policy and what we characterize as health services delivery
interventions. Those are things that are a bit more on the ground, like disease
management, the medical home, public reporting, pay for performance, those
kinds of changes.
Initially the module will operate independently. Over the long term we will
be able, because of the modeling architecture that we have built, the
microsimulation model, to connect that and have it interact with the model
structure that I showed you before.
One of COMPARE’s main commitments is to advancing policy analysis through
transparency. The results, and had we had more time, I would have bored you
with it through the website, but in your copious free time you can take a look
at the results, which are publicly available on the website, randcompare.org.
Also on the website you can find a white paper that describes the
microsimulation methodology. This is what we refer to internally as the short
form. It is only 40 pages long. Federico has a blockbuster 400-page version
coming soon to a theater near you. It will have quite a bit more detail.
There is also an aspect of the website that lets users interact with
basically the modeling results which has a data set underneath and people can
change parameters and assumptions, and look at how that affects the results of
these different policy options. So I think of it as fantasy policy maker.
Here are some of the strengths and limitations of the data sets that we are
using. I’m not going to spend a lot of time on the strengths, because I think
those are pretty well known. Let me just highlight some of the limitations that
we have encountered with the SIPP.
A big issue there is that it is not very timely. It was last fielded in
2004. We are using the 2001 with a 2002 panel update. It doesn’t within that
data set have reliable information on utilization, and it doesn’t have any
The MEPS-HC, which is a terrific data set, is a little bit problematic
because it is smaller than the SIPP and the CPS, which is difficult when you
get to much more subunit analysis. There is not unrestricted access to state
codes, so this has caused some challenges when there are important regulatory
differences state to state, and even with some of the Medicaid structure. So
for modeling some of that, one has to get a bit creative.
I will talk in a minute about why we use the Kaiser HRET as the main source
of information on benefit design, but it is quite small, and it is limited to
aggregate information on employees. So you don’t have any individual level
information that helps you understand some things about firm behavior.
Let me talk a little bit about data sets we had some challenges with. The
first is the MEPS HC-IC, which is a data set that is available through
restricted access. It is the kind of thing we use for modeling the choices that
are made by employees.
The problem we had with it is, it is fairly small. The data are quite noisy
and not particularly nationally representative. After spending quite a bit of
time working with this data set, we ended up using a simpler analysis than we
had anticipated, which meant that we couldn’t really take full advantage of the
information about health benefits. So we like a lot of other people assign
people to an average type plan, and you can’t answer questions about what if we
did this, that or the other thing to mandates, for example. So that kind of
fine level of analysis.
We had pretty timely access to the data through AHRQ, but it is still
inconvenient to have to go out to AHRQ to use it. You can’t import data
directly into the simulation model. You can only export regression
coefficients, so it is a fairly in efficient way to interact with the data.
Then the final plea for equipment upgrade. Our programmers complain that the
screens were small, the computers were outdated, and they therefore could only
spend limited amounts of time there before they were going home with migraine
headaches. So maybe in the stimulus package we can buy AHRQ some new computers,
I don’t know.
Then let me turn to the MEPS IC. Again, a restricted access data set. It is
a very large and unique data set, but it takes a long time to get access to it.
You have to put a proposal in.
Interestingly enough and certainly not helpful, you can’t get approval to
do analyses that somebody else is already doing, so there is actually no
opportunity to have groups run head to head with each other and learn from each
other by replicating analyses that are being done.
You have to propose doing something that benefits the Census. It is not
always clear that the things that would be useful in terms of health policy
analysis are things that you can — although we are creative — arguably
benefit the Census.
Again, this is inconvenient access. You have to go to a Census office. You
not only can’t import data into the simulation. They are very picky. In fact,
Federico recommends for those of you who have not reading the page that
describes access to the MEPS IC as something that tells you the difficulties.
They are not even really happy about you taking tables away. They would prefer
that you just take regression coefficients away, and that is problematic for
I think it took us about a year to get access to the data. We have just
gotten access after we have put our initial round of results out, so now we are
trying to figure out how we can make use of that.
Then to fantasy data sets we wish existed. It would be great to have a very
large cross sectional population survey that gave us an ability to understand
and model people’s preferences with respect to choice of insurance. We think
this can be fielded about every five years. It could include scenarios and
vignettes. It is the kind of thing that would have helped us with the utility
maximization and some of the other work that we have done.
Another data set that would be quite helpful is what we call the birth to
death HRS, which gives us an ability to understand transitions over the life
course, and would be particularly helpful for some of the health estimates that
we are producing out of the microsim, which we base on the HRS and then we do
lots of creative things to age it down to younger ages. But we could certainly
improve a lot of those health outcomes estimates that we are producing right
now with a much more detailed data set at younger ages.
We would love to have a large linked employee-employer data set that has
information about the choices that people face. This is something that we would
put together all in one place, things that we cludged together from multiple
data sources, that would allow us to better understand firm behavior. Right now
you have just aggregate data on the employees and firms, and you have a hard
time looking at individual behavior within firms. So this kind of a data set
would help with that.
Then the final thing that I think a lot of people are struggling with as we
look at policies like the National Health Insurance Exchange is understanding
something about the non-group market today, which is small but growing. There
isn’t any decent data out there, so I think all of us who are operating this
base are probably creatively making things up.
Then finally, as we look at the provider module, it would be helpful to
have a physician longitudinal data set that would enable us to model behavioral
responses to different kinds of incentives. It would be helpful to have that
mapped together with some understanding of physician characteristics and be
able to observe changes over time that are linked to those kinds of
characteristics, and that would allow you to link other kinds of data sets in
like Medicare data.
I worry a lot that policy options are being modeled today strictly on the
Medicare population. We don’t have much of an understanding of what goes on
outside of the Medicare population. From some of the analyses we have done at
the provider level, I can assure you that it is probably quite different
behavior within the commercial market than it is in the Medicare market.
So I will end with these three priorities. We really do support the letter
yesterday, which is improving access to data sets that already exist, and in
many cases to help with the documentation around using some of those data sets.
We would like to have data available in a more timely manner. We would like the
data that becomes publicly accessible to be no more than a couple of years old.
We hope that there is a priority on developing new data sets that allow us
to look at changes to the health care system beyond just coverage expansions. I
think there is a big focus on that. That is where a lot of the microsimulation
modeling has been, whether or not you believe that we are about to get
universal coverage in the next 90 days or whatever. I think that will only then
begin the next round of things we need to look at. We believe microsimulation
modeling can be helpful for that, but the data sets for doing those kinds of
analysis are much more limited.
DR. SCANLON: Thanks very much, Beth. If it works for everyone, I think what
we are going to do is have everyone present, and then we will have a joint
discussion for the first panel.
So Jim Baumgardner from CBO, you’re on next.
MR. BAUMGARDNER: Today I am going to primarily just talk about CBO’s
central model for examining proposals to expand health insurance coverage.
We call it HISim, the health insurance simulation model. We have been
developing the model since early in the decade and have used it for looking at
some kinds of proposals since around 2004.
Essentially it is aimed at looking at proposals aimed at the non-elderly
population and proposals that are aimed to affect the amount of health
insurance coverage. Obviously one of the effects we are interested in also is
effects on the federal budget, since that is our primary responsibility,
although looking at coverage and premiums and also looking at some
distributional effects is also something we are interested in doing.
Recently CBO released two big volumes on health in December of 2008. One of
those is called Budget Options. About a dozen of the proposals in there used
this health insurance simulation model to develop ten year federal cost
estimates as well as effects on coverage.
There is a December 2007 paper that we have that is on our website,
CBO.gov, that gives more details on what I am going to talk about today in
terms of the model itself.
It is a microsimulation model. SIPP is the platform for the model.
Underlying everything are the underlying observations from SIPP. When I get to
the next slide you will see how we have used a lot of other data sets and
imputed onto SIPP people from other data sets.
We allow for the multiple coverages that some people have that you see in
SIPP. We convert SIPP people into being part of health insurance units. In some
cases you will get a family within SIPP that has more generations than you
could plausibly have in one health insurance unit. Like, you might have a child
who is in their upper 20s, let’s say, who couldn’t really be on the parents’
policy. We may have grandparents present. So we do divide these into what seem
to be coherent health insurance units.
It is what we call an elasticity based model. We start with a baseline with
people in the coverage category, and then there will be some policy that comes
in and in essence changes the effective premium that is available to that unit.
Then people respond to those changes in premiums. So their coverage in the next
period depends on where they start from, and then depends on the change in
prices they see. We use elasticities from the literature to develop the
behavioral responses. Many of the papers are referenced in that 2007 document
that I mentioned before.
We also put each worker within SIPP as a synthetic firm created around
them. So we have both firms making decisions to offer coverage or not, and then
we have individuals, if they are offered coverage from the firm, they make a
decision whether to take it or not. Then they also have an option to go to a
non-group market, and we are developing other options besides the non-group
This is a rough schematic of how the model is set up and the changes that
go on. In the upper left are the main data sets that are involved. So again it
is a SIPP platform, but we use Bureau of Labor Statistics, some firm data they
have to help us with our wage distribution in these synthetic firms.
CBO projections are relevant. Obviously we are dealing with a major federal
program such as Medicaid and the SCHIP program, and we need to have people
assigned in our model, so that the numbers that we say are in Medicaid in our
model have to match what CBO says is in the baseline. So there are those kinds
of calibrations that are done. We use data from MEPS. That helps us with things
like health status spending, the contribution of employers, MEPS IC helps us
We use the MBER tax simulation model to get marginal tax rates. Of course
most people have employer sponsored insurance, so most people’s insurance is
excluded from taxes. That has an implication for what effectively the price of
insurance is after you take those tax effects into account, and that depends on
the marginal tax rate that a person is in. So we do need that kind of tax
Then the national health expenditure accounts are also used to calibrate
up, so that we want to match total premium numbers that are in the national
In your base case, you have a person who is maybe in a family, who maybe
are not, who has a firm that is offering insurance. We have their family
income. Then a proposal may come along with could for instance be a voucher or
some other subsidy or change in the tax treatment of insurance or a regulatory
change that is going to affect the premiums people face, or you may ultimately
have mandates on either individuals or employers to have coverage or to offer
it. That is going to change the after tax premiums that a person faces in the
employer sponsored market, and it will also change the alternative premium
outside of the employer based market. Then as I mentioned before, both firms
and individuals are responding then in this model.
Then ultimately we get changes in coverage, changes in the source of
coverage. We can do distributional analyses by poverty level, and ultimately we
want to be looking at budgetary effects. Federal budget effects is of course a
focus of ours.
I will talk about four kinds of things we have used this model for which
are proposals or potential proposals which we have in this health options
volume that we released in December. Then I will end by saying where things are
going down the road.
One such proposal we looked at was maybe what we would call in the category
of a regulatory change. One of the things that we do in building up premiums in
the model, if there is employer based insurance involved, the person is
assigned to a firm, and then we have a synthetic cohort of fellow workers, and
then we calculate based on information on health status and conditions and
hospital utilization that we can pick up in SIPP. We get the expected health
costs of each person in the firm, and then along with an administrative load we
calculate a premium for everyone in that synthetic firm.
Now, for people who are in a small firm we look at the state regulations on
insurance in the state the person is assigned to and take into account the rate
compression regulations in that state. We also do this in the non-group market
for people. A small set of states, a very small set, have full community
rating. A somewhat bigger set of states have a modified community rating, and
then many states in the non-group market don’t have a rating restriction like
that. But we model that to then get the estimated premium that the person would
face in the non-group market.
This slides gives you an idea for non-group premiums in 2009 by reported
health status of people. We have broken out the range of premiums that they
face. This is the national look. Some people in a community rate state are just
going to see a much narrower, or one premium.
So anyhow, that is what we do in non-group.
This first option that we looked at was a regulatory change that would
allow people to buy a non-group policy from any insurer, not necessarily one
that was licensed in the state the person lives in. What this does is, it
undoes community rating, in the sense that a young healthy person in New York
who is facing a high premium relative to their expected cost because of the
community rating in New York, could buy a policy from an insurer that is
licensed in Arizona and could probably find a lower premium there.
So you get a situation where some high expected cost people in what I will
call more regulated states ultimately would face a higher premium as the lower
cost people in those states unwind and go to other insurers outside of the
state where they could buy a less expensive policy. So you crank all this
Also there is an issue of mandates to cover in vitro fertilization and an
assorted list of things. That is also something one could avoid the rules in
one state if this were the policy.
So in the end we calculate about a half a million reduction in the
uninsured, but you do get this differential by health risk with some higher
health risk people in more regulated states seeing higher premiums. The biggest
premium reductions are for the people in those same states who had lower
expected health costs.
We also get some firms dropping insurance, in this case because as more
employees find a cheaper option in the non-group market, there is some impetus
for some firms at the margin to stop offering. That is part of the whole
A different option we looked at was a subsidy to purchase non-group
coverage. Roughly a subsidy was allowed of up to $1500 for a self only policy,
up to $3,000 for a family policy. That ends up being about a 40 or 50 percent
subsidy for the average uninsured person.
We set up this policy so that it would phase out starting at 200 percent of
the federal policy level, the subsidy would phase down to zero at 250 percent
of the poverty level. At most, 70 percent of the premium could be covered by
So what we got from the model then, about four million people ended up
using the voucher. About 2.2 million of those would have been otherwise
uninsured people. We calculated about an $8.6 billion budget cost in 2014. That
is a one year budget cost. So that is another kind of policy we looked at.
Then a third proposal we looked at would extend the tax deduction for all
non-group premiums. The currently self employed people can take a deduction,
but other people in the non-group market cannot, and this would extend the tax
deduction for anyone buying a non-group policy. In that case, we estimated that
we would get about 700,000 newly insured people as a 2014 number, as well as a
$6.3 billion budget cost.
If you do division and look at how many newly insured people do you get per
budget dollar of expenditure, this policy doesn’t do as well by that metric as
that previous one, the vouchering to low income people. Part of what goes on is
targeting, targeting, targeting. To just have a tax deduction, obviously people
at higher income are more likely to pay taxes if the deduction is worth more to
them, but that is where you have fewer of the uninsured people. So that is what
is going on there.
Yet another option we looked at was to look at this notion of federally
subsidized reinsurance for high cost claims. So the policy here was when a
person’s claims exceeded $50,000, the federal government would pay 75 percent
of the rest of their claim costs above that.
I think we looked at two options in our volume. I am only going to talk
about the one that was aimed at smaller firms, those with fewer than 100
employees in the non-group market, but I believe we also had an everybody
policy we put into the book.
We concluded though that this would be about a 14 percent subsidy,
obviously higher for individuals that had higher than expected health costs and
for firms that had a group of workers and covered family members that had
higher expected costs.
We used the MEPS, but we also used — and here is a data issue. As we
talked to actuaries about trying to model something like this, I think the
consensus view was that within MEPS it is the right tail of the spending
distributions then compared to what you would actually see in claims data. So
we used some Society of Actuaries data to fill that tail in some more and to do
So ultimately we estimated that the $2.1 million reduction in the number of
uninsured has a very large budget cost of $32 billion as the 2014 number, so
that is just an annual number. In essence you are buying a very large base out
in this proposal. Obviously under current law lots of people to have health
insurance, and that health insurance is covering all the claims. Here you are
taking 14 percent and have the government pay that.
Then finally, we also started to experiment with an employer pay or play
type of proposal. The proposal we looked at had a $500 fine per employee for
employers if they did not offer health coverage. This only was to larger firms,
firms with more than 50 employees. There was a requirement that they cover at
least 50 percent of the premium, those larger firms.
We estimated to begin with that there were about ten million uninsured
people, independents who are in firms that are large enough to be affected by
this type of pay or play system. In the end, we calculated that there would be
about a three or four percent reduction of uninsurance within that ten million
group, so pretty small actually, is the reduction of the number of uninsured
under that policy. The proposal does make some money for the federal government
though, and that is mainly firms paying the fine.
Of course, with a larger — $500 is pretty small. As you get up to what the
premium cost would be, you would expect much larger participation by firms.
So going forward, we of course are going to be in the heart of big
proposals to change the health insurance system. The other kinds of options,
some sort of subsidy proposal, some sort of subsidies probably targeted toward
low income, lower income populations, and other offerings possibly besides
non-group that is available today, those are probably the directions things are
Federico is here, but the brains behind our operation are working very hard
at a location south of here right now.
I can’t add much. I thought Beth did a nice job of covering some of the
data issues. I really can’t top that. We certainly would like everything. We
would like a huge sample of people and their employers and their fellow workers
and details on the insurance coverage all in one big sample, along with the
people’s health conditions and what their health spending had been. That is the
top-down approach to what we would like.
DR. SCANLON: Thanks very much, Jim. John, are you set? We will give him a
DR. GRUBER: Hi, sorry, I got dropped off just as Jim was finishing up. Are
you ready for me now?
DR. SCANLON: We are ready for you to go. We are trying to find your slides,
and then we will be set.
DR. GRUBER: Tell me when you are ready.
DR. SCANLON: Okay, we’re set.
DR. GRUBER: I have a less detailed presentation than what has been given so
far, but I think a lot of the things we have been discussing are pretty
similar. I’m glad you guys convened this meeting. I’m sorry I can’t be there. I
think it is very exciting that so many people are working this space and
thinking hard about putting numbers on questions which often are very hard to
put numbers on. I think the difficult issue for ASPE is going to be that even
when very smart, well meaning people get together and try to put the same
number of something, they are often going to differ. The difficulty for you is
how to work through that and figure out which numbers to rely on, so I think it
is useful to get a sense of what everybody is up to, so it is a useful meeting.
It is useful for me too to hear it.
I am going to talk about what I have been working on with my
microsimulation model. If you go to the second slide, there is a highly
detailed schematic of how the model works that I had to have my eighth grader
do, because I don’t know how to do this sort of thing.
The way the model works is not so different from a lot of other
microsimulation models. I have got two inputs, the data sources and the policy
parameters. This enormous black box which is 15,000 lines is data code, which
basically converts everything to price responses and makes behavioral
assumptions on how individuals and firms will respond. Then there is an output,
which is population flows and dollars. So I will talk a bit through how this
actually works in practice.
The next slide talks about the data. I thought Beth’s presentation was
excellent in terms of laying out the alternative data sources. Unlike Beth and
CBO, I used the CPS instead of the SIPP. I think there are pros and cons there.
One pro is, the CPS is a little bit more recent. There is a March 2005 CPS
which can be matched to February 2005 to get both coverage and employer
offering. So that is the data I used. On the other hand, the SIPP has the
problem that it doesn’t have dynamics, which is obviously important for
modeling things like state COBER subsidies. So that is a downside of the CPS.
The other upside of the CPS of course is that the samples are larger, and
for reasonably sized states you can do state specific estimates with reasonable
confidence. So that is another pro of the CPS, but I think that is something
that we need to discuss.
My model recalibrates the later CPS by age and income cells, recalibrating
insurance coverage in particular to where we are in the most recent CPS. Going
back to what Beth was saying, often if I do state work, states will often have
their own surveys. In Northern California this is true, in Connecticut this is
true. States have their own surveys, where they will have certain targets
states think are right for number of uninsured and such, so in those cases I
will recalibrate to those numbers, because on the state basis it is useful to
be using the same numbers the policy makers are using.
Employer premiums comes from the METS IC, from their public use data. They
have tables by state on premiums by firm size, so that is what I use there. I
also got untabulated data a few years ago from the METS IC folks on the
distribution of premiums within firm size, so what the tenth, 15th, et cetera
percentiles were. So when I assign firms a premium, I essentially assign them a
draw from the distribution for their firm size, and then renormalize by state.
Then for non-group premiums, what I do is take the MEPS micro data on
people and use that to get underlying health costs. I basically take people who
have employer insurance, say what do they spend in a year based on their age
and reported health status, and then I take that age reported health status, I
assign it to an underlying health care cost, and then I assign a load factor to
turn it into a non-group premium.
The tax rates are imputed using everyone’s favorite tool now, TaxSim, which
is pretty straightforward and that is the way you do tax stuff right now. Right
now the model is projected to 2009.
On to the next slide. What is in the black box is the key to the way I
think about things, let’s take everything and turn them into prices. So take
any proposal you want, or any market change and ask, how is that changing the
prices that face different people as you consider different avenues of
Here I should have had another nice diagram which I didn’t, but my model is
divided into four insurance groups, employer insured, non-group insured,
publicly insured, which is Medicaid because I am doing non-elderly, and the
uninsured. Then I have different ways they can get insurance through non-group,
through ESI or through public.
For any given policy shock, how does that policy shock affect the prices
each different group faces for each different avenue. So you can think of a
four by three matrix, four different insurance groups considering three
different avenues to buy insurance. We are saying, when there is a price shock,
how does that affect the prices that face each different group in each
So for example, if you give me a $1,000 credit towards non-group insurance,
it costs $10,000; that is a ten percent subsidy. Then one place I differ from
some other models is, I don’t try to internally estimate as part of the model
the key elasticities. I think that these elasticities are very hard to estimate
in cross sectional contexts. I think you need more precise empirical
strategies, like drawing on natural experiments and other things, to estimate
the key elasticity. So I draw the literature to estimate the best elasticities
to use for each response rather than internally to the model, estimating the
actual elasticities, which I think you can get misleading answers. So I draw on
and use the best estimates from health economics and more credible empirical
strategies to try to figure out what the right elasticities are to use.
The next slide gives an example of some of the various avenues one needs to
think about. Let’s say for example the government announced, as Bush had in his
budget for a number of years, we are going to have a credit towards non-group
insurance. One of the questions one has to think through is, why is this 15,000
lines of code and not ten lines of code. The answer is, because with every
given change there are a lot of different questions one needs to think about.
For example, to what extent would those who already have non-group
insurance take up this new entitlement? We know that people do not always take
up even money that is sitting on the table. With the earned income tax credit,
which is free money for people, take-up is about 85 percent. So we know that
take-up is less than complete, even among those who don’t have to change their
behavior much to take it. So that is the first question, how do you model the
take-up among those who already have non-group insurance.
The second question is, to what extent will the uninsured buy that
subsidized insurance, and how will that vary by their income and by their
health status. So I have a series of take-up equations which are going to say,
when we subsidize this non-group insurance, to what extent is that going to
cause the uninsured to take it.
Those take-up equations depend on a number of factors, the uninsured’s
health, their income, whether they are offered ESI. For example, we know from
work that Jessica has done and I have done and other people have done, Linda
Blumberg and Len Nichols, there is a whole series of articles which show that
people who are offered ESI but turn it down are not very price sensitive.
So what I have in my model is the notion that if you were already offered
employer sponsored insurance, you are uninsured, and you turn it down, you will
respond less to a non-group than if you weren’t offered, because you have
already revealed that you are not that interested in having health insurance.
So that is the kind of thing that is in those behavioral equations.
Then you have the question, to what extent will those who have employer
insurance decide to move to this new option. For most employer insurance, even
with $1,000 or $2,000 credit, it will still be cheaper to stay on ESI than to
go to the non-group market. But for others who are charged a lot for their
employer insurance and who are healthy so non-group is cheap, they might want
to move over. So that is another reaction you have to model.
Then probably most importantly is how do you model the employers to all
this, how will employers respond by dropping insurance or raising employee
contributions, or how they will respond by altering the generosity of what they
provide, which in turn feeds back to government budgets because of the tax
exclusion costs of employer provided health insurance.
This is just an incredibly brief overview of what is in this huge black
box, how do we convert these policies into price changes and how those price
changes change a move to behavior.
The next slide talks about something I feel pretty strongly about, which is
doing a fully integrated policy analysis. What we are seeing more and more, you
are seeing it with everything Baucus and Kennedy and Obama are talking about,
is not something that Bush talked about, let’s do a non-group credit and
These are very integrated policy analyses, with ten or 12 different moving
pieces, from non-group reforms to subsidized care to subsidized low income
pools to small business credits to pay or play tax on firms.
What is very important when you are doing things in price terms is to
consider these all at once. I think you can get in trouble if you stack these
and say, let’s first do this one, then this other one, so let’s first do the
effect of pay or play and then let’s do the effect of low income subsidies.
It turns out you can get a lot of what we call path dependence, and very
different answers, depending on which order you stack them in. It is very
important to do these things all at once, so you are not path dependent based
on what you are doing first or second.
The other thing that I think is important that I built in my model is not
to have — to try to minimize as possible knife-edged distinctions. Once the
credit gets to be X big, then firms will pay attention. I think that may in
fact be right, but it is incredibly sensitive to an unknown, which is, we don’t
really know what X is.
I will often get asked by policy makers, how big does the credit have to be
before people will respond. The answer is always, that is not a well defined
question. As the credit gets bigger, people respond more. I think it is very
important in my view, the way I do my modeling, to not have a knife-edged
cutoff, but rather have it be continuous, and reactions continuously increasing
The next slide talks about modeling firm behavior, similar to what a lot of
people are doing here. The model is based on individual data, but so much of
what we draw in a firm reaction. So that creates synthetic firms, where I have
data from the BLS for a given wage category. So for a given worker who earns
say $15,000 what is the distribution of earnings of their coworkers. I will
have it separately by firm size and locations in the country.
So I can say, given that you are in a firm of ten to 25 employees in the
West, given that you earn $10,000 to $15,000, I know the probability
distribution of the earnings of your coworkers. I know what share of your
coworkers are likely to earn $5,000 to $10,000, $10,000 to $15,000, $15,000 to
$20,000, all the way up to $100 plus.
So what I then do is go to the CPS and pluck out synthetic coworkers for
you to match that earnings distribution. So I have built up the synthetic firm
to match what I know is the underlying correlation of wages across firms. Then
once I have a synthetic firm I can say, now let’s ask how the firm reacts based
on the characteristics of workers in that firm. So that is what goes on with
the important part of modeling firm behavior.
The last slide, I want to talk about a couple of important caveats that I
think apply to all the models we are talking about today. We have talked some
so far about data as well, but I think we need to step back and think about the
The main thing you have to remember is that it is a garbage in-garbage out
process. We are just relying on a lot of assumptions here, we all are, and the
better the assumptions, the better the answer. I think that is why it is
critical to not worry just about the structure of what we do, but where we get
the elasticities that drive all the behavioral responses, and using the best
available evidence, using groups like this to discuss what the best available
evidence is on behavioral responses. Then when that is not available, figuring
out how you can tie to that evidence as closely as possible. So that is the
The second caveat which I think is true once again, it is certainly true
for my model and I would argue for most, is the effort is going to vary
depending on the magnitude of the change. I feel very confident that if you
tell me you are going to put in a $1,000 non-group credit, I can tell you with
reasonable precision what that is going to do. If you tell me you are going to
do a new pool where everybody gets subsidized for up to five times poverty,
financed by taking away the employer tax exclusion, I am on much thinner ice.
So once we move to grander changes of the type that we are talking about
now, I think we have to recognize that the confidence intervals go up, I
presume for all of our models, certainly for mine. The confidence intervals are
going to go up, and I think we need to recognize that uncertainty. Folks like
CBR aren’t allowed to do that; they have got to give one answer, but I think we
need to be internally honest, because we are going to be getting less precise
answers as the policies deviate further and further from where we are today.
Then that leads to my last point. As much as we can do in groups like this
to be transparent about modeling the process, I call my model a black box, but
in fact I have now put together a spreadsheet which lays out every single one
of the assumptions that is in the model, in a document which describes it all.
I think we all need to be as transparent as possible in what is going into the
models, so that ultimately folks like ASPE and CBO and others who need to use
these to make policy can understand why we are getting different answers and
what is going on.
So that is all I have.
DR. SCANLON: Thank you very much. Hopefully you can stay around for the
DR. GRUBER: Yes. I might hop off, but I see you have some more
presentations, and you are regrouping at noon for the discussion?
DR. SCANLON: No, actually probably about ten minutes or so. Bo Garrett is
going to present, and then we will have a discussion.
DR. GRUBER: Okay, good, I’ll just hang on then.
DR. SCANLON: Thanks. Bo, you’re on.
MR. GARRETT: Great, thank you. Good morning. Let me thank the committee for
inviting me today to talk about the health insurance policy simulation model
that I am creating with a team of my colleagues at the Urban Institute. I am
really here to talk today on behalf of the whole team.
For people who may have come in before I did, I brought handouts with me,
so there are handouts on the table. I am going to talk about our model. HIPSM
is a detailed microsimulation model of individuals, families or HIUs as they
technically are in the model, and employers making coverage decisions within a
health insurance market. You are going to be hearing a lot of variations on a
theme now, at this point in the morning, having heard about three other models.
This one builds on the Health Policy Center’s experience with the health
insurance reform simulation model, or HIRSM. That was used to model reforms in
Massachusetts in building the road map to coverage document, Blumberg, Holahan
This model takes a lot that we learned from that, but it is designed to be
faster and easier to tailor to new policy specifications and state specific
analyses. This was jointly developed with the Tax Policy Center.
To give you an overview of my talk, I am going to cover HIPSM’s
capabilities, the model structure, current applications, and I’ll just make a
few comments about data needs.
HIPSM is designed to cover a very wide range of policies and to be
expandable to ones that we haven’t yet put in. That, just to name off a few,
includes Medicaid and SCHIP eligibility expansions, individual and small group
market reforms, reinsurance, income related subsidies, purchasing pools,
individual employer mandates. Basically it is all the components that you need
to model something like a Massachusetts style expansion, and then there are
others as well.
The output, just to go straight to the end, the output of the
microsimulation model is a bunch of detailed tables of the estimated effects of
the reforms, including insurance coverage status of people and baseline and
reform. We have the four categories, uninsured, Medicaid, public, non-group and
ESI. The costs of the reforms to different groups, government employers and
individuals, changes in ESI premiums from sponsorship, a range of things, cut
by interesting factors like income, age, health status, firm size where
applicable. These tables, we have a specific set, but they can be modified and
extended for specific needs to look at very narrow populations that one wants
I like the word cludge you used for bringing all the different data sets
together. I love that word. We have had to use a lot of that for our database.
We need a data set, an individual level data set that resembles the health
insurance situation in the United States in 2009, or maybe some other year.
What we start with as our core file is the 2005 CPS, annual demographic file.
That is matched with the February file contingent work supplement to get
information on employer offer and some other variables. It is also linked to
the MEPS HC to bring in individual level expenses and other health status
variables, the survey of income, which has a lot of information on tax related
income, and tax variables from the Tax Policy Center’s tax model, which would
include marginal tax rates.
Given the goals of this talk, one thing to point out which is very clear is
that we have to bring all these variables together from different data sets. We
do our best to statistically match them in ways that preserve the relationships
among variables. But surely there are things that are lost. If I knew a
particular factor was lost, I would try to make sure the data reflected that in
the end, but you can’t look at everything. Ideally one would have this all on
one big file, one big survey.
Like some of the other models, workers are organized into synthetic firms,
defining what the population of firms looks like out there. We have used
information from both the MEPS IC and Statistics of U.S. Business. They both
have characteristics that the other ones don’t, and together they fill out a
picture of the distribution of firms.
Then these data need to be reweighted and adjusted to match various
benchmarks that we have for coverage, income, health care expenditures. Some of
the benchmarks come from work by Selden and Sing that does reconciliation of
the MEPS and the national health expenditure data, because we want the
aggregate amount of health care expenses in the model to be nationally
We have an aging module that gets us from the ’05 survey related to ’04
data for coverage and incomes, and bring those to 2009. When we do that, we can
use a variety of — first we will get it to ’07 data, because we know what the
’07 CPS is now, and then going beyond that we will make some additional
assumptions to get to ’09 and even further, as needed.
The premiums in HIPSM are endogenous as they were in HIRSM. They are built
up from the risk pools in their underlying health care costs. We apply typical
rating rules for a typical state; we don’t do each and every state’s rules at
this point, but we take typical rating rules in the individual and group
markets and then apply those to the national population within each risk pool,
whether it is an ESI risk pool or a non-group rating.
This results in expected — we can compute expected costs conditional on
the rating rules, multiply that by a loading factor for administrative load,
and then we get our premium estimates that would vary by firm size and the
different characteristics of synthetic firms and the different non-group rating
groups. We benchmark these with the ESI at least to values in the MEPs IC and
the Kaiser HRET employer benefits survey. In the latter case, if you look at
the tail data you can get a distribution of what the premiums out there look
like by firm size, for example. It helps us target those better.
Talking about the structure of the model, the behavioral effects in the
model, it uses a utility based approach. When we started this a few years ago,
we thought that would be very helpful in helping us — we intended for it to
help us get better estimates of the effects of reforms that are well outside
our historical experience, rather than an elasticity base model which is more
It also provides a general framework. In the Gruber model, all these
potential effects were converted into prices; we leave ours in their natural
form, because we specify this utility approach, so everything does get
eventually expressed into changes in relative utilities of the different
In the model, individuals choose the option that provides them the highest
utility as we determine it to be that firms offer. If there is total
willingness to pay, it sees the total cost of offering.
To get at things like people who look like they want to be in Medicaid but
they are not, you have to deal with these latent factors. We describe total
utility as a specified utility function which I will talk about, plus a latent
utility error term. That is because we need to assume that existing coverage is
optimal at baseline. We find you in this particular category with these
particular characteristics. For the model to work, that must have been what was
best for you at the time, at least in some sense.
The key to HIPSM’s mechanisms then are imputed error terms that do two
things. They have to insure that the baseline is optimal, and they have to
yield premium elasticity and take-up rates consistent with the targets. This is
where all the behavioral stuff is going on. This is our black box. This is
where we have to do calibrations.
What we do here, we have got the underlying utilities. We have got error
terms. Locally, for say a 20 percent change in the ESI premiums, we want firms
to quit offering if premiums were to rise by rates predicted by the best
elasticities from the literature. So we are targeting firm level take-up, firm
operated individual ESI take-up rates, non-group elasticities, Medicaid take-up
rates by whether you are uninsured, by whether you have ESI, which would yield
the assumptions we make for CREDA.
One of the things that the utility framework also does is, it automatically
will generate some heterogeneity in effects. We may have targeted an overall
aggregate elasticity for non-group, but then people who have higher health
expenses may be more likely to move because any change for them may be
magnified by the fact that it is all in a higher amount of dollars given their
The utility functions themselves are developed based on economic theory. I
believe the ones used in the RAND model are quite similar. It depends on the
expected out of pocket health care expenses, variants of out of pocket health
care expenses, the value of health care that you consume. That is going to
differ, depending on if you go from being uninsured to insured; you will have
an increase in your health care consumption, and that is going to add some
value. Out of pocket premiums, tax incentives, and we also have an
affordability component which is your expected out of pocket expenses of that
option over your income.
Let me walk you through how the flow goes in HIPSM. Reforms change the
available options to you. It may change rating rules, it may change relative
prices of ESI versus non-group coverage. This would also have potentially a
change in workers’ willingness to pay for ESI within the synthetic firms. Firms
will then react to these altered worker preferences.
Given what firms do, individuals or families will choose the new best
available coverage option given the firm’s decisions. The premiums need to
adjust to reflect the characteristics of the new risk pools. Then the model
will iterate until coverage changes across iterations and is stable.
To walk you through a particular example of Medicaid SCHIP. Let’s say
public coverage becomes available for new eligibles. I should say that the
current eligibility model in HIPSM is a detailed Medicaid eligibility
simulation model that we start with, but then we can specify an expansion as
covering everybody, like all children at 300 percent of poverty and adults at
150 percent of poverty. This may be particularly attractive for new eligibles
that had higher out of pocket costs. So this reduces the demand for ESI,
because now you have got this new option that is free or very low cost. Fewer
firms will offer ESI, so that will cause some people who may have had ESI and
aren’t eligible for Medicaid to lose it.
Overall, Medicaid SCHIP coverage should increase, uninsured and other types
of coverage would likely decrease. Government costs and total health care
spending rises. Private premiums will adjust, then we go through a round of
second order effects.
The model up to this point that I have described is mostly about coverage
changes. We can bring in alternative assumptions to the model results that take
into account various supply side considerations and others.
For example, we were looking at a particular proposal. It may have
proposals to invest in various cost containment strategies. You look at the
literature and consult experts and come to a conclusion about what kind of
effect those would have. Those can be then included in the model’s output.
Another example would be supply constraints that might occur if you have
universal coverage. The model may say costs are going to go up by this much
because it is initially assuming that there is going to be supply to meet
demand. That may not be true. If not, we can take that into account.
Our baseline assumptions would have consensus public program participation
rates from the literature. A proposal might include the fact that there is
going to be a huge outreach effort. Let’s say the best estimate of what that
would do from synthesizing the literature would be a ten percent rise; we can
include that in.
Also, because we have got this aging module, we can have various scenarios
for future wage, employment, health care costs and insurance premium growth.
Let me talk about some of the studies, well, I guess this is all of them
that we have underway. There are national studies and state specific studies.
The national studies, we are doing a variety of Medicaid SCHIP expansions for
Kaiser, where we are specifying different levels of expansion as a percentage
of FPL for children, parents and non-parents separately. We are doing this with
and without enhanced outreach efforts.
We also presented some preliminary analyses at the AEA meetings of a set of
reforms at the national level that would resemble the types of reforms that
would be contemplated. These build on each other in sequence in the way that we
present them. They are not sequential in the way they are applied, as John
Gruber was speaking of. They are integrated. These happen together when we are
specifying them as a group. So we looked at a Medicaid SCHIP expansion only if
you reformed the non-group and small group markets, like applying age rating
and giving low income premium subsidies, and what would happen if you had the
two of those. Then what would happen if on top of value added employer pay or
play mandate, et cetera.
We are also looking at what would happen if there is no reform. If there is
no reform, we can make various explicit assumptions about what the growth in
health care costs would be, if there is any growth in premiums over that, what
the growth in wages would be in the next five years by different quintile of
wages perhaps, and include that change in health care costs relative to changes
in the model which would reduce the amount of coverage and increase the number
of uninsured, but also have the demographic growth by age and gender cells, and
predict changes in employment and patterns like that.
We are also doing some state level analyses. I should talk about some of
the data issues particular to that in a little bit, but we are modeling a large
array of expansions for New York State. We are looking at a wide range of
things likely to be considered by a state or the federal government. We are
doing Massachusetts style expansions, looking at a single payor example. Within
the combined public-private approaches, we are doing everything from little
tweaks to much more comprehensive things.
We are also doing something like, if there is no health care reform, for
the Colorado Health Care Foundation and Medicaid SCHIP expansions and some
other things in Colorado.
I think Beth did a great job covering the data needs. I am just going to
make a few points. I’ll second the need for employer linked data. We would like
to have a large data set linking the two at the most superficial level for
better construction of synthetic firms, or it would be really synthetic at that
Also, for the testing of core hygienics about how firms make decisions
given how workers sort into firms and how they arrive at those decisions to
offer or not.
There is no national data set that has all the variables we need. There is
certainly no state level data set that has all the variables we need. Sample
size becomes a problem. A state specific database that is commingling all the
characteristics we need would be extremely useful.
I should say that we do the best we can. I think we do a reasonable job of
compensating for the data we don’t have. We will take several years of data in
a state or a region that we think can represent a state and then reweight that
data to look like particular data in a particular year, and bring in various
pieces of state information to try to take what we can know about the
characteristics of the state and make sure our database reflects that, before
we proceed to simulations. But obviously this would be much better if we had
real data to do those.
Thank you very much.
DR. SCANLON: Thank you, everyone. I think you gave us a lot to think about.
I’ll open this up for questions.
DR. STEINWACHS: Two of the modeling efforts use elasticities alone, if I
understand correctly. You are moving to utilities in RAND.
Could you say a little bit about two things I was interested in? One is how
much difference you see between using elasticities alone and using utilities.
Then, I wasn’t very clear about what data you use or what research findings you
use to get at the utilities themselves.
MR. GARRETT: In our case, we are probably looking at the same elasticities
that John Gruber would be looking at and John Gruber in many cases estimated,
and making sure that for a 20 percent change in non-group premiums, that we are
going to get in that particular case an effect similar to what a CBO study
says, in terms of what is going to be the percent change in the utility if
there is a 20 percent change in the non-group coverage.
So locally with that just one thing going on, we seek to be consistent with
what we determine to be the fast read of the literature. But then if you had
some really big change, if you had two changes at once, we wouldn’t just be
applying a particular loss in sequence; they would be interacting in the model
according to peoples’ utilities, so a different elasticity would come out than
the one you targeted to because the situation is different.
If you had a really big expansion, the intention of having this utility
approach is that you won’t get a constant utility as you go out or a constant
function going out. It will depend on who is being affected and what they value
in the model. We hope that that way we can get better answers to some of the
big what-if questions.
DR. STEINWACHS: So when you estimate a utility for a subgroup of the
population, how are you arriving at that estimate?
MR. GARRETT: That is a very good question, how do you start with utilities
to begin with. There, there were several papers in the literature that we
looked at, what would go in them. Some of the people at AHRQ did some papers.
Sherry Reid did a paper, Colleen Herring did some papers.
They would specify, here is what a utility function for different coverage
types might look like. There might be very particular reasonable coefficients
of risk aversion that one would have various assumptions. We would probably use
an assumption close to one Sherry Reid used in a model, about the value of
additional health care. So we will draw on those. We can experiment with them,
too. We will start with those. To the extent that that doesn’t yield what we
need in terms of cost, this is what the error times yield.
DR. GIROSI: I think Bo and I read the same papers.
DR. STEINWACHS: That sounds good. Evidence based approach, I like evidence
DR. GREEN: I’ve sure got a flood of questions, but I think I will save some
of them until the next session, too. Before I ask a couple here, as a
recovering physician, I’d like to thank you for the therapy session. Ever since
the publication of the Chasm work, you realize you go to school for a quarter
of a century, and then it appears you don’t exactly know how to answer the
patient’s question. It has been very reassuring that we are all in good
Do you guys do autopsies?
DR. STEINWACHS: On the health insurance system.
DR. GREEN: Once you have done some projections and then somebody makes a
decision, do you then go back and see how close your model was to what actually
happened? If so, do you have anything to say about any of these approaches,
which one is demonstrably more likely to hit the target?
MR. BAUMGARDNER: We go back on things if they actually happen. With the
non-elderly market, I can’t think of examples recently, but when we look at
Medicare and Medicaid projections at the end of the year, we have to then go
ahead with a baseline forecast for the next ten years, and are looking at what
happened in the last few years, how that turned out compared to what we had
expected at the time.
So there is that kind of low reactor that goes on. If we had more time, I
think we would do more of it, but we do some within the limits of the time we
have got to do it.
MS. MC GLYNN: We are looking at going back and looking at the Massachusetts
experience. I think that is the main big one we have, and look at how well we
would have predicted or replicated what happened. It does give you an
opportunity to learn what was unexpected when things get put into the real
world. We have a way to incorporate that back in.
We had to first do the step of putting the multiple options together, which
we have just done, because we previously had just looked at options one at a
time. Since Massachusetts was a multi option policy change, we weren’t quite
there yet. We are now pretty much at the point where we can — once we finalize
the utility maximization, we can go back and try to replicate that. But we
think that is a very important way of continuing to refine the capabilities of
the model. The big challenge as Jim said is that we don’t have that much real
world observation to draw from, but it is important.
MR. GARRETT: One thing about the Massachusetts expansion that would
certainly provide feedback to me is, before the individual mandate was in
place, a lot of people responded.
It might be more people than responded to a $500 penalty that hadn’t even
occurred or more, but the penalty hadn’t even occurred. If it does occur, now
that it has, is it commensurate with any old thousand dollar penalty of not
doing something. If there is an overwhelming response, you might think people
maybe are not being penalized much more than they would getting the equivalent
So you say, what to do with this fact. When we do our individual mandate,
we might just do it to assume straight out, rather than what is the response to
this financial penalty, for the individual mandate we might say let’s assume it
is 100 percent effective or 90 percent effective. Then how much would it cost,
and then you can do sensitivity analysis around that and create some different
DR. GREEN: I listened to all of your presentations, the title of the
hearing here, modeling insurance data including coverage, access, utilization,
quality and cost. But I was struck with the lack of consideration of results to
people, like emotional status, to cartoon it in this model. There would be an
expectation that more people would get better or fewer people would get better
or more people get worse or fewer people get worse, even measuring something
related to morbidity, comorbidity or functional status.
My question is, is that because it is out of scope and it was not what you
were asked to do? Or is it that you use it as a modifying or confounding
variable in the analysis at some point that I missed? Or is it a dependent
variable that is not available? Why do these types of analyses not drive it
home to the fact that all of this eventually needs to track to whether or not
MS. MC GLYNN: We actually do have health outcomes in our model. I greased
by that fairly quickly. I hoped to have Internet access where I could have
shown you those results.
We produce estimates that are primarily right now life expectancy
estimates. We hope over time to make those more detailed.
The challenge there frankly is the data limitations. When I talked about
the dream data set of the birth to death HRS, right now a lot of what we have
done is to build on work we have done previously called the future elderly
model, where we looked at changes in Medicaid policy and the health effects on
the population. We were able to draw that down with HRS with some comfort to
the 50 and older population, but then to look at health effects for those under
50 population, which fortunately have fewer deaths you have to worry about. It
is a bit more challenging.
But we actually do produce health effects, and that is something we are
going to continue to try to build on.
DR. GREEN: Can you help us understand those data limitations? Those are
really important to us. We are interested in data limitations.
MR. GARRETT: In a particular application, I think it would be reasonable in
many cases to say — our model at this point mostly is about the coverage
changes. That would be very reasonable if you were writing this up to say,
these are the coverage changes that we observed. Here is what we know about how
these will affect people’s life span, morbidity, et cetera, and talk about
those. If one had some marginal effects or something from the literature that
one wanted to add to that section, you could.
Of course it is a point that is very important to discuss. We would want to
highlight those in applications. It is not critical I think to getting to the
coverage estimates, but then once you have got the coverage estimates you can
talk about what further implications those would have.
DR. GIROSI: If you want to get into the health part of it, you do need to
follow people for a long time. The SIPP and the MEPS follow people, yes, but
for a relatively short amount of time. So if you want to see people developing
diabetes or developing a heart condition, you need to follow them for a number
of years. That is why the health and time survey is a wonderful survey which
follows people for up to 12 years biannually. It is a great source of
information of how people transition from one health state to another. You can
also correlate that with the insurance they get, the access to treatment they
get. That is why we were able to do some of that for the 50 and older,
basically. But the HRS does not do anything for age less than 50, 45, depending
on which ways you get it. So that is one of the data limitations.
DR. GREEN: I take it from those responses that — you mentioned this, Beth,
in your presentation too; is the key data limitation insufficient longitudinal
data about particular individuals?
DR. GIROSI: For doing things in health. You really need to follow people
for a long time.
MS. MC GLYNN: I guess I would also argue that one of the challenges I saw
when we were looking at the health results is, there isn’t a lot of good
detailed linkage about coverage, what that means for access, what that means
about whether people are getting the care they need, which I know from my own
work is not — coverage is not a guarantee; there is quite a bit of gap.
The thing that many of these things lack is any kind of real data about the
continuation of receipt. So we make lots of assumptions based on pretty narrow
bands of observation about, if I observe a patient who has been observed with
hypertension today and is getting hypertension meds today, we know what that is
going to look like in a year, five years, ten years.
So we made for chronic disease a lot of assumptions about the maintenance
of effort over a long period of time, which is probably the real key to the
ultimate health outcomes. We really have very little good national data. We
have lots of cross sectional data that we use to make those kinds of estimates,
but I think they are fairly heroic with respect to where we are really at. But
we are certainly trying to do the best we c with the data.
But I agree, if you are really interested in looking at health policy from
the perspective, as your committee, I presume does, of what the impact is on
the health of the population, which is ultimately what we ought to care about,
then I think we are sadly lacking data sets that help us with answering those
kinds of questions.
DR. SUAREZ: One of the interests I have is minority health and health
disparities. One of the things that I didn’t see and I wanted to explore that
in these microsimulation models and approaches, is to what extent — and there
is some literature out there about the difference in decision making by
minority groups of health insurance — to what extent the models can allow you
to look into the effect of policies on the decision and the insurance coverage
and other things on minority health groups, specific minority groups, then look
at the effect of those decisions on the outcomes of utilization, quality on
Is that something that your models can look into?
MR. GARRETT: I’ll just start. We have not looked at that much with the
model yet, but those data are there in the baseline file, and one could
tabulate the effects generated by the model by race and ethnicity.
We haven’t made specific tailoring to the utility function specifications
according to that. If there were some evidence to bear on that, we could.
The people would be found in different types of jobs with different
incomes, and to a large degree that ought to be reflected, and would have
different health statuses. To a certain degree that ought to be reflected in
the final results already. But that is an interesting question.
MS. MC GLYNN: We also haven’t done that much with it. We do have race data.
Often when we get into the analyses you end up with white/non-white being about
as discrete as you can get, even if you start with more detailed data than that
on the starting sample.
One of the things is a good capture in terms of all of the different places
people are getting care, how well we really understand. So we do know for a lot
of minority populations there are different avenues of seeking care. We don’t
think those are particularly well described in the data sets that we are using.
That would be helpful to have more data on that.
From some very small studies we have done, I suspect some of those suggest
some of those providers are delivering better care than what people are getting
in the middle. So the Medicaid population for instance may be less well served
than people who are going to free clinics, from some work we have done.
But it is something we certainly are interested in. We just haven’t gotten
very far with that.
MR. LAND: You mentioned the various problems with the data sets that are
available. They are not very timely, they don’t have the longitudinal approach,
they don’t have the data items you need and so forth, even down to the fact
that you don’t have very good computers that you can access.
I know the National Center goes hat in hand to other federal agencies to
solve some of their financial problems. Does AHRQ and Census Bureau and any
other data sets that you are using go to you all hat in hand and say, we could
do better if you could contribute to the cause, even to buying another computer
or to going to all the think tanks and research companies and saying, how about
kicking in some money? Does that happen? Is it possible? Would you participate?
We all know that the federal statistical agencies are challenged by
resources. I am just wondering if they have looked into other ways of helping
them out that would help you all out?
MS. MC GLYNN: As far as I know, the main solicitation that we get frankly
is to be asked to participate in efforts to lobby on the Hill for better
funding for these agencies. I did not use the word lobby. Strike that from the
But anyway, as far as I know, we have never been asked to do anything like
that, other than try to be a voice of support for the broader needs. I recently
got asked this by a committee staff on the Hill, about what the broader
research data needs were in a particular area.
So we get asked those questions, and we certainly try to lend support for
these kinds of efforts. For instance, on this longer HRS type longitudinal
data, we think there was something like that on the books, and then it got
cancelled. We certainly tried to speak in favor of that. We have not been asked
directly for contributions, old computers that we could offer.
DR. SELDEN: I think if there is any sort of issue like migraine related
with the screens that we have, I don’t know if that has been raised. But I
would be happy to talk to you about that, and see if we can —
DR. BANTHIN: AHRQ is not a — we cannot collect funds from people we serve.
We are prevented by law from doing that. But we can look into better computer
screens. We did buy a new server, a new high speed server recently for the data
DR. HORNBROOK: I was wondering whether in today’s world, with the emphasis
on loss of jobs and job loss and employers’ decisions about where to make jobs,
in foreign countries or in the States, how important is it for us to start
thinking about modeling the employers’ behavior under mandates versus under
totally being free of any employment related health insurance that goes into a
different kind of direct access through Medicare or Medicaid or some new
Beth indicated that the COMPARE model doesn’t talk much about the decisions
about jobs and loss of jobs. Do you have a sense of how important it would be
to focus on the employer behavior in these models?
MR. GARRETT: Our model also doesn’t endogenously have people losing their
jobs. We take the distribution of individuals as we get it from the CPS. We
will weight it differently to get to ’09, according to what happened or what we
believe happened, and farther out as well. But that is not having individuals
in the model lose their job or not.
All the models have people able to lose their health insurance coverage
should some other alternative become available, or lower cost alternative
MS. MC GLYNN: Colleagues of ours, Nera Sute and Danny Goldman, have
recently done an analysis which I think will be published in the next three
years in health services research, whatever the time lag is these days, that
looks at the effects on employment and productivity for industries related to
the proportion of employees covered by health insurance. It does an analysis
over two decades and does show an effect and does some comparisons to like
industries in Canada to try to get some sense of what it would look like in a
different policy framework.
So we have been talking to them about how we can take those kinds of
results and import them back into our model over time. We are just at the early
stages of talking about that. But it does look like that might be — that is a
way in general that we can bring things into the model to try to better
understand some of those more macro effects.
At the outset we tried to limit what we thought was possible. The general
feeling was that a lot of the macro stuff was harder to get your hands around
in the context of the microsim. But we are continuing to look at how we can do
that, particularly from where we started to today, the importance of that as a
policy context has certainly increased dramatically. Now we are re-upping those
efforts, plus we have gotten a lot of work under our belt so we can entertain
those possibilities. So we will see how that goes.
DR. HORNBROOK: One of the things with national health insurance would be an
auto rescue package.
MR. BAUMGARDNER: What we do, and I think I have to separate what is in the
health insurance model I talked about today from things that are done to put in
the HL settings.
We typically are dealing with a ten year budget window, so there is some
baseline coverage by coverage type that we are starting with every year, and
then you let the model run. How do we set that baseline insurance number, the
number of people without insurance every year? We do go through a separate
modeling exercise. Part of it is going on right now, where we do look at
changes in the unemployment rate within CBO’s macroeconomic forecasts. So some
of the issues you are referring to, it starts off with CBO’s macro forecast of
the unemployment rate, and then feeds into what we think the baseline
uninsurance will be, along with Census projections on changes in the
population, and at some level ethnicity also feeds in as well as our
projections based on the national health expenditures on premium growth. Both
evidence in the literature and some stuff we have done in house to put together
the population by demographics, and premium growth and unemployment rate
changes feed in to get baseline uninsurance amount.
Most large scale proposals, by the time they could possibly be implemented,
at least within CBO’s macroeconomic forecast these tough times we are in are
over by then, so you are back to a more normal rate of unemployment by the time
MR. J. SCANLON: Could I ask, how would you approach modeling some of these
relatively newer concepts at a conceptual level that were mentioned in the
various campaigns currently in the health policy community? Such things as a
national or regional health insurance connector or exchange, and then coupled
with that some sort of a public plan option. Is it possible to — have you done
any thinking about what assumptions would you make, what would it look like,
how would you model those parameters?
MS. MC GLYNN: We have a national insurance exchange as part of what we have
DR. GIROSI: This is something which is reasonable to do. It is within the
scope of the simulation. Things in the pipe within the range of all the models
that have been presented, I believe.
MR. J. SCANLON: You would just need specifications?
DR. GIROSI: Yes, specification work needs to be done, because that is the
problem; sometimes we don’t know exactly what does the plan look like. But that
is certainly within the scope.
MS. MC GLYNN: Our basic approach to that is, generally when we talk to
policy makers, getting detailed specifications is challenging, is that we
basically have a checklist of what we need to know in order to model this
option. We begin filling that in with what we have observed in other policy
proposals or what people we talk to or what the literature suggests might be
reasonable ways to think about it.
We do stress testing the limits of a design so that we think about things
that should push the model to do either disastrous or wonderful magical things
to try to understand how it behaves under different assumptions, and then have
available — and this is what we have done in all these policy options — a
fair amount of flexibility to mess with the parameters, I’m sure there is a
better word than that, do permutations of the combinations of the parameters in
order to come up with a sense of how that option would behave under lots of
different design scenarios. So that is the basic approach.
Then a lot of times, that starts a productive discussion about whether
these things make sense once you have seen these results. We have had this
experience, you show policy makers this is what it looks like when it is put
together the way you were thinking, and wow, that isn’t really what I was
hoping, what if we either add this element or jury rig with that element. That
is all stuff that we can do.
In fact, the design of what we put together was intended to allow people to
engage in that sort of what-if exercise and get a better sense. I think in
general what we know is that people’s priors are not always right. Not that the
modeling, to Jonathan’s earlier point, is perfect, particularly the more out of
nature we get with the modeling, but it begins helping have a discussion that
is certainly better than making stuff up, which is the more normal approach.
DR. SCANLON: Nancy, you get the last question before the break.
DR. BREEN: This is more about the wish list. If we were to obtain
additional data that would help with the modeling and have more evidence to
work with to maybe make fewer assumptions.
I am wondering, I think all of you used either the SIPP or the CPS as your
starting match file. I know that a lot of states are starting to collect data,
and a lot of them are telephone surveys, and they don’t have much identifying
information associated with them. I wondered, is that a limitation.
So my question is, what identifying information do you need in order to be
able to do the matches you need in order to model? Then a follow-up question to
those is, if you were able to model at a more local level, would that allow you
to bring more of these macro variables in and look at health disparities in a
more meaningful way than you are able to at the national level, just by virtue
of having better information and knowing what the political parameters are.
I was struck, Bo, when you said that you had a typical state insurance, and
I thought, what could that possibly be?
So anyway, I just wanted to know about the information you need on the base
file for matching, and then if you had more local data, would that allow for
better macro modeling and modeling of health disparity outcomes.
MR. GARRETT: In terms of matching, it is an interesting question. When a
state has a smaller survey of the residents, they may not have all the
covariates you want on there. The CPS would have the variables you need. You
could either learn what you can learn about coverage rates as the state sees it
from that file and bring that information over, and make sure that what you use
is related to reflect those assumptions. Or if you had enough variables you
could do a separate matching, if you had a really detailed state level survey.
In terms of identifiers, given that we are not matching the same individual
across these surveys, we are matching on the various covariates of age, health
status, gender, a laundry list.
DR. BREEN: All you need is a rich set of covariates, then. You don’t need
MR. GARRETT: When used as a rich set of covariates, ones that have real
data. The more covariates you have, the better. So for example if we had had
state available from the MEPS public use file that we used, we would have
matched on state when we matched the CPS as a factor.
But in terms of a typical state, in our case the typical state would have,
for the purpose of non-group payment, to answer your question, would have age
rating. It would have rating on health status, and it would not be community
rated. That is the assumption we make with the local state.
MS. BULLHEIMER: Could I just follow up? Do you match on race and ethnicity,
to Walter’s point? Is it possible to do that?
MR. GARRETT: In the MEPS? We do the match, yes. The short answer is no.
PARTICIPANT: Could you repeat the question?
MR. GARRETT: The question is, when we match observations from the CPS core
file that we start with, to the MEPS, do we match on race and ethnicity. The
answer is, we take race and ethnicity into account in the match. What we do is,
rather than have hundreds or thousands of cells for various characteristics, we
start with a few set of broad cells.
MS. TUREK: There are very different sized surveys, so how often is a MEPS
person matched to the CPS?
MR. GARRETT: They are not the exact same people, and some people get
reused. So a person who has a particular age and health status in the CPS is
going to get some person with the same in the — we define some very broad
cells on the basis of gender, age and income. Within each of those cells, we
have a very detailed regression of health expenditures that we estimate on MEPS
From that, one gets an expected value, an expected expense. That is like a
propensity score, but it is ultimately more of a continuous measure. Then we
match — within these broad cells, we match people who have similar expected
costs. Then arriving at those expected costs, we use race and ethnicity as
factors, as well as employment and a host of other things.
We do a fairly similar thing. Basically we started with a wish list of
covariates along which we want to match, age and health status and whether the
person works, the region where they live, race, ethnicity, gender.
So we start from a longer wish list and then we divide the data in cells.
But sometimes you can’t find someone in the same sets, especially because the
MEPS are kind of small. We are going to get two years of the MEPS.
So sometimes you can match along, take the race and ethnicity into account
if you are in a bigger cell, if you start from a bigger cell. But then if you
are looking at the non-group market, people aged 25 to 30, you are not going to
be able to match on race and ethnicity, because you don’t get enough people
there. There is not too much information there.
MR. GARRETT: I should mention this algorithm that smartly computes cells
based on a bunch of characteristics that will come out using some kind of
artificial intelligence to come up with that. We haven’t put that to use yet,
but I want to give them credit for assisting and making me known about it. We
expended it once, it was fairly comparable with what we had already, and we
didn’t pursue putting it in. But it is on the agenda.
DR. BANTHIN: Tom will talk about it when we are up next.
DR. SCANLON: Let me thank you all. John, thank you for hanging in with us
on the phone.
DR. GRUBER: I’m sorry I couldn’t participate more actively, but thanks for
DR. SCANLON: That’s okay, I understand. The phone is not the ideal
situation, but we really appreciate your contribution. Thanks again.
DR. GRUBER: You bet.
DR. SCANLON: Debbie is going to warn everyone about the new security. We
are going to take close to a 15-minute break. We will come back at five of 11
for the second panel. Thanks.
Agenda Item: Panel 2: Modeling Health Insurance Data,
Practice and Perspectives, Continued
DR. SCANLON: We had the dilemma in the first panel that we could have gone
on a lot longer in terms of questions. Now we are going to add to that dilemma,
I’m sure, through what we are going to hear.
Jessica, do you want to start?
DR. BANTHIN: Yes. Thank you very much. I will be talking with Tom Selden.
We both work within the Division of Modeling and Simulation. We are a small
group of six economists who work within the Center for Finance Access and Cost
Trends at AHRQ.
We wear many hats here, unlike some of the other presenters. We are
involved with collecting data, putting out public use files. My group in the
Division of Modeling and Simulation, we are very much involved in developing
augmented files and tools to help other modelers and simulators use the MEPS
data. Then we also develop and maintain our own simulation models.
Because we build models ourselves, we are very much aware of many of the
gaps, so we will talk about that in our presentation. We even have NISBE
statistical tools that will help us a lot in filling some of these gaps, and
these are available to others as well.
We are going to divide the presentation into two parts. We will first talk
about the household data and our household based models and tools, and then we
will talk about the MEPS insurance component, which is the employer survey, and
some of our work there. Again, it is public use files, data products. We also
do a lot of basic research to estimate some of the key parameters that are
inputs to many of the models around this room.
The MEPS household survey is a one stop data source for many of the key
elements in microsimulation models. At this point we have about 13,000
households, 35,000 persons. It is the non-institutionalized population from
1996. It is ongoing. We follow people for two years. We have monthly insurance
coverage. We have many insurance coverage variables as the committee knows, I’m
sure. We also have sections on employment offers, take-up access to care,
What is missing here is, we have detailed measures of health status, health
conditions. These are self reported data. They are valuable. Disability status,
work loss days, things like that. Some of the things that are missing though
that modelers, economists, would like to have. Although we have non-group
premiums and employee contributions to premiums, we are missing the employer
portion of the premium. That is really critical when you want to look at tax
exclusions and tax subsidies. We also don’t have premiums for those people who
have turned down coverage. Economists like to know that because they want to
know what someone has given up, what the price was when they gave it up.
Furthermore, except for one year, 1996, where we did collect policy
booklets and we abstracted detailed information on the benefits, we have not
repeated that data collection, so we are missing information on the generosity
of the coverage or the adequacy of the coverage that people do have.
This is an issue of growing importance. With rising health care costs,
employers are looking at all sorts of ways to cut back on the generosity of
that coverage in ways that people don’t notice or don’t mind. We have very
different policies out there in the market. There are catastrophic policies
that cover the high end and nothing much up front, but there are a growing
number of front end policies. This is like what Walmart provides, policies that
cover some preventive care, regular care, but you are not necessarily going to
be protected in the event of a catastrophic illness.
We do a lot of things to make the MEPS better to use for modeling and
simulation purposes. We have detailed income data which we then — we use the
MBR TaxSim and we compute federal income tax simulations.
Very importantly, we align our data. We periodically reconcile the MEPS
expenditure estimates with the national health accounts. We have done that
twice, the last time for 2002. We create a file where we have augmented the
MEPS and bumped it up so that it benchmarks to the NATA. We then use that file
to project forward to 2016. Those files are available for downloading by
request on our website. We are improving them. They will soon be available for
We also have developed imputed employer contributions from the MEPS IC.
Those are not available publicly yet, but we have them. In addition to
benchmarking our expenditures to NHEA, Tom has done some additional work to add
in provider tax subsidies spending that is not tied to specific patient events,
which is out of scope for MEPS, but in scope for a modeler who is trying to
benchmark to national health accounts data for this population.
We have immigration citizenship status that is imputed based on some
collected data, and then we impute documentation status, but only through ’05.
We have fully imputed jobs variables which are very useful. What is not here
is, we have a monthly expenditure file that some simulators like. It is being
used currently by people at Treasury. We have allocated the annual expenditures
in the MEPS to the 12 months of the year.
Let me just go back and talk about the importance of reconciling our
expenditure estimates to the NHEA. This is critical. I was involved in the 1993
efforts with the health care tax reform way back then, and I remember some of
the very early estimates where there were three different groups providing
preliminary estimates of what it would cost to reform the health care system.
There were three wildly different results. We realized that we had all started
at three very different places. We hadn’t aligned our benchmarks.
Simulation modelers have come a long way since then, so we undergo this
very time consuming detailed reconciliation of our expenditures to the national
health accounts. We work very closely with our colleagues at CMS to take things
out of the national health accounts as precisely as possible. These results are
published in a paper.
Since then we have updated this. Tom has updated it. We have a working
paper that is downloaded, Selden and Sing. You can download it from the Web, in
which we have done additional work to make up for example for the sim tail in
the MEPS. It seems likely we are not sampling really high expenditure cases.
These very sick people may be attriting from the survey, so we have done a
little statistical correction to make the tail look more like what we have seen
in some commercial claims data.
Then taxes. We have a detailed income and assets section in the MEPS. This
is an important part of our simulations, to look at the tax effects of any of
Then I just want to mention, these are a selection of some of the papers. I
included a handout with the full citations of selected papers, where we have
done some work that estimates some of the — we can make use of the MEPS
because it is such a rich database to estimate some of these key elasticities
and parameters that are key inputs into our models and other people’s models.
Before I turn it over to Tom, we also look at burdens of health care. MEPS
is particularly well suited for looking at financial out of pocket burdens when
we have good measures of income and we have good measures of out of pocket
spending. This is one of the strengths of the MEPS, and it is a key output that
policy makers really do care about, who is going to be affected by these
reforms, who wins, who loses. By modeling the effects of reforms within the
MEPS, we can get very accurate estimates of changes in burden across subgroups,
including minorities, although many of the key differences are along education,
income and source of insurance sources. But this output measure is very
important to our model.
DR. SELDEN: As Jessica mentioned, we are taking a two-pronged approach. We
are trying to develop and enhance the data products to facilitate these
modeling efforts that we have been hearing about. We also undertake our own
One of the key areas in which we have been providing estimates for quite a
while is through our KIDSIM model, where we have a set of spreadsheets with
program parameters. I think it is up to about 120,000 cells now of details
about all the different programs in all the different states to try to think
about — we merged this into MEPS to try to think about which kids are eligible
for which programs for health insurance. We developed then estimates of take-up
and we also look at the number of eligible uninsured children.
There are some differences among the different models. The CBO came down on
our side on that. One of the advantages that we have is that we start off a lot
closer in terms of the number of kids who are on Medicaid and public coverage
in general. So it is a real advantage when you go try to figure out who is
eligible and uninsured.
We use this to track progress of these programs over time, but also to
think about alternative scenarios for expanding eligibility or reducing it. We
have looked at the impacts that might have in terms of crowd-out and program
take-up. We have also estimated the cost that expansions have. It is not just
the budgetary costs, but there are also offsetting savings. If a child is
crowded out, then you don’t have to have the act subsidy for ESI. If a child
would have been uninsured but at risk of falling down into a Medicaid
spend-down, those costs are saved but they end up in the SCHIP budget. So there
are offsetting costs there that we have tried to take a look at. We are
currently updating those estimates. We are at 2007.
We have also expanded KIDSIM to include program eligibility for non-elderly
adults, so that is a major enhancement of providing some estimates to CRS, by
using this in the model.
One of the key aspects here is that we do have detailed information on
employment and disability status as well as insurance and expenditure by source
of payment. So we can triangulate to get at a very hard to model aspect of
this, which is eligibility through disability.
We have in the past had a much more integrated model where we would look at
specific health reform scenarios. We are not quite there. We are more looking
at ways that we can provide data products and basic research that assists other
programs, but we are also — as we build more and more, we are getting closer
and closer to having something like this in the future.
The big shift here in the talk is when we start talking about the MEPS IC.
People have mentioned it before. You have got a basic problem. You have a
household survey. You see a worker, and the worker’s spouse potentially, but
you don’t see all the other coworkers at that firm or even at that
establishment. If you have an establishment survey you see all the workers at
the firm, but you don’t even have any of their spouses.
You really have to have both pieces together. But it is very difficult to
do that. You can start off with a household survey and go talk to people and
their employers and find out about the employers, but then your sample response
rate is still way down.
So we have developed a methodology for bringing things together, the HC and
the IC. The IC is a really nice data source. It has got a very large sample. It
has got a very good response rate. Right now I think it is primarily being used
for modeling microsimulation efforts. It is primarily being used to some sample
means maybe by firm size that people are using to calibrate their model. But
there is so much more that can be done with that model add we are trying to do
A few years ago, an effort was made. That has lapsed, I won’t go into that
too much, but we are trying to resuscitate that by hopefully gaining access to
the IC data again so that we can populate these establishments with MEPS HC
One of the tools we use to do this, and it was really developed for other
things, but we have adapted this tool a bit, and it is available to everybody,
is this SASMAC we call HD call-up. It uses a lot of artificial intelligence to
think about the problem of how you match two data sources together when you
have thin cells, you have a very long wish list of dimensions on which you
would like to match up, and you have different sorts of parameters that you
want in terms of the minimum cell size and the rules of collapsing cells. You
can specify this stuff as a user and have this cranked through and do your
matching for you. That is available for anyone to use.
We use that here. We also use some raking plus stratification methods that
try to do the very best job we can on aligning the two data sets on about 20
different variables. What we end up with, we think, is a pretty sophisticated
model that then can at a minimum be used to produce the sorts of benchmarking
estimates that people really feel are most necessary.
So if you are dissatisfied with the MEPS IC benchmarks that you are
currently getting and you said, what I would really like is to look at people
in this wage range who don’t take up coverage but who are eligible. I want to
know what their premiums are, what was the lowest price option that they faced
that they turned down, and what does that distribution look like.
As soon as we have these two data sets merged together, we can provide that
benchmark that you are interested in. So if you have wish lists that you would
like to give us, we would love to hear them.
DR. BANTHIN: So we are in progress. We have many things in the works here,
but these are our main models, tools and plans.
DR. SCANLON: Thank you very much. John?
MR. SHEILS: Good morning. It is a pleasure to be here. These meetings where
modelers get together and talk are actually good therapy for all of us.
We have been using our model in various forms for about 22 years now. The
first time we used it was with the bipartisan Commission on Universal Coverage,
1988. That was the Pepper Commission, if some people remember that. It was a
great opportunity to do analyses of a few different alternative plans. A lot of
the plans we are talking about now are the same ones.
We just finished a study, we did a study last fall of comparing the impact
of President Obama’s proposal and the impact of Senator McCain’s proposal. In
that we looked at the public plan and what happens under different parameters,
different design issues.
With McCain, it was particularly challenging, because he took a very
different direction. It had a logic to it that you needed to explore if you
were going to do an objective job. You had to follow that logic and give it the
greatest credibility you can. But it was a plan where in effect you would have
an increase in medical underwriting. In states where they say you can’t turn
down somebody just because they have got a health condition, you can’t charge
them more just because they have a health condition, in his plan that would go
away, and you would have an enormous high risk pool and all kinds of
segmentation of risk and people shifting around.
You can’t answer that question, what its effects will be, unless you model
those flows. If you take the sick people out of the system, CREDAs go down for
the people you c help, so you get some takeout there just because of the way
you have broken it out.
So as I say, you have to think about those things pretty carefully.
We just finished something with the Commonwealth Fund, where we modeled ten
different Congressional plans. We have some detailed tables in addition to that
report, if anybody is interested in it.
We found that the key here is to get a baseline database that depicts the
characteristics of the population, individuals and employers. We found that the
best way to do that is through using the MEPS data. The MEPS data is
interesting because it will give you a lot of things that can be relevant. It
gives you month to month coverage data, it gives you a little bit of
information on utilization of health services over the course of a year, you
know when people have changed jobs.
There is also information on health conditions. That was very useful when
we had to start figuring out how to model underwriting. We could identify those
people who were going to be excluded. We could take them to the pool that they
are creating. And because we also have information on their costs, we can look
at what the pool’s costs are going to be. We can also look at what happens to
what remains in the individual market, people who are basically just healthy.
That brought down the premium quite a bit, which is what you see in states with
large high risk pools. The individual coverage actually turns out to be pretty
low cost because they don’t cover anybody who is sick.
So all of that information has just been so helpful to us over time. It
allows us to generate month by month income data, too, which is very useful for
modeling public assistance programs.
The model is set up to model Medicaid and SCHIP expansions, premium subsidy
programs such as those proposed by McCain through a tax credit or those
proposed by others through vouchers. We model the pay or play employer model.
There are at least three different variations on pay or play that we feel have
very different outcomes.
So every one of these proposals we look at, there is a great deal of review
of the assumptions for appropriateness of what we are modeling. So it is
important to remember that. There is individual coverage and so on. There are
individual mandates that have some weird effects on employers, too.
Insurance market regulations. We have a model that will simulate — and you
heard a similar one described earlier — will simulate the premiums for a group
based upon the characteristics of their population, and the rules that are
permitted in their state. There is age rating, for example; we will be able to
model that explicitly with an employer database. We also model that for the
individual market as well.
The databases we use here in HBSM start with the MEPS. That is the basic
household data file. We take the data for four years and get a large sample. We
also use CPS updates on coverage and population characteristics. This tends to
come out more recently so you want to stay current as possible. Optics are more
important than its impact on the actual results, but it is important. We also
use SIPP data in many places to develop some behavioral parameters,
participation rates in Medicaid and SCHIP, for example, is an example of what
we will look at.
The model is controlled to an array of program data. We rely very heavily
on data from the Office of the Actuary at CMS. There is once in history where
we felt they were wrong, and thank God, in the end it turns out we were a
little closer. You may remember that. We were doing the President’s analysis.
But we do control to that carefully.
We also go through this process to try to reconcile what is in the file
with what is in the accounts, keeping close tabs on what those numbers are,
because it will make a difference.
CBO projections are used where we can. CBO is not final because it is
infallible, but it is infallible because it is final. So we go with their
conclusions wherever possible.
We have a synthetic firm portion of the model. You know how certain
poisonous frogs are a bright color, to warn people not to go near them? That is
why we use the word synthetic, so you understand we are going into something
that we have to concoct from the available data out there.
We decided early on that it is very important to have a file that gives you
information on the employer, and then the characteristics and health
utilization features of everyone who is working in that firm plus their
dependents. To do that right, we feel you have to have a database that accounts
for a great many things, such as part time/full time status, wage level
certainly. There are a couple of other items such as how many people in the
firm have coverage, how many are eligible but turn it down, how many are
ineligible. So we go through a rather elaborate process of doing that, which
leads us on a strange trail here.
Our basic data on health plans is the 2006 Kaiser HREP data, which others
are using. But we match that with data from the 1997 RWJ survey of employers.
That is important, because it gives us a list of information on the
characteristics of the workers in the firm.
So for example, you could have a firm where most of the workers are low
wage females who work part time. When we match together and construct that,
those are the kinds of people who will be attached to that firm. If you have a
plan that treats part time workers differently, it will enable you to take that
into account, because for some individual firms that will have a profound
effect. At least, that is what we find with our numbers.
When we model proposals dealing with medical underwriting, with this match
we have the information on not just firm, but we have the information on their
health expenditures. That health expenditure information and their condition
data can be used to model health status rating where it is required in the
So we spend a lot of time doing that. We match between — the RWJF data is
quite old. Its health plan data is probably just short of useless because it is
so old, but we match the plans on the basis of just ranking of these health
plans by actuarial value. So the firms with good benefits get assigned to the
HEET firms that have good medical and so on.
I think I have already explained this. We use the METS data because it
gives us a database that has a great deal of detail in it. We don’t have to
worry so much. If you don’t have information on health conditions you have to
impute it to your file. You always have to worry about, did I impute it the
right way, did what I do make sense. We think it is more likely that it will
make sense if it comes from a database that contains its own unique joint
Another thing about the MEPS that we like is that it is conducive to a
month to month simulation methodology. This might be a distinction between ours
and some other peoples’ plans. We set the data up such that we have data for 12
months. For every individual we have their coverage in each of the 12 months,
thanks to the MEPS survey where they collect things.
We know about their employment. We can take their income and say their
earnings must have been distributed over these weeks where they worked and the
unemployment insurance while they were employed and so on. We are able to
construct a month by month accounting of their income and their employment
information. You can line that up with their coverage information, and that
coverage information is lined up with date of service for the services that
they use. So it is pretty powerful.
It is an enormous amount of housekeeping. It is no fun for the programmers
certainly, but we feel we can be a little bit more comfortable with it in that
The month by month simulation feature is very important as well from a
policy perspective. Earlier I think John was mentioning this. A lot of these
proposals that are coming up have many moving parts. Some affect employer
coverage and some affect Medicaid or the coverage for non-group.
If you are a person, and many people in the file are like this, where you
have somebody who is in for six months in Medicaid, they get a job, they
perhaps pick up coverage for an employer for the next six months or maybe work
for an employer and don’t take coverage, either way, it is interesting and it
is important, because those policies will affect what happens during those two
periods differently. Understanding how people are distributed across those
snippets of time will make a difference, part your eligibility, part your
costs, are important.
As I say, we merged the data for 2005.
Since you all are interested in data, I thought I would do something
interesting here and present some data. I am the only one who did this at this
point, so I hope I didn’t miss the mark too much.
This is a basic breakdown of some of the issues we face with just the
simple question of how many people are uninsured. Here are the data files we
With the MEPS 2005-2006, we get the first column of numbers, of 50 million
people uninsured at a point in time, 37 million uninsured a full 12 months, 68
million sometime over the year, and any time in 24 months we can do that as
well for MEPS and get about 83 million people.
Now, some of the other databases are a little more difficult to work with.
The CPS for example, if you take the questions literally, and the questions are
asked literally, what it tells you is the number of people who didn’t have
coverage from any source over 12 months, over a year. It is an all year number.
The problem is there is a great deal of under reporting of Medicaid
coverage in the CPS. If we like CPS because it gives us some state level data,
this is an important thing to hang onto. So we correct for under reporting, and
we do this by this month to month simulation approach, distributed income
across the months of work, earnings across the months of work, weeks of work,
months, and the various income sources. We construct month by month income
levels. It is very much like what they do in the TRIM model.
From there, we go one further step. We take the coverage from employment
and spread it over weeks of employment, of course. So you get a synthetic
representation of what the distribution of coverage was for each individual in
the file for each of 12 months. We correct for under reporting controlled to
the average number of people who are supposed to be on that program, who were
ever on the program.
The third column shows us what we wind up with. We get the number of people
uninsured all year. We get 34 million, which reflects the under count in the
CPS of Medicaid. The CPS gives us 47 million at a point in time, 62 million at
any time in the year.
The difference between that and MEPS is interesting, because I think the
MEPS under reports coverage a little bit for Medicaid, and we do make an
adjustment for that. So once you get through doing adjustments, you get some
We took the SIPP data and used that. We really got stuck on this for
awhile. These are the numbers for the 2004-2006 period, the 2004 panel. We have
done some reweighting to try and replicate the figures. We get a very different
set of answers. We only get 30 million uninsured all year, about the same
number of uninsured at any point in time, but only 30 million all year, only 58
million any time in the year, and only 74 million any time over 24 months.
Either the SIPP is right or these guys are wrong.
This chart shows what happens in the 2001 panel of SIPP and the 2004 panel
of SIPP. Same thing; average monthly of uninsured in 2001 is 43 million, but in
2004 we are getting 45 million. That makes sense; the number of uninsured we
know did increase over that period. And we get an increase in the number of
people who are uninsured all year; it goes from 24 million to 30 million. That
is rather a big jump, but it is in the right direction.
But then sometime in the year we are getting less now in 2004 than in 2001.
We are running the data through the state programs saying this and that, so it
is a red flag for us.
It also shows the number of people uninsured sometime in 24 months, but
down from 79 to 74 million people, and that is in the wrong direction, we
believe. There is a great deal of variability in the estimates it gives you on
a month to month basis.
We are concerned about the 24 month business, because we do see the MEPS
going in the opposite direction. The number of uninsured for 24 months has gone
up steadily; MEPS shows that it is going down.
I want to go back to this one thing here. Going back to the CPS table, a
lot of us know, many people have said, that it is not an estimate of the number
of people uninsured all year. It is really an estimate of the number of people
who are uninsured at that point in time of the survey.
A lot of people feel that fairly strongly. I think that is wrong, but we
went through this process of saying we are going to figure out what it really
means and correct for under reporting and try to do everything right. What we
wound up with was an estimate of the number of people who are uninsured at a
point in time, 47 million people.
Well, people have been saying that CPS says there are 46.5 million, almost
the same number who are uninsured at a point in time. Do you remember Emily
Latella from Saturday Night Live, Gilda Radnor? She would get all upset about
this issue, somebody out there, it is not a problem. She turns and goes, oh,
never mind. That is basically where I am going to end.
DR. SCANLON: Thank you.
MR. SHEILS: Actually there are a couple of suggestions of things we would
like. It would be great to have the state identifiers for MEPS. It is very
unsporting that you have them and we don’t.
I would like to see the SIPP data longitudinally stabilized. It is nerve
wracking. Program cost data by eligibility group is important for getting some
of our costs so we can get a little better data from the Medicaid program in
particular. It would be interesting if at some point we would get micro data on
employers from MEPS. I think you have already heard about that.
MS. TUREK: This is a joint presentation. I am the government project
officer, and Linda is the project officer at the Urban Institute. We did it
slightly different, because we were told 15 minutes. So we figured we would
give you a fast introduction to TRIM, talk about what we do in the health area,
and then if there is time we can give you a lot more detail.
TRIM is a modeling system. We model pretty much all of the major federal
tax and transfer programs. We apply rules at the individual record level,
basically the program rules, to see who is eligible, and then we talk about who
We do this because one of the big questions in earlier years was, this
person gets this many dollars from Medicaid, that many dollars from welfare;
how much money does this person or family get in total? And because the program
rules vary from program to program, the only way that you could get it was
through this kind of system where you modeled the individual and then could add
up the family or household level.
We primarily use the CPS, but there are a lot of databases. We statistical
match in the HIS and pretty much all the other databases everybody else uses.
CPS is in the process now of adding point in time health insurance
questions based on a bunch of research funded by Census, ASPE and SHAGAC, which
matched the surveys to Medicaid statistic information system records. That
would be a good benefit.
TRIM has been around almost as long as I have. I have been involved with it
since 1976 either as the project officer or when I was the division director
managing the project officer.
Urban Institute is our current contractor. We have it set up so that for
each TRIM product we have a technical representative in house that does the
review. So the health models are reviewed by Tom Musco and Don Cox because they
know what the programs are about.
It is a public use model. It is a client server system that you can get to
over the Web. We give you a complete detailed description of how everything
that is done in the model for every program. We give you our databases
annually, I think going back to ’93, and you can get them back to 1980 if we
use the CVs. We give you the program rules year by year and you can actually,
if you can figure out how to do it, go out and run your own simulations on
We had once wanted to fix it so that you could change the program
parameters. They had somebody who developed a straight language, but
unfortunately we ran out of money. To get out there and model, we don’t give
Urban any money to help you, so you have to be pretty good at guessing what to
do. Budgets are tight
The programs we model are health, Medicaid SCHIP and employer sponsored
insurance, the cash income programs, non-cash benefits and the taxes, and
medical tax credits would be included there.
What its main purposes are, are to identify people who are eligible for
means tested programs. As people have pointed out, the CPS under reports many
of them, and so we run people one by one through the eligibility rules from the
programs. We do this state by state. We can do all of our estimates state by
state, and we come up with a list of eligibles. Then we impute participation.
Then once we get a baseline that looks a lot like the world is today, where we
have each program matching what the administrative data says or close thereto
— there is never quite an exact match because the administrative data will
include people we don’t. Medicaid would have the people in long term care
institutions. But we use the baseline and then we do what-ifs, what if we
changed this, how does it change the world from today.
We also go for quality review. Right now the Urban Institute has — outside
of our TRIM team we have technical experts who review all deliverables. I think
you are one of them. We also had outside reviews of the model. John Treika
reviewed it a couple of years ago. They looked at the Medicaid and SCHIP
modeling, and we did in the late ’80s pay the National Academies of Science an
enormous amount of money to do a complete and entire review of the model. A lot
of what they said we could afford has been implemented.
I am going to turn it over to Linda, who will tell you about some of the
work we have done with the health models.
MS. GIANNARELLI: Thank you. I am just going to talk about the aspects of
TRIM that relate to today’s topic, so don’t worry, I’m not going to start
talking about the SSI model.
I want to talk about three different pieces of TRIM. The first, what is our
Medicaid and SCHIP modeling, the second, our employer sponsored insurance
modeling. This wasn’t specifically on our list of bullet points, but since
matching has come up quite a bit in today’s discussion, I also want to mention
our statistical match with the NHIS data.
First, our Medicaid and SCHIP modeling. As Joan mentioned, we simulate
eligibility and we simulate enrollment and we can do what-if modeling. In terms
of simulating eligibility, where all the other models were looking person by
person, household by household, person by person for everybody in the CPS data.
Like the Lewin model, we do this on a month by month basis. We also feel that
modeling eligibility month by month is very important because we do find a lot
of people who appear to be eligible for part of the year and not the entire
We use the same kinds of things that the Lewin model does as far as how do
we go about allocating annual income across the months of the year. I think
sometimes people think, all you have in the CPS is annual income, so you are
really making that up, but there is a great deal more in the CPS than maybe
someone would initially think of. You know for instance how many weeks people
have worked. You know how many spells of unemployment they have had. You know
the number of weeks of unemployment.
So there actually is quite a bit to go on. It is not just making it up,
although to continue with the frog analogy, it is a somewhat brightly colored
frog, but not a very brightly colored frog.
We do the eligibility modeling month by month, and we capture all the
detailed state specific rules for each type of eligibility, as you see on the
These are a few of our eligibility results, just to give you a sense.
Obviously we can cut this a lot of more detailed ways, but we like the Lewin
model can look at eligibility in the average month of the year or we can look
at eligibility on an annual basis, is someone ever eligible for Medicaid or
SCHIP in any month of the calendar year.
This chart is looking at eligibility by main reason for eligibility, by
user group. This particular slide is dividing it by mandatory due to SSI, other
mandatory SCHIP and some other optional categories, and medically needy, and
finding a total of 59 million people eligible for Medicaid or SCHIP in at least
one month of the year, and these estimates are for 2003.
After we do eligibility we also model enrollment. Again picking up on the
Lewin presentation, we are trying to correct for the under reporting that is in
the CPS data, not only for Medicaid and SCHIP, but for all the social programs.
In fact, for Medicaid and SCHIP it is somewhat less bad than it is for TANF and
some other programs.
So basically we are taking the people who are eligible for Medicaid and who
report that they are enrolled in Medicaid or SCHIP, but then we are taking a
subset of the eligible non-reporters to augment that caseload in order to come
acceptably close to size and characteristics of the actual caseload. As far as
that actual caseload, we are getting that from the EMSIS data. We are not using
the EMSIS data unadjusted; we are starting from the EMSIS data but we are
subtracting the institutionalized who wouldn’t be in the CPS data, adjusting
for cross state duplication, making some other adjustments in order to arrive
at our final targets. For SCHIP we are using targets from the SEDS data.
That gives us a baseline that can be used for a variety of different
purposes. One can compare eligibles to enrollment to estimate take-up rates.
One can examine the characteristics of eligible individuals, decompose changes
from the prior year. So for example, to what extent does it look like changes
in eligibility are due to changes in economic circumstances versus changes in
program rules. One can do estimates of uninsured people and eligible uninsured
people. One interesting use is to be able to do alternative poverty analyses,
for instance the National Academies of Sciences poverty proposals have
different thresholds depending on peoples’ insurance status, either uninsured,
publicly insured or privately insured. So by correcting for the under reporting
of Medicaid and SCHIP it allows one to implement those NAS poverty definitions
a little bit more accurately than using the unadjusted CPS data.
Some examples of Medicaid and SCHIP alternative policy simulations. One can
change one or more specific eligibility rules. One can play a what-if game, for
instance, what if all states have the eligibility rules of a particular state.
One can play other what-if games of applying rules from a prior year. You can
leave eligibility unchanged and modify take-up. Playing off of TRIM’s
comprehensive nature, one can look at cross program interactions. You can say
what if there were a particular change in SSI, for instance something like
eliminating the 209B rules so that everyone in SSI was automatically eligible,
what would that do to Medicaid enrollment. So you can look at those kinds of
In employer sponsored insurance, what the model does is impute premiums.
The Kaiser HRET data was mentioned earlier, and that is also what we are doing
for matching in premiums. We are matching in both the employer and the employee
premiums. That is what we are doing for the private. You can see on the slide
the data sources for the federal employees. So we are doing that match in order
to bring into our data the premiums as well as the plan type of a particular
person’s imputed coverage.
There is also an element of that model that imputes whether or not a person
has been offered health insurance, whether they are in a firm that offers to
some people but just not them individually, or whether they personally received
an offer but they chose not to take it. So that imputation is available for
potential alternative simulations.
The ESI model can be used for a couple of different kinds of uses. You can
analyze insurance status combined with whether or not insurance was offered or
accepted. You can do some what-if analyses, what if more individuals accepted
the offers. The imputed premiums are available for modeling tax credits as part
of TRIM’s federal income tax modeling.
I don’t have a slide for the NHIS match, but we do a statistical match with
the NHIS data every year. We bring in the ADL and IDL information and we also
bring in reasons for limitations. So for people who have activity limitations
we are bringing in the variables that indicate why they are having those
activity limitations. We also bring in through the NHIS match the private
non-group premiums. We used to have a different way of doing that, but now that
we are doing that NHIS match, we are getting those premiums from the NHIS data.
That information again is available for use in alternative simulations.
Interestingly, the CPS pretty soon is going to start having some limitation
data directly in the CPS, so that aspect of the match will no longer be
necessary because we will actually have real data on that in the CPS. So that
will be great.
Some capabilities in TRIM that were used for 1990’s health reform analysis.
We had in the 1990s some elements maintained in TRIM that allowed us to look at
changes in employer provision of ESI, look at the employee decision between
public and private coverage, and look at a broader range of health spending.
Those elements of the model have not been kept up to date.
Some ongoing and planned work. Something that picks up on a couple of the
comments made by other presenters is the modeling of eligibility and the
problems with under reporting in various surveys.
One thing we are doing right now under the TRIM contract is, we are looking
at that matched data file that matched to the CPS with the EMSIS data. We are
basically looking at people who we simulate to be eligible for Medicaid and/or
enrolled in Medicaid, are we finding them in the EMSIS data, and conversely for
who are enrolled in Medicaid according to the EMSIS data, do we find them to
be eligible. We are analyzing those results right now, and I think that is
going to be extremely helpful in terms of improving both our eligibility and
our enrollment modeling.
We are finishing up our 2006 baselines right now. There is a task planned
on the agenda for fiscal 2010 to bring some work by Rich Johnson on long term
care insurance, to bring that work into the TRIM framework.
Joan mentioned a lot of the ASPE uses of TRIM, so I won’t go into this too
much. There have been a lot of ASPE uses of TRIM in the health area. The second
bullet point there, looking at the effects of a non-group health insurance task
credit, and lots of different kinds of baseline analyses. Looking at
eligibility by AGI level and by immigrant status. Immigrant status is one thing
that is imputed into the CPS data as part of our pre-processing. At various
points we have done some very detailed tables of eligibility cut different ways
that has then been provided to folks doing detailed modeling at Treasury so
that they could feed those pieces of information into their modeling.
Lots of different non-ASPE uses of TRIM. CBO has used TRIM, has used our
longitudinal series of eligibility and enrollment flags from TRIM. CRS has used
TRIM to look at the impact of Medicaid and SCHIP restrictions on benefits to
non-citizens. I will let you scan the other bullet points.
I think some of the strengths of TRIM as far as health, its ability to do
detailed state specific eligibility modeling, the ability to look at cross
program interactions. One that I didn’t think to put on this slide is the
ability to model a package of changes, not just to one program, but to look at
a package of changes and to do distributional analyses. For instance, if there
is a legislative package that is including changes not only in let’s say
Medicaid eligibility but also taxes, also food stamps, to be able to model all
of that in one internally consistent framework, and then to look at
distributional impacts and winners and losers and other kinds of impacts of the
entire package of changes at the same time.
MS. TUREK: And you can easily look at minorities, because we have that data
on the CPS.
MS. GIANNARELLI: Right, you can cut results by whatever variable is
available in the CPS.
As Joan mentioned, the model is available online. Trim.urban.org is the
website. Registered data users can extract variables, so you can extract
eligibility flags once they have been made public use and put on the server. If
you request simulation access, then you can run simulations.
There is a not so pretty picture of Medicaid documentation available online
so you can read the documentation of the various modules online. This is just a
picture of looking at a particular household through the micro data viewer
DR. TUREK: This is basically what we can do. This shows you the steps that
TRIM goes through when it is doing the simulation. I think the ability to look
at immigrants is very important because there will be programs that do not
relate to illegal’s or will depend on when the immigrant arrived and their
status. We incorporated that capability quite awhile ago. We used them to do a
lot of analysis of immigrant programs.
Although the last charts are not very pretty, all you have got to do is go
to Trim.urban.org and you will get these charts. It is very comprehensive
documentation that explains in great detail how each model is developed. We
believe that models used for policy analysis need to be transparent. So if
somebody else wants to reproduce your results, they have the information they
need to do it.
We were fortunate when we went from a mainframe computer to a client server
system that my boss was a systems programmer. So TRIM was built not only by the
analyst, but by computer professionals who established it to standards in the
industry for computer programs. It is a C++ model, and it is updated daily
pretty much. But I like it. It is kind of fun. We are always adding something
new to it.
DR. SCANLON: Thank you very much. Let’s open it up for some questions
before we have the committee discussion.
DR. SUAREZ: This is a question for Art and maybe the other members of the
panel can join in as well.
We heard this morning different models built on utility based approaches,
model built on elasticities based approaches, models that use CPS or the
infrastructure of data and some that use SIPP. I was wondering if there has
been an evaluation of the various models in a way that shows where the
different outputs are and why those differences are.
An example. This is an evaluation of the various concepts using a common
policy. So we have at one end a policy situation and you drive the policy
through the various models and come up with all these results. Where are the
differences in those results, and why are those differences happening? So that
at the end of the day maybe we get a better understanding of where the research
DR. TUREK: The National Academies study did look across models. I don’t
really know of anything since then that has taken that kind of a comprehensive
DR. BANTHIN: Linda Bellheimer can speak to this because she was the project
officer in an extensive comparison funded by Robert Wood Johnson.
DR. BELLHEIMER: Yes, we actually did this with Dr. Jerry Glied at Columbia.
DR. TUREK: When?
DR. BELLHEIMER: It was about five years ago. We took three of the best
known national models, all of which you have heard from today, three of the
seven that you have heard about today were involved in this.
We started with a meeting in which we brought all of the microsimulation
modelers into a room, and we spent a whole day going through assumptions that
underlie models. Sherry did a wonderful job of getting agreement on, these are
the parameter assumptions on which there is agreement. These are the
assumptions on which there is not agreement. This is a gray area in the middle
where people are going to adapt depending upon the issue.
So we had that as the starting point. Then we designed a fairly plain
vanilla health reform with some public coverage expansion, some tax subsidies,
and we asked three different groups to model it, and to document where their
assumptions did not follow the standard set of assumptions on which we had
agreed to, where they were in the gray areas or where they firmly had their own
It was a very interesting exercise. The three models did not come out with
the same sine on the impact on coverage. When we peeled back the layers of the
onion to find out why, there were two things that drove this, and you have
heard a lot about both of them today. One was the assumption about the
employer’s response. The behavioral response of employers was critical in the
assumptions. That was where there had been a lot of difference in the meeting
that we had talked about, about where people were.
The other was about premiums in the non-group market. I think there was a
fundamental difference between models in which premiums were established
endogenously within the model and models in which premiums were imported from
We subsequently had another meeting in which because of the interest in the
premiums in the non-group market, we brought a whole lot of members of the
insurance industry in to talk about this. They were very perplexed by the
notion of endogenous determination of premiums and said, this is not the way we
think at all.
I think it was a very interesting meeting. We talked all day. I think
probably by the end of the day there was a sense, over time, yes, what you are
saying is happening in your models would probably work its way out in the
marketplace, but if you are looking for 2009, 2010, 2011, that this would be
the response that you would see in the marketplace, probably not.
But I think what we learned from this was that there is an enormous amount
to be learned from having multiple modeling groups modeling. We learned more
about the things that you really need to be sensitive to in health reform by
looking at what grows the differences than we would any other way.
So I think thinking about transparency and for the committee to be thinking
about how you get multiple modelers to look at a proposal and then share the
reason for the differences is probably going to tell us more about health
reform than any one model looking at it alone.
DR. SUAREZ: It sounds like there is an opportunity for repeating that
experience, but maybe even with a larger group of models.
DR. BELLHEIMER: Yes, I think it would be an excellent thing to do.
DR. SUAREZ: Thank you.
MR. SHEILS: Once you have done an analysis of the plan, it is good to try
and reconcile them with the other guy’s numbers if you can. We have done that
in the past and have learned a lot. I think the other guys learned something
too, but we learned an awful lot about it.
We do have this joke at work, though. If somebody else gets a modeling
result that looks like ours, we declare that they must be making the same
DR. GREEN: I would like to explore with any of the panelists two quick
areas pretty quickly here. Before I do, an unanticipated consequence of the
hearing this morning for me as a member of the committee is to be reminded how
unnecessarily complex the system is. The take-home message for me is, the
models would be a lot better.
The first area has to do with something that I believe Linda brought up. It
has to do with where politics and policy often mix, and that is in geography.
You mentioned geospatial data; it is my impression that people working to take
full advantage of these investments in public data and national data sources
spend a great deal of money geocoding information that is in them. I am
wondering about your opinions about the sufficiency or the strategy that we
should be contemplating about these data sources arriving geocoded before
distribution versus, it depends what we are going to use them for, and
everybody just needs to plan on that being a cost of business.
DR. SELDEN: We have files available for people to use in house that go to
quite narrow detail on where people are located. But the problem is that we
cannot even release a file to you here to use outside the data center that has
state on it, because the sample sizes are so small. We are talking about
medical conditions and specific procedures that are done and total amounts of
money spent, and the timing of the event. It is very easy to use that
information with state identifiers to violate confidentiality. So we just have
real limitations there.
DR. BANTHIN: Let me just add, we do geocode the data in MEPS. You can
access it through the data center. You have to submit a proposal. If you bring
in other data like location of hospitals, we will match it and then give you a
file back. Sometimes you have to use it within the center if you are going to
continue to use the identifiers, but if not you can take it away.
DR. GREEN: That set up my second area. I wanted to ask you your opinions
about privacy considerations that restricted the use of the data to answer
urgent policy issues.
MR. SHEILS: I think it would be a disaster if people somehow got the idea,
I mean the public got the idea that their material won’t be treated
confidentially. Reporting in all these surveys could disintegrate, sort of like
our economy disintegrated. That I think is the worst fear.
But I have got to say, I would love to see the state code on MEPS. It would
be interesting to see if there is some way to — I guess you could do that by
eliminating some of the other variables, but it would have to be for all of the
Maybe there is some way to prioritize on what types of geographic
disclosure is preferable. Urban-rural is important to a lot of researchers, but
state code might be important to another group of researchers. You can’t
release the files both ways. You have to pick one or the other, or else you
defeat the purpose of doing it.
So it would be interesting if there is some forum for figuring out what the
prioritization should be of those things for research purposes.
DR. SCANLON: Let me follow up on this with reference to the IOM report that
came out at the beginning of the month. Even though it is not dealing with
survey data per se, I think it has applicability to the issue of survey data.
This was about HIPAA and privacy rules and how that impacts on research.
They came up with three sets of recommendations, two of which in some respects
are what are going to be the governmental rules. The third one though referred
to researchers. It said, the researchers need to think about how they can
increase the security for data that they are using, and in some respects how
they can make the public confident that they have increased the security.
I guess I find this interesting, because it is saying that there is going
to be an obligation here on both sides. If we are thinking about changing the
rules by which data that are potentially identifiable would be available, what
is going to be the quid pro quo.
The reason this becomes particularly important is because we have had
hearings before for the full committee and for the work group where the premise
that all data are going to become identifiable at some point in time, that
there is not going to be any core that you can’t match to some other data set
and through that identify an individual. So the public use files are becoming
leaner and leaner under the current rules.
So the idea would be, if we have to relax those rules, how do we do it with
confidence. So both panels, the question is, what can we say in terms of
response to this kind of recommendation that is going to convince the public
that we can think about changing rules for disclosure?
DR. TUREK: I think if you are talking about the kind of policy analysis
that is being done in response to program changes, you also have to think about
timeliness. If it takes you six months to a year to get the data through some
kind of long research agreement, by the time you get it you won’t need it
anymore. So I think timeliness is critical.
DR. BANTHIN: It seems to me that one ought to be able to come up with some
kind of certification standards, and that there ought to be — we are required
internally to go through fairly strong steps to protect data. We actually
collect a lot of primary data, so we pay a lot of attention to security of
those data, because we are also a defense contractor. It is hyper ridiculous
actually, but it is very secure, and there are very strong standards there.
There are people in charge of that. There are plans in place. Those get audited
and checked. There are limitations on who can have their hands on what data.
There are significant consequences to employees who violate any aspect of that.
So I would wonder whether it would be possible to create some certification
program and allow people to become certified because they have a program in
place that allows them to handle those data in a secure manner, and somehow cut
through some of the processes that are in place today that are aimed at
protecting privacy, but that get in the way of nimbleness or utility of the
data for policy research, particularly policy research that is trying to be
responsive as things are under discussion. So I offer that as one possibility.
MR. SHEILS: One thought too is that a lot of these services are done year
after year after year. We could I suppose do a regional identifier for a couple
of years, and then when you get the new samples of people, this time you use
state and don’t put out the rural.
The researchers will always have access to a data file that has a geocode.
It may be two or three years old, but at least you will have some degree of
DR. SUAREZ: Just a quick follow-up on that. I am a member of the Privacy
and Security Subcommittee of the NCVHS. Thank you for that recommendation. I
think that is an excellent concept, the development of a certification program.
But for purposes of the discussion, all the databases that I have heard of,
I’m not sure they fall under HIPAA, in the following sense. I don’t think AHRQ
is a covered entity and MEPS is not a HIPAA covered data source. The reason why
MEPS cannot release state level data comes from other sources, not HIPAA.
So there are issues beyond HIPAA with respect to the release of identifiable
information on public databases or databases held by government agencies.
DR. BANTHIN: We are sensitive to this issue, because we do as AHRQ
employees have access to all these variables, and we don’t like to be called as
having unfair advantage, unsporting. So we do try to make efforts to release
stuff. But it is hard. We have not — our agency or center has not devoted
creative energy to thinking about this. We are limited in our funds, so we have
this data center, we now have agreements with the Census so that people don’t
have to come to Rockville. They can go anywhere Census has a data center and
use our data.
The MEPS IC, the employer data, that is the most difficult to get of all.
It has everything to do with the IRS. They approve all those proposals. We
don’t even have access to these data, and AHRQ has funded the collection of the
data. We will wait seven to eight months, and I am crossing my fingers, to get
our proposal approved. IRS reviews those. They don’t want anybody to think that
your tax records are being used irresponsibly.
DR. SCANLON: Walter, you were absolutely right, in terms of — this is not
HIPAA data. One of the things about thinking about research and privacy is that
we need to think about across the board what are all the rules, how do we
harmonize them in a way that we are both protective of the privacy interests
that we care about as well as promoting the research. It is thinking very
broadly in this process.
DR. BREEN: Another thing that Tom referred to briefly but I think we need
to be mindful of is that a lot of these data sets are just too small. The
sample size is just too small to be able to look at regional or particularly
state or even large metropolitan areas. So I think we really need to think
about investing more resources in these data sets.
I just want to point out the MEPS, which is very efficient, it is done on
top of the National Health Interview Survey, but it is constrained by the
sample size of the National Health Interview Survey. The resources used for
that survey have diminished over the last ten, 15 years. So that needs to be
corrected as well.
Then I think there could be more centralization among the data collection
as well. You have got a number of surveys, some of which are redundant, and
those could be explored more and the resources could be used better. But I
think we need more resources also on some of these key surveys which are used
to build for modeling purposes.
DR. SCANLON: We did try to address that yesterday with a letter to that
MR. GARRETT: I believe there are people like in the LAHD program who have
tried to take confidential data with the right access to that and blur it, but
preserve as many characteristics as you can. I don’t know how good this can be
done or not, but there might be a blurred real data set that is far superior to
matched data sets.
DR. BELLHEIMER: We are blurring data. In our linked databases where we link
our surveys through administrative data, clearly once you have got a survey
that is linked to Medicare or to Social Security, the chances of identification
become much higher. There is a great demand for public use files from these
We have started what we call data perturbation, which is an evolving
science, which involves perturbing the data very slightly and maintaining the
internal structure of the variance/covariance matrices.
We have put out our first perturbed data files. There is a very recent
article in the American Journal of Epidemiology about the methods that are
being used in this. Some researchers feel very comfortable, and we have on the
website a table which shows you the variance/covariance from the perturbed data
to see how close they are. Some are comfortable using those, others
uncomfortable and want to use the original data.
So what we suggest is that they use the public use files to go on the
website, design their analyses, and then send their final programs to us and we
will run them on the original data so you feel more comfortable. So they get a
tool quickly that they can use, and they can develop their programs, and then
they can get a fast turnaround on the actual data if they don’t want to use the
But I predict that over time, researchers will start feeling more
comfortable just using the blurred data and not wanting necessarily to get the
DR. GREEN: So building off of that, the committee members will correct
this, but I think we folks on the committee are struggling to contend with our
hopes that the population and individuals in the population reap benefits with
these investments of public money into developing data sets and organizing them
so that huge policy decisions that affect everybody can be well informed.
With that hope on one side of our discussion, then we have these fears on
the other side that we will do harm, and that in the process of doing this, we
subject individuals and the population to risk for harm. This is something that
will not go away. We can title the hearing whatever we want to, and the issue
is resources, it seems to me.
So I would welcome the panelists’ suggestions as to what are the next
steps. Is there an urgent need to systematically undergird progress on the
nation’s health statistics system if we only could reframe it, get the right
people in the room for a day. But again, maybe that is overstating it, but I
feel like we are struggling to think fresh and anew about this in ways that
really get the benefit. Meanwhile, we have precious few examples of where harm
has been done.
PARTICIPANT: By researchers.
DR. GREEN: By researchers, right, by investigators. So if you have got
ideas, they would be well received.
DR. TUREK: I wanted to say something about synthetic data. It is very
important that the covariates that relate to the programs you are interested in
don’t change much. The one synthetic data that I saw greatly changed the
covariate on the disabled who are a major group getting federal programs.
So I think I am going to be looking, because I wouldn’t have used synthetic
data. I would have gotten enormously different results. So the question is, can
you do it in such a way that the data that people use to estimate the impact on
public programs doesn’t change; which variables do you select to look at.
DR. STEINWACHS: I have a question, but there was an oversight. Dale pointed
out to me that we need to give Linda Bellheimer credit for helping identify all
of you and getting you here to this panel. Also, Roshita Dorsey was not in the
room when I gave her thanks for doing this, too. Dale keeps telling me he just
fronts for the operation, but he is a very good front person, I think.
I would like to go back to the discussion earlier about the health
outcomes, because ultimately if you are asking the nation to invest a lot of
money in health insurance, and that is always what comes back in the messages
on this, is the so-what question. Some of the so-whats are, we cut
administrative cost to other things, but the other is the quality and outcome
The point was made before that to do it well, we need some longitudinal
data sets to be able to observe that pattern of insurance, uninsurance over
time, and recognizing that that is not going to be here for anything in the
near term other than what you have taken advantage of already.
There is a lot of cross sectional data that talks particularly around
associations like potentially preventable admissions or ambulatory care
sensitive admissions. You could go sometimes and link the failure to get
certain services to certain expected outcomes. For example, for most common
chronic diseases, you could make some estimates. I guess that is what the RAND
model does in terms of mortality. So I was interested in trying to hear from
you, are there some things that with the currently available data you see could
be done to build out another piece of these models that tries to answer the
question of, if we were fully insured versus how we are currently, what would
be expected over time in terms of improved health, productivity, longevity,
other measures that we don’t provide now?
PARTICIPANT: What kind of insurance are you talking about, Don?
DR. STEINWACHS: Offline Beth and I were talking about how insurance gets
characterized. She was pointing out, it is the actuarial value of the
insurance, and is that a good proxy for coverage. But a lot of what the public
sees is, you are insured or uninsured.
A lot of the research when you get down to it looks at, are you insured,
are you uninsured, and some of those results are rather striking when you say
what happens to use of services and what happens to some chronic disease
What I was looking for was your insights for what could be done to build
out in directions that would help answer people’s questions, saying what are we
going to get for it in terms of better health and a more productive nation.
DR. SELDEN: I don’t have any answers to those questions, but I think you
could actually expand the question in one other dimension. The outcomes are not
just health outcomes and productivity outcomes, but also financial outcomes in
terms of what financial stresses households are facing, not just whether they
are insured or not, but what happens to the frequency with which families have
ten or 20 percent out of pocket burdens as a percentage of their income.
So it is something that is relatively low-hanging fruit for most of these
models. It would be great to see result tabulated. That is an easier question
DR. STEINWACHS: It ties also to the bankruptcy. You hear every so often
about people going into bankruptcy.
DR. SELDEN: Exactly.
DR. BANTHIN: I just want to add one thing. We know that even in countries
with universal coverage, there still remain disparities in terms of chronic
conditions, use of preventive services, use of the health care system, that are
generally along socioeconomic lines.
MEPS has been used. Thomas applied some World Bank measures of equity in
delivery and use of health care using that data. I am looking at the role of
assets, exogenous changes in wealth, controlling for insurance and their
effects on use of preventive care.
DR. STEINWACHS: So Jessica, could you give us some news about what is going
to happen to us if we lost some of our wealth recently?
DR. BANTHIN: I have preliminary results that say if you have an increase in
wealth and you go get that colonoscopy and those dental checkups, the
implication is, if you lose all that wealth, you may postpone your colonoscopy
and your dental checkups.
MR. SHEILS: There are lots of confounding factors. For example, a lot of
the uninsured, they just don’t want to buy any insurance because they are
healthy, their kids are healthy, and they have other needs, very serious needs.
Did I really say you shouldn’t get the car fixed so you can have insurance,
because you have got to get to work.
So there are a number of factors in play there, but to the extent that the
uninsured are somehow self selecting into that situation, it is quite possible
that they are healthy and they are not going to be a better health outcome.
It is unfortunate, but there are little bits of work you can do. I know the
IOM did a study on the uninsured, trying to establish the case. They were
unable to demonstrate any conclusive productivity. They couldn’t find the
surveys, they couldn’t find the data to say productivity goes up.
We took the NHIS and we fooled around with it one day, and we got an
estimate of what the increase in worker productivity is, or rather the
reduction in the number of work lost days. It was something like .4 work lost
days. It wasn’t that hard to do that. But you may not be terribly pleased with
the output. I’m not sure that productivity will be affected, or even our
international competitors would be affected.
DR. STEINWACHS: It won’t be the first time I haven’t been pleased by some
of the outputs. The other point that I think you are making is that it is not
just receiving services, it is quality too. Some of the things that we are
concerned about are conditions that are untreated, undiagnosed and therefore
you don’t get a benefit.
MR. SHEILS: I just want to say, we almost always show that the uninsured
are spending more, not less, as a consequence of reform, because a lot of them
are now required to get them.
MS. MC GLYNN: We find the same thing. I guess it is important to ask if
that is the right question. Honestly, to use the favored buzz word in the
nation today, one might construct a comparative effectiveness frame and ask if
you are interested in advancing the health of the population, what are the
tools at your disposal, and where frankly does insurance fall within that.
If you look at just in terms of contributors to mortality, we know the
health system is a relatively small contributor compared to things like
behaviors that aren’t clearly influenced by insurance or not. So we might be
better off investing in education, investing in putting recess back into
schools, investing in teaching our kids nutrition.
I think there are a lot of those things that insurance is coming in and
trying to pick up the mess that we already created. So I think these things are
important to look at. There is a fairly strong critical religious element that
doesn’t want to hear that, that thinks they know the answer. I think that poses
some challenges when either you do a document and you come up with not finding
a big effect, or you suggest that insurance isn’t the only mechanism.
So I think those things ought to be increasingly going hand in hand with
models that look at just the financial aspects. When I talk about quality of
care stuff, I really do think that what we have seen is that — and this is
true not only in this country but in other countries — just getting people
coverage and assets does not guarantee that once you walk in the door of the
doctor’s office, you are going to get what you need.
That frankly is the bigger determinant of health outcomes. There are
certainly a lot of factors that are tightly linked to the factors that we know
are true for some subsets of the uninsured, that may make them either at higher
risk or higher benefit from also having coverage. But we have to look more
comprehensively at putting a system together that delivers the care that is
So just fixing the problem of uninsurance isn’t going to fix all these
problems as well. Not that it is unimportant, but it is not the only thing.
DR. TUREK: The Daschle book, many of the examples there were of people who
had been employed with private insurance who ended up with something like
cancer couldn’t work and lost their insurance, and ended up going bankrupt.
One of the questions here is what happens to insurance if you get really,
MR. HITCHCOCK: I was just going to note that we have been here for over
four hours, and I haven’t heard anything about the concept of uninsured and
then the concept of employers choosing to raise deductibles or copayments and
the effect that that might have on enrollment. I don’t really know if we need
to discuss it, but I would just like to note that I was surprised that those
concepts weren’t more part of our discussion today.
DR. SCANLON: That may be a question for all of our modelers, to what extent
are the characteristics of policies built into the models. I think we heard a
little bit of that.
The other thing I didn’t hear much of is the whole issue of insurer
behavior. I think there is this potential assumption that health insurance is
going to involve some rigid set of rules and all insurance is going to conform
to that. Well, health insurance could involve modification of current rules
where there is latitude, and what insurers do may be an important part,
particularly if it is not across the country, but we have some state options in
there, what insurers do may then affect the outcome.
So maybe talking a little bit about the insurance offerings and the
MR. SHEILS: That is a really good point. Selection effects are a difficult
thing to model. What actuaries will do is say, we just assume you are getting
20 percent sicker people, because that is what we have seen elsewhere. That is
valid from an insurance perspective.
I’m not sure how well any of the models follow that. We model the selection
that comes from the obvious; we try to keep it here for you rather than over
here. We can model that.
An example. There was this guy down in Florida who was trying to enroll
people in his Medicare HMO. He did it by having street dances. The theory was,
if they are healthy enough to go dance in the street, they are a better risk.
How do you legislate that option?
So we are at this point are a little bit — maybe other people here have
got this mastered, but we are a little bit stumped on how we are actually going
to model it. But at the micro level the way we do it, this kind of selection
without resorting to an actuarial mechanism.
MS. MC GLYNN: I don’t think we have much data on insured behavior, because
we don’t largely observe who gets denied, who applies. A lot of that is just
not in the data sets in any way that helps you model insurer behavior. So with
big scale changes that is a little bit out of the model right now.
MR. GARRETT: What is in the model? At least with the utility based
framework, we have a standardized ESI type of package and we have a
standardized non-group type of package, and those are very different from each
other in terms of their coinsurance that we convert it to, and deductibles.
So to the extent that you have different types of coverage, it is not
always the same thing, we attempt to reflect that. When there is a reform, it
is not what is cheapest in our models, it is what is the value for you, given
various cost sharing arrangements and how it is going to affect the out of
DR. SELDEN: Just quickly on details of what coverage people hold, it has
been quite a long time since a natural forum for collecting that information
and presenting nationally representative information about that, and how it
interplays with the burdens that people experience and whatnot, would be in
MEPS. It has been a long time since we have collected and abstracted those. I’m
not sure what it would cost. But it is fairly low-hanging fruit.
Another piece of low-hanging fruit might be to ask questions regarding
medical debt in the MEPS.
DR. SUAREZ: I don’t know how this plays into the models that I didn’t hear
today was these so-called non-traditional health coverage approaches like HSAs
and MSAs. Is that something that is part of the consideration in those models?
Health service accounts.
MR. SHEILS: We do model it. I can’t tell you we do a great job of it, but
we do model it. We model the savings effects. One of the things that you find
is that, they talk about a high deductible plan, like $1,000. If you break you
ankle, you will spend $1,000 in 20 minutes.
It turns out that about 84 percent of health spending occurs above $1,000
deductible. If you take everybody who spends more than $1,000, add up the
amount in excess of $1,000, 84 percent. Once you have walked into an emergency
room, your insurers are gone. That is too harsh, but it is an aspect. It is not
going to have quite the effect that one might hope for.
MR. J. SCANLON: In all of your models, I think this is true, when you are
estimating, you take a baseline case in your current policy and you run it into
the future, and then you apply the potential policy changes.
Can you estimate the dynamics as well? Movement from an employer sponsored
plan to another plan, movement into Medicaid or SCHIP or something like that?
You can show that dynamic among and in between those enrollments as well?
MR. SHEILS: Right. A public plan is an opportunity to cover your workers at
a lower price. It models that. There is also the crowd-out issue. If your
workers become eligible for all these benefits, with the firm file, we have got
all the characteristics on these people, so we will figure out what their
premiums would be and if they are eligible. Then you can compare that to what
the premiums would be if the employer provides it.
You find a lot of cases where it is just cheaper to buy it. You have
identified a group that has a high risk of crowding out.
MR. J. SCANLON: Can I ask John the same question I asked the previous
panel? You have simulated — I know you did work for Commonwealth where you
tried to simulate a health insurance exchange and a public plan option. What
sort of assumptions did you make about a public plan?
MR. SHEILS: We did it for the Presidential comparisons. When Hilary Clinton
and when Senator Edwards talked about a public plan, they were very clear it
would be modeled on Medicare. Senator Obama was very careful not to say that.
We had no idea.
I had to cover my butt, so I just took a rate that was halfway between
private and Medicare, which was higher than any payment rate under the federal
programs today except FCHPP. We used that, and on the basis of that we were
able to identify those groups under the proposal with all of the other factors
considered now, it is better to go into the public plan.
The decision to move through the plan, not everybody is going to — I can
save ten cents, I am going over here. We had plan change elasticities that we
used to model that, but that is an extrapolation with probably very different
DR. BANTHIN: You can add inertia factors. We did that in our ’99 paper.
Economic theory predicts the long run equilibrium. That can take ten years to
happen, and politicians and policy makers want to know what is going to happen
next year and the year after. So you can work in some of these effects that
slow down changes.
Agenda Item: Wrap Up, Final Comments
DR. SCANLON: We are getting close to one o’clock, and people probably have
some commitments. According to our agenda, we are about 50-some minutes past
the time when the committee was going to resolve these issues that we are
struggling with. I think we are fortunate that we are 50 minutes past that, so
now we can put off resolving it to another day.
I want to thank everybody, and then open it up for last comments from the
committee members. I think this has been a valuable day in reinforcing a number
of things and adding a lot of new information to what we need to think about.
Our challenge, Larry portrayed it very accurately in terms of balancing
various perspectives. The challenge is to think about how to do that, how to
communicate that. We don’t make policy, we present information to the people
that do make policy; how to present the information on what seems to us like a
reasonable balance, how that balance does not involve unnecessary risk and yet
provides significant benefits, because that is what they make decisions upon
when they are thinking about policy.
So I’ll open it up to questions and comments.
DR. HORNBROOK: I just want to remind us that the data lag here is really
bad. We look at Oregon just for example. Our economy is going down the tube,
school districts are cutting back, the state is cutting back. We watch the
enrollment figures for health plans. They are losing enrollment. Those people
aren’t leaving the state, they are going on the uninsured roles. So there is
some very, very recent changes that are very marked now in certain parts of the
country, especially where the economic downturn has been very serious. It is
affecting families, and of course the mortgage crisis is affecting them. So we
are going to see people coming in on Medicaid. The Medicaid programs are going
to have their expenditures go up, but you are going to see the state budgets
can’t afford it. It is going to be a strange simulation that is way beyond some
of the dynamics in the models we have now.
DR. GREEN: This has been a week like no other in my personal life. I came
in Monday, and we started a pre-hearing before the NCVHS, and we are now having
a post hearing after the NCVHS meeting, and they have all had different names.
DR. STEINWACHS: Just to give you variety.
DR. GREEN: Well, actually I think it is to give us the illusion of variety.
It has been a very convincing week. We actually have a fairly desperate
emerging need. Mark crystallized it again for us. The way we go about doing
statistics and analytical work with the intention of influencing the directions
we take is insufficient. We don’t seem to have any dissent about that.
We are getting glimpses of the new world in the information age where data
and information flow differently, connect differently, are available in
different weights and different scale and different velocity. I am beginning to
think that, as is often the case in policy, we may not have it framed right. I
am looking forward to continue to work with the committee about us trying to
get our minds around how this needs to be framed in order that we can get to a
well ordered solution.
MS. MC GLYNN: I just want to make a comment on that. I really do feel like
this is a fundamental issue in health care that for decades now, we are trying
to do stuff with an unbelievably inadequate data infrastructure. Until we bite
the bullet and start with what would a functioning health industry look like
that had data like almost any other industry you can look at to make real time
So forget about research and simulation and modeling; what about just
making decisions day after day by individuals, by physicians, by insurance
companies, by firms, et cetera? What would that information infrastructure look
like? In some ways, that needs to be the focus. I say this at some hazard of
having my health research union card pulled, but almost more of that is
important than fixing the research survey database, or at least it ought to
also be on the table.
I think that until you get to that point, the ability to respond to nuanced
ideas about changes in how we pay for health care, nuanced ideas about
comparative effectiveness information available, none of that can get
translated into really different behaviors on the top of the information
infrastructure that we have today.
So I think you are absolutely on target in terms of saying, are you asking
the right question, and what would it look like if you came at it from a
different direction. We are starting over today, and we are going to build a
system where every day in real time I have information on how many people died,
how many people have disease X, how many people got treatment Y, how much did I
pay for that, and that it looks like models that we can see in retail, so that
you can have response in real time, solving today’s problems, not problems that
we now well understood five years ago.
Sorry, one of my soapboxes.
DR. SUAREZ: I was actually doing a little bit of modeling here with the
wish list that you all provided us. I think that is a great place. I think it
all builds into the comments that have been made.
I was looking at the various suggestions, and I want to point out a few of
them very briefly here. We need to increase the size of population surveys, and
we need to have better longitudinal data sets, like longitudinal coverage data,
creating a long term longitudinal survey.
We talked a lot about state specific data and trying to add that level of
analysis and trying to drill down into better understanding of the demographic
at that sub-national level. We also talked about linking employer and employee
data sets. We talked about having a better employer micro data set, and even
getting some better non-group market data. RAND has been developing a provider
data set, a longitudinal larger perspective data set.
Then I think we heard a few other major themes, like the privacy issue and
the concerns around what privacy will continue to do, and might even do more as
we get into the next iteration of privacy protections, and this concept of
creating a certification program.
We talked about the need to look at subsets of population like minority
health and health disparities and the effect of policies on minority population
choices for health coverage.
But the most fascinating thing was your comment about, five years ago doing
these analyses of the most prominent microsimulation models, and trying to
perhaps go back and do on in a larger scale, and perhaps use that as a way to
look at your question, Larry, are we looking at this in the right way, or
should we be thinking of taking a different approach altogether.
So I think it has been a rich morning. I really want to thank you also for
coming and giving us all this wonderful information.
DR. STEINWACHS: I just want to thank all of you. This has been to me very
insightful and useful. You did discourage me away from my health outcome model.
It was part of my comparative effectiveness approach. Beth has just taken it
right on and said, forget it.
I think this has helped, really. You have heard the questions from the
committee, to think about what are the data needs and how do we begin to meet
those. This really ties into a vision of the Nationwide Health Information
Network and health information infrastructure, of which statistics is a piece,
but not the whole. So we keep talking about it and we keep having this vision,
but at the same time feel that we are moving all too slowly to connect those
pieces together, and we need to. So again, I thank you very, very much.
I thank all the people that helped make this possible, Roshita and Dale.
(Whereupon, the meeting was adjourned at 1:00 p.m.)