Anderson, David F. with David P. McCaffrey, Paul McCold and Doa Hoon Kim,"Regression and Case Studies of Public Programs: Discrepant Findings and a Suggested Bridge", 1984

Online content

Fullscreen
179

REGRESSION AND CASE STUDIES OF PUBLIC PROGRAMS:

DISCREPANT FINDINGS AND A SUGGESTED BRIDGE*

David P. McCaffrey
David F. Andersen
Paul McCold

Doa Hoon Kim

Nelson A. Rockefeller College of
Public Affairs and Policy

State University of New York at Albany

omments of Fred Thompson, Irene Lurie, and Richard Nunez are
gratefully acknowledged. Thi

This was a fully collaborative effort.

ABSTRACT

Case studies of regulatory and social programs suggest that policy

systems are dynamic. In the systems described, outcomes depend on how

numerous variables interact over time, and feedback among variables

"simultaneous" causation over multiple time periods -- is more a rule

than an exception. However, the most influential evaluations of public

prograns are studies using multiple regression. A recognized limitation

of multiple regression is its relative insensitivity to multiperiod

strategies, feedback among variables, and other dynamics. Accordingly,

we maintain, the findings of regression-based and case studies commonly
conflict. Simulation modelling can serve as a methodological bridge
between case studies and regression-based studies of policy systems,
fuiproving) CheOEeEICAL wodery GF thelayatans aid preViaing @ way te

evaluate the of i

models. The results of

some early experiments along these lines are presented.

SECTION 1: INTRODUCTION

Case studies of regulatory and social programs suggest that policy
systems are dynamic. In the systems described, outcomes depend on how
e

numerous variables interact over time, and feedback among variables --

‘simultaneous causation over multiple time periods--is more a rule than

an exception. For example, weak

er poor
often prompt opposition leading to strong regulation and management
reform; "unlikely" coalitions form routinely; and so forth (McCaffrey,
1982: 402-403; Bacow, 1980: ch. 5; Policy Studies-Urban Systems
Vol. 11).

1981:

How

r, the most influential evaluations of public programs are not
case studies, but rather studies using miltiple regression. A recognized
limitation of multiple regression is its relative insensitivity to

multiperiod strategies, feedback among variables, and other dynamics.

Statistical and data problems tightly constrain the number of variables,

and feedbacks that can be i ly built into

regression-based analyses of public programs (Meadows, 1980: 40-46).

Also, the research i

with i i

cross- "chin" i

of large numbers of
individual cases, rather than longitudinal, "thick" description of fewer
cases. Such distant examination of even many cases will detect only the
most obvious dynamic relationships, and even then not provide data that

are sufficiently rich to explain them.

An assumption of regre:

ion analysis is that variable interactions

feedbacks, nonlinearities, and other dynamics are not so important as

180

oe

to prevent the necessarily simplified regression model from picking up
the strongest signals in the policy system. The aspiration of regression
analysis is to detect central factors that dominate other relationships,
however complex.

Arguably, the method works quite well, for it appears to uncover
strong, stable signals in numerous policy systems. Regression-based
studies find, fairly consistently, that Occupational Safety and Health
Administration inspections do not reduce industrial injury rates or
increase investment in safety-related capital equipment (Smith, 1979;
Viscusi, 1979; McCaffrey, 1983). They find, fairly consistently, that
state certificate of need (CON) regulation of hospital investment does
not reduce capital investment (Steinwald and Sloan, 1980; Sloan and
Steinwald, 1980; Joskow, 1981; Fuangchen and McCaffrey, 1983); that
variability in school expenditures or teaching methods are not associated

istically with variati

in student (Hanushek, 1981);
and that the organizational characteristics of medical care, or

or ethnic i of indivi

are iated in
only trivial statistical ways with use of physicians’ services (Mechanic,
1978). The list of stable findings from regression analyses could be
lengthened almost at will.

But what is striking is that these stable statistical findings
commonly conflict directly with analyses using case study or other

"thick" descriptive technique

As we will see below, intensive studies

of firms suggest that OSHA does improve safety performance and increase

y case studies suggest that

~3-

certificate of need regulation does affect the medical technology market
and hospital investment (E.g., see U.S. Office of Technology Assessment,
1981: 32-38; Howell, 1981); that school characterististics do affect
student performance (Rutter et al., 1979); and chat organizational and
Psychosocial factors do substantially influence physician utilization
(Mechanic, 1979). Analysts sometimes label the case studies as
"interesting" anomalies, but maintain that greater veight should be given
* regression analysis of carefully selected samples (E.g., see Sloan,
1982: 200). We question this argument because the discrepancies between
case studies and regression results are too common and systematic to
reflect only the anomalous qualities of the cases.

Instead, this paper maintains that regression-based studies, however

stable their results, are not definitive. Practically none of the

Tegression based studies of occupational safety regulation, and very few
of the studies in other areas, are grounded in credible theories of how
the policy systems work. Although the empirical results are consistent,
it ie therefore difficult to know what the results mean. The problem
exists because the data used to operationalize variables me: jure the
variables only grossly or ambiguously. Or, worse, the variables
themselves have no theoretical meaning, serving only as rough proxies for
variables or processes that are undefined or unknown, but Suspected to be
somehow important. the gaps between data and variables, or the
ambiguities of the variables themselves, mean that even stable empirical
results tell us little.

181

-4-

‘ rT
Section Il of this paper illustrates these theoretical and empirical
problems in detail, focussing on occupational safety regulation. Other
policy areas could have been easily used (F¢ of
licy areas could hat n y used (For extended discussions

: ‘ ; A
lar problems in the research on hospital regulation, education, an

physician utilization, see McCaffrey, Andersen, McCold, and “ 11984).
Section III outlines a research strategy, relying on dynamic

simlation modelling, that could ease some of these problems. The first
element of this strategy is to construct simulation models of a given
policy system using both quantitative and qualitative data. Simulation
modelling is a method designed specifically to examine and structure
dynamic causal relationships; its distinctive strengths are those
qualities in which regression methods are weak. The method is also a way
to structure in theoretically coherent ways what otherwise would be

disparate, noncomparable case studies. The second element of the

strategy uses the simulation model-~and variations on it--to generate
synthetic data on the simulated “reality.” Regression models can ven be
used to estimate properties of the simlation model, and the relative
success of alternative regression specifications in describing the known
model can be noted. The two elements of the research strategy are
related, but each has a distinctive advantage. The first element would
improve causal mapping of the policy system; the second element would

of i ion models

provide a way to test the

prior to data collection for an evaluation project.

ape 182
SECTION 11: RESEARCH ON OCCUPATIONAL SAFETY REGULATION AND OTHER POLICY

SESEARCH ON OCCUPATIONAL SAFETY REGULATION AND OTHER POLICY

SYSTEMS: THEORETICAL AND EMPIRICAL ISSUES

SESTEMS: THEORETICAL AND EMPIRICAL ISSUES

OCCUPATIONAL SAFETY REGULATION

Since 1971 the Occupational Safety and Health Administration (OSHA)
has enforced standards for working conditions using inspections and
fines. Most of the empirical studies evaluating the impact of the
Occupational Safety and Health Administration present a "theory of work
injuries" (Smith, 1979: 147-151); @ “conceptual analysis of occupational
health and safety regulation" (Viscusi, 1979: 118-122); or a “model of
injury rate determination" (Mendeloff, 1979: 98-102). These theoretical
sections serve as interpretations of the results, if the results are in
the predicted direction and statistically significant. The theories
focus.on the associations between occupational injury rates (or injury
rate changes) and OSHA inspections, industry, the size of firms, firms’
employment changes, and firm-specific factors. Some studies have also
examined the impact of workers’ compensation benefit increases on injury
rates, but for a variety of reasons these studies have not been
integrated into the study of OSHA's impact.

Empirical findings regarding these variables are quite stable, with a
mix of positive, negative, and negligible associations. But either the
data used to measure the variables are quite incomplete, or the variables

themselves have no clear theoretical relationship to injury rates.

. -6-

Therefore, when the studies try to estimate how a policy variable--
usually the OSHA inspection variable--is related to injury rates,.the
relationship is very ambiguous. Or, when analysts only try to "control"

for variables so that the independent effects of the policy variable can
be

sessed, they cannot in fact assume that control variables do not
interact in unmeasured ways with the policy variable because they have
only the vaguest idea as to the dynamics of either the control or policy
variables (See Footnote 1 above). Consider the results for each

variable.

Safety and Health Admini i i

Analysts are interested in the penalties that firms expect for
violating OSHA's standards for working conditions. OSHA could induce
firms to invest more heavily in safety and health capital by increasing
the probability of inspections and/or the penalties associated with

standards violations (Mendeloff, 1979: 94, 108; Viscu:

» 1979: 123-125;

mith, 1979: 149; McCaffrey, 1983: 132-133). ‘The studies examine how the
rates of i ions in i the of i ions in

firms, or penalty levels affect injury rates:

The obvious ways to assess the impacts of OSHA inspections, since
reliable pre-OSHA and post-OSHA data are unavailable, are to compare the
injury rates of industries with relatively high and low rates of
inspections, or the injury rates of inspected firms with those of
uninopected firms. Several studies have found that high inspection rates
in industries are not associated with relative injury rate reductions:

Several analysts are skeptical of the velue of results for an aggregate

-7-

category like "industry." Robert Smith, for example, suggested that it is
relatively futile to estimate the effects of OSHA from industrial data
because the potential effect of OSHA is relatively emall and "likely

confined" to inspected firms (1979: 147, 169); thui

analysts prefer to

use firm-specific data rather than industrial date.

One cannot meaningfully compare inspected and uninspected firms.
Comparing them would require that inspected and uninspected firms behave
in similar ways, but because inspections can be triggered by employees’
complaints, catastrophes, growing hazards, labor-management problems, and

so forth, inspected firms may be qualitatively different from uninspected

firms. Thu:

analysts have opted to compare rates of firms that have

been inspected at different times of a given yea

jeually March-April
inspectees with November-December i ¢

P
are used rather than

ry y 90 that the i i

vould not have been prompted by temporary increases in injuries reflected

in the prior year's injury rate). Although the late inspectees would

presumably be similar in other respects to the early inepectees, their
annual injury rate would not be affected significantly by the November or
December inspection. Thus, the impact of the inspection on the annual

injury rate can be Quite

ly, the i i
variable is not associated with relative declines in injury rates.

The inspection variable for firms comes from the Bureau of Labor
Statistics survey of injuries and illnesses, and represents the month of
the first inspection that a firm had in the year (Zero if there was no

inspection). This inspection variable does not articulate possibly

-8-

important variations in enforcement. The variable does not indicate if
the inspection was a routine “general schedule" inspection, or whether it
was prompted by a major accident, an employee complaint, or was a
follow-up visit to check for abatement of violations cited in previous
inspections. The variable does not record multiple inspections, nor does
it indicate if the inspection resulted in any penalties. .
Finally, even. though the inspection data are reported by month,
variables which one wants to correlate with the inspection data are
reported only by year, and so one cannot measure temporal sequences.
Studies test for associations between changes in injury rates and the
occurrence of inspections. However, the resulting associations do not
indicate whether (1) the inspection altered injury rates (e.g., by
reducing the injury rate, or by increasing the injury rate by increasing
the reporting of injuries); (2) whether the inspection was induced by a
prior increase in injuries; or (3) whether both (1) and (2) occurred in
the firm, Indeed, it is not at all unreasonable to suspect that all of
these tendencies are real. Interpreting non-significant statistical
effects is difficult when, for example, inspections might improve the
level of safety in the firm but simultaneously sensitize workers and
those who record injuries to safety problems. (In a related observation,
|. Kip Viscusi suggested that increases in safety-health investment may
not reduce injury rates as workers will respond to gafer conditions by

“diminish{ing] their level of safety-enhancing agtions...that affect

either the probability of an accident or the size of the loss...For

example, workers may get more careless if the company adds guards or

-9-

184

safety cables to the machine” [p. 118]. The point, however, does not
figure in his statistical estimates of the net impact of OSHA).

The ambiguities of the inspection data become troublesome when one
contrasts studies of OSHA's impacts which use multiple regression with
those using other methods. The conclusion commonly drawn from the null
regression results for the inspection variable is that the expected

penalties for violating OSHA's standards are too low to justify

safety i + W. Kip Viscusi, in fact,

analyzed a ion of i ies'

y ated ii and
reported that a greater frequency of inspections was not associated with

higher i in

y capital from 1972 to 1975
(1979). “Another conclusion commonly drawn is that OSHA's standards may

not address true safety problems of firms.

However, studies which directly ask management or workers about
OSHA's effects on safety conditions or opetations, or look at
disaggregated investment data, imply that OSHA does affect firns'

behavior and their level of safety. After studying plants unionized by

the International Association of Machinists, Kochan, Dyer, and Lipsky
wrote that since OSHA's establishment "management has assigned a higher
priority to plant safety, the ability of the union to influence

management decision making on safety issues has increased, and the role

of

fety

has been bolstered." They reported

that 69 percent of the workers, and 48 percent of the managers, said that
OSHA had a "strong" or "very strong" impact on safety consciousness and

responsiveness in the plants (1977: 5, 36, 76-77). A survey of chemical

= "10 -

workers by Cambridge Research Reports for the Shell Oi] Company reported
that 64 percent of the workers said that safety conditions had improved
in their plants in the 1972-1978 period (1978: 110), and that OSHA was a
contributing factor. (See also Freedman, 1981).
Also, an Arthur Andersen and Company survey of 48 large firms, which

is the most rigorous study of regulatory costs to date, found that the 48
firms spent $68 million in capital outlays directly related to OSHA, and
$184 million in OSHA-related expenditures overall, in 1977 (1979: Section
8, p. 5). The report added that the firms had higher OSH-related capital
costs in earlier years, which included the period studied by Viscusi
(1972-1975). The Arthur Andersen study, which involved a detailed
analysis of 48 firms’ investment decisions in 1977, and less detailed
analysis of earlier decisions, directly conflicts with Professor

Viscusi's, which involved pooled

of
industry data, with inspection frequency being the OSHA policy variable.

Thus, multivariate studies ini

industry i »
injury rate, and inspection data, or data on firms’ injury rates and the
occurrence of inspections, find that OSHA's activities are not associated

with injury rate changes. Those studies examining particular groups of

firms in great detail, focussing on a variety of practices and decisions
within firms, find just as consistently that OSHA's activities do affect

firms’ behavior and the level of safety in firms. (The “level of safety"

is not the same as “injury rates" because the reporting of injuries may

vary for reasons other than the level of safety. But this would mean

that focussing solely on injury rates would be an inappropriate way to

evaluate OSHA's impact on occupational safety).

-~a.b-

One could argue that these effects may be limited to specific classes

of firms--e.g., the firms unionized by the i i

of
Machinists, chemical workers, or the large firms studied by Arthur
Andersen and Company. But this would mean that multivariate studies are

mis

pecified, leaving out important variables that do magnify OSHA's

effects. The argument would also--for several other re:

jons--be weak.
(For example, the 48 firms studied by Arthur Andersen accounted for
nineteen percent of all capital investments made in 1977 in the United

States; this would be a very large “atypical” cla

of firms [Arthur
Andersen Report Executive Summary, p. 7]).

Another possibility is that regulation affects firms' operations and
behavior in more complex ways than can be realistically described by
wultivariate regression studies of national data, or of rather skeletal
firm-specific data. Regardless of whether one regarde OSHA's possible
effects as desirable or perverse, it appears that one will not understand
their dynamics except through alternative methods.

Examination of the other variables included in studies of OSHA
underscores this point. These are typically. entered as control variables;
however, as alluded to earlier (See footnote 1), their connections to
injury rates ‘are so ill-specified that we cannot assume that they are
independent of OSHA's effects. These other variables are now discussed.
Industry Variables

Occupational safety hazards vary by industry. Some technologies are
tore hazardous than others. The marginal costs and benefits of safety

controls also vary by industry. Time series data which indicate the

-12-

dominant type of technology, or safety controls’ costs and benefits, are
often not available at the industry level and hardly ever available at
the level of the firm.

Also, trade associations, insurance companies, or other

organizations, for any number of reasons, attend to safety probleme in

some industries more than in others. For example, the chemical industry

is reportedly criticized far more heavily for--and is more conscious
of--toxic waste disposal problems than the primary metals industry,
although the industries’ waste outputs are similar, because of the

chemical industry's attachment to Love Canal and s.

Street Journal, June 30, 1983: 58).

‘A study by Lawrence Bacow found that

1 safety are i more vi in the
auto industry than in the steel industry for several reasons, although
doth

industries have formally strong agreements (Bacow, 1980: ch. 5).
Detailed comparative studies of industry attention to safety matters

(such as Bacow's) are rare. Other industry effects, in addition to

industry technology or varying attention to safety matters, may be

consequential as well.

Analysts try to control for industry effects by entering dumy
variables for industry (usually the two-digit Standard Industrial

Classification level). Note that this procedure does not indicate what

“industry effects" are; it only lets us assess whether or not some

non-specified industriel factors are associated with injury rates. It

does not let us

sess whether OSHA contributes to industry effects
by--for example-~giving unions in certain industries a "club" to enforce

safety agreement:

or by affecting the cost and effectiveness of

fety

eq

equipment by stimulating

for certain i
or through several other plausible channels, none of which are adequately
described by an inspection variable.
Employment: Le
Injury rates vary by employment size in a curvilinear way. Small
firms (1-99 workers) and large firms (500 or more workers) tend to have
lower injury rates than firms of intermediate size. One explanation
commonly noted in the quantitative research is that small firms closely
monitor safety conditions, and that large firms are likely to have
injury-reducing safety programs. No one really knows if this explanation
is accurate. Alternatively, true injury rates could be inversely related
to firm size because of economies of scale in safety work, but small

firms might have distincti

injury di Data on

safety record keeping violations in the product safety area

re

consistent with this view (Linneman, 1981: 474).

The size category var:

bles do not indicate the size-related factors
that influence firms' injury rates. Rather, the size variables only
indicate whether or not some non- or ill-defined size-related factors
appear to influence injury rates. If, for example, OSHA reduced true
injury rates in distinctively hazardous small firms, but simultaneously

led the firms to tighten up distincti i

injury di
one could easily find no net change in injury rates. To conclude that

OSHA did not affect firms’ behavior would be an error.

are i

ly with

Changes in employment in firms are usually

-W-
represented by a ratio of employment in a year t+1 to employment in a

year t (EMP,,,/EMP,); changes in employment in industries are often

represented by new hire rates. Again, while these variables appear to

"control" for employment increase effects, there is very little research
on what these effects are. Employment increases may be associated with
an influx of accident-prone or inexperienced workers; with an influx of

young workers who are more willing to complain about job hazards, report

injuries, and request inspections; with an increase in the ratio of

workers to i worker:

¢ in firms (increasing
rates of injuries per 100 employees relative to other firms); and other

factors. While time and time again employment incre:

e effects appear
important, there are no studies that would give us a reasonable way to

choose among these or other interpretations of the employment increase

variable.
'Firm-specific" tors

Firms in certain industries or size categories, with certain
employment level changes, or inspected by OSHA, undoubtedly have other
characteristics associated with injury rates. Particularly outmoded

equipment, 1 di ies, or ieti

of the
plant's immediate location are possibilities. These factors should be
reflected in each year's injury rate. One way to control for such
factors in a year t is to enter the lagged injury rate as a dependent
variable (Smith, 1979: 149). Consistently, this variable accounts for

65-90 percent of the variance "explained" in studies of dccupational

injury rates, with the figure increasing with sample size (Smith, 1979;

= 15 -

McCaffrey, 1983. Data on the proportion of explained variance due to
lagged injury rates are available from McCaffrey). Such a variable does
not indicate which plant-specific conditions tend to influence injury

rates; it simply "controls" for such conditions. Thus, note that the
variable which the results of i etudies of

i afety 1 is a proxy variable with no

referent 1. Imagine any plausible interaction between unique firm

and OSHA's exi

a6 proximity to an OSHA area

office and the ion it distri iss xii

8 who
commonly threaten to "call in OSHA," managers who for reasons of personal

biography stress safety, etc.

and these interactions could contribute to
thie variable in ways not picked up by an OSHA inspection variable, or an

inspection-lagged injury rate variable.

A final factor thought to affect injury rates is the liberalization

of states’ workers' compensation la

although the effects are
ambiguous. Benefit liberalization may decrease injury rates as employers
strive to reduce injuries to avoid higher insurance premiums.
Conversely, benefit liberalization may increase injury rates as vorkers
become more willing to report injuries, claiming compensation. Studies
of the association between benefit liberalization ard injury rates
report, fairly consistently, that benefit liberalization is associated
with increased injury rates (Worrall and Appel, 1982; Butler and Worrall,
1983; McCaffrey, 1983). None of these studies have looked closely at the

dynamics ing the isti iatii

her the

187

See

first effect does not exist; whether the first effect exists but is

somewhat offset by the second; or whether other factors might explain the

benefit liberali injury rate

Nor have the studies

the workers’

variable with OSHA,

F ii, Be iabl

Thus, we really do not know how

workers’ compensation liberalization ought to enter into a theory of
occupational safety regulation.
Summary

The results of research on occupational safety regulation are quite
stable, showing a mix of positive, negative, and negligible
associations. However, the results say surprisingly little about the
dynamics of occupational injury regulation. The data used approximate
only ambiguously the variables of interest. In most cases the operative
variables themselves are not specified or are ambiguously specified, eo
it is extremely difficult to interpret their individual effects or know

ible interactions with other variables,

of their po:
Porter, Connolly, Heikes, and Park (1981) outlined four types of
demands that might be made of regression. We might ask regression to
describe the characteristics of a single data set; to predict the
characteristics of future behavior in the policy system; to causally
model (or "explain") the patterns of behavior in the policy system; or to
causally predict how future manipulated shifts in one or more variables
in the eyatem will affect other variables in the system. Where does the
research on occupational injury rates leave us in terms of these levels

of policy analysis?

-.7-

The research does not describe the characteristics of single data
sets, because it does not describe the specific factors underlying injury
rate patterns in single data sets. Consequently, although one can
predict how statistical associations will turn out in additional data
sets, one is not predicting the dynamics of injury rates at all. One is
predicting only that the current ambiguities will persist. A fortiori,

the area has no basis on which to develop

models or causal

models of

1 safety ion, or i
occupational safety regulation policy.

The problems of research in occupational safety and health regulation
appear in numerous other policy areas.” These common problems involve
the relative is of

dyn

ice in i (or
even longitudinal) regression studies; the extent to which regression
results are not linked uniquely to theories of how policy systems
operate, or do not: suggest what adequate theories might be; and
aggregation of variables in ways that may level important sources of
variation among variables.

Detailed case studies, which distinctively avoid these problems (and,
taken alone, have other problems) often directly conflict with the
regression-based research in other areas, just as the two types of
research on OSHA conflict. The case studies indicate that certain

variables are related strongly in ways that are somewhat complex and

unfold only over time. Regression analyses of cross-sectional or even of
thin longitudinal data detect few of these relationships (McCaffrey,

Andersen, McCold, and Kim, 1984), It is reasonable to suspect that the

18 -

discrepancies between the two sets of findings reflect regression's
relative insensitivity to dynamic or longitudinal relationships.
Certainly the evidence for this view is not definitive. However, the
evidence is strong enough to lead us to believe that the marginal payoffs
of shifting some effort to process-sensitive analytical methods may be
quite high. As will be discussed in the next section, the payoffs would

be both conceptual and technical. The conceptual payoff would be more

modelling. The technical payoff

would be a pi to

how well ion models

do describe the characteristics of policy systems.

SECTION III: RESPONSES TO THE PROBLEMS: SIMULATION MODELLING AND

SYNTHETIC DATA EXPERIMENTS

This section discusses two ways to cope with the methodological
problems discussed in this paper. The first activity involves extending
case-study analyses through the creation of dynamic simulation models of
policy systems. ‘The purpose of this activity is to describe more
adequately how events, constraints, and feedbacks emerge in policy
systems like occupational safety or hospital regulation, and to identity
underlying structural properties of the systems. The second activity
involves an attempt to estimate experimentally the conditions under which

dynanic feedback driven systems may present problems for regression

analysis. "

-19-

CONSTRUCTING DYNAMIC SIMULATION MODELS OF POLICY SYSTEMS

Case studies and other types of qualitative research suggest that
policy systems. are dynamic and ought to be. observed longitudinally.
However, regression based methods do not handle such complexity well.
Because the use of regression analysis requires large numbers of
observations, analysts are necessarily constrained in the time period and
richness of the data collected for any single case. Analysts tend to
restrict exploration of causal behavior in policy systems, and modelling
of the systems, to accomodate what can be roughly tested with these quite
skeletal data. Even where the theoretical models are more fully
developed, the data are not sufficiently rich to allow adequate tests of
the models. Thus, the theoretical and empirical work on the policy
systems are, at best, loosely coupled. 7

Dynamic simulation modelling is a method that can link
regression-based results and the large amounts of case material in these
areas in a theoretically disciplined way. Like case studies, dynamic
similation modelling uses diverse sources of information. The method
uses the numerical data available to regression analysis, but the method

also draws on i iews, ici rg and other

non-numerical data in constructing models of how the policy systems

function. The method goes beyond case studies, however, in providing

some clear principles for i

the i ion into a i
model of the system. These principles are derived from feedback loop
theory (Forrester, 1961, 1980; Randers, 1980), and force analysts to

specify clearly--in both graphic and mathematical form—the multiple

= 20-

of variables and

in the system, the time periods
in which these effects play out, and how the multiple connections in the
system either amplify or dampen certain behaviors. The behavior of the

model can be projected into the future through computer simulation.

There is a constant interaction between the implications of the model and

behavior in the "

eal" world; discrepancies force the analyst to refine

the theoretical structure of the model.

Once the evidence gathered through essentially case study methods has
been assembled into a formal similation model, several features not

available through simple case studies are available. First, most of the
theoretical richness of the case study can be retained.

This is because

the relatively flexible ical form of si i

models allows for
the inclusion of a large

number of "hard" and "soft" effects. Second,
the simulation model will contain a mathematically explicit causal
structure. Whereas case studies can be rich in descriptive detail,
simulation models force the analyst to posit explicit causal hypotheses
concerning how the policy system operates. These causal hypotheses can
then be tested further via regression or other forms of statistical or
empirical analysis. A third feature of the simulation model is that all
assumptions of the model are available readily for inspection since they
have been cast into an unambiguous mathematical form. The model (and
hence the causal theory being tested) must be absolutely consistent and
logically complete. These features, often neglected in simple case
studies, must appear in a formal simulation model or the model simply

will not compile and run on the computer.

- ae

A fourth feature of the simulation model is that the behavioral
consequences over time of the causal hypotheses can be known with
certainty. That is, by actually running the model and creating a
simulated scenario, the model itself will generate synthetic time series
data that arise solely from the caueal assumptions built into the model.
By running the model and observing if the simulated behavior is
reasonable, the analyst can gain some additional insight into the

and logical of the

built into the

model. A fifth feature of a case study recast as a simlation model is

that analysts can easily insert alternative causal hypotheses into the

model, and then examine the implications of such changes over time.?
Thus, the first benefit of simlation modelling is that it can use

diverse sources of i ion, including ipti

rich case
studies, to theoretically model policy systems in a rigorous way.

USING DATA EXPERIMENTS TO TEST A k MODEL'S ROBUSTNESS

The second benefit of simulation modelling is that synthetic data

experiments, using the simulation model, can explore in mathematically

precise ways the and itivity of models.
The basic ideas behind ic data are, in princi
simple. First, the sii ion model--or a plausi liminary model--

is run a large number of times under stochastically varying conditions to

create a large number of observations. These observations are called

“synthetic data." Because the structure and stochastic character of the
model is known, the characteristics of the data generated by the model

are known as well. The experimenter ought to be able to design a

190

-22-

ying on si i lagged iabl
and variable transformations if necessary--to recover the known structure
and parameters of the data generating model.

Tne nature of the synthetic "reality" is easily changed by altering
the simvlation model; similarly, the specification of the regression
model may be changed easily. These changes may include “corrupting” the
data generating model by inserting amounts of measurement error,

inserting odd lations within i iabl

planting

plausible misspecifications in the regression model, and other probl
that are suspected to exist in real data and analysis. Because the
structure of the data generating model is known, the experimenter can
monitor in a mathematically precise way how well alternative regression

models recover the si ion model's and

under

these imposed and ification problems. ion of one
such experiment usually takes from five minutes to one half hour to
complete.

Such synthetic data experimente bear a family resemblance to Monte
Carlo simulations and often prove interesting because the statistical
properties of feedback-driven systems differ from those assumed by
ordinary least squares regression and most of its derivatives (Andersen,
1981). For example, Peter Senge (1977) has used this experimental design

to inv

tigate how well regression models can recover parameters from a
dynamic eyatem when the observed independent variables have been

corrupted by relatively small amounts of measurement error. For feedback

driven systems, the estii

tion models performed well for no or minute

-23.

quantities of measurement error, but the estimation model's performance
degenerated rapidly under small to moderate amounts of measurement error
(See also Johnson, 1980).

Also, Mass and Senge have used a similar design to investigate
several puzzles that arise when tests of statistical significance are

used to include or exclude variables when estimati from a

dynamic feedback-driven system. They concluded that simple tests of
significance may be poor indicators of which variables are important and
unimportant in determining the behavior of policy systems. (Mass and
Senge, 1980) Similarly, Luecke and McGinn (1975) used a modified Markov
chain to investigate how well regression might estimate schooling
effectiveness. Luecke and McGinn discovered that even when expenditures
for educational inputs such as better teachers and schools are known to
have an effect on student achievement in a simulated reality, estimation
models used in major evaluations of education still produced null
results

We have begun to extend this method to the study of regulatory
problems; the first results, dealing with a very simple model of OSH
regulation, are now discussed.

‘An OSH Data Experiment: Early Results

The B:

ic Model. A simplified simulation model of accident
generation within firms was created. According to the model, accident
rates within @ firm were an additive combination of factors related to a
firm's size, industry type, and safety programs, and a firm-specific

constant. As discussed below, one version of the model assumed that OSHA

191

- 24 =

inspections influenced the impact of firm's safety programs on accident
rates; another version of the model assumed that OSHA inspections did not
influence safety programs and accident rates. We explicitly set the basic
model to resemble closely the model specified by most regression studies
of OSHA's impact.

The model generated monthly accident rates for 1000 firms, and
accumulated these monthly rates into yearly total rates. The generating
formula used was:

Acc =K+(5*S8) +24 (541) +E
where

‘Acc = monthly accident rate;

K = a firm specific constant (value 0 through 1);

$ = the cumulative accident factor determined by a firm's safety

program (value 0 - 2) (Henceforth; this particular factor is called

the "safety factor);

2 © the accident factor determined by the size of the firm (value

0-1)

1 = the accident factor determined by the firm's industry type (value

0- 2)5

E = a random error term (value 0 - 1).

Firms fell into one of five categories of increasing size. If size
was 1 or 5, Z = 0; if size was 2 or 4, 2 = .5; andsif size was 3, 2 = 1.
‘Thus, the size-related accident factors followed the curvilinear pattern

reported in the literature. Two versions of the model generated

'  25-

data sets. In one version, it we

specified that OSHA inspections would

reduce the actual safety factor to zero for five month:

; this is called
the "OSHA-effective" model. In the second version, OSHA inspections did

not affect the actual safety factor; this

8 called the "OSHA-

ineffective"

model. Initially, all values were selected randomly for the
firms. A firm's safety factor, size, and industry were allowed to vary
between months using a Markov process of varying degrees of stability.
In the model, OSHA inspected ali firms in the second year (months
13-24), half early and half late in the year (in March-April or in
November-December). The data output for each of the 1000 firms included
the yearly accident rate for both years, firm size and type at the end of
the second year, and the month of inspection. The inspection variable was
coded 1 if the firm w

inspected early, and 0 if the firm was inspected
late.

A generalized multiple linear regression model was used to evaluate
the effects of OSHA inspections on yearly accident rates. Following the
earlier studies, the regression equation attempted to fit the accident
rate for the second year to the previous year's accident rate, firm size
and type, and whether the firm was inspected early or late. The
regression coefficient and significance level for the inspection variable
were recorded for the data sets generated by the OSHA-effective and
OSHA-inef fective versions of the model.

If the ion model

the data generated by

the simlation model, then the regression coefficients would match the

difference between accident rates of the two years. The actual impact of

192

~ 26 -

OSHA inspections on accident rates could be determined by simply

annual accid within the simul

where OSHA had an
effect from annual accidents within the simulation where OSHA had no
effect. This is because it all other ways, including random sequences,
these two model variations were identical. Furthermore, the F value for
the inspection variable would be highly significant for the
OSHA-ef fective simulation and would be insignificant for the

OSHA-ineffective simulation. A more detailed description of both the
data generating models (programmed in BASIC) and the regression models
(using ordinary least squares in SPSS) is reported elsewhere (McCaffrey,

Andersen, McCold, and Kim: 1984).

Basic Model Test. The ion’ model

the
simulated reality under a wide variety of stochastic conditions. The
relative size of the random error term, E, and the “stability” of the
Markov process that changed safety program's effects, firm size, and
industry type were varied substantially (A perfectly "stable" Markov
process has all unit values on the diagonal of the state transition
matrix-~a value as low as .5 was tested). Under stochastic processes of
varying stability, the regression model detected that OSHA reduced
accidents in the OSHA-effective simulation, and that OSHA did not reduce
accidents in the OSHA-ineffective simlation. Bear in mind, however,
that none of the empirical and theoretical problems discussed in Section
TI above were yet built into the data generating model. Two of these

problems were then introduced into the simulation, one at a time.

- 27

Misspecification of the OSHA ion Variable. As di in

Section II above, numerous empirical and interpretive problems surround
the OSHA inspection variable. In the OSHA-effective simlation earlier,
OSHA inspections always increase the effectiveness of safety programs; a
value of 1 for the inspection variable is a perfect and unique predictor

of systematic improvement. We relaxed this

sumptions slightly by
introducing varying degrees of mismeasurement to the OSHA variable. The

involved "miscoding" the i

variable for certain
varying percentages of the firms, rendering the variable inaccurate for
those firms. The results of this experiment are presented in Figure 1.
The horizontal axis reports the percent of the inspection variables
that have been miscoded in the OSHA-effective simulation. The vertical
axis presents the F statistic for the OSHA inspection variable (Scale 1)
and the predicted effect (B value) of OSHA inspections (Scale 2). The

horizontal dashed line reads on scale 2, and shows the "real" effect of

OSHA derived not by statistical inference but by directly comparing the

and

For no measurement error (i.e., no misspecification of the OSHA

variable) the i

model predicts OSHA's by

roughly a factor of 2. This result is highly statistically significant.

When the i ion variable for in

30% of the i firms

has been mi ified the ion model's icti

the correct magnitude of effect for OSHA inspections.
Both the predicted effect of OSHA inspections and the associated F

statistic decline ically as the mi

At

+= 28 -

roughly 45%

specification, the predicted OSHA effect is not
statistically significant, although OSHA is known to be effective in the
model.

The results suggest that measurement error in the OSHA variable
reduced the predicted effect of OSHA on accident rates, and degraded the
F statistic. Also, an estimate under considerable measurement error
matched the actual simlated effect of OSHA inspections.

OSHA Interact with Reporting Behavior. In a second

experiment, we specified in the simulation that an OSHA inspection
increased the reporting of accidents in the firm. We specified that,
Prior to the inspection, only some fraction of the accidents were

recorded; an OSHA i ion led to ing of all acci In the

real world such an effect is plausible, because inspections could

increase awareness of safety issues, improve recordkeeping lest

violations of reporting be cited by and so

forth.

Figure 2 presents the results of this experiment. The Percent of
Accidents Reported Prior to OSHA Inspection is recorded on the horizontal
axis. The vertical Scale 1 and Scale 2 are analogous to those scales in

Figure 1. Again, the actual

effect of OSHA i ions in
reducing accidents is computed by comparing the OSHA-effective to the

OSHA-ineffective simlation, and is shown by a horizontal dashed line

labelled

‘actual effect." The line labeled “apparent effect" is derived

by comparing the ive to the i ive sii i

This

line represents the pure effect of increases in accident reporting and

- 29 -

194

actual accident reductions due to inspections. Of course, such a curve

could never be computed in a real data analysis situation since the true
effect of OSHA on safety, and its effect through changes in reporting ere
deeply intertwined.

If it is assumed that all accidents were reported prior to OSHA
inspection, the regression model predicts an OSHA effect that it roughly
twice the actual effect. When roughly 96 percent of accidents had been
reported prior to OSHA's inspections, the regression model's estimate
coincides with the actual effect of OSHA on accidents reduction. At
roughly 91% of accidents reported prior to inspection, the regression
model predicts no effect of OSHA in the simulated world where OSHA is
known to have an effect. The OSHA effect is not statistically
significant from roughly 88% to 93% of accidents reported prior to
inspection. For lower prior reporting rates, the model predicts that
OSHA inspections actually increase accitaeice (Again, this is in the
simulation where OSHA is known a priori to make safety programs perfectly
effective).

These results map the statistical implications of a plausible
situation in which OSHA inspections increase accident reporting, and
simultaneously reduce accidents. Of course, the exact ranges of
significance reported here are artifacts of the simulation model being
used. The important point’ is not that prior accident reporting rates of
between 88 and 93% will produce null results nor that a 96% prior

reporting rate produces unb:

ed estimates of OSHA effect. Rather, the

point is that these types of experiments can alert the analyst to a

- 30 -

possibly important effect before detailed and expensive data collection
begins. Also, such experiments can be used to probe the sensitivity of
regression-based results to complications suggested in case studies.
Thus, the simulation models provide a bridge between case studies and
regression studies via synthetic data experiments.

Next Steps. In the experiments reported above, the simulation model
was built so as to conform to the specification most commonly used in

regression studies of OSHA effectiveness. The structure of the model wi

relatively simple, containing only a few lags, non-linearities and

feedback effects. A more ambitious task would be to base the simulation
model directly on the richer and more detailed insights gained through
case studies. Plausible theoretical and empirical complications would be
initially built into the model, rather than tacked on as in the simple
illustrative experiments described above.

Furthermore, the simulation model, should be subjected to more
extensive analysis and sensitivity testing before being used in synthetic
data experiments. For example, the estimated parameters from the
synthetic world should closely resemble in size, magnitude and
significance those. observed from real data sets. Such a full range of
tests were not performed in the results discussed above.

Finally, insights into how such synthetic data experiments should be
performed need to be refined by replicating these experiments in several
other policy areas (McCaffrey, Andersen, McCold, and Kim, 1984). Such a

research ‘program would generate guidelines for using synthetic data

experiments to bridge the gap between case studies and regre:

on studies

2a 195

of the impact of government programs. These experiments, and the

theoretical simulation modelling that precede them, could begin to remedy

problems iated with ical and poorly measured

variables, and excessively simplified causal structures.

~32-

FIGURE 1
Simulated Effect
of Measurement Error
Scale Scale
1 2
100-20
90
80
45 Predicted
70 << effect (B value)
60 (change due to
w“ inspections)
50-104
40
30
F value of

= 25 Predicted effect ~~ *™ NN
10 crwea

‘i of F (p05) ————->

10 20 - 30 40 30

Percent of Inspections Misspecified

NOTE: Scale 1. F value of inspection vas
‘Scale 2-

~33-

FIGURE 2

Simulated Effect of
Reporting Error

Scale Scale
“uo 36
120 3.0)
100 25
F value of
80 20 — predicted effect
60 15
40 1.0
205
oo SE)
\ \ee Pscoes dn teporting rate)
40 woh 7 Grey tas Be apace
1S \e fu effect
-2.0 : :

80 82 84 86 88 90 92 94 96 98 100

Percent of Accidents Reported

Prior to 0.S.H.A. Inspections

NOTE: Scale 1-F value of inspection variable
Scale

2: mean change in accidents due to O.S.H.A. inspections
{accidents per year per firm)

196

yp
Footnotes

1. Often these variables are entered only as "controls" for

environmental effects, and so it might seem unreasonable to expect

specification of their i h

lati ip to the
dependent variable. We do not think it is unreasonable for two

reasons. First, analysts commonly do assert a relationship between a

control variable and the dependent variable; the problem is that

there is usually no basis for the assertion. Second, having only

vague (or no) sense of the meaning of control variables can lead us

Fe overlook how they interact with policy variables. Usually the

descriptive data on the possible associations among control, policy,
and dependent variables are poor. (For example, only sparse
descriptive information on how OSHA inspections filter into firms’

behavior, or how hospitals make investment decisions [Sloan and

Steinwald, 1980: 18) are available). If a alysts had a reasonable

understanding of the relationship of the control variables to the
Policy and dependent variables they could assert knowledgeably that
control variables are theoretically independent of policy variables.

When they have only the slightest, if any, idea of the control

variables’ dynamics they cannot know when control variables are
independent of policy variables. Nor can they just look at

correlations between control and policy var:

bles, for if they do not

have well founded ideas about theoretical connections of control and

le

i

197

-35-

(continued).

policy variables they are not able to interpret the correlations.
(Nonsignificant correlations could reflect offsetting dynamics or
unmeasured interactions as easily as null associations).

An extended discussion of how the theoretical and empirical problems
discussed in this paper apply to hospital certificate of need
regulation, and also hospital rate setting regulation, is found in
McCaffrey, Andersen, McCold and Kim (1984).

However, the i

of dynamic
modelling are bought at a price. Such models are typically beset by

complex

Parameter estimation (Graham, 1980) as
well as overall model validation (Forrester and Senge, 1980). A trade
off emerges between the relative strengths of regression and dynamic

simulation modelling. Take parameter estimation as a case in point.

Simulation modelling uses a of subjecti

data sources both in formulating a rich causal structure as well as

in estimating the i with the st On the

ether hand, because regression analyses are restricted to av:

lable
longitudinal or cross-sectional data sets, they usually have less
richness of causal structure, but many fewer questions concerning

par due to the

of formal tests of

The data

reported

below represent one of the several possible ways that case studie

dynamic simulation models, and regression modele can be combined to

capitalize on the relative strengths of all three approaches.

- 36>

REFERENCES

Andersen, David

1981 "Kalman filter estimates of system states compared to the general
regression probl. International Journal of Policy Analysis and
Information Systems, 5: 95-109 (No. 2).

Arthur Andersen and Company

1979 Cost of Government Regulation: Study for the Business Roundtable
(Three Volumes). New York: The Business Roundtable.

Bacow, Lawrence

1980 Bargaining for Job Safety and Health. Cambridge: MIT Press.

Butler, Richard, and John Worrall

1983 ‘Markers’ compensation: Benefit and injury clains rates in the

" Review of 65: 580-589.

Cambridge Research Reports, Inc.

1978 Public and Worker Attitudes Toward Carcinogens and Cancer Risk.
ridge: Cambridge Reports, Inc.

Forrester, Jay
MIT Press.

1961 Industrial Dynamics. Cambridge:

1980 "Information sources for modeling the national e

onomy." Journal
of the American Statistical Association, 75: 555-566 (September).
Forrester, J., and P. Senge F

1980 Steet for Confidence in System Dynamics Models." Pp. 209-228 in
A.A. Legasto, J.W. Forrester, and J.M. Lynch (Eds.), System
Dynamics, TIMS Studies in the Management Sciences, Vol. 14.

‘Austerdam: North Holland Publishing.

Freedman, Audrey

to Health Risk. New York: The Conference Board.

1981 Industry Respons:

Fuangchen, Somporn, and David McCaffrey

e impact of certificate of need and rate setting regulation on
Paper presented to Fifth Annual Researc!
lation for Public Policy Analysis and ianspenesit®s

1983"
—— technology."
mference of Associa


-37- . - 38+

198
Graham, Alan McCaffrey, David
1980 "Parameter Estimation in System Dynamics Modeling." Pp. 143-161. in l9s2" and 1 Toward explaini
J. Randers (Ed.), Elements of the System Dynamics Method. a discrepancy." Administrative Science Quarterly, 27: 398-419
Cambridge: MIT Press. (September).
Hanushek, Eric 1983 "An assessment of OSHA's recent effects on injury rates."
: Journal of Human Resources, 18: 131-146 (Winter).
1981 "Throwing money at school! Journal of Policy Analysis and
Management, 1: 19-41 (Fall). McCaffrey, D., D. Andersen, P. McCold, D.H. Kim
Howell, Julianne 1984 "Dynamic Systems, Ambiguous Variables, and Discrepant Findings.”
Working Paper, Rockefeller College of Public Affairs and Policy,
1981 Regulating Hospital Capital Investment: The Experience in SUNY-Albany
Massachusetts. NCHSR Research Summary Series, DDHS Publ. No.
(PHS) 81-3298 (March). Washington: U.S. Department of Health Meadows, Donella
and Human Services.
: 1980 "The Unavoidable A Priori." Pp. 23-57 in J. Randers (Ed.), Elements
Johnson, Curtis of the System Dynamics Method. Cambridge; MIT Press.
1980 "Some effects of data errors in econometric models." TIMS Mechanic, David
Studies in the Management Sciences, 14: 143-159.
1979 “Correlates of physician utilization: ny ae major pobetutriere
Joskow, Paul studies of physician utilization find tr: pschos and
organizational effects?” Journal of Health ia Social Behavior,
1981 Controlling Hospital Costs. Cambridge: MIT Press. 20: 387-396.
Kochan, T., L. Dyer, and D. Lipsky Mendeloff, John
1977 The i of Safety Committ! 1979 Regulating Safety. Cambridge: MIT Press.
W.E. Upjohn Institute for Employment Research. °
“Policy Analysis-Urban Systems, Inc.
Linneman, Peter
1981 Eraluation of the Effects of Certificate of Need Programs Three
1980 “The effects of consumer safety standards: The 1973 mattress mes). Washington: U.S. Department of Health and Human
flammability standard." Journal of Law and Economics, 23: setts,
461-479 (October).
Porter, A., T. Connolly, R. Heikes, and C. Park
Luecke, Daniel, and Noel McGinn
1981 "Misleading indicators: the limitations of multiple linear
1975 " ion analysis and i i Can ion in ion of policy Policy
they be trusted?" Harvard Educational Review, 44: 325-350. Sciences, 13: 397-418,

Mass, N. and P. Senge Randers, Jorgen (Ed.)

1980 “Alternative Tests for ing Model Variables." Pp. 203-223 in
J. Randers (Ed.), Elements of the System Dynamics Method.
Cambridge: MIT Pre:

1980 Elements of the System Dynamics Method. Cambridge: MIT Press.
Rutter, H. B. Maugham, P. Mortimére, and J. Ouston

1979 Fifteen Thousand Hours: Secondary Schools and Their Effects on
Children. Cambridge: Harvard University Press.

-39-

Senge, Peter

1977 "Statistical estimation of feedback models." Simulation (June),

pp. 177-184.
Sloan, Frank

1982 "Government and the regulation of hospital care." American
Economic Review Papers and Proceedings, 72: 196-201.

Sloan, Frank, and Bruce Steinwald

1980 Insurance, Regulation, and Hospital Costs. Lexington: Lexington
Books.

th, Robert

1979 "The impact of OSHA inspections on manufacturing injury rates."
The Journal of Human Resources, 14: 145-170 (Spring).

Steinwald, Bruce, and Frank Sloan

1980 "Regulatory approaches to hospital cost containment: A synthesis
of the empirical evidence." Pp. 274-308 in M. Olson (ed.), A New
Approach to the Economics of Health Care. Washington: American
Enterprise Institute.

U.S. Office of Technology Assessment

1981 Policy Implications of the Computed Tomography (CT) Scanner.
Washington: U.S. Office of Technology Assessment.

Viscusi, W. Kip

1979 “The impact of OSHA inspections on manufacturing injury rates.
Bell Journal of Economics, 10: 117-140 (Spring).

Worrall, John, and David Appel

e and benefit utilization in workers’
The Journal of Risk and Insurance,

1982 "The wage replacement
compensation insurance
September: 361-371.

199

Metadata

Resource Type:
Document
Rights:
Date Uploaded:
March 27, 2026

Using these materials

Access:
The archives are open to the public and anyone is welcome to visit and view the collections.
Collection restrictions:
Access to this collection is unrestricted unless otherwide denoted.
Collection terms of access:
https://creativecommons.org/licenses/by/4.0/

Access options

Ask an Archivist

Ask a question or schedule an individualized meeting to discuss archival materials and potential research needs.

Schedule a Visit

Archival materials can be viewed in-person in our reading room. We recommend making an appointment to ensure materials are available when you arrive.