Download as pdf or txt
Download as pdf or txt
You are on page 1of 15

271

Article

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
Elevating the Scientific and Public Policy
Discourse on Crime Prevention: Taking
Stock of the ‘What Works’ Report’s
Influence 20 Years on
Brandon C. Welsh and Andrea B. Wexler

Abstract In 1997, the Office of Justice Programs published Preventing Crime: What Works, What Doesn’t, What’s
Promising (Sherman, L. W., Gottfredson, D. C., MacKenzie, D. L., Eck, J. E., Reuter, P., and Bushway, S. D. (1997).
Preventing Crime: What Works, What Doesn’t, What’s Promising. Washington, DC: Office of Justice Programs). The
report was commissioned by the US Congress and was prepared by a team of criminologists from the University of
Maryland. It aspired to be a methodologically rigorous and comprehensive review of the effectiveness of crime
prevention programmes, ranging from prenatal home visits to community policing to parole. This 20-year review
of the ‘what works’ report finds that it has been influential in elevating both the scientific and public policy discourse
on crime prevention. It did this on three main fronts. First, it reaffirmed that not all evaluation designs are equally
valid and made clear that only designs that provide confidence in observed effects should contribute to the evidence
base. Secondly, it advanced the equally important task of assessing research evidence and, despite some limitations,
adopted a more rigorous method for this purpose. Thirdly, undergirding all of this was the report’s commitment to the
communication of science for the benefit of all parties: policymakers, practitioners, researchers, and the public.
Implications for policy—with special reference to evidence-based policing—and research are discussed.

Introduction rigorous and comprehensive review of crime pre-


It is not an exaggeration to say that the ‘what works’ vention programmes, ranging from prenatal home
report—officially titled Preventing Crime: What visits to community policing to parole. The report
Works, What Doesn’t, What’s Promising (Sherman concluded that there was sufficient scientific evi-
et al., 1997)—was a landmark publication in the dence to establish a provisional list of programme
field of criminology and criminal justice (Gest, modalities that worked, did not work, and were
2001; Greenwood, 2006). Commissioned by the inconclusive but showed promise in reducing
US Congress, it aspired to be a methodologically crime. In many respects, the report represented

School of Criminology and Criminal Justice, Northeastern University, Churchill Hall, 360 Huntington Avenue, Boston, MA
02115, USA. E-mail: b.welsh@northeastern.edu
The authors wish to thank the journal editor and the anonymous reviewer for insightful comments.

Advance Access publication: 6 February 2019


Policing, Volume 13, Number 3, pp. 271–285
doi:10.1093/police/paz006
ß The Author(s) 2019. Published by Oxford University Press. All rights reserved.
For permissions please e-mail: journals.permissions@oup.com
272 Policing Article B. C. Welsh and A. B. Wexler

the beginning of the US interest and the interest of 1998). The aims of the programme were two-fold:

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
many other Western countries, in an evidence- (a) ‘to better inform policymakers on the import-
based approach to crime policy (Fagan and ance of basing crime prevention programs on sci-
Buchanan, 2016; Farrington, 2000). entific evidence’ and (b) ‘to advance the body of
Like other influential policy works before it and knowledge upon which sound crime prevention de-
since, timing had something to do with its influ- cisions should be made’ (Sherman et al., 2006, p. 3).
ence. Published in February 1997, the report fol- Intriguingly, the most consequential immediate
lowed passage of the 1994 Violent Crime Act and influence of the report took place outside of the
became part of the lingering, vitriolic debate on USA. In 1997, the newly elected British government
prevention versus punishment in the USA initiated a comprehensive spending review of all
(Currie, 1994); it appeared as the US crime drop government services, including criminal justice
of the 1990s was in full swing (Blumstein and and other responses to crime. The review by the
Wallman, 2006); and it arrived on the cusp of the Home Office, including its report, titled Reducing
evidence-based movement in the social and beha- Offending (Goldblatt and Lewis, 1998), was mod-
vioural sciences (Chalmers, 2003). elled on the Preventing Crime report—down to how
Whatever role timing may have played, it is what the research was organized and assessed. This
the report said that carried far more weight. At review led directly to the government’s ‘Crime
center stage was the report’s focus on the science Reduction Programme’, a three-year (1999–2002),
of crime prevention and its relevance for public £400 million (approximately $640 million at the
policy, and this translated into an immediate influ- time) initiative. According to Dhiri et al. (2001, p.
ence on several fronts. This began with coverage in 181), it represented the ‘biggest single investment in
the press. For example, the New York Times called an evidence-based approach to crime reduction to
the report ‘[t]he most comprehensive study ever of have ever taken place in any country’.2
crime prevention’ (Butterfield, 1997, p. A20). U.S. As important as some of these events were, they
News & World Report referred to the report as ‘au- were rather short-lived and far from impactful on
thoritative’ and stated that it ‘makes the fervent the scientific and policy understanding of the pre-
ideological battles over anticrime policy seem vention of crime. When viewed over a longer time
beside the point’ (Gest, 1997, p. 38).1 The report frame, however, a much different view of the influ-
was the subject of four Congressional hearings, and ence of the report takes shape. It is the thesis of this
its authors were invited on many occasions to brief article that the ‘what works’ report proved influen-
policymakers and other leaders across the USA and tial in helping to elevate both the scientific and
abroad. It also attracted the attention of charitable public policy discourse on crime prevention. It
foundations and private donors interested in sup- did this on three main fronts.
porting a continuation of the report’s work. These First, the report reaffirmed that not all evaluation
grants and donations were used to establish the designs are equally valid and made clear—in no
Crime Prevention Effectiveness Program at the uncertain terms—that only designs that provide a
University of Maryland’s Department of high degree of confidence in observed effects (i.e.
Criminology and Criminal Justice (Sherman et al., experimental and high-quality quasi-experimental
1
Gest (1997, p. 38) added that, ‘Dud programs are scattered across the spectrum, from summer jobs for youths (a liberal
favorite) to military-style boot camps for lesser offenders (a centrist-conservative favorite).’
2
It is important to note that the Crime Reduction Programme was criticized for failing to deliver on its primary objective of
using a ‘research-driven’ approach to guide policy and practice on what works best to reduce crime (Maguire, 2004). This had
a great deal to do with poor government management and problems with the implementation of programs at the local level
(Nutley and Homel, 2006).
Scientific and public policy discourse on crime prevention Article Policing 273

designs) should contribute to the evidence base. detractors, and the criticisms covered many fronts

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
These views were by no means original (Cook and (Haggerty, 2008; Stafford and Silverlock, 2000). We
Campbell, 1979). What was original was the au- reviewed as many critiques as we could locate, and
thors’ use of an objective and understandable meas- these are discussed throughout the article. Relatedly,
ure, known as the Scientific Methods Scale (SMS), we were also interested in more general critiques of
to rate evaluation designs on overall internal valid- the evidence-based approach (Sampson, 2010;
ity. The bottom line is that evaluation designs need Sampson et al., 2013), and these were consulted.
to be guided by the research question at-hand, not Lastly, we drew upon the growing body of crimino-
the other way around. logical literature as it pertains to the evidence-based
Secondly, for the equally important task of assess- approach. This includes writings on theory, methods
ing the research evidence (i.e. all of the included of experimentation and systematic review, and re-
evaluation studies), the report moved beyond the search utilization (Fagan and Buchanan, 2016;
traditional narrative review and adopted an empir- Mears, 2007, 2010; Petrosino and Lavenberg, 2007;
ical review method. The report also established rules Weisburd et al., 2016; Zane and Welsh, 2018).
for drawing conclusions about what works, what The article is organized around the three afore-
does not work, and what is promising. Crucially, mentioned major areas of influence of the report:
as part of this process, the report called attention (a) rating evaluation designs; (b) assessing research
to the provisional nature of science: scientific con- evidence; and (c) communicating science. We
clusions drawn today can be altered—even up- begin with some background on the report.
ended—with the results of new studies over time.
Third, and undergirding all of this, was the re-
port’s commitment to the communication of sci- Background: the ‘what works’
ence for the benefit of all parties: policymakers,
report
practitioners, researchers, and the public. As noted
by Shonkoff and Bales (2011, p. 17), ‘There is a In documenting the beginnings of the ‘what works’
widespread assumption that science has a role to report, one is struck by the short yet intense period
play in advising policymakers on crafting responses of time leading up to its public release in February
to complex social problems . . . There are also many 1997. The starting point was in April 1996, when
countervailing forces that impede the fulfillment of the 104th US Congress enacted a conference report
that role.’ The report’s efforts to engage in commu- that required the Attorney General to provide
nicating science took many different forms, and this Congress with an independent review of the effect-
may very well be one of its most important legacies. iveness of state and local crime prevention pro-
For example, elements of this have been central to grammes funded by the federal justice
the advancement of evidence-based policing (Lum department. To help achieve these ends, the
and Koper, 2017; Sherman, 2013). review was required to use ‘rigorous and scientific-
In an effort to review the potential influence of ally recognized standards and methodologies’ (U.S.
the ‘what works’ report, we consulted a wide range House of Representatives, 1996, p. 103).
of published documentation, including peer-re- At the direction of the Assistant Attorney
viewed journal articles, essays, books, and reports. General for the Office of Justice Programs, the
This began with the original report (Sherman et al., National Institute of Justice (NIJ) was tasked with
1997; see also Sherman et al., 1998) and subsequent issuing a ‘competitive solicitation for proposals’.
related works ( MacKenzie, 2000; Sherman et al., This was done in coordination with several other
2002, 2006). It is important to note that the report federal agencies, including the Bureau of Justice
and subsequent related works were not without Assistance, the Office of Juvenile Justice and
274 Policing Article B. C. Welsh and A. B. Wexler

Delinquency Prevention, and the Executive Office the full range of techniques or measures that have

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
of Weed and Seed (Sherman, 1997a, chap. 1, p. 4). been used to try to prevent crime—from prenatal
In June 1996, the NIJ released the solicitation, and home visits to community policing to parole.
within short order, a panel of experts appointed by Herein, the authors argued that ‘[t]he value of a
the agency began the process of evaluating pro- broad framework for analyzing crime prevention
posals. By August 1996, the panel had completed policies is its focus on the whole forest rather than
its review of proposals and selected the one sub- on each tree’ (Sherman, 1997b, chap. 2, p. 4).
mitted by a team of six scholars from the
University of Maryland’s Department of Classification
Criminology and Criminal Justice.3 The deadline Rather than draw upon one of the more traditional
for completing the report was 27 January 1997. taxonomies that have been used to classify or or-
With only 6 months to complete the report, the ganize the wide array of crime prevention pro-
University of Maryland team set in motion a rather grammes ( Brantingham and Faust, 1976; Tonry
Herculean task, which included reviewing more and Farrington, 1995; van Dijk and de Waard,
than 500 evaluations of crime prevention pro- 1991), the authors elected to focus on the institu-
grammes and producing a 565-page report. In add- tional setting or domain in which programmes take
ition to the six scholars, the project was aided by at place. Altogether, seven major institutional settings
least 22 graduate students and a scientific advisory were articulated: families; schools; communities;
panel of distinguished scholars from criminology labor markets; places (e.g. urban centers, homes);
and affiliated fields. police agencies; and courts and corrections. In
The scope of the report was defined as such: ‘a acknowledging that the definitions of these settings
critical assessment, based on a growing body of sci- were quite broad and that there was some overlap
ence, of the effectiveness of a wide range of crime across some settings, the authors seemed content in
prevention strategies, operated at the local level, finding that ‘much of the crime prevention litera-
with and without the support of Federal funds’ ture fits quite neatly into [these] seven major insti-
(Sherman, 1997a, chap. 1, p. 5). Within this tutional settings’ (Sherman, 1997b, chap. 2, p. 8).
scope, four key features served as the foundation
of the report and guided its conclusions: definition Scientific basis
of crime prevention; classification of crime preven- Perhaps the most important feature of the report is
tion; science of crime prevention; and methods for that it was firmly anchored in the science of pro-
searching for studies and assessing effectiveness. An gramme evaluation. Important to this is that there
overview of each of these features follows. are ‘widely agreed-upon rules for assessing the level
of certainty that a conclusion in any one test is
Definition correct’ (Sherman et al., 1998, p. 3). These rules
For the purposes of the report, crime prevention was pertain to the methodological quality of the evalu-
defined in its widest sense by its outcome: the pre- ation design, which depends on three main criteria:
vention of a future criminal act. By not considering statistical conclusion validity; internal validity; and
the character or function of prevention (Foxcroft, construct validity (Cook and Campbell, 1979;
2014), this definition allowed for the inclusion of Shadish et al., 2002).4 ‘Validity refers to the
3
The team included all six authors of the eventual report: Lawrence Sherman, Denise Gottfredson, Doris MacKenzie, John
Eck, Peter Reuter, and Shawn Bushway.
4
External validity is often included as a fourth criterion. In the present context of the validity of a single study, it is not
possible to make an assessment of the generalizability of the study’s observed effects. This requires at least a second test or
replication of the study.
Scientific and public policy discourse on crime prevention Article Policing 275

correctness of inferences about cause and effect’ for conclusions to be made about what works, what

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
(Shadish et al., 2002, p. 34). All things being does not work, and what is promising5 in the seven
equal, designs with higher validity provide more settings, there was an effort to make the approach
confidence in the observed effects. In recognition somewhat robust. At its core, the authors relied on
that not all evaluations are equally valid and want- the vote-count review method, which tallies-up the
ing to include only the highest quality designs, the ‘number of studies with statistically significant
authors developed an instrument to rate evaluation findings in favor of the hypothesis and the
designs: the SMS. (The SMS and validity are dis- number contrary to the hypothesis’ (Wilson,
cussed in more detail in the section on rating evalu- 2001, p. 73). Two elements helped to improve
ation designs.) upon the vote-count review: (a) it integrated the
SMS and (b) a clear set of rules was established
Methodology for the different conclusions (i.e. what works and
The research methods used in the report relate to so on). For conclusions about what works and what
two key functions: searching for studies and assess- does not work, this had the effect of providing some
ing the research evidence. A fair amount is known measure of external validity. (These rules are dis-
about the process that was used to identify studies. cussed in the section on assessing research
This began with consulting recent secondary re- evidence.)
views. If the authors determined the secondary
review to be reliable, it was used to make a ‘deter-
mination of the strength of the evidence about the
category [or program type]’ (Sherman and Rating evaluation designs
Gottfredson, 1997, appendix, p. 3). If no secondary One of the most noteworthy features of the ‘what
review was available, a ‘primary evidence analysis’ works’ report was its use of a scale to rate the meth-
was carried out. This involved an effort to identify odological quality of each study’s evaluation design.
‘every evaluation it [the chapter research team] can Known as the SMS, the authors set out to devise a
locate from all sources’ (Sherman and Gottfredson, tool that would (a) provide a sound measure of
1997, appendix, p. 3). Much less is known about the overall internal validity and (b) easily be commu-
search techniques that were used to identify studies nicated to stakeholders. The scale is as follows, ran-
as part of the primary evidence analysis. For ex- ging from level 1 (weakest on overall internal
ample, were electronic bibliographic databases validity) to level 5 (highest):
used? Were the references of eligible studies used
to identify additional studies? Was forward Level 1: Correlation evidence: low of-
searching used with eligible studies—a way to iden- fending correlates with the intervention.
tify more recent follow-ups? These are some of the Level 2: Nonequivalent control group
search techniques that are required in the conduct or one-group pre-post design: treat-
of systematic reviews (Petrosino and Lavenberg, ment group compared with nonequi-
2007). What is known is that the search for studies valent control group; treatment group
was limited mostly to the published literature, measured before and after the interven-
which raises the additional concern of publication tion (with no control group).
bias (Wilson, 2009). Level 3: Equivalent control group
On the matter of assessing the research evidence design: treatment group compared
(i.e. the more than 500 studies collected) to allow with comparable control group,

5
A fourth category, what is unknown, was also discussed, but it received far less systematic attention.
276 Policing Article B. C. Welsh and A. B. Wexler

including pre-post and treatment-con- were in use in the medical and social and beha-

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
trol comparisons. vioural sciences, which relied on a summation of
Level 4: Control of extraneous vari- scores (e.g. 0–100) on a number of specific criteria
ables: treatment group compared with (Brounstein et al., 1997; Juni et al., 1999; Moher
control group, with control of extrane- et al., 1995). As noted by Farrington et al. (2002,
ous influences on the outcome (e.g., by p. 15), one of the concerns with these types of scales
matching, prediction or propensity is that ‘it is hard to know what meaning to attach to
scores, or statistical controls). any score, and the same score can be achieved in
Level 5: Randomized experiment: units many different ways’.
assigned at random to treatment and The SMS has been the subject of some criticism.
control groups. For example, Hope (2005), focusing only on the
With respect to overall internal validity, the SMS chapter on community crime prevention, argues
is heavily influenced by the pioneering theoretical that the ‘research designs that score highly on the
and empirical work on experimentation by Donald SMS are likely to suffer from Type II Error—that is,
Campbell and Thomas Cook, especially their classic they are likely to be interpreted as inferring errone-
book (Cook and Campbell, 1979; see also Campbell ously that treatment effects do not exist when in
and Stanley, 1966; Shadish et al., 2002). As one ex- fact they do’ (p. 276, emphasis in original). The
ample, level 3 is considered to be the minimum core of his argument is that key threats to internal
interpretable design by Cook and Campbell validity (i.e. selection, history), of which the experi-
(1979). This is because it rules out many threats mental method attempts to control, are the ‘social,
to internal validity, including history, maturation/ collective “mechanisms” that they [the community
trends, instrumentation, testing, and mortality. It interventions] activate in order to bring about out-
was also regarded by the authors of the ‘what works’ comes’ (Hope, 2005, p. 276). A rebuttal to this view
report as the minimum design that was adequate could take on any number of issues, including what
for drawing conclusions about what works and appears to be a dismissal of sophisticated experi-
what does not work. mental and quasi-experimental designs to account
While the SMS focused exclusively on internal for some of the unique challenges confronting the
validity, it is important to note that all evaluation evaluation of area-based studies (e.g. mixed units of
projects were also rated on statistical conclusion analysis, small sample size) and oversight to evi-
validity and on construct validity. Specifically, the dence of an indirect relationship between research
following four aspects of each study were rated: ‘For design and study outcomes in criminological stu-
statistical conclusion validity: (a) Was the statistical dies (Weisburd et al., 2001; Welsh et al., 2011).
analysis appropriate? (b) Did the study have low Another criticism of the SMS was directed at the
statistical power to detect effects because of small threshold—the minimum research design (i.e. level
samples? (c) Was there a low response rate or dif- 3)—that was used to include studies in the report.
ferential attrition? For construct validity: (d) What Hough (2010) contends that it is too generic and
was the reliability and validity of measurement of should be adjusted in consideration of the body of
the outcome?’ (Farrington et al., 2002, pp. 17–18). work being reviewed: ‘The threshold for inclusion
The authors also opted for this simple five-point of studies should be set individually for each review,
scale to facilitate ease of communication. It was in light of available evidence.’ No mention is made
intended that each point on the scale should be if this pertains to the level of methodological rigor
understandable. This was in sharp contrast to of the studies under review, the quantity of the
many other scales of methodological quality that studies, or both.
Scientific and public policy discourse on crime prevention Article Policing 277

In the context of reviews of the literature, includ- level 3 as the threshold of methodological quality,

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
ing systematic reviews carried out under the aus- Sherman and his colleagues took a ‘middle road
pices of the Campbell Collaboration, Farrington between reaching very few conclusions with great
and Petrosino (2001, p. 42) note that the criterion certainty and reaching many conclusions with little
of methodological quality that is used for including certainty’ (MacKenzie, 2000, p. 461).
studies is likely the ‘most important and controver-
sial’. It is important because it speaks to the degree
of confidence that one can have in reported effects Assessing research evidence
of individual studies and conclusions drawn from Just as it is crucial to use the highest quality evalu-
the aggregation of those studies. It is controversial ation designs to investigate the effects of crime pre-
because setting the bar too high may end up exclud- vention programmes, it is also important that the
ing all or many studies on a topic, leaving nothing most rigorous methods be used to assess the avail-
to say and giving the impression that the studies are able research evidence. There are several different
flawed (Farrington and Petrosino, 2001). This is types of review methodology, with the main ones
especially problematic for area-based studies in being narrative, vote-count, and systematic/meta-
general, which are less likely to use designs that analytic. The systematic review, which often in-
equate to levels 4 and 5 on the SMS (Farrington cludes a meta-analysis, is considered to be the
and Welsh, 2006). most exacting method and is recommended above
Even more problematic in using what amounts all others (Petrosino and Lavenberg, 2007;
to a rigid threshold approach (e.g. only including Petticrew and Roberts, 2006). As noted by
randomized experiments) is that it dismisses the Petrosino et al. (2001, p. 2), ‘the foremost advan-
strength of the causal inference that can be achieved tage of systematic reviews is that when done well
from a wide range of other rigorous designs (i.e. and with full integrity, they provide the most reli-
levels 3 and 4 on the SMS), whether it be for indi- able and comprehensive statement about what
vidual- or area-based studies (Nagin and Weisburd, works’. In short, the systematic review uses rigorous
2013). In the only known empirical study on the methods for locating, appraising, and synthesizing
subject, Stockard and Wood (2017) found that use evidence from prior evaluation studies, and it is
of a rigid ‘threshold approach’ (i.e. only including exhaustive in its coverage of the literature (pub-
randomized experiments), when compared with an lished and unpublished).
‘inclusive approach’ (i.e. in accordance with the The ‘what works’ report used the vote-count
work of Campbell and Cook), resulted in the exclu- method, not the systematic review method, to
sion of a large proportion of identified studies, make an assessment of the research evidence. The
leading to the concern that ‘policy recommenda- main reason was the relatively short time frame that
tions [would be] based on only part of the available was available to carry out the review. The vote-
literature’ (p. 472). A further analysis showed that count review method falls short of the systematic
effect sizes from the inclusive approach were ‘more review on a number of grounds, but in the present
precise and closer to those of other reviews’ circumstances, the most pressing matter had to do
(Stockard and Wood, 2017, p. 471). with the limited depth in the coverage of the litera-
In the case of the ‘what works’ report, there was a ture. Sherman and Gottfredson (1997, appendix,
clear trade-off between the level of certainty in the p. 1) capture this point with a view to the larger
answers that can be given about programme effects mission of the project: ‘Our approach strikes a
and the level of useful information that can be compromise between breadth and depth, without
gleaned from the available science. By adopting any compromise in scientific integrity.’
278 Policing Article B. C. Welsh and A. B. Wexler

Importantly, the vote-count review offers some Critiques of the report’s efforts to assess the

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
advantages over the traditional narrative review, research evidence center on methodological limita-
the latter being the most commonly used method tions of the vote-count review and concerns
at the time and the method the authors were about the generalizability of the conclusions.
attempting to improve upon. For instance, the Perhaps the most important methodological limi-
vote-count incorporates a quantitative element by tation of the vote-count review has to do with its
considering statistical significance. There is also focus on statistical significance. The main problem
some effort to overcome researcher bias. This with using statistical significance is that it depends
can include the adoption of a methodology for partly on sample size and partly on effect size. For
searching for studies, which can help mitigate example, a significant result may reflect a small
self-selection. Furthermore, as noted earlier, two effect in a large sample or a large effect in a small
elements helped to improve upon the vote-count sample. As noted by Wilson (2001, p. 73), at the
method. One was the integration of the SMS.6 The root here is that ‘[l]arger studies, all else being
other was the establishment of a set of rules for equal, provide more precise estimates of the rela-
drawing conclusions about the research evidence, tionship of interest and thus should be given greater
with all programme types classified into one of four weight in a review.’
categories: what works, what does not work, what is With respect to concerns about the generalizabil-
promising, and what is unknown.7 ity of the conclusions, Tilley (2000, p. 124) takes
The end product was a list showing that 15 pro- issue with the seemingly infinite conditions for pro-
gramme types were effective, 23 programme types gramme effectiveness (‘population demographics,
were ineffective, 30 programme types were promis- program elements, and social context’), leading
ing, and as many as 68 programme types were of him to conclude that the four categories used to
unknown effectiveness (Sherman et al., 1997, classify programme types are ‘highly misleading’.
1998). Importantly, the list was considered provi- He goes on to say that they should be replaced
sional, which was, in the words of the authors, ‘just with more accurate renderings: ‘What has been
as all scientific knowledge is provisional’ (Sherman found to work somewhere’ (for what works);
et al., 1998, p. 3). This acknowledgement stands out ‘What has been found not to work somewhere’
as one of the key contributions of the report. (for what does not work); ‘What seems to have
Clearly, providing a list of programme types that worked somewhere’ (for what is promising); and
work, for example, is compelling and readers want ‘What may work somewhere’ (for what is un-
to know this information. But the list itself is far less known) (Tilley, 2000, p. 124). Stafford and
important than all of the considerations—the Silverlock (2000, p. 92) take this criticism one
process—that went into deriving the list. step further, arguing that the report’s ‘findings are

6
This had the effect of addressing a serious limitation: equal weight given to all studies irrespective of methodological quality.
7
What works: These are program types that prevent crime. Program types coded as working must have at least two level-3 to
level-5 evaluations with tests of significance showing effectiveness and the preponderance of all available evidence supporting
the same conclusion. What does not work: These are program types that fail to prevent crime. Program types coded as not
working must have at least two level-3 to level-5 evaluations with tests of significance showing ineffectiveness and the
preponderance of all available evidence supporting the same conclusion. What is promising: These are program types
where the level of certainty from available evidence is too low to support generalizable conclusions but where there is
some empirical basis for predicting that further research could support such conclusions. Program types were coded as
promising if they were found to be effective in significance tests in one level-3 to level-5 evaluation and in the preponderance
of the remaining evidence. What is unknown: Any program type not classified in one of the other three categories is defined as
having unknown effects.
Scientific and public policy discourse on crime prevention Article Policing 279

of limited use to the practitioner’. This is because Sciences, Engineering, and Medicine, 2017, pp. 1–

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
there is no recipe for ‘how’ to replicate the findings. 2). In their effort to communicate the science of
The authors were deliberate in acknowledging crime prevention to policymakers, practitioners,
concerns about the generalizability of the report’s researchers, and the public, the authors of the
conclusions for what works and what does not ‘what works’ report adopted two structured
work: approaches.8 The first, known as ‘aggregation and
translation’, involves ‘one-way communication’
The weakest aspect of this classification
(National Academies of Sciences, Engineering,
system is that there is no standard
and Medicine, 2017, p. 40). This took the form of
means for determining external valid-
ity: exactly what variations in program the report itself, a summary of the report published
content and setting might affect the by the NIJ (Sherman et al., 1998), a government
generalizability of findings from exist- website from which the report could be down-
ing evaluations. In the current state of loaded, and numerous Congressional hearings
science, that can be accomplished only and meetings. Admittedly rather straightforward,
by the accumulation of many tests in this approach was about getting the information
many settings with all major variations out there and in multiple understandable formats
on the program theme. None of the for the different consumers.
programs reviewed for this report A second approach used by the authors of the
have accumulated such a body of report to communicate science is referred to as
knowledge so far. (Sherman et al., ‘brokering’. As described by the National
1998, p. 6) Academies of Sciences, Engineering, and Medicine
(2017, p. 41), a ‘broker in this context is an inter-
With this health warning in mind, and no appar- mediary (an organization or a person) that bridges
ent solution available, it does raise questions about science and policy making by providing informa-
the public policy utility of the lists; in addition, it tion while developing relationships within net-
brings us back to the provisional nature of the lists. works and linking those who produce knowledge
Indeed, it would take the findings of systematic re- with those who may use it’. It seems safe to say that
views of the Campbell Collaboration and others this approach was a product of the widespread
over the next two decades to bring clarity to some interest in the report, something that happened in
of the conclusions reached by the report (Weisburd the immediate aftermath of its release. Initially, this
et al., 2016). took the form of the Crime Prevention
Effectiveness Program, which was established at
the authors’ university. While the original concep-
Communicating science tion of the programme was never truly realized,
The communication of science has been defined as part of its mission—to advance an evidence-based
the ‘exchange of information and viewpoints about approach to preventing crime—continued for
science to achieve a goal or objective such as foster- many years after in the form of an annual meeting
ing greater understanding of science and scientific that brought together researchers and policy-
methods or gaining greater insight into diverse makers. Known as the Jerry Lee Crime Prevention
public views and concerns about the science related Symposium, with financial backing from the phil-
to a contentious issue’ (National Academies of anthropist Jerry Lee, the annual meeting initiated
8
National Academies of Sciences, Engineering, and Medicine (2017, pp. 40–43) delineates three main approaches to science
communication in a policy-making context: (a) aggregation and translation, (b) brokering, and (c) partnerships.
280 Policing Article B. C. Welsh and A. B. Wexler

academic and policy research and policy briefings communication of science was about much more

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
with elected officials and legislative staff on Capitol than generating lists of what works, what does not
Hill. Also, from the first meeting (in April 2001), it work, and so on. In the context of programme
established a close partnership with the newly evaluation, the authors saw it as vital to convey
formed Campbell Collaboration, the leading inter- the provisional nature of science; that is, knowledge
national organization dedicated to the advance- is ‘always becoming more refined, and therefore no
ment of evidence-based social policy (Farrington conclusion is permanent’ (Sherman et al., 1998, p.
and Welsh, 2001). 3). While the scientific community understands
Beyond the approaches used, the communica- this principle, it is often overlooked in reviews of
tion of science was at the heart of the mission of the literature, whereby, conclusions can be touted
the project; specifically, it was central to the core as final. More importantly, this principle is not
components of the report: rating evaluation designs well-understood outside of the scientific commu-
and assessing research evidence. For the rating of nity and can be a cause for concern with science in
evaluation designs—through the use of the SMS— general. To these latter points, the words of
it was imperative that the scale be understandable Bromme and Beelmann (2018, p. 351) are instruct-
to target audiences outside of academia. As was ive: ‘The provisional nature of scientific findings is
later noted, ‘the SMS can be criticized, but it has not an exception or fault within the progress of
the virtue of simplicity. It can be improved, but at science. Revising the present state of knowledge
the cost of simplicity’ (Farrington et al., 2002, p. is . . . the everyday business of researchers.’
21). In short, the key was not to dilute the central Moreover, in explaining the provisional nature
message: ‘What is essential is to use some kind of of science, it was done neither to temper the fact
methodological quality scale in order to communi- that the vote-count review had some deficiencies
cate to scholars, policymakers, and practitioners nor that there were some concerns about external
that not all research is of the same quality, and validity. Instead, the authors were clear-eyed and
that more weight should be given to higher quality transparent about the provisional nature of science,
evaluation studies’ (Farrington et al., 2002, p. 21). noting in the 1998 summary of the report: ‘All con-
The report’s effort to communicate scientific clusions reported in this Research in Brief reflect
concepts in a straightforward and understandable the state of scientific knowledge as of late 1996.
way to the public and other stakeholders is by no By the time this Research in Brief is published [in
means unique. Other projects, which may take the July 1998], new research results may be available
form of reviews or larger platforms, continuously that would modify the conclusions’ (Sherman
grapple with striking a balance between communi- et al., 1998, p. 3).
cating scientific information in an accurate and
simple way. For example, Harvard University’s
Center on the Developing Child’s long-term project
to communicate the science of child development
Discussion and conclusions
to policymakers and the public is guided by the It has been 20 years since the publication of the
following principle: ‘science can be served credibly ‘what works’ report by Sherman et al. (1997). The
and the public can be better informed without ser- report aspired to be a methodologically rigorous
ious distortions of complex concepts even as they and comprehensive review of the effectiveness of
are shortened and translated into lay terms’ crime prevention programmes, and it had an im-
(Shonkoff and Bales, 2011, pp. 17–18). mediate influence in several areas. In the interven-
With respect to the assessing research evidence ing years since its release, the report’s relevance has
component of the ‘what works’ report, here the not appeared to wane. For example, the SMS
Scientific and public policy discourse on crime prevention Article Policing 281

continues to receive widespread application by (National Research Council, 2005). In fact, this

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
scholars and organizations (e.g. Campbell view flies in the face of the contribution to policy
Collaboration, Washington State Institute for made by high-quality quasi-experimental methods
Public Policy) engaged in research contributing to (Nagin and Weisburd, 2013).
evidence-based policy. Also, as of August 2018, the High-quality designs to evaluate programme ef-
Web of Science shows that the report (along with fects are also crucial for informing decisions about
the 1998 summary) has been cited more than spending on crime prevention. The reason is that an
12,000 times in over 9,000 different published economic analysis (i.e. the technique used to meas-
works. Furthermore, the ‘what works’ report has ure economic efficiency) is an extension of an out-
been recognized as the beginning of interest in an come evaluation and is only as defensible as the
evidence-based approach to crime policy in the evaluation on which it is based. It has long been
USA and abroad. recommended that economic analyses be limited
This article set out to take stock of the ‘what to programmes that have been evaluated with an
works’ report’s potential influence over the last “experimental or strong quasi-experimental de-
two decades. On the basis of a review of a wide sign” (Weimer and Friedman, 1979, p. 264).
range of documentation on the report and in Increasingly, this is becoming a fixture in economic
related subject areas, we argue that the report analyses with a direct bearing on public policy. For
proved influential in elevating both the scientific example, the use of high-quality designs is a criter-
and public policy discourse on crime prevention ion of the Washington State Institute for Public
and that it did this on three main fronts. First, it Policy’s benefit-cost modelling work (Lee et al.,
reaffirmed that not all evaluation designs are 2012), one of the most rigorous and consequential
equally valid and made clear that only designs evidence-based policy initiatives of the last 15 years.
that provide confidence in observed effects should Another issue has to do with the need for greater
contribute to the evidence base. Secondly, it attention to challenges facing programme imple-
advanced the equally important task of assessing mentation (Fagan, 2017; Fixsen et al., 2009; Telep
research evidence and, despite some limitations, and Weisburd, 2016). The importance of imple-
adopted a more rigorous method for this purpose. mentation to the evidence-based model is best cap-
Thirdly, undergirding all of this was the report’s tured by the following: ‘Evidence-based policing
commitment to the communication of science for assumes that experiments alone are not enough.
the benefit of all parties: policymakers, practi- Putting research into practice requires just as
tioners, researchers, and the public. much attention to implementation as it does to
Each of these three points plays a central role in controlled evaluations’ (Sherman, 1998, p. 7).
current efforts in advancing an evidence-based ap- On the matter of assessing research evidence, it is
proach to crime policy, and it is in this context that important to note that the ‘what works’ report did
we explore implications for policy and research. On not in any way influence the wave of systematic
the matter of advocating for the use of the highest reviews of the literature that would shortly follow
quality designs to evaluate programme effects, sev- under the auspices of the Campbell Collaboration
eral key issues warrant attention. One is the or those conducted by other researchers. But it can
acknowledgement that this has nothing to do be argued that the report demonstrated the need for
with only using randomized controlled experi- a narrower focus and greater rigor in the reviewing
ments. The ‘what works’ report and, more broadly, process. Despite all of the efforts to improve the
the body of work by Campbell, Cook, and other vote-count method (i.e. by integrating the SMS
leading experimentalists in the social and beha- and establishing a set of rules for the conclusions),
vioural sciences say nothing to this effect it stands out as the weakest component in an
282 Policing Article B. C. Welsh and A. B. Wexler

otherwise rigorous report. We think the main mes- (blueprintsprograms.com; formerly Blueprints for

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
sage here is that future reviews of the literature need Violence Prevention), was started around the same
to adopt the methodology of systematic reviews. In time of the release of the ‘what works’ report.
the case of intervention research, this includes a Another registry, Crime Solutions (crimesolu-
much narrower focus. Long gone are the days of tions.gov/), was started in 2011 and draws heavily
sweeping reviews of what works, unless they are on the report. A recent critical review of these and
founded on systematic reviews (Weisburd et al., other registries finds that there is large-scale vari-
2016). ability across registries, with special concerns about
There is also a need for systematic reviews to be discrepancies in the criteria used to identify what
updated in a timely manner. The Campbell works and in the lack of agreement on the pro-
Collaboration, for example, recommends that au- grammes that get selected for dissemination
thors update reviews every 5 years. In addition to (Fagan and Buchanan, 2016). The review authors
attending to the provisional nature of science, this call for a number of changes to improve on this
has much to do with making reviews relevant to the state of affairs, and these changes should be heeded.
times. If findings of systematic reviews are to have At the same time, there needs to be a programme
any influence on policy, they need to report on the of research dedicated to the science of communi-
latest research. cating the science of crime prevention. As noted by
When it comes to the communication of science, Shonkoff and Bales (2011, p. 30), ‘knowledge trans-
the ‘what works’ report seems to have been ahead of fer from research to policy and service delivery’ is a
its time or, more correctly, the authors were well complex process that merits study in its own right.
aware of the necessity to convey scientific informa- Other fields in the social, behavioural, and natural
tion in a way that is understandable to all the key sciences have embarked on this task (National
stakeholders. It is clear that the report was not writ- Academies of Sciences, Engineering, and
ten to simply publicize the research findings; it was Medicine, 2017; Shonkoff and Bales, 2011). Fagan
also written to communicate the science of prevent- and Buchanan’s (2016) critical review is an import-
ing crime in order to foster more effective policy ant step for the prevention of crime, and we think it
and practice moving forward. is high time for criminology and criminal justice
The importance of this approach has given rise to more broadly to follow suit.
several key developments. In the context of evi-
dence-based policing, this has helped to foster
increased receptivity of police officers to this poli-
References
cing philosophy (Telep, 2017; Telep and Lum,
Blumstein, A. and Wallman, J. (eds). (2006). The Crime
2014). It has also fostered the establishment of nu-
Drop in America, Rev. edn. New York: Cambridge
merous professional societies dedicated to the ad- University Press.
vancement of evidence-based policing, both on an Brantingham, P. J. and Faust, F. L. (1976). A Conceptual
international (Society of Evidence-Based Policing) Model of Crime Prevention’. Crime & Delinquency 22(3):
and national (Australia/New Zealand, Canada, and 284–296.
Bromme, R. and Beelmann, A. (2018). Transfer Entails
USA) stage (Martin and Mazerolle, 2015).
Communication: The Public Understanding of (Social)
Another key development in communicating the Science as a Stage and a Play for Implementing
science of preventing crime has been crime preven- Evidence-Based Prevention Knowledge and Programs’.
tion registries. These registries are designed to as- Prevention Science 19(3): 347–357.
Brounstein, P. J., Emshoff, J. G., Hill, G. A., and Stoil, M. J.
semble, analyse, and promote effective crime
(1997). ‘Assessment of Methodological Practices in the
prevention programmes. One of the registries, Evaluation of Alcohol and Other Drug (AOD) Abuse
Blueprints for Healthy Youth Development Prevention’. Journal of Health and Social Policy 9(2): 1–9.
Scientific and public policy discourse on crime prevention Article Policing 283

Butterfield, F. (1997). Most Efforts to Stop Crime Fall Far Gest, T. (1997). A Taxpayer’s Guide to Crime and

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
Short, Study Finds. New York Times, 16 April 1997, p. Punishment. U.S. News & World Report, 21 April 1997,
A20. p. 38.
Campbell, D. T. and Stanley, J. C. (1966). Experimental and Gest, T. (2001). Crime and Politics: Big Government’s Erratic
Quasi-Experimental Designs for Research. Chicago: Rand Campaign for Law and Order. New York: Oxford
McNally. University Press.
Chalmers, I. (2003). ‘Trying to Do More Good Than Harm Goldblatt, P and Lewis, C. (eds). (1998). Reducing Offending:
in Policy and Practice: The Role of Rigorous, Transparent, An Assessment of Research Evidence on Ways of Dealing
Up-To-Date Evaluations’. Annals of the American with Offending Behaviour. London: Home Office.
Academy of Political and Social Science 589: 22–40. Greenwood, P. W. (2006). Changing Lives: Delinquency
Cook, T. D. and Campbell, D. T. (1979). Quasi- Prevention as Crime-Control Policy. Chicago: University
Experimentation: Design and Analysis Issues for Field of Chicago Press.
Settings. Chicago: Rand McNally. Haggerty, K. D. (2008). ‘Book Review of Evidence-Based
Currie, E. (1994). ‘What’s Wrong with the Crime Bill’. The Crime Prevention’. Theoretical Criminology 12(1): 116–
Nation 258(4): 118–121. 121.
Dhiri, S., Goldblatt, P., Brand, S., and Price, R. (2001). Hope, T. (2005). ‘Pretend It Doesn’t Work: The “Anti-
Evaluation of the United Kingdom’s ‘Crime Reduction Social” Bias in the Maryland Scientific Methods Scale’.
Programme’: Analysis of Costs and Benefits. In Welsh, European Journal on Criminal Policy and Research
B. C., Farrington, D. P., and Sherman, L. W. (eds), 11(3–4): 275–296.
Costs and Benefits of Preventing Crime. Boulder, CO: Hough, M. (2010). ‘Gold Standard or Fool’s Gold? The
Westview Press, pp. 179–201. Pursuit of Certainty in Experimental Criminology’.
Fagan, A. A. (2017). ‘Illuminating the Black Box of Criminology & Criminal Justice 10: 11–22.
Implementation in Crime Prevention’. Criminology & Juni, P., Witschi, A., Bloch, R., and Egger, M. (1999). ‘The
Public Policy 16(2): 451–455. Hazards of Scoring the Quality of Clinical Trials for Meta-
Fagan, A. A. and Buchanan, M. (2016). What Works in Analysis’. Journal of the American Medical Association 282:
Crime Prevention? Comparison and Critical Review of 1054–1060.
Three Crime Prevention Registries’. Criminology & Lee, S., Aos, S., Drake, E. K. et al. (2012). Return on
Public Policy 15(3): 617–649. Investment: Evidence-Based Options to Improve Statewide
Farrington, D. P. (2000). ‘Explaining and Preventing Outcomes. Olympia, WA: Washington State Institute for
Crime: The Globalization of Knowledge’. Criminology Public Policy.
38(1): 1–24. Lum, C. M. and Koper, C. S. (2017). Evidence-Based
Farrington, D. P., Gottfredson, D. C., Sherman, L. W., and Policing: Translating Research into Practice. New York:
Welsh, B. C. (2002). The Maryland Scientific Methods Oxford University Press.
Scale. In Sherman, L. W., Farrington, D. P., Welsh B. MacKenzie, D. L. (2000). ‘Evidence-Based Corrections:
C., and MacKenzie, D. L. (eds), Evidence-Based Crime Identifying What Works’. Crime & Delinquency 46:
Prevention. New York: Routledge. 457–471.
Farrington, D. P. and Petrosino, A. (2001). ‘The Campbell Maguire, M. (2004). ‘The Crime Reduction Programme in
Collaboration Crime and Justice Group’. Annals of the England and Wales: Reflections on the Vision and the
American Academy of Political and Social Science 578: Reality’. Criminal Justice 4(3): 213–237.
35–49. Martin, P. and Mazerolle, L. (2015). ‘Police Leadership in
Farrington, D. P. and Welsh, B. C. (eds). (2001). ‘What Fostering Evidence-based Agency Reform’. Policing 10:
Works in Preventing Crime? Systematic Reviews of 34–43.
Experimental and Quasi-Experimental Research’. Annals Mears, D. P. (2007). ‘Towards Rational and Evidence-Based
of the American Academy of Political and Social Science Crime Policy’. Journal of Criminal Justice 35(6): 667–682.
578: 1–173. Mears, D. P. (2010). American Criminal Justice Policy: An
Farrington, D. P. and Welsh, B. C. (2006). ‘A Half-Century Evaluation Approach to Increasing Accountability and
of Randomized Experiments on Crime and Justice’. Crime Effectiveness. New York: Cambridge University Press.
and Justice: A Review of Research 34: 55–132. Moher, D., Jadad, A. R., Nichol, G. et al. (1995). ‘Assessing
Fixsen, D. L., Blase, K. A., Naoom, S. F., and Wallace, F. the Quality of Randomized Controlled Trials: An
(2009). ‘Core Implementation Components’. Research on Annotated Bibliography of Scales and Checklists’.
Social Work Practice 19(5): 531–540. Controlled Clinical Trials 16(1): 62–73.
Foxcroft, D. R. (2014). ‘Can Prevention Classification Be Nagin, D. S. and Weisburd, D. (2013). ‘Evidence and Public
Improved by Considering the Function of Prevention?’. Policy: The Example of Evaluation Research in Policing’.
Prevention Science 15(6): 818–822. Criminology & Public Policy 12: 651–679.
284 Policing Article B. C. Welsh and A. B. Wexler

National Academies of Sciences, Engineering, and Sherman, L. W., Farrington, D. P., Welsh, B. C., and

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
Medicine (2017). Communicating Science Effectively: A MacKenzie, D. L. (eds). (2006). Evidence-Based Crime
Research Agenda. Washington, DC: National Academies Prevention, rev. edn. New York: Routledge.
Press. Sherman, L. W. and Gottfredson, D. C. (1997). Appendix:
National Research Council (2005). Improving Evaluation of Research Methods. In Sherman, L. W., Gottfredson, D.
Anticrime Programs. Washington, DC: National C., MacKenzie, D. L., Eck, J. E., Reuter, P., and Bushway
Academies Press. S. D. Preventing Crime: What Works, What Doesn’t,
Nutley, S. and Homel, P. (2006). ‘Delivering Evidence-Based What’s Promising. Washington, DC: Office of Justice
Policy and Practice: Lessons from the Implementation of Programs, pp. 1–8.
the UK Crime Reduction Programme’. Evidence and Sherman, L. W., Gottfredson, D. C., MacKenzie, D. L., Eck,
Policy 2(1): 5–26. J. E., Reuter, P., and Bushway, S. D. (1997). Preventing
Petrosino, A., Boruch, R. F., Soydan, H., Duggan, L., and Crime: What Works, What Doesn’t, What’s Promising.
Sanchez-Meca, J. (2001). ‘Meeting the Challenges of Washington, DC: Office of Justice Programs.
Evidence-Based Policy: The Campbell Collaboration’. Sherman, L. W., Gottfredson, D. C., MacKenzie, D. L., Eck,
Annals of the American Academy of Political and Social J. E., Reuter, P., and Bushway, S. D. (1998). Preventing
Science 578: 14–34. Crime: What Works, What Doesn’t, What’s Promising.
Petrosino, A. and Lavenberg, J. (2007). ‘Systematic Reviews Research in Brief, July. Washington, DC: National
and Meta-Analyses: Best Evidence on “What Works” for Institute of Justice.
Criminal Justice Decision Makers’. Western Criminology Shonkoff, J. P. and Bales, S. N. (2011). ‘Science Does Not
Review 8: 1–15. Speak for Itself: Translating Child Development Research
for the Public and Its Policymakers’. Child Development
Petticrew, M. and Roberts, H. (2006). Systematic Reviews in
82(1): 17–32.
the Social Sciences: A Practical Guide. Oxford, UK:
Blackwell. Stafford, J. and Silverlock, L. (2000). What Works and What
Makes What Works Work? In Ballintyne, S, Pease, K., and
Sampson, R. J. (2010). ‘Gold Standard Myths: Observations
McLaren, V. (eds), Secure Foundations: Key Issues in
on the Experimental Turn in Criminology’. Journal of
Crime Prevention, Crime Reduction and Community
Quantitative Criminology 26(4): 489–500.
Safety. London: Institute for Public Policy Research, pp.
Sampson, R. J., Winship, C., and Knight, C. (2013). 91–101.
‘Translating Causal Claims: Principles and Strategies for
Stockard, J. and Wood, T. W. (2017). ‘The Threshold and
Policy-Relevant Criminology’. Criminology & Public
Inclusive Approaches to Determining “Best Available
Policy 12: 587–616.
Evidence”: An Empirical Analysis’. American Journal of
Shadish, W., Cook, T. D., and Campbell, D. T. (2002). Evaluation 38(4): 471–492.
Experimental and Quasi-Experimental Designs for
Telep, C. W. (2017). ‘Police Officer Receptivity to Research
General Causal Inference. Boston: Houghton-Mifflin.
and Evidence-Based Policing: Examining Variability
Sherman, L. W. (1997a). Introduction: The Congressional within and across Agencies’. Crime & Delinquency 63:
Mandate to Evaluate. In Sherman, L. W., Gottfredson, D. 976–999.
C., MacKenzie, D. L., Eck, J. E., Reuter, P., and Bushway Telep, C. W. and Lum, C. M. (2014). ‘The Receptivity of
S. D. Preventing Crime: What Works, What Doesn’t, Officers to Empirical Research and Evidence-Based
What’s Promising, chap. 1. Washington, DC: Office of Policing: An Examination of Survey Data from Three
Justice Programs, pp.1–22. Agencies’. Police Quarterly 17(4): 359–385.
Sherman, L. W. (1997b). Thinking about Crime Prevention. Telep, C. W. and Weisburd, D. (2016). Policing. In
In Sherman, L. W., Gottfredson, D. C., MacKenzie, D. L., Weisburd, D., Farrington, D. P., and Gill, C. E. (eds),
Eck, J. E., Reuter, P., and Bushway S. D. Preventing Crime: What Works in Crime Prevention and Rehabilitation:
What Works, What Doesn’t, What’s Promising, chap. 2. Lessons from Systematic Reviews. New York: Springer,
Washington, DC: Office of Justice Programs, pp.1–32. pp. 137–168.
Sherman, L. W. (1998). Evidence-Based Policing. Tilley, N. (2000). The Evaluation Jungle. In Ballintyne, S.,
Washington, DC: Police Foundation. Pease, K., and McLaren, V. (eds), Secure Foundations: Key
Sherman, L. W. (2013). The Rise of Evidence-Based Issues in Crime Prevention, Crime Reduction and
Policing: Targeting, Testing, and Tracking. In Tonry, M. Community Safety. London: Institute for Public Policy
(ed), Crime and Justice in America: 1975–2025. Crime and Research, pp. 115–130.
Justice: A Review of Research, vol. 42. Chicago: University Tonry, M. and Farrington, D. P. (1995). Strategic
of Chicago Press, pp. 377–451. Approaches to Crime Prevention. In Tonry, M. and
Sherman, L. W., Farrington, D. P., Welsh, B. C., and Farrington, D. P. (eds), Building a Safer Society:
MacKenzie, D. L. (eds). (2002). Evidence-Based Crime Strategic Approaches to Crime Prevention. Chicago:
Prevention. New York: Routledge. University of Chicago Press, pp. 1–20.
Scientific and public policy discourse on crime prevention Article Policing 285

U.S. House of Representatives (1996). Conference Report Justice?’. Annals of the American Academy of Political

Downloaded from https://academic.oup.com/policing/article-abstract/13/3/271/5308392 by Pontifica Universidad Catolica de Peru user on 21 July 2020
104-378. 104th Congress, 1st Session, 27 April 1996. and Social Science 578: 50–70.
van Dijk, J. J. M. and de Waard, J. (1991). ‘A Two- Welsh, B. C., Peel, M. E., Farrington, D. P., Elffers, H., and
Dimensional Typology of Crime Prevention Projects: Braga, A. A. (2011). ‘Research Design Influence on Study
With a Bibliography’. Criminal Justice Abstracts 23: Outcomes in Crime and Justice: A Partial Replication
483–503. with Public Area Surveillance’. Journal of Experimental
Weimer, D. L. and Friedman, L. S. (1979). Efficiency Criminology 7(2): 183–198.
Considerations in Criminal Rehabilitation Research: Wilson, D. B. (2001). ‘Meta-Analytic Methods for
Costs and Consequences. In Sechrest L, White, S. O., Criminology’. Annals of the American Academy of
and Brown, E. D. (eds), The Rehabilitation of Criminal Political and Social Science 578: 71–89.
Offenders: Problems and Prospects. Washington, DC: Wilson, D. B. (2009). ‘Missing a Crucial Piece of the Pie:
National Academy of Sciences, pp. 251–272. Simple Document Search Strategies Inadequate for
Weisburd, D., Farrington, D. P., and Gill, C. E. (eds). Systematic Reviews’. Journal of Experimental
(2016). What Works in Crime Prevention and Criminology 5(4): 429–440.
Rehabilitation: Lessons from Systematic Reviews. New Zane, S. N. and Welsh, B. C. (2018). ‘Toward an “Age of
York: Springer. Imposed Use”? Evidence-Based Crime Policy in a Law
Weisburd, D., Lum, C. M., and Petrosino, A. (2001). ‘Does and Social Science Context’. Criminal Justice Policy
Research Design Affect Study Outcomes in Criminal Review 29(3): 280–300.

You might also like