Download as pdf or txt
Download as pdf or txt
You are on page 1of 67

A Guide to Assessments That Work 2nd

Edition John Hunsley (Editor)


Visit to download the full and correct content document:
https://ebookmass.com/product/a-guide-to-assessments-that-work-2nd-edition-john-h
unsley-editor/
A GUIDE TO ASSESSMENTS THAT WORK
A GUIDE TO ASSESSMENTS
THAT WORK

S e c o n d E di t i o n

EDITED BY

John Hunsley and Eric J. Mash

1
1
Oxford University Press is a department of the University of Oxford. It furthers
the University’s objective of excellence in research, scholarship, and education
by publishing worldwide. Oxford is a registered trade mark of Oxford University
Press in the UK and certain other countries.

Published in the United States of America by Oxford University Press


198 Madison Avenue, New York, NY 10016, United States of America.

© Oxford University Press 2018

First Edition published in 2008

All rights reserved. No part of this publication may be reproduced, stored in


a retrieval system, or transmitted, in any form or by any means, without the
prior permission in writing of Oxford University Press, or as expressly permitted
by law, by license, or under terms agreed with the appropriate reproduction
rights organization. Inquiries concerning reproduction outside the scope of the
above should be sent to the Rights Department, Oxford University Press, at the
address above.

You must not circulate this work in any other form


and you must impose this same condition on any acquirer.

CIP data is on file at the Library of Congress


ISBN 978–​0–​19–​049224–​3

1 3 5 7 9 8 6 4 2
Printed by Sheridan Books, Inc., United States of America
Contents

Foreword to the First Edition Part II Attention-​Deficit and Disruptive


by Peter E. Nathan vii Behavior Disorders

Preface xi 4. Attention-​Deficit/​Hyperactivity Disorder 47


CHARLOTTE JOHNSTON
About the Editors xv SARA COLALILLO

Contributors xvii 5. Child and Adolescent Conduct Problems 71


PAUL J. FRICK
Part I Introduction ROBERT J. McMAHON

1. Developing Criteria for Evidence-​Based Part III Mood Disorders and Self-​Injury
Assessment: An Introduction to Assessments
That Work 3 6. Depression in Children and Adolescents 99
JOHN HUNSLEY LEA R. DOUGHERTY
ERIC J. MASH DANIEL N. KLEIN
THOMAS M. OLINO
2. Dissemination and Implementation of
Evidence-​Based Assessment 17 7. Adult Depression 131
AMANDA JENSEN-​DOSS JACQUELINE B. PERSONS
LUCIA M. WALSH DAVID M. FRESCO
VANESA MORA RINGLE JULIET SMALL ERNST

3. Advances in Evidence-​Based Assessment: 8. Depression in Late Life 152


AMY FISKE
Using Assessment to Improve Clinical ALISA O’RILEY HANNUM
Interventions and Outcomes 32
ERIC A. YOUNGSTROM
ANNA VAN METER 9. Bipolar Disorder 173
SHERI L. JOHNSON
CHRISTOPHER MILLER
LORI EISNER
vi Contents

10. Self-​Injurious Thoughts and Behaviors 193 Part VI Schizophrenia and Personality
ALEXANDER J. MILLNER Disorders
MATTHEW K. NOCK
20. Schizophrenia 435
Part IV Anxiety and Related Disorders SHIRLEY M. GLYNN
KIM T. MUESER
11. Anxiety Disorders in Children and
Adolescents 217 21. Personality Disorders 464
SIMON P. BYRNE STEPHANIE L. ROJAS
ELI R. LEBOWITZ THOMAS A. WIDIGER
THOMAS H. OLLENDICK
WENDY K. SILVERMAN Part VII Couple Distress and Sexual
Disorders
12. Specific Phobia and Social Anxiety
Disorder 242 22. Couple Distress 489
KAREN ROWA DOUGLAS K. SNYDER
RANDI E. MCCABE RICHARD E. HEYMAN
MARTIN M. ANTONY STEPHEN N. HAYNES
CHRISTINA BALDERRAMA-​DURBIN
13. Panic Disorder and Agoraphobia 266
AMY R. SEWART 23. Sexual Dysfunction 515
MICHELLE G. CRASKE NATALIE O. ROSEN
MARIA GLOWACKA
14. Generalized Anxiety Disorder 293 MARTA MEANA
MICHEL J. DUGAS YITZCHAK M. BINIK
CATHERINE A. CHARETTE
NICOLE J. GERVAIS Part VIII Health-​Related Problems

15. Obsessive–​Compulsive Disorder 311 24. Eating Disorders 541


SHANNON M. BLAKEY ROBYN SYSKO
JONATHAN S. ABRAMOWITZ SARA ALAVI

16. Post-​Traumatic Stress Disorder in Adults 329 25. Insomnia Disorder 563
SAMANTHA J. MOSHIER CHARLES M. MORIN
KELLY S. PARKER-​GUILBERT SIMON BEAULIEU-​BONNEAU
BRIAN P. MARX KRISTIN MAICH
TERENCE M. KEANE COLLEEN E. CARNEY

Part V Substance-​Related and Gambling 26. Child and Adolescent Pain 583
Disorders C. MEGHAN McMURTRY
PATRICK J. McGRATH
17. Substance Use Disorders 359
DAMARIS J. ROHSENOW 27. Chronic Pain in Adults 608
THOMAS HADJISTAVROPOULOS
18. Alcohol Use Disorder 381 NATASHA L. GALLANT
ANGELA M. HAENY MICHELLE M. GAGNON
CASSANDRA L. BONESS
YOANNA E. McDOWELL Assessment Instrument Index 629
KENNETH J. SHER
Author Index 639
19. Gambling Disorders 412 Subject Index 721
DAVID C. HODGINS
JENNIFER L. SWAN
RANDY STINCHFIELD
Foreword to the First Edition

I believe A Guide to Assessments that Work is the right There is also much to admire within the pages of the
book at the right time by the right editors and authors. volume. Each chapter follows a common format pre-
The mental health professions have been intensively scribed by the editors and designed, as they point out,
engaged for a decade and a half and more in establish- “to enhance the accessibility of the material presented
ing empirically supported treatments. This effort has led throughout the book.” First, the chapters are syndrome-​
to the publication of evidence-​based treatment guidelines focused, making it easy for clinicians who want help in
by both the principal mental health professions, clinical assessing their patients to refer to the appropriate chapter
psychology (Chambless & Ollendick, 2001; Division 12 or chapters. When they do so, they will find reviews of the
Task Force, 1995), and psychiatry (American Psychiatric assessment literature for three distinct purposes: diagno-
Association, 1993, 2006). A substantial number of books sis, treatment planning, and treatment monitoring. Each
and articles on evidence-​ based treatments have also of these reviews is subjected to a rigorous rating system
appeared. Notable among them is a series by Oxford that culminates in an overall evaluation of “the scientific
University Press, the publishers of A Guide to Assessments adequacy and clinical relevance of currently available
that Work, which began with the first edition of A Guide measures.” The chapters conclude with an overall assess-
to Treatments that Work (Nathan & Gorman, 1998), now ment of the limits of the assessments available for the syn-
in its third edition, and the series includes Psychotherapy drome in question, along with suggestions for future steps
Relationships that Work (Norcross, 2002) and Principles to confront them. I believe it can well be said, then, that
of Therapeutic Change that Work (Castonguay & this is the right book by the right editors and authors.
Beutler, 2006). But is this the right time for this book? Evidence-​based
Now we have an entire volume given over to evidence-​ treatments have been a focus of intense professional atten-
based assessment. It doesn’t appear de novo. Over the tion for many years. Why wouldn’t the right time for this
past several years, its editors and like-​minded colleagues book have been several years ago rather than now, to
tested and evaluated an extensive series of guidelines for coincide with the development of empirically supported
evidence-​based assessments for both adults and children treatments? The answer, I think, reflects the surprisingly
(e.g., Hunsley & Mash, 2005; Mash & Hunsley, 2005). brief history of the evidence-​based medical practice move-
Many of this book’s chapter authors participated in these ment. Despite lengthy concern for the efficacy of treat-
efforts. It might well be said, then, that John Hunsley, Eric ments for mental disorders that dates back more than
Mash, and the chapter authors in A Guide to Assessments 50 years (e.g., Eysenck, 1952; Lambert & Bergin, 1994;
that Work are the right editors and authors for this, the first Luborsky, Singer, & Luborsky, 1976; Nathan, Stuart, &
book to detail the assessment evidence base. Dolan, 2000), it took the appearance of a Journal of the
viii Foreword to the First Edition

American Mental Association article in the early 1990s that currently lack empirical support. I agree. As with
advocating evidence-​based medical practice over medi- a number of psychotherapy approaches, there remain a
cine as an art to mobilize mental health professionals to number of understudied assessment instruments whose
achieve the same goals for treatments for mental disor- evidence base is currently too thin for them to be con-
ders. The JAMA article “ignited a debate about power, sidered empirically supported. Like the editors, I believe
ethics, and responsibility in medicine that is now threat- we can anticipate enhanced efforts to establish the limits
ening to radically change the experience of health care” of usefulness of assessment instruments that haven’t yet
(Patterson, 2002). This effort resonated widely within the been thoroughly explored. I also anticipate a good deal
mental health community, giving impetus to the efforts of of fruitful discussion in the professional literature—​and
psychologists and psychiatrists to base treatment decisions likely additional research—​on the positions this book’s
on valid empirical data. editors and authors have taken on the assessment instru-
Psychologists had long questioned the uncertain reli- ments they have evaluated. I suspect their ratings for
ability and utility of certain psychological tests, even “psychometric adequacy and clinical relevance” will be
though psychological testing was what many psychologists extensively critiqued and scrutinized. While the resul-
spent much of their time doing. At the same time, the tant dialogue might be energetic—​even indecorous on
urgency of efforts to heighten the support base for valid occasion—​as has been the dialogue surrounding the evi-
assessments was limited by continuing concerns over the dence base for some psychotherapies, I am hopeful it will
efficacy of psychotherapy, for which many assessments also lead to more helpful evaluations of test instruments.
were done. Not surprisingly, then, when empirical sup- Perhaps the most important empirical studies we might
port for psychological treatments began to emerge in the ultimately anticipate would be research indicating which
early and middle 1990s, professional and public support assessment instruments lead both to valid diagnoses and
for psychological intervention grew. In turn, as psycho- useful treatment planning for specific syndromes. A dis-
therapy’s worth became more widely recognized, the tant goal of syndromal diagnosis for psychopathology has
value of psychological assessments to help in the plan- always been diagnoses that bespeak effective treatments. If
ning and evaluation of psychotherapy became increas- the system proposed in this volume leads to that desirable
ingly recognized. If my view of this history is on target, the outcome, we could all celebrate.
intense efforts that have culminated in this book could I congratulate John Hunsley and Eric Mash and their
not have begun until psychotherapy’s evidence base had colleagues for letting us have this eagerly anticipated
been established. That has happened only recently, after a volume.
lengthy process, and that is why I claim that the right time Peter E. Nathan
for this book is now. (1935–2016)
Who will use this book? I hope it will become a favor-
ite text for graduate courses in assessment so that new
generations of graduate students and their teachers will References
come to know which of the assessment procedures they American Psychiatric Association. (1993). Practice guidelines
are learning and teaching have strong empirical support. for the treatment of major depressive disorder in adults.
I also hope the book will become a resource for practitio- American Journal of Psychiatry, 150 (4 Supplement),
ners, including those who may not be used to choosing 1–​26.
assessment instruments on the basis of evidence base. To American Psychiatric Association. (2006). Practice guidelines
the extent that this book becomes as influential in clinical for the treatment of psychiatric disorders: Compendium,
psychology as I hope it does, it should help precipitate a 2006. Washington, DC: Author.
change in assessment test use patterns, with an increase in Castonguay, L. G., & Beutler, L. E. (2006). Principles of thera-
the utilization of tests with strong empirical support and a peutic change that work. New York: Oxford University
Press.
corresponding decrease in the use of tests without it. Even
Chambless, D. L., & Ollendick, T. H. (2001). Empirically
now, there are clinicians who use assessment instruments
supported psychological interventions: Controversies
because they learned them in graduate school, rather than and evidence. In S. T. Fiske, D. L. Schacter, & C.
because there is strong evidence that they work. Now, a Zahn-​Waxler (Eds.), Annual review of psychology
different and better standard is available. (Vol. 52, pp. 685–​716). Palo Alto, CA: Annual Review.
I am pleased the editors of this book foresee it provid- Division 12 Task Force. (1995). Training in and dissemina-
ing an impetus for research on assessment instruments tion of empirically-​validated psychological treatments:
Foreword to the First Edition ix

Report and recommendations. The Clinical Psychologist, Mash, E. J., & Hunsley, J. (Eds.). (2005). Developing
48, 3–​23. guidelines for the evidence-​based assessment of child
Eysenck, H. J. (1952). The effects of psychotherapy: An eval- and adolescent disorders (special section). Journal of
uation. Journal of Consulting Psychology, 16, 319–​324. Clinical Child and Adolescent Psychology, 34(3).
Hunsley, J., & Mash, E. J. (Eds.). (2005). Developing guide- Nathan, P. E., & Gorman, J. M. (1998, 2002, 2007). A guide
lines for the evidence-​based assessment (EBA) of adult dis- to treatments that work. New York: Oxford University
orders (special section). Psychological Assessment, 17(3). Press.
Lambert, M. J., & Bergin, A. E. (1994). The effectiveness Nathan, P. E., Stuart, S. P., & Dolan, S. L. (2000). Research
of psychotherapy. In S. L. Garfield & A. E. Bergin on psychotherapy efficacy and effectiveness: Between
(Eds.), Handbook of psychotherapy and behavior change Scylla and Charybdis? Psychological Bulletin, 126,
(4th ed., pp. 143–​189). New York: Wiley. 964–​981.
Luborsky, L., Singer, B., & Luborsky, L. (1976). Comparative Norcross, J. C. (Ed.). (2002). Psychotherapy relationships
studies of psychotherapies: Is it true that “everybody has won that work: Therapist contributions and responsiveness to
and all must have prizes?” In R. L. Spitzer & D. F. Klein patients. New York: Oxford University Press.
(Eds.), Evaluation of psychological therapies (pp. 3–​22). Patterson, K. (2002). What doctors don’t know (almost every-
Baltimore, MD: Johns Hopkins University Press. thing). New York Times Magazine, May 5, 74–​77.
Preface

BACKGROUND clinical psychology journals (e.g., Arbisi & Beck, 2016;


Jensen-​Doss, 2015). The evidence base for the value of
Evidence-​based practice principles in health care systems monitoring treatment progress has increased substantially,
emphasize the importance of integrating information as have calls for the assessment of treatment progress to
drawn from systematically collected data, clinical exper- become standard practice (e.g., Lambert, 2017). There
tise, and patient preferences when considering health is also mounting evidence for assessment as a key com-
care service options for patients (Institute of Medicine, ponent for engaging clients in effective mental health
2001; Sackett, Rosenberg, Gray, Haynes, & Richardson, services (Becker, Boustani, Gellatly, & Chorpita, 2017).
1996). These principles are a driving force in most health Unfortunately, some long-​ standing problems evident
care systems and have been endorsed as a necessary foun- in the realm of psychological assessment remain. Many
dation for the provision of professional psychological ser- researchers continue to ignore the importance of evaluat-
vices (American Psychological Association Presidential ing the reliability of the assessment data obtained from
Task Force on Evidence-​Based Practice, 2006; Dozois their study participants (e.g., Vacha-​Haase & Thompson,
et al., 2014). As psychologists, it is difficult for us to imag- 2011). Despite the demonstrated impact of treatment
ine how any type of health care service, including psycho- monitoring, relatively few clinicians systematically and
logical services, can be provided to children, adolescents, routinely assess the treatment progress of their clients
adults, couples, or families without using some type of (Ionita & Fitzpatrick, 2014), although it appears that stu-
informal or formal assessment methods. Nevertheless, dents in professional psychology programs are receiving
until relatively recently, there was an almost exclusive more training in these assessment procedures than was the
focus on issues related to developing, disseminating, and case in the past (e.g., Overington, Fitzpatrick, Hunsley,
providing evidence-​based interventions, with only cursory & Drapeau, 2015). All in all, though, when viewed from
acknowledgment of the role that evidence-​based assess- the vantage point of the early years of the 21st century, it
ment (EBA) activities play in the promotion of evidence-​ does seem that steady progress is being made with respect
based services. to EBA.
Fortunately, much has changed with respect to EBA As was the case with the first edition, the present vol-
since the publication of the first edition of this volume in ume was designed to complement the books published
2008. A growing number of publications are now available by Oxford University Press that focus on bringing the best
in the scientific literature that address the importance of of psychological science to bear on questions of clini-
solid assessment instruments and methods. Special sec- cal importance. These volumes, A Guide to Treatments
tions on EBA have been published in recent issues of top that Work (Nathan & Gorman, 2015) and Psychotherapy
xii Preface

Relationships that Work (Norcross, 2011), address inter- to (a) understanding the patient’s or client’s needs and
vention issues; the present volume specifically addresses (b) accessing the scientific literature on evidence-​based
the role of assessment in providing evidence-​based ser- treatment options. We also recognize that many patients
vices. Our primary goal for the book was to have it address or clients will present with multiple problems; to that end,
the needs of professionals providing psychological services the reader will find frequent references within a chapter
and those training to provide such services. A secondary to the assessment of common co-​occurring problems that
goal was to provide guidance to researchers on scientifi- are addressed in other chapters in the volume. To be opti-
cally supported assessment tools that could be used for mally useful to potential readers, we have included chap-
both psychopathology research and treatment research ters that deal with the assessment of the most commonly
purposes. Relatedly, we hope that the summary tables pro- encountered disorders or conditions among children,
vided in each chapter will provide some inspiration for adolescents, adults, older adults, and couples.
assessment researchers to try to (a) develop instruments Ideally, we want readers to come away from each chap-
for specific assessment purposes and disorders for which, ter with a sense of the best scientific assessment options
currently, few good options exist and (b) expand our lim- that are clinically feasible and useful. To help accomplish
ited knowledge base on the clinical utility of our assess- this, we were extremely fortunate to be able to assemble a
ment instruments. stellar group of contributors for this volume. The authors
are all active contributors to the scientific literature on
assessment and share a commitment to the provision of
ORGANIZATION EBA and treatment services.
To enhance the accessibility of the material presented
All chapters and tables in the second edition have been throughout the book, we asked the authors, as much as pos-
revised and updated by our expert authors to reflect recent sible, to follow a common structure in writing their chap-
developments in the field, including the publication of ters. Without being a straitjacket, we expected the authors
the fifth edition of the Diagnostic and Statistical Manual to use these guidelines in a flexible manner that allowed for
of Mental Disorders (DSM-​ 5; American Psychiatric the best possible presentation of assessment work relevant
Association, 2013). For the most part, the general cover- to each disorder or clinical condition. The chapter format
age and organization of the first edition, which our read- generally used throughout the volume is as follows:
ers found useful, has been retained in the second edition. Introduction: A brief overview of the chapter content.
Consistent with a growing developmental psychopathol- Nature of the Disorder/​ Condition: This section
ogy perspective in the field, the scope of some chapters includes information on (a) general diagnostic consid-
has expanded in order to provide more coverage of assess- erations, such as prevalence, incidence, prognosis, and
ment issues across the lifespan (e.g., attention-​ deficit/​ common comorbid conditions; (b) evidence on etiology;
hyperactivity disorder in adults). The most important and (c) contextual information such as relational and
changes in organization involve the addition of two new social functioning and other associated features.
chapters, one dealing with the dissemination and imple- Purposes of Assessment: To make the book as clinically
mentation of EBA (Chapter 2) and the other dealing with relevant as possible, authors were asked to focus their
new developments in EBA (Chapter 3). The contents of review of the assessment literature to three specific assess-
these chapters highlight both the important contributions ment purposes: (a) diagnosis, (b) case conceptualization
that assessment can make to the provision of psychological and treatment planning, and (c) treatment monitoring
services and the challenges that mental health profession- and evaluation. We fully realize the clinical and research
als face in implementing cost-​effective and scientifically importance of other assessment purposes but, rather than
sound assessment strategies. attempting to provide a compendium of assessment mea-
Consistent with evidence-​ based psychology and sures and strategies, we wanted authors to target these
evidence-​based medicine, the majority of the chapters three key clinical assessment purposes. We also asked
in this volume are organized around specific disorders authors to consider ways in which age, gender, ethnicity,
or conditions. Although we recognize that some clients and other relevant characteristics may influence both the
do not have clearly defined or diagnosable problems, the assessment measures and the process of assessment for the
vast majority of people seeking psychological services disorder/​condition.
do have identifiable diagnoses or conditions. Accurately For each of the three main sections devoted to spe-
assessing these disorders and conditions is a prerequisite cific assessment purposes, authors were asked to focus on
Preface xiii

assessment measures and strategies that either have demon- issues without having to make frequent detours to discuss
strated their utility in clinical settings or have a substantial psychometrics.
likelihood of being clinically useful. Authors were encour- At the conclusion of each of these three main sec-
aged to consider the full range of relevant assessment meth- tions there is a subsection titled Overall Evaluation that
ods (interviews, self-​report, observation, performance tasks, includes concise summary statements about the scientific
computer-​based methods, physiological, etc.), but both sci- adequacy and clinical relevance of currently available
entific evidence and clinical feasibility were to be used to measures. This is where authors comment on the avail-
guide decisions about methods to include. ability (if any) of demonstrated scientific value of follow-
Assessment for Diagnosis: This section deals with ing the assessment guidance they have provided.
assessment measures and strategies used specifically for Conclusions and Future Directions: This final section
formulating a diagnosis. Authors were asked to focus in each chapter provides an overall sense of the scope
on best practices and were encouraged to comment on and adequacy of the assessment options available for the
important conceptual and practical issues in diagnosis disorder/​condition, the limitations associated with these
and differential diagnosis. options, and possible future steps that could be taken to
Assessment for Case Conceptualization and Treatment remedy these limitations. Some authors also used this sec-
Planning: This section presents assessment measures tion to raise issues related to the challenges involved in
and strategies used to augment diagnostic information trying to ensure that clinical decision-​making processes
to yield a full psychological case conceptualization that underlying the assessment process (and not just the assess-
can be used to guide decisions on treatment planning. ment measures themselves) are scientifically sound.
Specifically, this section addresses the domains that the
research literature indicates should be covered in an EBA
to develop (a) a clinically meaningful and useful case con- ACKNOWLEDGMENTS
ceptualization and (b) a clinically sensitive and feasible
service/​treatment plan (which may or may not include To begin with, we express our gratitude to the authors. They
the involvement of other professionals). diligently reviewed and summarized often-​ voluminous
Assessment for Treatment Monitoring and Treatment assessment literatures and then presented this informa-
Outcome: In this third section, assessment measures and tion in a clinically informed and accessible manner. The
strategies were reviewed that can be used to (a) track the authors also worked hard to implement the guidelines we
progress of treatment and (b) evaluate the overall effect of provided for both chapter structure and the ratings of vari-
treatment on symptoms, diagnosis, and general function- ous psychometric characteristics. Their efforts in construct-
ing. Consistent with the underlying thrust of the volume, ing their chapters are admirable, and the resulting chapters
the emphasis is on assessment options that have support- consistently provide invaluable clinical guidance.
ing empirical evidence. We also thank Sarah Harrington, Senior Editor for clini-
Within each of the three assessment sections, standard cal psychology at Oxford University Press, for her continued
tables are used to provide summary information about interest in the topic and her ongoing support for the book.
the psychometric characteristics of relevant instruments. We greatly appreciate her enthusiasm and her efficiency
Rather than provide extensive psychometric details in throughout the process of developing and producing this
the text, authors were asked to use these rating tables to second edition. We are also indebted to Andrea Zekus,
convey information on the psychometric adequacy of Editor at Oxford University Press, who helped us with the
instruments. To enhance the utility of these tables, rather process of assembling the book from start to finish. Her assis-
than presenting lists of specific psychometric values for tance with the myriad issues associated with the publication
each assessment tool, authors were asked to make global process and her rapid response to queries was invaluable.
ratings of the quality of the various psychometric indices Finally, we thank all the colleagues and contributors
(e.g., norms, internal reliability, and construct validity) to the psychological assessment and measurement litera-
as indicated by extant research. Details on the rating sys- tures who, over the years, have shaped our thinking about
tem used by the authors are presented in the introductory assessment issues. We are especially appreciative of the
chapter. Our goal is to have these tables serve as valuable input from those colleagues who have discussed with us
summaries for readers. In addition, by using the tables to the host of problems, concerns, challenges, and promises
present psychometric information, the authors were able associated with efforts to promote greater awareness of the
to focus their chapters on both conceptual and practical need for EBA within professional psychology.
xiv Preface

References practice and usage of progress monitoring measures.


Canadian Psychology, 55, 187–​196.
American Psychiatric Association. (2013). Diagnostic and sta-
Jensen-​Doss, A. (2015). Practical, evidence-​ based clinical
tistical manual of mental disorders (5th ed.). Arlington,
decision making: Introduction to the special series.
VA: American Psychiatric Publishing.
Cognitive and Behavioral Practice, 22, 1–​4.
American Psychological Association Presidential Task Force on
Lambert, M. J. (2017). Maximizing psychotherapy outcome
Evidence-​Based Practice. (2006). Evidence-​based prac-
beyond evidence-​ based medicine. Psychotherapy and
tice in psychology. American Psychologist, 61, 271–​285.
Psychosomatics, 86, 80–​89.
Arbisi, P. A., & Beck, J. G. (2016). Introduction to the special
Nathan, P. E., & Gorman, J. M. (Eds.). (2015). A guide to
series “Empirically Supported Assessment.” Clinical
treatments that work (4th ed.). New York, NY: Oxford
Psychology: Science and Practice, 23, 323–​326.
University Press.
Becker, K. D., Boustani, M., Gellatly, R., & Chorpita, B. F.
Norcross, J. C. (Ed.). (2011). Psychotherapy relationships
(2017). Forty years of engagement research in children’s
that work: Evidence-​ based responsiveness (2nd ed.).
mental health services: Multidimensional measure-
New York, NY: Oxford University Press.
ment and practice elements. Journal of Clinical Child
Overington, L., Fitzpatrick, M., Hunsley, J., & Drapeau, M.
& Adolescent Psychology. Advance online publication.
(2015). Trainees’ experiences using progress monitor-
Dozois, D. J. A., Mikail, S., Alden, L. E., Bieling, P. J.,
ing measures. Training and Education in Professional
Bourgon, G., Clark, D. A., . . . Johnston, C. (2014).
Psychology, 9, 202–​209.
The CPA Presidential Task Force on Evidence-​Based
Sackett, D. L., Rosenberg, W. M. C., Gray, J. A. M., Haynes,
Practice of Psychological Treatments. Canadian
R. B., & Richardson, W. S. (1996). Evidence based
Psychology, 55, 153–​160.
medicine: What it is and what it is not. British Medical
Institute of Medicine. (2001). Crossing the quality chasm: A
Journal, 312, 71–​72.
new health system for the 21st century. Washington,
Vacha-​Haase, T., & Thompson, B. (2011). Score reliabil-
DC: National Academies Press.
ity: A retrospective look back at 12 years of reliability
Ionita, F., & Fitzpatrick, M. (2014). Bringing science to
generalization studies. Measurement and Evaluation in
clinical practice: A Canadian survey of psychological
Counseling and Development, 44, 159–​168.
About the Editors

John Hunsley, PhD, is Professor of Psychology in the Eric J. Mash, PhD, is Professor Emeritus in the
School of Psychology at the University of Ottawa and Department of Psychology at the University of Calgary. He
is a Fellow of the Association of State and Provincial is a Fellow of the American Psychological Association, the
Psychology Boards and the Canadian Psychological Canadian Psychological Association, and the American
Association. He has served as a journal editor, an edito- Psychological Society. He has served as an editor, edito-
rial board member for several journals, and an editorial rial board member, and consultant for many scientific
consultant for many journals in psychology. He has pub- and professional journals and has written and edited many
lished more than 130 articles, chapters, and books related books and journal articles related to child and adolescent
to evidence-​based psychological practice, psychological mental health, assessment, and treatment.
assessment, and professional issues.
Contributors

Jonathan S. Abramowitz, PhD: Department of Simon P. Byrne, PhD: Yale Child Study Center, Yale
Psychology and Neuroscience, University of North School of Medicine, New Haven, Connecticut
Carolina at Chapel Hill, Chapel Hill, North Carolina
Colleen E. Carney, PhD: Department of Psychology,
Sara Alavi: Eating and Weight Disorders Program, Icahn Ryerson University, Toronto, Ontario, Canada
School of Medicine at Mt. Sinai, New York, New York
Catherine A. Charette: Département de psychoé-
Martin M. Antony, PhD: Department of Psychology, ducation et de psychologie, Université du Québec en
Ryerson University, Toronto, Ontario, Canada Outaouais, Gatineau, Quebec, Canada

Christina Balderrama-​Durbin, PhD: Department of Sara Colalillo, MA: Department of Psychology,


Psychology, Binghamton University—State University of University of British Columbia, Vancouver, British
New York, Binghamton, New York Columbia, Canada

Simon Beaulieu-​Bonneau, PhD: École de psychologie, Michelle G. Craske, PhD: Department of Psychology,
Université Laval, Quebec City, Quebec, Canada University of California at Los Angeles, Los Angeles,
California
Yitzchak M. Binik, PhD: Department of Psychology,
McGill University, Montreal, Quebec, Canada Lea R. Dougherty, PhD: Department of Psychology,
University of Maryland, College Park, Maryland
Shannon M. Blakey, MS: Department of Psychology and
Neuroscience, University of North Carolina at Chapel Michel J. Dugas, PhD: Département de psychoéducation
Hill, Chapel Hill, North Carolina et de psychologie, Université du Québec en Outaouais,
Gatineau, Québec, Canada
Cassandra L. Boness, MA: Department of Psychological
Sciences, University of Missouri, Columbia, Missouri Lori Eisner, PhD: Needham Psychotherapy Associates,
LLC
xviii Contributors

Juliet Small Ernst: Cognitive Behavior Therapy and Richard E. Heyman, PhD: Family Translational Research
Science Center, Oakland, California Group, New York University, New York, New York

Amy Fiske, PhD: Department of Psychology, West David C. Hodgins, PhD: Department of Psychology,
Virginia University, Morgantown, West Virginia University of Calgary, Calgary, Alberta, Canada

David M. Fresco, PhD: Department of Psychological John Hunsley, PhD: School of Psychology, University of
Sciences, Kent State University, Kent, Ohio; Department Ottawa, Ottawa, Ontario, Canada
of Psychiatry, Case Western Reserve University School of
Medicine, Cleveland, Ohio Amanda Jensen-​Doss, PhD: Department of Psychology,
University of Miami, Coral Gables, Florida
Paul J. Frick, PhD: Department of Psychology, Louisiana
State University, Baton Rouge, Louisiana; Learning Sheri L. Johnson, PhD: Department of Psychology,
Sciences Institute of Australia; Australian Catholic University of California Berkeley, Berkeley, California
University; Brisbane, Australia
Charlotte Johnston, PhD: Department of Psychology,
Michelle M. Gagnon, PhD: Department of University of British Columbia, Vancouver, British
Psychology, University of Saskatchewan, Saskatoon, Columbia, Canada
Saskatchewan, Canada
Terence M. Keane, PhD: VA Boston Healthcare System,
Natasha L. Gallant, MA: Department of Psychology, National Center for Posttraumatic Stress Disorder, and
University of Regina, Regina, Saskatchewan, Canada Boston University School of Medicine, Boston, Massachusetts

Nicole J. Gervais, PhD: Department of Psychology, Daniel N. Klein, PhD: Department of Psychology, Stony
University of Toronto, Toronto, Ontario, Canada Brook University, Stony Brook, New York

Maria Glowacka: Department of Psychology and Eli R. Lebowitz, PhD: Yale Child Study Center, Yale
Neuroscience, Dalhousie University, Halifax, Nova School of Medicine, New Haven, Connecticut
Scotia, Canada
Kristin Maich, MA: Department of Psychology, Ryerson
Shirley M. Glynn, PhD: VA Greater Los Angeles University, Toronto, Ontario, Canada
Healthcare System and UCLA Department of Psychiatry
and Biobehavioral Sciences, David Geffen School of Brian P. Marx, PhD: VA Boston Healthcare System,
Medicine, Los Angeles, California National Center for Posttraumatic Stress Disorder,
and Boston University School of Medicine, Boston,
Thomas Hadjistavropoulos, PhD: Department of Massachusetts
Psychology, University of Regina, Regina, Saskatchewan,
Canada Eric J. Mash, PhD: Department of Psychology, University
of Calgary, Calgary, Alberta, Canada
Angela M. Haeny, MA: Department of Psychological
Sciences, University of Missouri, Columbia, Missouri Randi E. McCabe, PhD: Anxiety Treatment and
Research Clinic, St. Joseph’s Healthcare, Hamilton, and
Alisa O’Riley Hannum, PhD, ABPP: VA Eastern Department of Psychiatry and Behavioral Neurosciences,
Colorado Healthcare System, Denver, Colorado McMaster University, Hamilton, Ontario, Canada

Stephen N. Haynes, PhD: Department of Psychology, Yoanna E. McDowell, MA: Department of Psychological
University of Hawai’i at Mānoa, Honolulu, Hawaii Sciences, University of Missouri, Columbia, Missouri
Contributors xix

Patrick J. McGrath, PhD: Centre for Pediatric Thomas H. Ollendick, PhD: Department of Psychology,
Pain Research, IWK Health Centre; Departments Virginia Polytechnic Institute and State University,
of Psychiatry, Pediatrics and Community Health & Blacksburg, Virginia
Epidemiology, Dalhousie University; Halifax, Nova
Scotia, Canada Kelly S. Parker-​Guilbert, PhD: Psychology Department,
Bowdoin College, Brunswick, ME and VA Boston
Robert J. McMahon, PhD: Department of Healthcare System, Boston, Massachusetts
Psychology, Simon Fraser University, Burnaby, British
Columbia, Canada; BC Children’s Hospital Research Jacqueline B. Persons, PhD: Cognitive Behavior
Institute, Vancouver, British Columbia, Canada Therapy and Science Center, Oakland, California and
Department of Psychology, University of California
C. Meghan McMurtry, PhD: Department of Psychology, at Berkeley, Berkeley, California
University of Guelph, Guelph; Pediatric Chronic Pain
Program, McMaster Children’s Hospital, Hamilton; Vanesa Mora Ringle: Department of Psychology,
Department of Paediatrics, Schulich School of Medicine University of Miami, Coral Gables, Florida
& Dentistry, Western University, London; Ontario,
Canada Damaris J. Rohsenow, PhD: Center for Alcohol and
Addiction Studies, Brown University, Providence, Rhode
Marta Meana, PhD: Department of Psychology, Island
University of Nevada Las Vegas, Las Vegas, Nevada
Stephanie L. Rojas, MA: Department of Psychology,
Christopher Miller, PhD: VA Boston Healthcare System, University of Kentucky, Lexington, Kentucky
Center for Healthcare Organization and Implementation
Research, and Harvard Medical School Department of Natalie O. Rosen, PhD: Department of Psychology and
Psychiatry, Boston, Massachusetts Neuroscience, Dalhousie University, Halifax, Nova Scotia,
Canada
Alexander J. Millner, PhD: Department of Psychology,
Harvard University, Cambridge, Massachusetts Karen Rowa, PhD: Anxiety Treatment and Research
Clinic, St. Joseph’s Healthcare, Hamilton, and
Charles M. Morin, PhD: École de psychologie, Department of Psychiatry and Behavioral Neurosciences,
Université Laval, Quebec City, Quebec, Canada McMaster University, Hamilton, Ontario, Canada

Samantha J. Moshier, PhD: VA Boston Healthcare Amy R. Sewart, MA: Department of Psychology, University
System and Boston University School of Medicine, of California Los Angeles, Los Angeles, California
Boston, Massachusetts
Kenneth J. Sher, PhD: Department of Psychological
Kim T. Mueser, PhD: Center for Psychiatric Sciences, University of Missouri, Columbia, Missouri
Rehabilitation and Departments of Occupational
Therapy, Psychological and Brain Sciences, and Wendy K. Silverman, PhD: Yale Child Study Center,
Psychiatry, Boston University, Boston, Massachusetts Yale School of Medicine, New Haven, Connecticut

Matthew K. Nock, PhD: Department of Psychology, Douglas K. Snyder, PhD: Department of Psychology,
Harvard University, Cambridge, Massachusetts Texas A&M University, College Station, Texas

Thomas M. Olino, PhD: Department of Psychology, Randy Stinchfield, PhD: Department of Psychiatry,
Temple University, Philadelphia, Pennsylvania University of Minnesota, Minneapolis, Minnesota
xx Contributors

Jennifer L. Swan: Department of Psychology, University Lucia M. Walsh: Department of Psychology, University
of Calgary, Calgary, Alberta, Canada of Miami, Coral Gables, Florida

Robyn Sysko, PhD: Eating and Weight Disorders Thomas A. Widiger, PhD: Department of Psychology,
Program, Icahn School of Medicine at Mt. Sinai, University of Kentucky, Lexington, Kentucky
New York, New York
Eric A. Youngstrom, PhD: Department of Psychology
Anna Van Meter, PhD: Ferkauf Graduate School of and Neuroscience, University of North Carolina at
Psychology, Yeshiva University, New York, New York Chapel Hill, Chapel Hill, North Carolina
Part I

Introduction
1

Developing Criteria for


Evidence-​Based Assessment:
An Introduction to Assessments That Work

John Hunsley
Eric J. Mash

For many professional psychologists, assessment is clients’ problems and strengths. Whether construed as
viewed as a unique and defining feature of their expertise individual client monitoring, ongoing quality assurance
(Krishnamurthy et al., 2004). Historically, careful atten- efforts, or program evaluation, assessment is central to
tion to both conceptual and pragmatic issues related to efforts to gauge the impact of health care services pro-
measurement has served as the cornerstone of psychologi- vided to ameliorate these problems (Brown, Scholle, &
cal science. Within the realm of professional psychology, Azur, 2014; Hermann, Chan, Zazzali, & Lerner, 2006).
the ability to provide assessment and evaluation services Furthermore, the increasing availability of research-​
is typically seen as a required core competency. Indeed, derived treatment benchmarks holds out great promise
assessment services are such an integral component of for providing clinicians with meaningful and attainable
psychological practice that their value is rarely questioned targets for their intervention services (Lee, Horvath, &
but, rather, is typically assumed. However, solid evidence Hunsley, 2013; Spilka & Dobson, 2015). Importantly,
to support the usefulness of psychological assessment is statements about evidence-​ based practice and best-​
lacking, and many commonly used clinical assessment practice guidelines have begun to specifically incor-
methods and instruments are not supported by scientific porate the critical role of assessment in the provision
evidence (e.g., Hunsley, Lee, Wood, & Taylor, 2015; of evidence-​ based services (e.g., Dozois et al., 2014).
Hunsley & Mash, 2007; Norcross, Koocher, & Garofalo, Indeed, because the identification and implementation
2006). Indeed, Peterson’s (2004) conclusion from more of evidence-​based treatments rests entirely on the data
than a decade ago is, unfortunately, still frequently provided by assessment tools, ignoring the quality of
true: “For many of the most important inferences profes- these tools places the whole evidence-​based enterprise in
sional psychologists have to make, practitioners appear jeopardy.
to be forever dependent on incorrigibly fallible inter-
views and unavoidably selective, reactive observations as
primary sources of data” (p. 202). Furthermore, despite DEFINING EVIDENCE-​BASED ASSESSMENT
the current emphasis on evidence-​based practice, profes-
sional psychologists report that the least common purpose There are three critical aspects that should define
for which they use assessment is to monitor treatment evidence-​
based assessment (EBA; Hunsley & Mash,
progress (Wright et al., 2017). 2007; Mash & Hunsley, 2005). First, research findings
In this era of evidence-​based health care practices, and scientifically supported theories on both psycho-
the need for scientifically sound assessment methods pathology and normal human development should be
and instruments is greater than ever (Barlow, 2005). used to guide the selection of constructs to be assessed
Assessment is the key to the accurate identification of and the assessment process. As Barlow (2005) suggested,

3
4 Introduction

EBA measures and strategies should also be designed to accuracy (sensitivity, specificity, predictive power, etc.)
be integrated into interventions that have been shown to of cut-​scores for criterion-​referenced interpretation (cf.
work with the disorders or conditions that are targeted Achenbach, 2005). Furthermore, there should be sup-
in the assessment. Therefore, while recognizing that porting evidence to indicate that the EBAs are sensitive
most disorders do not come in clearly delineated neat to key characteristics of the individual(s) being assessed,
packages, and that comorbidity is often the rule rather including characteristics such as age, gender, ethnic-
than the exception, we view EBAs as being disorder-​or ity, and culture (e.g., Ivanova et al., 2015). Given the
problem-​specific. A problem-​specific approach is consis- range of purposes for which assessment instruments
tent with how most assessment and treatment research is can be used (i.e., screening, diagnosis, prognosis, case
conducted and would facilitate the integration of EBA conceptualization, treatment formulation, treatment
into evidence-​ based treatments (cf. Mash & Barkley, monitoring, and treatment evaluation) and the fact that
2007; Mash & Hunsley, 2007; Weisz & Kazdin, 2017). psychometric evidence is always conditional (based on
This approach is also congruent with the emerging sample characteristics and assessment purpose), support-
trend toward personalized assessment and treatment ing psychometric evidence must be considered for each
(e.g., Fisher, 2015; Ng & Weisz, 2016; Sales & Alves, purpose for which an instrument or assessment strategy is
2016; Seidman et al., 2010; Thompson-​Hollands, Sauer-​ used. Thus, general discussions concerning the relative
Zavala, & Barlow, 2014). Although formal diagnostic sys- merits of information obtained via different assessment
tems provide a frequently used alternative for framing the methods have little meaning outside of the assessment
range of disorders and problems to be considered, com- purpose and context. Similarly, not all psychometric ele-
monly experienced emotional and relational problems, ments are relevant to all assessment purposes. The group
such as excessive anger, loneliness, conflictual relation- of validity statistics that includes specificity, sensitivity,
ships, and other specific impairments that may occur in positive predictive power, and negative predictive power
the absence of a diagnosable disorder, may also be the is particularly relevant for diagnostic and prognostic
focus of EBAs. Even when diagnostic systems are used assessment purposes and contains essential information
as the framework for the assessment, clinicians need to for any measure that is intended to be used for screening
consider both (a) the potential value of emerging trans- purposes (Hsu, 2002). Such validity statistics may have
diagnostic approaches to treatment (Newby, McKinnon, little relevance, however, for many methods intended to
Kuyken, Gilbody, & Dalgleish, 2015) and (b) that a nar- be used for treatment monitoring and/​or evaluation pur-
row focus on assessing symptoms and symptom reduction poses; for these purposes, sensitivity to change is a much
is insufficient for treatment planning and treatment eval- more salient psychometric feature (e.g., Vermeersch,
uation purposes (cf. Kazdin, 2003). Many assessments are Lambert, & Burlingame, 2000).
conducted to identify the precise nature of the person’s Finally, even with data from psychometrically
problem(s). It is, therefore, necessary to conceptualize strong measures, the assessment process is inherently
multiple, interdependent stages in the assessment pro- a decision-​ making task in which the clinician must
cess, with each iteration of the process becoming less iteratively formulate and test hypotheses by integrating
general in nature and increasingly problem-​specific with data that are often incomplete or inconsistent. Thus,
further assessment (Mash & Terdal, 1997). In addition, a truly evidence-​based approach to assessment would
for some generic assessment strategies, there may be involve an evaluation of the accuracy and usefulness
research to indicate that the strategy is evidence-​based of this complex decision-​making task in light of poten-
without being problem-​specific. Examples of this include tial errors in data synthesis and interpretation, the costs
functional assessments (Hurl, Wightman, Haynes, & associated with the assessment process, and, ultimately,
Virues-​Ortega, 2016) and treatment progress monitoring the impact that the assessment had on clinical out-
systems (e.g., Lambert, 2015). comes. There are an increasing number of illustrations
A second requirement is that, whenever pos- of how assessments can be conducted in an evidence-​
sible, psychometrically strong measures should be based manner (e.g., Christon, McLeod, & Jensen-​Doss,
used to assess the constructs targeted in the assess- 2015; Youngstrom, Choukas-​ Bradley, Calhoun, &
ment. The measures should have evidence of reli- Jensen-​Doss, 2015). These provide invaluable guides
ability, validity, and clinical utility. They should also for clinicians and provide a preliminary framework that
possess appropriate norms for norm-​ referenced inter- could lead to the eventual empirical evaluation of EBA
pretation and/​or replicated supporting evidence for the processes.
Developing Criteria for Evidence-Based Assessment 5

FROM RESEARCH TO PRACTICE: USING instrument is scientifically sound (cf. Streiner, Norman,
A “GOOD-​E NOUGH” PRINCIPLE & Cairney, 2015). Unfortunately, this is of little aid to the
clinicians and researchers who are constantly faced with
Perhaps the greatest single challenge facing efforts to the decision of whether an instrument is good enough,
develop and implement EBAs is determining how to scientifically speaking, for the assessment task at hand.
start the process of operationalizing the criteria we just Prior to the psychometric criteria we set out in the
outlined. The assessment literature provides a veritable first edition of this volume, there had been attempts to
wealth of information that is potentially relevant to EBA; establish criteria for the selection and use of measures for
this very strength, however, is also a considerable liability, research purposes. Robinson, Shaver, and Wrightsman
for the size of the literature is beyond voluminous. Not (1991), for example, developed evaluative criteria for
only is the literature vast in scope but also the scientific the adequacy of attitude and personality measures, cov-
evaluation of assessment methods and instruments can be ering the domains of theoretical development, item
without end because there is no finite set of studies that development, norms, inter-​ item correlations, internal
can establish, once and for all, the psychometric proper- consistency, test–​retest reliability, factor analytic results,
ties of an instrument (Kazdin, 2005; Sechrest, 2005). On known groups validity, convergent validity, discriminant
the other hand, every single day, clinicians must make validity, and freedom from response sets. Robinson and
decisions about what assessment tools to use in their prac- colleagues also used specific psychometric criteria for
tices, how best to use and combine the various forms of many of these domains, such as describing a coefficient
information they obtain in their assessment, and how to α of .80 as exemplary. A different approach was taken by
integrate assessment activities into other necessary aspects the Measurement and Treatment Research to Improve
of clinical service. Moreover, the limited time available Cognition in Schizophrenia Group to develop a consen-
for service provision in clinical settings places an onus sus battery of cognitive tests to be used in clinical trials
on using assessment options that are maximally accurate, in schizophrenia (Green et al., 2004). Rather than setting
efficient, and cost-​effective. Thus, above and beyond the precise psychometric criteria for use in rating potential
scientific support that has been amassed for an instru- instruments, expert panelists were asked to rate, on a nine-​
ment, clinicians require tools that are brief, clear, clini- point scale, each proposed tool’s characteristics, includ-
cally feasible, and user-​friendly. In other words, they need ing test–​retest reliability, utility as a repeated measure,
instruments that have clinical utility and that are good relation to functional outcome, responsiveness to treat-
enough to get the job done (Barlow, 2005; Lambert & ment change, and practicality/​tolerability. An American
Hawkins, 2004; Weisz, Krumholz, Santucci, Thomassin, Psychological Association Society of Pediatric Psychology
& Ng, 2015; Youngstrom & Van Meter, 2016). task force used a fairly similar strategy. The task force
As has been noted in the assessment literature, there efforts, published at approximately the same time as the
are no clear, commonly accepted guidelines to aid clini- first edition of this volume, focused on evaluating psycho-
cians or researchers in determining when an instrument social assessment instruments that could be used in health
has sufficient scientific evidence to warrant its use (Kazdin, care settings (Cohen et al., 2008). Instrument character-
2005; Sechrest, 2005). The Standards for Educational and istics were reviewed by experts and, depending on the
Psychological Testing (American Educational Research available empirical support, were evaluated as promising,
Association, American Psychological Association, & approaching well-​established, or well-​established. These
National Council on Measurement in Education, descriptors closely resembled those that had been used to
2014) sets out generic standards to be followed in devel- identify empirically supported treatments.
oping and using psychological instruments but is silent Clearly, any attempt to develop a method for deter-
on the question of specific psychometric values that an mining the scientific adequacy of assessment instruments
instrument should have. The basic reason for this is that is fraught with the potential for error. The application
psychometric characteristics are not properties of an of criteria that are too stringent could result in a solid
instrument per se but, rather, are properties of an instru- set of assessment options, but one that is so limited in
ment when used for a specific purpose with a specific number or scope as to render the whole effort clinically
sample. Quite understandably, therefore, assessment worthless. Alternatively, using excessively lenient criteria
scholars, psychometricians, and test developers have been could undermine the whole notion of an instrument or
reluctant to explicitly indicate the minimum psycho- process being evidence based. So, with a clear awareness
metric values or evidence necessary to indicate that an of this assessment equivalent of Scylla and Charybdis, a
6 Introduction

decade ago we sought to construct a framework for the regarding critical aspects of the client’s biopsychosocial
chapters included in the first edition of this volume that functioning and context that are likely to influence the
would employ good-​enough criteria for rating psycho- client’s adjustment), (e) treatment design/​planning (i.e.,
logical instruments. In other words, rather than focus- selecting/​developing and implementing interventions
ing on standards that define ideal criteria for a measure, designed to address the client’s problems by focusing on
our intent was to provide criteria that would indicate the elements identified in the diagnostic evaluation and the
minimum evidence that would be sufficient to warrant case conceptualization), (f) treatment monitoring (i.e.,
the use of a measure for specific clinical purposes. We tracking changes in symptoms, functioning, psychologi-
assumed, from the outset, that although our framework cal characteristics, intermediate treatment goals, and/​or
is intended to be scientifically sound and defensible, it is variables determined to cause or maintain the problems),
a first step rather than the definitive effort in designing a and (g) treatment evaluation (i.e., determining the effec-
rating system for evaluating psychometric adequacy. Our tiveness, social validity, consumer satisfaction, and/​or cost-​
framework, described later, is unchanged from the first effectiveness of the intervention).
edition because there have been no developments in the The chapters in this volume provide summaries of
measurement and assessment literatures that have caused the best assessment methods and instruments available
us to reconsider our earlier position. Indeed, as we indi- for commonly encountered clinical assessment purposes.
cate in the following sections of this chapter, several criti- While recognizing the importance of other possible
cal developments have served to reinforce our views on assessment purposes, chapters in this volume focus on
the value of the framework. (a) diagnosis, (b) case conceptualization and treatment
In brief, to operationalize the good-​enough principle, planning, and (c) treatment monitoring and treatment
specific rating criteria are used across categories of psycho- evaluation. Although separable in principle, the purposes
metric properties that have clear clinical relevance; each of case conceptualization and treatment planning were
category has rating options of adequate, good, and excel- combined because they tend to rely on the same assess-
lent. In the following sections, we describe the assessment ment data. Similarly, the purposes of treatment monitor-
purposes covered by our rating system, the psychometric ing and evaluation were combined because they often,
properties included in the system, and the rationales for but not exclusively, use the same assessment methods
the rating options. The actual rating system, used in this and instruments. Clearly, there are some overlapping
volume by all authors of disorder/​problem-​oriented chap- elements, even in this set of purposes; for example, it is
ters to construct their summary tables of instruments, is relatively common for the question of diagnosis to be
presented in two boxes later in this chapter. revisited as part of evaluating the outcome of treatment.
In the instrument summary tables that accompany each
chapter, the psychometric strength of instruments used
ASSESSMENT PURPOSES for these three main purposes are presented and rated.
Within a chapter, the same instrument may be rated for
Although psychological assessments are conducted for more than one assessment purpose and thus appear in
many reasons, it is possible to identify a small set of inter- more than one table. Because an instrument may possess
related purposes that form the basis for most assessments. more empirical support for some purposes than for others,
These include (a) diagnosis (i.e., determining the nature the ratings given for the instrument may not be the same
and/​or cause[s]‌of the presenting problems, which may or in each of the tables.
may not involve the use of a formal diagnostic or catego- The chapters in this volume present information on
rization system), (b) screening (i.e., identifying those who the best available instruments for diagnosis, case con-
have or who are at risk for a particular problem and who ceptualization and treatment planning, and treatment
might be helped by further assessment or intervention), monitoring and evaluation. They also provide details on
(c) prognosis and other predictions (i.e., generating pre- clinically appropriate options for the range of data to col-
dictions about the course of the problems if left untreated, lect, suggestions on how to address some of the challenges
recommendations for possible courses of action to be commonly encountered in conducting assessments, and
considered, and their likely impact on the course of suggestions for the assessment process. Consistent with
the problems), (d) case conceptualization/​ formulation the problem-​specific focus within EBA outlined previ-
(i.e., developing a comprehensive and clinically rele- ously, most chapters in this volume focus on one or more
vant understanding of the client, generating hypotheses specific disorders or conditions. However, many clients
Developing Criteria for Evidence-Based Assessment 7

present with multiple problems and, therefore, there possible. The precise nature of what constituted adequate,
are frequent references within a given chapter to the good, and excellent varied, of course, from category to cat-
assessment of common co-​occurring problems that are egory. In general, however, a rating of adequate indicated
addressed in other chapters in the volume. To be opti- that the instrument meets a minimal level of scientific
mally useful to potential readers, the chapters are focused rigor; good indicated that the instrument would generally
on the most commonly encountered disorders or condi- be viewed as possessing solid scientific support; and excel-
tions among children, adolescents, adults, older adults, lent indicated there was extensive, high-​quality support-
and couples. With the specific focus on the three critical ing evidence. Accordingly, a rating of less than adequate
assessment purposes of diagnosis, case conceptualization indicated that the instrument did not meet the minimum
and treatment planning, and treatment monitoring and level set out in the criteria. A rating of not reported indi-
treatment, within each disorder or condition, the chapters cated that research on the psychometric property under
in this volume provide readers with essential information consideration had not yet been conducted or published.
for conducting the best EBAs currently possible. A rating of not applicable indicated that the psychomet-
ric property under consideration was not relevant to the
instrument (e.g., inter-​ rater reliability for a self-​
report
PSYCHOMETRIC PROPERTIES symptom rating scale).
AND RATING CRITERIA When considering the clinical use of a measure, it
would be desirable to only use those measures that would
Clinical assessment typically entails the use of both idio- meet, at a minimum, the criteria for good. However,
graphic and nomothetic instruments. Idiographic mea- because measure development is an ongoing process, the
sures are designed to assess unique aspects of a person’s rating system provides the option of the adequate rating
experience and, therefore, to be useful in evaluating in order to fairly evaluate (a) relatively newly developed
changes in these individually defined and constructed measures and (b) measures for which comparable levels
variables. In contrast, nomothetic measures are designed of research evidence are not available across all psycho-
to assess constructs assumed to be relevant to all indi- metric categories in the rating system. In several chapters,
viduals and to facilitate comparisons, on these constructs, authors explicitly commented on the status of some newly
across people. Most chapters include information on developed measures, but by and large, the only instru-
idiographic measures such as self-​monitoring forms and ments included in chapter summary tables were those
individualized scales for measuring treatment goals. For that had adequate or better ratings on the majority of the
such idiographic measures, psychometric characteristics psychometric dimensions. Thus, the instruments pre-
such as reliability and validity may, at times, not be easily sented in these tables represent only a subset of available
evaluated or even relevant (but see Weisz et al., 2011). It assessment tools.
is crucial, however, that the same items and instructions Despite the difficulty inherent in promulgating sci-
are used across assessment occasions—​without this level entific criteria for psychometric properties, we believe
of standardization it is impossible to accurately determine that the potential benefits of fair and attainable criteria
changes that may be due to treatment (Kazdin, 1993). far outweigh the potential drawbacks (cf. Sechrest, 2005).
The nine psychometric categories rated for the instru- Accordingly, reasoned arguments from respected psycho-
ments in this volume are norms, internal consistency, metricians and assessment scholars, along with summaries
inter-​rater reliability, test–​retest reliability, content valid- of various assessment literatures, guided the selection of
ity, construct validity, validity generalization, sensitivity criteria for rating the psychometric properties associated
to treatment change, and clinical utility. Each of these with an instrument. Box 1.1 presents the criteria used in
categories is applied in relation to a specific assessment rating norms and reliability indices; Box 1.2 presents the
purpose (e.g., case conceptualization and treatment plan- criteria used in rating validity indices and clinical utility.
ning) in the context of a specific disorder or clinical con-
dition (e.g., eating disorders, self-​injurious behavior, and
Norms
relationship conflict). Consistent with our previous com-
ments, factors such as gender, ethnicity, and age must be When using a standardized, nomothetically based
considered in making ratings within these categories. For instrument, it is essential that norms, specific criterion-​
each category, a rating of less than adequate, adequate, related cutoff scores, or both are available to aid in
good, excellent, not reported, or not applicable was the accurate interpretation of a client’s test score
8 Introduction

BOX 1.1 Criteria at a Glance: Norms and BOX 1.2 Criteria at a Glance: Validity and Utility
Reliability
C O N T E N T VA L I D I T Y
NORMS
Adequate = The test developers clearly defined
Adequate = Measures of central tendency and distribu- the domain of the construct being assessed and
tion for the total score (and subscores if relevant) based ensured that selected items were representative of
on a large, relevant, clinical sample are available. the entire set of facets included in the domain.
Good = Measures of central tendency and distribution Good = In addition to the criteria used for an
for the total score (and subscores if relevant) based adequate rating, all elements of the instrument
on several large, relevant samples (must include (e.g., instructions and items) were evaluated
data from both clinical and nonclinical samples) are by judges (e.g., by experts or by pilot research
available. participants).
Excellent = Measures of central tendency and distribu- Excellent = In addition to the criteria used for a
tion for the total score (and subscores if relevant) good rating, multiple groups of judges were
based on one or more large, representative samples employed and quantitative ratings were used by
(must include data from both clinical and nonclini- the judges.
cal samples) are available.
C O N S T R U C T VA L I D I T Y

I N T E R NA L C O N S I S T E N C Y Adequate = Some independently replicated evidence


Adequate = Preponderance of evidence indicates of construct validity (e.g., predictive validity, con-
α values of .70–​.79. current validity, and convergent and discriminant
Good = Preponderance of evidence indicates α values validity).
of .80–​.89. Good = Preponderance of independently repli-
Excellent = Preponderance of evidence indicates cated evidence, across multiple types of validity
α values ≥ .90. (e.g., predictive validity, concurrent validity, and
convergent and discriminant validity), is indica-
I N T E R -​R AT E R R E L I A B I L I T Y tive of construct validity.
Excellent = In addition to the criteria used for a good
Adequate = Preponderance of evidence indicates κ rating, there is evidence of incremental validity
values of .60–​.74; the preponderance of evidence with respect to other clinical data.
indicates Pearson correlation or intraclass correla-
tion values of .70–​.79. VA L I D I T Y G E N E R A L I Z AT I O N
Good = Preponderance of evidence indicates κ val-
Adequate = Some evidence supports the use of this
ues of .75–​ .84; the preponderance of evidence
instrument with either (a) more than one specific
indicates Pearson correlation or intraclass correla-
group (based on sociodemographic characteristics
tion values of .80–​.89.
such as age, gender, and ethnicity) or (b) in mul-
Excellent = Preponderance of evidence indicates κ
tiple contexts (e.g., home, school, primary care set-
values ≥ .85; the preponderance of evidence indi-
ting, and inpatient setting).
cates Pearson correlation or intraclass correlation
Good = Preponderance of evidence supports the
values ≥ .90.
use of this instrument with either (a) more than
one specific group (based on sociodemographic
T E S T –​R E T E S T R E L I A B I L I T Y
characteristics such as age, gender, and eth-
Adequate = Preponderance of evidence indicates nicity) or (b) in multiple settings (e.g., home,
test–​retest correlations of at least .70 over a period school, primary care setting, and inpatient
of several days to several weeks. setting).
Good = Preponderance of evidence indicates test–​ Excellent = Preponderance of evidence supports the
retest correlations of at least .70 over a period of use of this instrument with more than one specific
several months. group (based on sociodemographic characteristics
Excellent = Preponderance of evidence indicates such as age, gender, and ethnicity) and across mul-
test–​retest correlations of at least .70 over a period tiple contexts (e.g., home, school, primary care set-
of a year or longer. ting, and inpatient setting).
Developing Criteria for Evidence-Based Assessment 9

Regardless of the population to which comparisons are


BOX 1.2 Continued to be made, a normative sample must be truly represen-
T R E AT M E N T S E N S I T I V I T Y tative of the population with respect to demographics
and other important characteristics (Achenbach, 2001;
Adequate = Some evidence of sensitivity to change Wasserman & Bracken, 2013). Ideally, whether con-
over the course of treatment. ducted at the national level or the local level, this would
Good = Preponderance of independently replicated involve probability-​sampling efforts in which data are
evidence indicates sensitivity to change over the obtained from the majority of contacted respondents. As
course of treatment. those familiar with psychological instruments are aware,
Excellent = In addition to the criteria used for a good such a sampling strategy is rarely used for the devel-
rating, evidence of sensitivity to change across dif- opment of test norms. The reliance on data collected
ferent types of treatments. from convenience samples with unknown response rates
reduces the accuracy of the resultant norms. Therefore,
CLINICAL UTILITY at a minimum, clinicians need to be provided with
Adequate = Taking into account practical consid- an indication of the quality and likely accuracy of the
erations (e.g., costs, ease of administration, avail- norms for a measure. Accordingly, the ratings for norms
ability of administration and scoring instructions, required, at a minimum for a rating of adequate, data
duration of assessment, availability of relevant from a single, large clinical sample. For a rating of good,
cutoff scores, and acceptability to clients), the normative data from multiple samples, including non-
resulting assessment data are likely to be clini- clinical samples, were required; when normative data
cally useful. from large, representative samples were available, a rat-
Good = In addition to the criteria used for an ing of excellent was applied.
adequate rating, there is some published evi-
dence that the use of the resulting assessment Reliability
data confers a demonstrable clinical benefit
(e.g., better treatment outcome, lower treatment Reliability is a key psychometric element to be considered
attrition rates, and greater client satisfaction with in evaluating an instrument. It refers to the consistency of
services). a person’s score on a measure (Anastasi, 1988; Wasserman
Excellent = In addition to the criteria used for an & Bracken, 2013), including whether (a) all elements
adequate rating, there is independently replicated of a measure contribute in a consistent way to the data
published evidence that the use of the resulting obtained (internal consistency), (b) similar results would
assessment data confers a demonstrable clinical be obtained if the measure was used or scored by another
benefit. clinician (inter-​ rater reliability),1 or (c) similar results
would be obtained if the person completed the measure a
second time (test–​retest reliability or test stability). Not all
reliability indices are relevant to all assessment methods
(American Educational Research Association, American and measures, and the size of the indices may vary on the
Psychological Association, & National Council on basis of the samples used.
Measurement in Education, 2014). For example, Despite the long-​standing recognition of the central-
norms can be used to determine the client’s pre-​and ity of reliability to all forms of psychological measure-
post-​treatment levels of functioning and to evaluate ment, there is a persistent tendency in psychological
whether any change in functioning is clinically mean- research to make unwarranted assumptions about reli-
ingful (Achenbach, 2001; Kendall, Marrs-​Garcia, Nath, ability. For example, numerous reviews have found that
& Sheldrick, 1999). Selecting the target population(s) almost three-​fourths of research articles failed to provide
for the norms and then ensuring that the norms are information on the reliability estimates of the measures
adequate can be difficult tasks, and several sets of norms completed by participants in the studies (e.g., Barry,
may be required for a measure. One set of norms may be Chaney, Piazza-​ Gardner, & Chavarria, 2014; Vacha-​
needed to determine the meaning of the obtained score Haase & Thompson, 2011). Inattention to reliability, or
relative to the general population, whereas a different the use of an inappropriate statistic to estimate reliability,
set of norms could be used to compare the score to spe- has the potential to undermine the validity of conclu-
cific subgroups within the population (Cicchetti, 1994). sions drawn from research studies. Concerns have been
10 Introduction

raised about the impact of these errors in a broad range of of state-​like variables and life stress inventories), so test–​
research domains, including communication (Feng, 2015), retest reliability was only rated if it was relevant. A rating
psychopathology (Rodebaugh et al., 2016), and clinical of adequate required evidence of correlation values of .70
diagnosis (Chmielewski, Clark, Bagby, & Watson, 2015). or greater, when reliability was assessed over a period of
As emphasized throughout this volume, a careful consider- several days to several weeks. We then faced a challenge
ation of reliability values is essential when selecting assess- in determining appropriate criteria for good and excellent
ment instruments for clinical services or clinical research. ratings. In order to enhance its likely usefulness, the rating
With respect to internal consistency, we focused on system should be relatively simple. However, test–​retest
coefficient alpha (α), which is the most widely used index reliability is a complex phenomenon that is influenced by
(Streiner, 2003). Although there have been repeated (a) the nature of the construct being assessed (i.e., it can be
calls to abandon the use of coefficient α in favor of more state-​like, trait-​like, or influenced by situational variables),
robust and accurate alternatives (e.g., Dunn, Baguley, (b) the time frame covering the reporting period instruc-
& Brunsden, 2014; Kelley & Pornprasertmanit, 2016), tions (i.e., whether respondents are asked to report their
it is rare to find an internal consistency coefficient current functioning, functioning over the past few days,
other than α used in the clinical assessment literature. or functioning over an extended period, such as general
Recommendations in the literature for what constitutes functioning in the past year), and (c) the duration of the
adequate internal consistency vary, but most authorities retest period (i.e., whether the time between two admin-
seem to view .70 as the minimum acceptable value for α istrations of the instrument involved days, weeks, months,
(e.g., Cicchetti, 1994), and Charter (2003) reported that or years). In the end, rather than emphasize the value of
the mean internal consistency value among commonly increasingly large test–​retest correlations, we decided to
used clinical instruments was .81. Accordingly, a rating of maintain the requirement for .70 or greater correlation
adequate was given to values of .70–​.79; a rating of good values but require increasing retest period durations of
required values of .80–​.89; and, finally, because of cogent (a) several months and (b) at least 1 year for ratings of
arguments that an α value of at least .90 is highly desirable good and excellent, respectively.
in clinical assessment contexts (Nunnally & Bernstein,
1994), we required values ≥ .90 for an instrument to be
Validity
rated as having excellent internal consistency. Note that it
is possible for α to be too (artificially) high, as a value close Validity is another central aspect to be considered when
to unity typically indicates substantial redundancy among evaluating psychometric properties. Recent editions of
items (cf. Streiner, 2003). the Standards for Educational and Psychological Testing
These value ranges were also used in rating evidence (American Educational Research Association, American
for inter-​rater reliability when assessed with Pearson corre- Psychological Association, & National Council on
lations or intraclass correlations. Appropriate adjustments Measurement in Education, 1999, 2014) explicitly state
were made to the value ranges when κ statistics were used, that validity is a unitary concept and that it is not appro-
in line with the recommendations discussed by Cicchetti priate to consider different types of validity. Despite these
(1994; see also Charter, 2003). Note that although a num- admonitions, research on validity continues to use con-
ber of statistics are superior to κ, it continues to be the cepts such as content validity, predictive validity, and
most commonly used inter-​rater reliability statistic (Xu incremental validity. Setting aside the wide range of con-
& Lorber, 2014). Importantly, evidence for inter-​ rater ceptual and practical issues associated with the lack of
reliability could only come from data generated among consensus on the framing of test validity (for a detailed
clinicians or clinical raters—​estimates of cross-​informant discussion, see Newton & Shaw, 2013), there is a very
agreement, such as between parent and teacher ratings, simple reason for incorporating several types of validity
are not indicators of reliability. into the rating system used in this book: The vast majority
In establishing ratings for test–​retest reliability values, of the literature on clinical assessment, both historically
our requirement for a minimum correlation of .70 was and currently, does not treat validity as a unitary concept
influenced by summary data reported on typical test–​ (cf. Strauss & Smith, 2009). To strike a balance between
retest reliability results found with clinical instruments the unitary approach advocated by the Standards and
(Charter, 2003) and trait-​ like psychological measures the multiplicity of validity types used in the literature,
(Watson, 2004). Of course, not all constructs or measures we focused on content validity, construct validity, validity
are expected to show temporal stability (e.g., measures generalization, and treatment sensitivity. In the following
Developing Criteria for Evidence-Based Assessment 11

paragraphs, we explain further the rationale for our use of validity was also required. As was the case when we intro-
four types of validity. As is readily apparent by reviewing duced the rating system in the first edition of this book, we
the summary tables of instruments in the following chap- were unable to find any clearly applicable standards in the
ters, the extent and strength of research evidence across literature to guide us in developing criteria for validity gen-
these types of validity can vary substantially for a given eralization or treatment sensitivity (a dimension rated only
assessment instrument. for instruments used for the purposes of treatment moni-
Foster and Cone (1995) drew an important distinc- toring and treatment evaluation). Therefore, adequate
tion between “representational” validity (i.e., whether a ratings for these dimensions required some evidence of,
measure really assesses what it purports to measure) and respectively, the use of the instrument with either more
“elaborative” validity (i.e., whether the measure has any than one specific group or in multiple contexts and evi-
utility for measuring the construct). Attending to the dence of sensitivity to change over the course of treatment.
content validity of a measure is a basic, but frequently Consistent with ratings for other dimensions, good and
overlooked, step in evaluating representational valid- excellent ratings required increasingly demanding levels
ity (Haynes, Richard, & Kubany, 1995). As discussed by of evidence in these areas.
Smith, Fischer, and Fister (2003), the overall reliability
and validity of scores on an instrument is directly affected
Utility
by the extent to which items in the instrument adequately
represent the various aspects or facets of the construct the It is also essential to know the utility of an instrument for
instrument is designed to measure. Assuming that repre- a specific clinical purpose. The concept of clinical util-
sentational validity has been established for an assessment ity, applied to both diagnostic systems (e.g., Keeley et al.,
purpose, it is elaborative validity that is central to clini- 2016; Kendell & Jablensky, 2003; Mullins-​Sweatt, Lengel,
cians’ use of a measure. Accordingly, replicated evidence & DeShong, 2016) and assessment tools (e.g., di Ruffano,
for a measure’s concurrent, predictive, discriminative, Hyde, McCaffery, & Bossuyt, 2012; Yates & Taub, 2003),
and, ideally, incremental validity (Hunsley & Meyer, has received increasing attention in recent years. Although
2003) should be available to qualify a measure for con- definitions vary, they have in common an emphasis on gar-
sideration as evidence based. We have indicated already nering evidence regarding actual improvements in both
that validation is a context-​sensitive concept—​inattention decisions made by clinicians and service outcomes experi-
to this fact can lead to inappropriate generalizations being enced by clients. Unfortunately, despite thousands of studies
made about a measure’s validity. Thus, there should be on the reliability and validity of psychological instruments,
replicated elaborative validity evidence for each purpose there is only scant attention paid to matters of utility in most
of the measure and for each population or group for which assessment research studies (McGrath, 2001). This has
the measure is intended to be used. This latter point is directly contributed to the current state of affairs in which
especially relevant when considering an instrument for there is very little replicated evidence that psychological
clinical use, and thus it is essential to consider evidence assessment data have a direct impact on improved provision
for validity generalization—​that is, the extent to which and outcome of clinical services. Currently, therefore, for
there is evidence for validity across a range of samples and the majority of psychological instruments, a determination
settings (cf. Messick, 1995; Schmidt & Hunter, 1977). of clinical utility must often be made on the basis of likely
In the rating system used in subsequent chapters, rat- clinical value rather than on empirical evidence.
ings of content validity evidence required explicit consider- Compared to the criteria for the psychometric dimen-
ation of the construct facets to be included in the measure sions presented thus far, our standards for evidence of
and, as the ratings increased, involvement of content clinical utility were noticeably less demanding. This was
validity judges to assess the measure (Haynes et al., 1995). necessary because of the paucity of information on the
Unlike the situation for reliability, there are no commonly extent to which assessment instruments are acceptable
accepted summary statistics to evaluate construct validity to clients, enhance the quality and outcome of clinical
(but see Markon [2013] and Westen & Rosenthal [2003]). services, and/​or are worth the costs associated with their
As a result, our ratings were based on the requirement of use. Therefore, we relied on authors’ expert opinions to
increasing amounts of replicated evidence of elements of classify an instrument as having adequate clinical utility.
construct validity such as predictive validity, concurrent The availability of any supporting evidence of utility was
validity, convergent validity, and discriminant validity; in sufficient for a rating of good, and replicated evidence of
addition, for a rating of excellent, evidence of incremental utility was necessary for a rating of excellent.
12 Introduction

The instrument summary tables also contain one final For information on important developments on rating
column, used to indicate instruments that are the best systems used in many areas of health care research, the
measures currently available to clinicians for specific pur- interested reader can consult the website of the Grading
poses and disorders and, thus, are highly recommended of Recommendations Assessment, Development and
for clinical use. Given the considerable differences in Evaluation (GRADE) working group (http://​www.grade-
the state of the assessment literature for different disor- workinggroup.org).
ders/​conditions, chapter authors had some flexibility in The second issue has to do with the responsible clini-
determining their own precise requirements for an instru- cal use of the guidance provided by the rating system.
ment to be rated, or not rated, as highly recommended. Consistent with evaluation and grading strategies used
However, to ensure a moderate level of consistency in through evidence-​ based medicine and evidence-​ based
these ratings, a highly recommended rating could only be psychology initiatives, many of our rating criteria relied
considered for those instruments having achieved ratings on the consideration of the preponderance of data rel-
of good or excellent in the majority of its rated psycho- evant to each dimension. Such a strategy recognizes both
metric categories. Although not required in our system, if the importance of replication in science and the fact
several instruments had comparable psychometric merits that variability across studies in research design elements
for a given assessment purpose, some chapter authors con- (including sample composition and research setting) will
sidered the cost and availability of an assessment instru- influence estimates of these psychometric dimensions.
ment when making this recommendation (see also Beidas However, we hasten to emphasize that reliance on the
et al., 2015). preponderance of evidence for these ratings does not
imply or guarantee that an instrument is applicable for all
clients or clinical settings. Our intention is to have these
SOME FINAL THOUGHTS ratings provide indications about scientifically strong mea-
sures that warrant consideration for clinical and research
We are hopeful that the rating system described in this use. As with all evidence-​based efforts, the responsibility
chapter, and applied in each of the chapters of this book, rests with the individual professional to determine the
will continue to aid in advancing the state of evidence-​ suitability of an instrument for the specific setting, pur-
based psychological assessment. We also hope that it will pose, and individuals to be assessed.
serve as a stimulus for others to refine and improve upon Third, as emphasized throughout this volume, focus-
our efforts (cf. Jensen-​Doss, 2011; Youngstrom et al., in ing on the scientific evidence for specific assessment tools
press). Whatever the possible merits of the rating system, should not overshadow the fact that the process of clinical
we close this chapter by drawing attention to three critical assessment involves much more than simply selecting and
issues related to its use. administering the best available instruments. Choosing
First, although the rating system used for this volume the best, most relevant, instruments is unquestionably an
is relatively simple, the task of rating psychometric proper- important step. Subsequent steps must ensure that the
ties is not. Results from many studies must be considered instruments are administered in an appropriate manner,
in making such ratings, and precise quantitative standards accurately scored, and then individually interpreted in
were not set for how to weight the results from studies. accordance with the relevant body of scientific research.
Furthermore, in the spirit of evidence-​based practice, it However, to ensure a truly evidence-​based approach to
is also important to note that we do not know whether assessment, the major challenge is to then integrate all the
these ratings are, themselves, reliable. Reliance on indi- data within a process that is, itself, evidence-​based. This
vidual expert judgment, no matter how extensive and will likely require both (a) a reframing of the assessment
current the knowledge of the experts, is not as desirable process within the larger health and social system context
as basing evidence-​based conclusions and guidance on in which it occurs and (b) the use of new technologies
systematic reviews of the literature conducted according to enable complex decision-​making and integration of
to a consensually agreed upon rating system. However, large amounts of assessment information in both tradi-
for all the potential limitations and biases inherent in our tional and nontraditional health service delivery settings
approach, reliance on expert review of the scientific lit- (Chorpita, Daleiden, & Bernstein, 2015). Much of our
erature is the current standard in psychology and, thus, focus in this chapter has been on evidence-​based meth-
was the only feasible option for the volume at this time. ods and instruments, in large part because (a) methods
Developing Criteria for Evidence-Based Assessment 13

and specific measures are more easily identified than are review of seven journals. Health Education & Behavior,
processes and (b) the main emphasis in the assessment lit- 41, 12–​18.
erature has been on psychometric properties of methods Beidas, R. S., Stewart, R. E., Walsh, L., Lucas, S., Downey,
and instruments. As we indicated early in the chapter, an M. M., Jackson, K., . . . Mandell, D. S. (2015). Free,
evidence-​based approach to assessment should be devel- brief, and validated: Standardized instruments for
low-​resource mental health settings. Cognitive and
oped in light of evidence on the accuracy and usefulness
Behavioral Practice, 22, 5–​19.
of this complex, iterative decision-​making task. Although
Brown, J., Scholle, S. H., & Azur, M. (2014). Strategies for
the chapters in this volume provide considerable assis- measuring the quality of psychotherapy: A white paper
tance for having the assessment process be informed by to inform measure development and implementation
scientific evidence, the future challenge will be to ensure (Prepared for the U.S. Department of Health and
that the entire process of assessment is evidence based. Human Services). Retrieved from https://​aspe.hhs.
gov/​report/​strategies-​measuring-​quality-​psychotherapy-​
white- ​ p aper- ​ i nform- ​ m easure- ​ d evelopment- ​ a nd-​
Note implementation
Charter, R. A. (2003). A breakdown of reliability coeffi-
1. Although we chose to use the term “inter-​rater reli-
cients by test type and reliability method, and the clini-
ability,” there is some discussion in the assessment literature
cal implications of low reliability. Journal of General
about whether the term should be “inter-​rater agreement.”
Psychology, 130, 290–​304.
Heyman et al. (2001), for example, suggested that because
Chmielewski, M., Clark, L. A., Bagby, R. M., & Watson, D.
indices of inter-​rater reliability do not contain information
(2015). Method matters: Understanding diagnostic reli-
about individual differences among participants and only
ability in DSM-​IV and DSM-​5. Journal of Abnormal
contain information about one source of error (i.e., differ-
Psychology, 124, 764–​769.
ences among raters), they should be considered to be indices
Chorpita, B. F., Daleiden, E. L., & Bernstein, A. D. (2015).
of agreement, not reliability.
At the intersection of health information technol-
ogy and decision support: Measurement feedback
systems . . . and beyond. Administration and Policy in
References Mental Health and Mental Health Services Research,
Achenbach, T. M. (2001). What are norms and why do 43, 471–​477.
we need valid ones? Clinical Psychology: Science and Christon, L. M., McLeod, B. D., & Jensen-​Doss, A. (2015).
Practice, 8, 446–​450. Evidence-​based assessment meets evidence-​based treat-
Achenbach, T. M. (2005). Advancing assessment of chil- ment: An approach to science-​informed case conceptu-
dren and adolescents: Commentary on evidence-​based alization. Cognitive and Behavioral Practice, 22, 74–​86.
assessment of child and adolescent disorders. Journal Cicchetti, D. V. (1994). Guidelines, criteria, and rules
of Clinical Child and Adolescent Psychology, 34, of thumb for evaluating normed and standardized
541–​547. assessment instruments in psychology. Psychological
American Educational Research Association, American Assessment, 6, 284–​290.
Psychological Association, & National Council on Cohen, L. L., La Greca, A. M., Blount, R. L., Kazak, A.
Measurement in Education. (1999). Standards for E., Holmbeck, G. N., & Lemanek, K. L. (2008).
educational and psychological testing. Washington, Introduction to the special issue: Evidence-​ based
DC: American Educational Research Association. assessment in pediatric psychology. Journal of Pediatric
American Educational Research Association, American Psychology, 33, 911–​915.
Psychological Association, & National Council on di Ruffano, L. F., Hyde, C. J., McCaffery, K. J., & Bossuyt, P.
Measurement in Education. (2014). Standards for M. M. (2012). Assessing the value of diagnostic tests: A
educational and psychological testing. Washington, framework for designing and evaluating trials. British
DC: American Educational Research Association. Medical Journal, 344, e686.
Anastasi, A. (1988). Psychological testing (6th ed.). New York, Dozois, D. J. A., Mikail, S., Alden, L. E., Bieling, P. J.,
NY: Macmillan. Bourgon, G., Clark, D. A., . . . Johnston, C. (2014). The
Barlow, D. H. (2005). What’s new about evidence based CPA presidential task force on evidence-​based practice
assessment? Psychological Assessment, 17, 308–​311. of psychological treatments. Canadian Psychology, 55,
Barry, A. E., Chaney, B. H., Piazza-​ Gardner, A. K., & 153–​160.
Chavarria, E. A. (2014). Validity and reliability report- Dunn, T. J., Baguley, T., & Brunsden, V. (2014). From alpha
ing in the field of health education and behavior: A to omega: A practical solution to the pervasive problem
14 Introduction

of internal consistency estimation. British Journal of Jensen-​Doss, A. (2011). Practice involves more than treat-
Psychology, 105, 399–​412. ment: How can evidence-​based assessment catch up to
Feng, G. C. (2015). Mistakes and how to avoid mistakes evidence-​based treatment? Clinical Psychology: Science
in using intercoder reliability indices. Methodology, and Practice, 18, 173–​177.
11, 13–​22. Kazdin, A. E. (1993). Evaluation in clinical prac-
Fisher, A. J. (2015). Toward a dynamic model of psycho- tice: Clinically sensitive and systematic methods of
logical assessment: Implications for personalized care. treatment delivery. Behavior Therapy, 24, 11–​45.
Journal of Consulting and Clinical Psychology, 83, Kazdin, A. E. (2003). Psychotherapy for children and adoles-
825–​836. cents. Annual Review of Psychology, 54, 253–​276.
Foster, S. L., & Cone, J. D. (1995). Validity issues in clinical Kazdin, A. E. (2005). Evidence-​based assessment of child and
assessment. Psychological Assessment, 7, 248–​260. adolescent disorders: Issues in measurement develop-
Green, M. F., Nuechterlein, K. H., Gold, J. M., Barch, D. ment and clinical application. Journal of Clinical Child
M., Cohen, J., Essock, S., . . . Marder, S. R. (2004). and Adolescent Psychology, 34, 548–​558.
Approaching a consensus cognitive battery for clinical Keeley, J. W., Reed, G. M., Roberts, M. C., Evans, S. C.,
trials in schizophrenia: The NIMH-​ MATRICS con- Medina-​Mora, M. E., Robles, R., . . . Saxena, S. (2016).
ference to select cognitive domains and test criteria. Developing a science of clinical utility in diagnostic
Biological Psychiatry, 56, 301–​307. classification systems field study strategies for IDC-​11
Haynes, S. N., Richard, D. C. S., & Kubany, E. S. (1995). mental and behavioral disorders. American Psychology,
Content validity in psychological assessment: A func- 71, 3–​16.
tional approach to concepts and methods. Psychological Kelley, K., & Pornprasertmanit, S. (2016). Confidence inter-
Assessment, 7, 238–​247. vals for population reliability coefficients: Evaluation of
Hermann, R. C., Chan, J. A., Zazzali, J. L., & Lerner, D. methods, recommendations, and software for composite
(2006). Aligning measure-​based quality improvement measures. Psychological Methods, 21, 69–​92.
with implementation of evidence-​ based practices. Kendall, P. C., Marrs-​Garcia, A., Nath, S. R., & Sheldrick, R.
Administration and Policy in Mental Health and Mental C. (1999). Normative comparisons for the evaluation of
Health Services Research, 33, 636–​645. clinical significance. Journal of Consulting and Clinical
Heyman, R. E., Chaudhry, B. R., Treboux, D., Crowell, J., Psychology, 67, 285–​299.
Lord, C., Vivian, D., & Waters, E. B.. (2001). How Kendell, R., & Jablensky, A. (2003). Distinguishing between
much observational data is enough? An empirical test the validity and utility of psychiatric diagnoses. American
using marital interaction coding. Behavior Therapy, 32, Journal of Psychiatry, 160, 4–​12.
107–​123. Krishnamurthy, R., VandeCreek, L., Kaslow, N. J., Tazeau,
Hsu, L. M. (2002). Diagnostic validity statistics and the Y. N., Miville, M. L., Kerns, R., . . . Benton, S. A.
MCMI-​III. Psychological Assessment, 14, 410–​422. (2004). Achieving competency in psychological assess-
Hunsley, J., Lee, C. M., Wood, J., & Taylor, W. (2015). ment: Directions for education and training. Journal of
Controversial and questionable assessment techniques. Clinical Psychology, 60, 725–​739.
In S. O. Lilienfeld, S. J. Lynn, & J. Lohr (Eds.), Science Lambert, M. J. (2015). Progress feedback and the OQ-​
and pseudoscience in clinical psychology (2nd ed., pp. system: The past and the future. Psychotherapy, 52,
42–​82). New York, NY: Guilford. 381–​390.
Hunsley, J., & Mash, E. J. (2007). Evidence-​based assess- Lambert, M. J., & Hawkins, E. J. (2004). Measuring out-
ment. Annual Review of Clinical Psychology, 3, 57–​79. come in professional practice: Considerations in select-
Hunsley, J., & Meyer, G. J. (2003). The incremental valid- ing and using brief outcome instruments. Professional
ity of psychological testing and assessment: Conceptual, Psychology: Research and Practice, 35, 492–​499.
methodological, and statistical issues. Psychological Lee, C. M., Horvath, C., & Hunsley, J. (2013). Does it
Assessment, 15, 446–​455. work in the real world? The effectiveness of treatments
Hurl, K., Wightman, J., Haynes, S. N., & Virues-​Ortega, J. for psychological problems in children and adoles-
(2016). Does a pre-​interventions functional assessment cents. Professional Psychology: Research and Practice,
increase intervention effectiveness? A meta-​analysis of 44, 81–​88.
within-​subjects interrupted time-​series studies. Clinical Markon, K. E. (2013). Information utility: Quantifying the
Psychology Review, 47, 71–​84. total psychometric information provided by a measure.
Ivanova, M. Y., Achenbach, T. M., Rescorla, L. A., Turner, Psychological Methods, 18, 15–​35.
L. V., Ahmeti-​Pronaj, A., Au, A., . . . Zasepa, E. (2015). Mash, E. J., & Barkley, R. A. (Eds.). (2007). Assessment of
Syndromes of self-​ reported psychopathology for ages childhood disorders (4th ed.). New York, NY: Guilford.
18–​59 in 29 societies. Journal of Psychopathology and Mash, E. J., & Hunsley, J. (2005). Evidence-​based assessment
Behavioral Assessment, 37, 171–​183. of child and adolescent disorders: Issues and challenges.
Developing Criteria for Evidence-Based Assessment 15

Journal of Clinical Child and Adolescent Psychology, 34, tools. Clinical Psychology: Science and Practice, 23,
362–​379. 265–​283.
Mash, E. J., & Hunsley, J. (2007). Assessment of child and Schmidt, F. L., & Hunter, J. E. (1977). Development of a
family disturbance: A developmental systems approach. general solution to the problem of validity generaliza-
In E. J. Mash & R. A. Barkley (Eds.), Assessment of tion. Journal of Applied Psychology, 62, 529–​540.
childhood disorders (4th ed., pp. 3–​ 50). New York, Sechrest, L. (2005). Validity of measures is no simple matter.
NY: Guilford. Health Services Research, 40, 1584–​1604.
Mash, E. J., & Terdal, L. G. (1997). Assessment of child and Seidman, E., Chorpita, B. F., Reay, B., Stelk, W., Kutash,
family disturbance: A behavioral-​ systems approach. K., Mullican, C., & Ringeisen, H. (2010). A framework
In E. J. Mash & L. G. Terdal (Eds.), Assessment of for measurement feedback to improve decision-​making
childhood disorders (3rd ed., pp. 3–​ 68). New York, in mental health. Administration and Policy in Mental
NY: Guilford. Health and Mental Health Services Research, 37,
McGrath, R. E. (2001). Toward more clinically relevant 128–​131.
assessment research. Journal of Personality Assessment, Smith, G. T., Fischer, S., & Fister, S. M. (2003). Incremental
77, 307–​332. validity principles in test construction. Psychological
Messick, S. (1995). Validity of psychological assess- Assessment, 15, 467–​477.
ment: Validation of inferences from persons’ responses Spilka, M. J., & Dobson, K. S. (2015). Promoting the interna-
and performances as scientific inquiry into score mean- tionalization of evidence-​based practice: Benchmarking
ing. American Psychologist, 50, 741–​749. as a strategy to evaluate culturally transported psycho-
Mullins-​Sweatt, S. N., Lengel, G. J., & DeShong, H. L. logical treatments. Clinical Psychology: Science and
(2016). The importance of considering clinical util- Practice, 22, 58–​75.
ity in the construction of a diagnostic manual. Annual Strauss, M. E., & Smith, G. T. (2009). Construct valid-
Review of Clinical Psychology, 12, 133–​155. ity: Advances in theory and methodology. Annual
Newby, J. M., McKinnon, A., Kuyken, W., Gilbody, S., & Review of Clinical Psychology, 5, 89–​113.
Dalgleish, T. (2015). Systematic review and meta-​ Streiner, D. L. (2003). Starting at the beginning: An intro-
analysis of transdiagnostic psychological treatments for duction to coefficient alpha and internal consistency.
anxiety and depressive disorders in adulthood. Clinical Journal of Personality Assessment, 80, 99–​103.
Psychology Review, 40, 91–​110. Streiner, D. L., Norman, G. R., & Cairney, J. (2015).
Newton, P. E., & Shaw, S. D. (2013). Standards for talking Health measurement scales: A practical guide to their
and thinking about validity. Psychological Methods, 18, development and use (5th ed.). New York, NY: Oxford
301–​319. University Press.
Ng, M. Y., & Weisz, J. R. (2016). Building a science of per- Thompson-​Hollands, J., Sauer-​Zavala, S., & Barlow, D. H.
sonalized intervention for youth mental health. Journal (2014). CBT and the future of personalized treatment: A
of Child Psychology and Psychiatry, 57, 216–​236. proposal. Depression and Anxiety, 31, 909–​911.
Norcross, J. C., Koocher, G. P., & Garofalo, A. (2006). Discredited Vacha-​Haase, T., & Thompson, B. (2011). Score reliabil-
psychological treatments and tests: A Delphi poll. ity: A retrospective look back at 12 years of reliability
Professional Psychology: Research and Practice, 37, 515–​522. generalization studies. Measurement and Evaluation in
Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric the- Counseling and Development, 44, 159–​168.
ory (3rd ed.). New York, NY: McGraw-​Hill. Vermeersch, D. A., Lambert, M. J., & Burlingame, G.
Peterson, D. R. (2004). Science, scientism, and professional M. (2000). Outcome Questionnaire: Item sensitiv-
responsibility. Clinical Psychology: Science and Practice, ity to change. Journal of Personality Assessment, 74,
11, 196–​210. 242–​261.
Robinson, J. P., Shaver, P. R., & Wrightsman, L. S. (1991). Wasserman, J. D., & Bracken, B. A. (2013). Fundamental
Criteria for scale selection and evaluation. In J. P. psychometric considerations in assessment. In J. R.
Robinson, P. R. Shaver, & L. S. Wrightsman (Eds.), Graham & J. A. Naglieri (Eds.), Handbook of psychol-
Measures of personality and social psychological atti- ogy: Volume 10. Assessment psychology (2nd ed., pp.
tudes (pp. 1–​16). New York, NY: Academic Press. 50–​81). Hoboken, NJ: Wiley.
Rodebaugh, T. L., Sculling, R. B., Langer, J. K., Dixon, D. Watson, D. (2004). Stability versus change, dependability
J., Huppert, J. D., Bernstein, A., . . . Lenze, E. J. (2016). versus error: Issues in the assessment of personality over
Unreliability as a threat to understanding psychopa- time. Journal of Research in Personality, 38, 319–​350.
thology: The cautionary tale of attentional bias. Journal Weisz, J. R., Chorpita, B. F., Frye, A., Ng, M. Y., Lau, N.,
of Abnormal Psychology, 125, 840–​851. Bearman, S. K., & Hoagwood, K. E. (2011). Youth Top
Sales, C. M. D., & Alves, P. C. G. (2016). Patient-​centered Problems: Using idiographic, consumer-​guided assess-
assessment in psychotherapy: A review of individualized ment to identify treatment needs and to track change
16 Introduction

during psychotherapy. Journal of Consulting and kappa. Journal of Consulting and Clinical Psychology,
Clinical Psychology, 79, 369–​380. 82, 1219–​1227.
Weisz, J. R., & Kazdin, A. E. (Eds.). (2017). Evidence-​based Yates, B. T., & Taub, J. (2003). Assessing the costs, ben-
psychotherapies for children and adolescents (3nd ed.). efits, cost-​
effectiveness, and cost–​benefit of psycho-
New York, NY: Guilford. logical assessment: We should, we can, and here’s how.
Weisz, J. R., Krumholz, L. S., Santucci, L., Thomassin, K., & Psychological Assessment, 15, 478–​495.
Ng, M. Y. (2015). Shrinking the gap between research Youngstrom, E. A., Choukas-​Bradley, S., Calhoun, C. D., &
and practice: Tailoring and testing youth psychothera- Jensen-​Doss, A. (2015). Clinical guide to the evidence-​
pies in clinical care contexts. Annual Review of Clinical based approach to diagnosis and treatment. Cognitive
Psychology, 11, 139–​163. and Behavioral Practice, 22, 20–​35.
Westen, D., & Rosenthal, R. (2003). Quantifying construct Youngstrom, E. A., & Van Meter, A. (2016). Empirically sup-
validity: Two simple measures. Journal of Personality ported assessment of children and adolescents. Clinical
and Social Psychology, 84, 608–​618. Psychology: Science and Practice, 23, 327–​347.
Wright, C. V., Beattie, S. G., Galper, D. I., Church, A. Youngstrom, E. A., Van Meter, A., Frazier, T. W., Hunsley,
S., Bufka, L. F., Brabender, V. M., & Smith, B. L. J., Prinstein, M. J., Ong, M.-​L., & Youngstrom, J. K.
(2017). Assessment practices of professional psy- (in press). Evidence-​based assessment as an integrative
chologists: Results of a national survey. Professional model for applying psychological science to guide the
Psychology: Research and Practice, 48, 73–​78. voyage of treatment. Clinical Psychology: Science and
Xu, S., & Lorber, M. F. (2014). Interrater agreement statistics Practice.
with skewed data: Evaluation of alternatives to Cohen’s
17

Dissemination and Implementation


of Evidence-​Based Assessment

Amanda Jensen-​Doss
Lucia M. Walsh
Vanesa Mora Ringle

During the past two decades, there has been a major Burns, 2003; Pogge et al., 2001) suggests that improved
push to increase the use of evidence-​based practices in diagnostic assessment could have a positive effect across
clinical settings. The American Psychological Association the treatment process. As detailed throughout this book,
Presidential Task Force on Evidence-​ Based Practice evidence-​based diagnostic assessment typically involves
(2006) defines evidence-​ based practice in psychology the use of structured diagnostic interviews or rating scales.
(EBPP) as “the integration of the best available research Studies have demonstrated that when clinicians use
with clinical expertise in the context of patient character- structured diagnostic interviews, they assign more accu-
istics, culture, and preferences” (p. 271) and states that rate diagnoses (Basco et al., 2000), better capture comor-
the goal of EBPP is to improve public health through bidities (Matuschek et al., 2016), assign more specific
the application of research-​ supported assessment, case diagnoses (Matuschek et al., 2016), reduce psychiatrist
formulation, therapeutic relationship, and treatment evaluation time (Hughes et al., 2005), and decrease the
approaches. Although EBPP is defined broadly, many likelihood that a psychiatrist will increase a patient’s medi-
efforts to improve practice have focused on treatment, cation dose (Kashner et al., 2003).
with less attention paid to other aspects of practice. There Using EBA for progress monitoring can support clinical
is a particular need to focus on increasing use of evidence-​ judgment by creating an ongoing feedback loop to support
based assessment (EBA), as assessment cuts across all of ongoing case conceptualization (Christon et al., 2015) and,
these other areas of practice. For example, assessment if data suggest clients are at risk for treatment failure, revise
results should form the foundation of case conceptualiza- the treatment plan (Claiborn & Goodyear, 2005; Lambert,
tion; inform decisions about which treatments to use; and Hansen, & Finch, 2001; Riemer, Rosof-​Williams, & Bickman,
provide data about whether treatment is working, whether 2005). Gold standard progress monitoring of this nature typi-
therapy alliance is strong, and when to end treatment. cally involves administering rating scales every session or two
There are several reasons why a focus on EBA will and then incorporating the feedback into clinical decisions.
increase the likelihood that EBPP will lead to improved This differs from what we refer to here as “outcome monitor-
public health. First, EBA can improve the accuracy of ing,” or administering outcome measures before and after
diagnoses, which is one important component of case treatment to determine treatment effectiveness. Although use-
conceptualization and treatment selection (Christon, ful for many purposes, this type of outcome monitoring does
McLeod, & Jensen-​Doss, 2015). Research linking diag- not support clinical decision-​making during service provision.
nostic accuracy to improved treatment engagement and Several monitoring and feedback systems (MFSs) have
outcomes (Jensen-​Doss & Weisz, 2008; Klein, Lavigne, been developed to support ongoing progress monitoring;
& Seshadri, 2010; Kramer, Robbins, Phillips, Miller, & they typically include a battery of progress measures and

17
18

18 Introduction

generate feedback reports that often include warnings monitoring. As discussed in the following sections, these
if a client is not on track for positive outcomes (Lyon, gaps have important implications for the accuracy of
Lewis, Boyd, Hendrix, & Liu, 2016). Extensive research clinician-​generated diagnoses and the accuracy of clini-
with adult clients suggests that clinician use of MFSs can cian judgments about treatment progress.
improve outcomes, particularly for those “not on track”
for positive outcomes (Krägeloh, Czuba, Billington,
Research–​Practice Gaps in Diagnostic Assessment
Kersten, & Siegert, 2015; Shimokawa, Lambert, & Smart,
2010); similar results have been found for youth clients As detailed throughout the chapters in this book,
(Bickman, Kelley, Breda, De Andrade, & Riemer, 2011; evidence-​based diagnostic assessment for most disorders
Stein, Kogan, Hutchison, Magee, & Sorbero, 2010), relies on standardized diagnostic interviews and/​or rat-
although effects vary based on organizational support for ing scales. Unfortunately, surveys of training programs
progress monitoring (Bickman et al., 2016). suggest that clinicians are not being prepared to conduct
The purpose of this chapter is to make the case that these assessments during their graduate training. Several
significant work is needed to encourage the dissemina- surveys of psychology programs have been conducted in
tion of information about EBA and its implementation in the past three decades (e.g., Belter & Piotrowski, 2001;
clinical practice. First, we discuss “assessment as usual,” Childs & Eyde, 2002), with the two most recent (Mihura,
how it differs from EBA, and reasons for these differences. Roy, & Graceffo, 2016; Ready & Veague, 2014) finding
Then, we describe efforts to increase use of EBA through that training in assessment has generally remained con-
dissemination and implementation efforts. Finally, we stant, but there has been an increase in training focused
present some ideas for future work needed to further on assessment of treatment outcomes, psychometrics,
advance the use of EBA. Consistent with the other chap- and neuropsychology. However, these two studies found
ters in this book, we focus on assessment of psychopathol- inconsistent results regarding the use of clinical inter-
ogy and its application to clinical diagnosis and progress viewing. Ready and Veague reported only half to three-​
monitoring. Although similar issues likely exist for other fourth of programs included clinical interviewing as a
forms of assessment, such as psychoeducational assess- focus of training. However, Mihura et al. found that 92%
ment, discussion of those is beyond the scope of this book. of programs queried included clinical interviewing as a
Finally, although assessment tools to support case con- required topic. These differences might reflect a change
ceptualization are described in the subsequent chapters during the 3 years that passed between the two studies.
of this book, most of the literature studying clinician case However, it is also likely that the two studies also obtained
conceptualization practices and how to improve them has information from different programs, as Mihura and col-
focused on whether clinicians can apply specific theoreti- leagues included more programs than Ready and Veague
cal models to client data and the validity of case concep- and each study only obtained data from approximately
tualizations (e.g., Abbas, Walton, Johnston, & Chikoore, one-​third of all of the American Psychological Association
2012; Flinn, Braham, & das Nair, 2015; Persons & (APA)-​accredited programs.
Bertagnolli, 1999) rather than on how to collect and inte- Two studies have examined training in diagnostic
grate EBA data to generate a case conceptualization. As assessment specifically. Ponniah et al. (2011) surveyed
mentioned previously, both diagnostic assessment and clinical training directors from social work, clinical psy-
progress monitoring can support case conceptualiza- chology PhD and PsyD, and psychiatric residency pro-
tion; therefore, much of the literature we discuss here grams regarding training in structured assessment based
has implications for case conceptualization. However, in on Diagnostic and Statistical Manual of Mental Disorders
the Future Directions section, we address steps needed to (DSM) criteria. Only one-​ third of surveyed programs
advance assessment-​based case conceptualization. reported providing both didactics and supervision in
diagnostic assessment, with clinical psychology PhD and
psychiatry residency programs being most likely to do so
IS THERE A RESEARCH–​P RACTICE and social work programs the least likely. These results are
GAP IN ASSESSMENT? concerning because master’s level clinicians represent the
majority of those providing services to those with mental
Despite the proliferation of excellent assessment tools, health disorders in the United States (Garland, Bickman,
available data suggest there are significant training and & Chorpita, 2010). Focusing on clinical psychology PhD
practice gaps in both diagnostic assessment and progress and PsyD programs exclusively, Mihura et al. (2016)
19

Dissemination and Implementation of Evidence-Based Assessment 19

found that less than half of the programs required a course different types of doctoral programs (e.g., counseling vs.
and less than one-​fourth required an applied practicum PsyD; Overington, Fitzpatrick, Hunsley, & Drapeau,
on any structured diagnostic interview. Differences in 2015) and training program models (e.g., practitioner-​
required structured diagnostic interview courses were scholar models vs. clinical-​scientist models; Overington
found between training models: 73% of clinical science et al., 2015), although little is known about progress moni-
and 63% of scientist-​ practitioner programs required a toring training in master’s programs. Similar to training
course, whereas only 35% of practitioner-​ focused pro- programs, fewer than half of internship directors report
grams had a similar requirement. having their trainees use progress monitoring (Ionita,
Not surprisingly based on these training gaps, available Fitzpatrick, Tomaro, Chen, & Overington, 2016; Mours,
data suggest that clinicians are not engaged in EBA for Campbell, Gathercoal, & Peterson, 2009), and nearly
diagnostic assessment. Existing surveys across a range of one-​third of directors have never heard of progress moni-
clinicians indicate that unstructured interviews are com- toring measures (Ionita et al., 2016).
monly relied on for diagnosis (e.g., Anderson & Paulosky, Similar to diagnostic assessment, there are low rates
2004), that evidence-​based tools are infrequently used of progress monitoring among practicing clinicians.
(e.g., Gilbody, House, & Sheldon, 2002; Whiteside, Much of the research in this area is focused on outcome
Sattler, Hathaway, & Douglas, 2016), and diagnostic monitoring (Cashel, 2002; Hatfield & Ogles, 2004), with
practices often do not map on to best practice guidelines relatively less focus given to ongoing progress monitor-
(e.g., Demaray, Schaefer, & Delong, 2003; Lichtenstein, ing. Surveys suggest that many clinicians report track-
Spirito, & Zimmermann, 2010). ing client progress (Anderson & Paulosky, 2004; Gans,
Unfortunately, these gaps between “best practice” Falco, Schackman, & Winters, 2010). However, many
and “as usual” assessment practices have implications for of them are not using validated measures, instead rely-
the accuracy of diagnoses generated in routine practice. ing on tools developed within the clinic, unstructured
Studies comparing clinician-​generated diagnoses to those interviews, reports from clients, and clinical judgment
generated through comprehensive, research-​ supported (Anderson & Paulosky, 2004; Gans et al., 2010; Johnston
methods consistently find low rates of agreement between & Gowers, 2005). This finding has been supported in
the two (Rettew, Lynch, Achenbach, Dumenci, & two recent large surveys of psychologists and master’s
Ivanova, 2009; Samuel, 2015). Studies examining the level clinicians, in which fewer than 15% of clinicians
validity of clinician-​generated diagnoses also suggest that engaged in ongoing progress monitoring (Ionita &
these diagnoses are less valid than the evidence-​based diag- Fitzpatrick, 2014; Jensen-​Doss et al., 2016). Clinicians
noses (Basco et al., 2000; Jewell, Handwerk, Almquist, & who do use progress monitoring measures appear to be
Lucas, 2004; Mojtabai, 2013; Samuel et al., 2013; Tenney, using these measures to track progress internally and
Schotte, Denys, van Megen, & Westenberg, 2003). for administrative purposes, but they rarely report using
them to plan treatment or monitor progress (Garland,
Kruse, & Aarons, 2003).
Research–​Practice Gaps in Progress Monitoring
This lack of formal progress monitoring is concern-
Most of what is known about graduate training in progress ing in light of data showing that it is difficult for clini-
monitoring focuses on trainee psychologists. As described cians to accurately judge client progress based on clinical
throughout this volume, most progress monitoring tools judgment alone. For example, when Walfish, McAlister,
are standardized rating scales, so many of the assessment O’Donnell, and Lambert (2012) asked a multidisciplinary
training gaps discussed previously also are relevant for sample of clinicians to rate their own level of skills, none of
progress monitoring. However, other surveys have focused them ranked themselves as below average, and one-​fourth
on whether trainees are trained in utilizing these scales of them rated themselves at the 90th percentile of clinical
for ongoing progress monitoring, rather than for diagnos- skill relative to their peers. These therapists estimated that
tic assessment. With regard to APA accredited psychol- three-​fourths of their clients improved in therapy, with
ogy programs, Ready and Veague (2014) found that only less than 5% deteriorating; nearly half of them said that
approximately half of programs offer courses focused on none of their clients ever deteriorated. The study authors
progress monitoring, and Mihura et al. (2016) found only point out that these estimates deviated wildly from pub-
10% of programs require their students to routinely use lished estimates of improvement and deterioration rates.
outcome measures in their practical. Differing rates of Interestingly, available data suggest that even when clini-
training in progress monitoring have been found between cians are trained to engage in progress monitoring, this
20

20 Introduction

does not improve their ability to rate progress based on Catchpoole, 2006; Overington et al., 2015). Clinicians also
clinical judgment alone. Hannan and colleagues (2005) report anxiety about progress monitoring data being used
removed feedback reports from a setting that had been for performance evaluation, use of these measures ruining
using an MFS and asked clinicians to predict their clients’ rapport, concern regarding how to present results correctly
outcomes. Those clinicians underestimated how many to clients, and a general lack of knowledge about progress
clients would deteriorate or not improve, and they over- monitoring (Ionita et al., 2016; Johnston & Gowers, 2005;
estimated how many would improve. Meehan et al., 2006). In a study that separately asked about
attitudes toward the practice of progress monitoring and
attitudes toward standardized progress measures, clinicians
WHY IS THERE A RESEARCH–​P RACTICE GAP reported very positive attitudes toward the idea of monitor-
IN ASSESSMENT? ing progress but more neutral attitudes toward the mea-
sures themselves (Jensen-​Doss et al., 2016), suggesting that
As detailed previously, lack of training is likely one factor clinicians are open to engaging in the practice if their con-
that contributes to clinicians’ not utilizing best assessment cerns about the measures themselves can be addressed.
practices. In addition, research has identified other clini- Consistent with research on diagnostic assessment, more
cian and organizational variables that might be contribut- positive attitudes toward progress monitoring are associ-
ing to this research–​practice gap. ated with higher rates of self-​reported progress monitor-
Several studies have focused on clinician attitudes ing practices (Hatfield & Ogles, 2004; Jensen-​Doss et al.,
that might be driving assessment practices. Jensen-​Doss 2016; Overington et al., 2015).
and Hawley (2010, 2011) conducted a national, multi- A number of organizational barriers and facilitators of
disciplinary survey to assess clinicians’ attitudes toward EBA have also been identified in the literature. Lack of
standardized assessment tools, with a particular focus on organizational support is a barrier frequently mentioned by
diagnostic assessment. On average, clinicians reported clinicians, including both active discouragement from super-
neutral to positive attitudes toward standardized assessment visors and administration regarding the use of measures and
tools, although this varied by discipline, with psychologists little guidance given by organizational leaders of when and
reporting more positive attitudes compared to psychiatrists, how often to use them (Connors, Arora, Curtis, & Stephan,
marriage and family therapists, social workers, and mental 2015; Gilbody et al., 2002; Ionita et al., 2016; Overington
health counselors (Jensen-​Doss & Hawley, 2010). Attitudes, et al., 2015). Many of the practical concerns described previ-
particularly beliefs about the practicality of standardized ously also speak to organizational factors, such as the amount
assessment tools, predicted self-​reported use of these tools. of time clinicians are allowed to spend on assessment and the
Other studies have found that clinicians have concerns that budget available for purchasing assessment tools. Clinicians
structured diagnostic interviews would be unacceptable also often report that administrators are more interested in
to clients (Bruchmüller, Margraf, Suppiger, & Schneider, tracking administrative outcomes (e.g., length of wait list, cli-
2011), although data gathered directly from clients do not ent turnover, and number of sessions) than outcomes such
support this view (Suppiger et al., 2009). as functioning and symptom reduction (Gilbody et al., 2002;
Studies have also examined clinician attitudes toward Johnston & Gowers, 2005). Conversely, clinicians who indi-
progress monitoring. Across studies, attitudes toward these cate their organizations have policies or rules about assess-
measures have varied from neutral to positive, although ment are more likely to report using progress monitoring
concerns regarding the validity of the measures (e.g., (Jensen-​Doss et al., 2016). Clinician assessment practices
whether they accurately reflected client progress) are com- also vary across organizational settings; providers working
mon (Cashel, 2002; Gilbody et al., 2002; Hatfield & Ogles, in private practice settings are less likely to use standardized
2007; Ionita et al., 2016; Johnston & Gowers, 2005). As diagnostic and progress monitoring tools than are those work-
with diagnostic assessment, clinicians often report practi- ing in other settings (Jensen-​Doss et al., 2016; Jensen-​Doss &
cal concerns about progress monitoring, including lim- Hawley, 2010).
ited access to affordable measures, measures being too
long, difficulties reaching clients to fill out measures, and
little time to administer measures and keep track of when EFFORTS TO IMPROVE ASSESSMENT PRACTICES
to fill out measures (Gleacher et al., 2016; Hatfield &
Ogles, 2004; Ionita et al., 2016; Johnston & Gowers, 2005; The studies reviewed previously indicate that although
Kotte et al., 2016; Meehan, McCombes, Hatzipetrou, & effective assessment tools exist, they often are not making
21

Dissemination and Implementation of Evidence-Based Assessment 21

their way into practice settings. As such, several efforts that assess outcomes for various health domains that pro-
have been made to bridge this research–​practice gap, mote and facilitate outcome and progress monitoring
some focused on specific evidence-​based measures (e.g., (http://​www.healthmeasures.net/​explore-​measurement-​
rating scales for trauma; National Child Traumatic Stress systems/​promis; Cella et al., 2010; HealthMeasures,
Network, 2016) and others focused on EBA processes 2017; Pilkonis et al., 2011). In a novel approach to dis-
(e.g., using an MFS to gather data and using feedback to semination, the APA has recently funded a grant to update
make decisions about treatment; Bickman et al., 2016). assessment pages on Wikipedia, with a focus on assess-
Efforts to improve clinician assessment practices can be ments that are freely available (Youngstrom, Jensen-​Doss,
divided into dissemination efforts, or efforts to inform Beidas, Forman, & Ong, 2015–​2016).
clinicians about EBA tools, and implementation efforts
that seek to support clinicians in their use of such tools.
Training Efforts
Implementation efforts can be subdivided into those
focused on training clinicians in EBA; those focused on Some groups are moving beyond dissemination to pro-
implementing EBA in individual organizations; and those vide training in EBA to clinicians. Relative to the numer-
focused on integrating EBA into mental health systems, ous studies focused on training clinicians in treatments
such as state public mental health systems. Although a (Herschell, Kolko, Baumann, & Davis, 2010), there are
comprehensive review of all of these efforts is beyond the fewer EBA training studies. Documented EBA training
scope of this chapter, we highlight some illustrative exam- efforts to date have consisted of workshops, workshops
ples of each approach. plus ongoing consultation, and courses. Didactic training
workshops have helped improve clinician progress moni-
toring attitudes (Edbrooke-​Childs, Wolpert, & Deighton,
Dissemination Efforts
2014; Lyon, Dorsey, Pullmann, Silbaugh-​ Cowdin, &
Assessment-​focused dissemination efforts have typically Berliner, 2015), self-​ efficacy (Edbrooke-​Childs et al.,
created sources for clinicians to identify evidence-​based 2014), and use (Persons, Koerner, Eidelman, Thomas, &
measures or guides for them to engage in EBA processes. Liu, 2016).
This volume is an example of an EBA dissemination Another training approach is to follow workshops with
effort, as are publications in EBA special journal issues ongoing consultation. For example, a training effort in
(Hunsley & Mash, 2005; Jensen-​Doss, 2015; Mash & Washington state included 6 months of expert-​led phone
Hunsley, 2005) and review papers, such as Leffler, Riebel, consultation and found that training impacted clini-
and Hughes’ (2015) review of structured diagnostic inter- cian attitudes, skill, and implementation of standardized
views for clinicians. The DSM board has also embarked assessment tools (Lyon et al., 2015).
on efforts to improve diagnostic practices and accuracy by Finally, online training has recently been applied to
outlining steps for diagnosis and creating decision trees EBA training. For example, Swanke and Zeman (2016)
to support differential diagnosis (First, 2013). Although created an online course in diagnostic assessment for mas-
these dissemination efforts have typically focused on ter’s level social work students. The course was based on a
what clinicians should do, Koocher and Norcross have problem-​based learning approach wherein students were
also published articles identifying discredited assessment given diagnostic problems to solve by identifying symp-
methods (Koocher, McMann, Stout, & Norcross, 2015; toms and matching symptoms to DSM diagnoses. At the
Norcross, Koocher, & Garofalo, 2006). end of the course, the average student grade on content
There are also efforts to disseminate EBA informa- quizzes was 78.7% and the class was well-​received by the
tion online. For example, there is a website dedicated students, although students’ levels of knowledge prior to
to information about measures relevant to the assess- the course are not know, so it is difficult to determine
ment of traumatized youth (http://​www.nctsn.org/​ whether the course actually increased knowledge.
resources/​online-​research/​measures-​review; National
Child Traumatic Stress Network, 2016), a repository of
Organizational-​Level Implementation Efforts
information about assessment tools relevant to child wel-
fare populations (http://​www.cebc4cw.org/​assessment-​ Another approach to increasing use of EBA is for orga-
tools/​measurement-​tools-​highlighted-​on-​the-​cebc; The nizations to attempt to change assessment practices
California Evidence-​ Based Clearinghouse for Child organization-​wide. Several examples of such efforts have
Welfare, 2017), and the PROMIS website with measures been documented in the literature, including studies
2

22 Introduction

examining the impact of organizations incorporating requiring evidence-​based assessment. System-​level imple-
structured diagnostic interviews (e.g., Basco et al., 2000; mentations documented in the literature have primarily
Lauth, Levy, Júlíusdóttir, Ferrari, & Pétursson, 2008; focused on progress monitoring. An early example was
Matuschek et al., 2016) and progress monitoring systems the state of Michigan’s use of the Child and Adolescent
(e.g., Bickman et al., 2011, 2016; Bohnenkamp, Glascoe, Functional Assessment Scale (CAFAS; Hodges & Wong,
Gracey, Epstein, & Benningfield, 2015; Strauss et al., 1996). As described by Hodges and Wotring (2004), clini-
2015; Veerbeek, Voshaar, & Pot, 2012). cians in the public mental health system were required
One illustrative example of organizational-​ level to use the CAFAS to track client outcomes. Data were
implementation work focused on progress monitoring is then used to provide clinicians and agencies feedback on
the work of Bickman and colleagues. Following an ini- individual client and agency-​wide outcomes, including
tial successful randomized effectiveness trial showing that comparison to agency and state averages.
using an MFS called Contextualized Feedback System Hawaii, which has been a pioneer in the advancement
(CFS) improved client outcomes (Bickman et al., 2011), of evidence-​based treatments (EBTs) in the public sec-
Bickman and colleagues (2016) conducted a second ran- tor, has supported these efforts by developing and imple-
domized trial within two mental health organizations. All menting an MFS that is used statewide (Higa-​McMillan,
clinicians within the agencies were required to administer Powell, Daleiden, & Mueller, 2011; Kotte et al., 2016;
CFS, and cases were randomly assigned to receive feed- Nakamura et al., 2014). To date, both clinicians and case-
back as soon as measures were entered into the system workers across various agencies in the state have been
(i.e., clinicians immediately received feedback reports trained in and are implementing the MFS. In an effort to
summarizing the CFS data) or to receiving feedback encourage the use of EBA, Higa-​McMillan et al. reported
every 6 months. Before the trial began, the investiga- the use of “Provider Feedback Data Parties” during which
tors conducted a “pre-​implementation contextualization client progress and clinical utilization of the data are dis-
phase,” during which they held workgroups to understand cussed. Other studies on Hawaii’s EBA efforts observed
existing clinic procedures and brainstorm about how CFS that the fit between the MFS and case manager character-
would fit into those procedures. Training and ongoing istics facilitated MFS implementation, whereas provider
consultation in CFS was provided to clinicians and to concerns about the clinical utility and scientific merit of
agency administrators to ensure both clinical (i.e., using it the MFS were reported as barriers (Kotte et al., 2016).
with individual clients) and organizational (e.g., ongoing Internationally, system-​ level efforts to implement
review of aggregated data to identify problems with CFS progress monitoring have been reported in the United
implementation) use of CFS. After finding that only one Kingdom and Australia. Efforts to implement routine
clinic demonstrated enhanced outcomes with CFS, the monitoring throughout the United Kingdom have been
authors determined that the two agencies differed in their ongoing for well over a decade (Fleming, Jones, Bradley,
rates of questionnaire completion and viewing of feed- & Wolpert, 2016; Hall et al., 2014; Mellor-​Clark, Cross,
back reports. To better understand these findings, they Macdonald, & Skjulsvik, 2016). The Child Outcomes
then conducted qualitative interviews with the participat- Research Consortium (CORC; http://​www.corc.uk.net),
ing clinicians (Gleacher et al., 2016). Clinicians at the a learning and planning collaboration of researchers, ther-
clinic with better implementation and outcomes reported apists, managers, and funders, has spearheaded most of
more barriers to using CFS with their clients than did this work. CORC has made valid, reliable, brief, and free
clinicians at the other clinic, perhaps because they were measures available to all clinicians working in the United
using it more often. However, they also reported fewer Kingdom, provided training in the measures, and created
barriers at the organizational level and more support from an MFS to support their use. These measures are reported
their organizational leadership. The authors concluded to be widely implemented, but not at an optimal level
that organizational factors are strong drivers of implemen- (Mellor-​Clark et al., 2016), so efforts are now focused on
tation success. adopting more theory-​driven approaches to implement-
ing the system (Mellor-​Clark et al., 2016; Meyers, Durlak,
& Wandersman, 2012). In Australia, efforts to implement
System-​Level Efforts
progress monitoring have been ongoing since the late
Another approach to implementation is for mental health 1990s and include training and development of computer
systems, such as state public mental health agencies or systems to support data collection and analysis (Meehan
agencies like the Veteran’s Administration, to enact policies et al., 2006; Trauer, Gill, Pedwell, & Slattery, 2006).
23

Dissemination and Implementation of Evidence-Based Assessment 23

Outcome data are collected at all public clinics and are to attain competence in diagnosis of clients via mea-
aggregated at a national level to be used for comparison surement and interviews and to assess treatment effec-
by local clinics. tiveness, but they gave little guidance regarding what
Finally, note that policies focused on other aspects of constitutes appropriate assessment (APA, 2006; Canadian
care can also have implications for assessment. For exam- Psychological Association, 2011; Ponniah et al., 2011).
ple, the Precision Medicine Initiative (The White House A similar picture exists in accreditation guidelines for
Office of the Press Secretary, 2015) focuses on increasing mental health counseling (American Mental Health
personalized medical treatments that take individual dif- Counselors Association, 2011), marriage and family
ferences in genes and environment into account. Such therapy (Commission on Accreditation for Marriage
tailored approaches are likely going to require increased and Family Therapy Education, 2014), and bachelor’s
use of psychosocial assessment in health care settings. and master’s level social work programs (Commission on
Similarly, the US Medicare and Medicaid system is mov- Accreditation & Policy, 2015), although these guidelines
ing increasingly toward value-​ based payment, where do include training in progress monitoring as a way to per-
reimbursement is based on quality, rather than quantity, form program evaluation.
of care (Centers of Medicare & Medicaid Services, 2016). In January 2017, a new set of accreditation guide-
As such, assessment of quality indicators within publicly lines for the APA went into effect that include EBA as
funded behavioral health settings is going to become a core competency (APA Commission on Accreditation,
increasingly important. Finally, initiatives to implement 2015). A recent Canadian task force focused on increas-
EBTs often lead to the development of assessment pro- ing EBP use (Task Force on Evidence-​Based Practice of
cesses to support those treatments, as evidenced by the Psychological Treatments; Dozois et al., 2014) empha-
Hawaii initiative described previously. sized monitoring progress and outcomes throughout treat-
ment. However, the Canadian Psychological Association
accreditation guidelines for doctoral programs have not
FUTURE DIRECTIONS been updated to reflect this change as of this publication.
2. Increase “best practice” training strategies in EBA
As we hope this review has made clear, the literature dissemination and implementation efforts. Although
on EBA contains both good and bad news. On the one exceptions exist (Lyon et al., 2015), the primary approach
hand, a number of excellent EBA tools exist and some that has been taken to training clinicians in EBA is
efforts are underway to encourage clinician use of those what is sometimes referred to as a “train and pray”
tools. On the other hand, significant gaps continue to approach: Bring clinicians together for a workshop and
exist between assessment best practices and what the then hope they take what they have learned and apply it
average clinician does in practice. To address these gaps, in practice. The literature on training in EBTs suggests
we have several suggestions for future directions the field that such an approach is unlikely to lead to sustained
should take. practice changes (Herschell et al., 2010). Rather, train-
1. Increase graduate-​level training in evidence-​based ing needs to involve active learning strategies, ongoing
diagnostic assessment and progress monitoring. Most of the consultation in the practice, and attention to contextual
training and implementation efforts described previously variables such as whether clinicians have adequate orga-
have primarily focused on retraining clinicians whose nizational support to continue using the practice (Beidas
graduate training likely did not include in-​depth training & Kendall, 2010; Herschell et al., 2010). Examples of
in structured diagnostic assessment or progress monitor- strategies that could be incorporated into EBA trainings
ing. Researchers focused on EBTs have called for an include engaging clinicians in behavioral rehearsal dur-
increased focus on training at the graduate level because ing training (Beidas, Cross, & Dorsey, 2014); providing
training people well at the outset is likely easier and more ongoing consultation after initial training (e.g., Bickman
cost-​effective than trying to retrain them (e.g., Bearman, et al., 2016; Lyon et al., 2015); increasing sustainability of
Wadkins, Bailin, & Doctoroff, 2015). assessment practices through “train the trainer” models
One avenue for improving graduate training is increas- that train agency supervisors to provide ongoing supervi-
ing the specificity of accreditation guidelines for training sion assessment (Connors et al., 2015); and incorporat-
programs (Dozois et al., 2014; Ponniah et al., 2011). For ing all levels of an agency into training through learning
both psychology and psychiatry training programs, past collaborative models that address implementation at the
accreditation standards stressed the need for students clinician, supervisor, and administrator levels (e.g., Ebert,
24

24 Introduction

Amaya-​Jackson, Markiewicz, Kisiel, & Fairbank, 2012; reporting requirements. Lyon and Lewis (2016) point out
Nadeem, Olin, Hill, Hoagwood, & Horwitz, 2014). these shifts provide an opportunity to increase the use
3. Increase our focus on pragmatic assessment. Studies of progress monitoring. In a recent review, Lyon, Lewis,
conducted with clinicians consistently suggest that Boyd, Hendrix, and Liu (2016) identified 49 digital MFSs
perceived lack of practicality is a major barrier to clini- that could be used by clinicians with access to comput-
cian use of EBA (e.g., Ionita et al., 2016; Jensen-​Doss & ers or tablets to administer progress measures and rap-
Hawley, 2010). In addition, the fact that many clinicians idly receive feedback. However, fewer than one-​third of
who do gather assessment data do not actually incorpo- those were able to be directly incorporated into electronic
rate that data into clinical decisions (Garland et al., 2003; health care records, and Lyon and colleagues concluded
Johnston & Gowers, 2005) suggests that they may not that additional work is needed to develop digital MFSs
find the data clinically useful. Glasgow and Riley (2013) that can be incorporated into the daily workflow of prac-
have called for the field to focus on pragmatic measures, tice in a way that is sustainable.
which they define as measures “that [have] relevance to Another technological advance with great potential
stakeholders and [are] feasible to use in most real-​world to enhance assessment is smartphone technologies that
settings to assess progress” (p. 237). They propose criteria support data collection. Researchers have developed
for determining whether a measure is pragmatic, includ- applications to support real-​time data collection (Trull
ing that is it important to stakeholders, such as clients, & Ebner-​Priemer, 2009) and have begun to examine the
clinicians, or administrators; that it is low burden to com- clinical utility of such applications for gathering informa-
plete; that it generates actionable information that can be tion such as mood (e.g., Schwartz, Schultz, Reider, &
used in decision-​making; and that it is sensitive to change Saunders, 2016) or pain ratings (Sánchez-​Rodríguez, de
over time. Expanding our reviews of EBA tools to include la Vega, Castarlenas, Roset, & Miró, 2015). Such applica-
dimensions such as these might help identify measures tions could facilitate self-​monitoring of symptoms between
most likely to make their way into practice. One example sessions or efficient collection and scoring of progress
of such a review was conducted by Beidas and colleagues monitoring data in session. Many smartphone applications
(2015), who identified brief, free measures and rated their to track psychological well-​being are already commercially
psychometric support for a range of purposes, including available (e.g., a November 15, 2016, search of the Google
screening, diagnosis, and progress monitoring. Play store yielded more than 50 results for “mood track-
Another opportunity for increasing the practical- ing”), and an important next step is to determine how
ity of assessment is to take advantage of recent policies these applications can be ethically developed and incorpo-
emphasizing increased data collection and accountability rated into clinical practice (Jones & Moffitt, 2016).
in health care settings (e.g., the “Patient Protection and 5. Develop theoretical models of organizational support
Affordable Care Act,” 2010). Lyon and Lewis (2016) point for EBA. Despite numerous studies suggesting that orga-
out the opportunity that these policies provide for increas- nizational context is critical to EBA (e.g., Gleacher et al.,
ing use of progress monitoring. As agencies increasingly 2016; Jensen-​Doss et al., 2016), there is a need for concep-
incorporate health information technologies, such as tual models that can guide organizational approaches to
electronic medical records, into their settings to meet improving assessment practices. Models of organizational
data reporting requirements, there is an opportunity to culture and climate have been developed to explain use
integrate electronic MFSs into these systems (Lyon et al., of EBTs (e.g., Williams & Glisson, 2014) and have been
2016). If progress monitoring can be built into the daily translated into organizational interventions that improve
workflow of clinicians, this greatly increases its practicality. EBT uptake and client outcomes (e.g., Glisson, Williams,
4. Leverage technology to increase the use of EBA. Hemmelgarn, Proctor, & Green, 2016). Many aspects of
Another avenue for increasing the practicality of assess- these models are likely applicable to the use of EBA, but
ment is to incorporate technologies such as electronic the constructs within them may need to be elaborated.
health care records platforms and smartphone applica- Although existing models might be helpful to guide
tions into the assessment process. With the rise of policies EBA implementation in agency settings such as clinics or
emphasizing increased data collection and accountabil- schools, these models are not as applicable to clinicians
ity in health care settings (e.g., “Patient Protection and working in private practice, who seem to be the clinicians
Affordable Care Act,” 2010), mental health settings are least likely to engage in EBA (Jensen-​Doss et al., 2016).
increasingly relying on health information technolo- Additional work is needed to understand the needs of this
gies, such as electronic health care records, to meet data population.
Another random document with
no related content on Scribd:
sentiva spuntarsi una lacrima. Non poteva a meno di rievocare il
tempo in cui que’ suoi cari rallegravano il tetto domestico e
scherzavano sulle ginocchia materne. Ahimè, ormai la madre era
morta da un pezzo.... Involontariamente l’occhio del colonnello si
posava sulla parete ove sotto i ritratti di Vittorio Emanuele, di
Umberto, di Garibaldi, di Lamarmora, di Napoleone III, ecc., ecc.,
c’era un gruppo di fotografie di famiglia. La più antica e sbiadita era
appunto quella di sua moglie, una donna esile, dall’aria stanca e
sofferente. A fianco di lei Bedeschi in persona, in uniforme, con la
sua medaglia al valor militare sul petto, con la sua mano bravamente
piantata sull’elsa della sciabola. Qualche linea più sotto l’effigie di tre
giovinotti, Vittorio, Augusto, ed un terzo, minore di tutti.
Sicuro, c’era un terzo figliuolo, Federico, ed egli solo non s’era fatto
vivo in quel giorno, e da Londra, dove si trovava, non aveva spedito
nè una lettera, nè un dispaccio. Quando Bedeschi pensava a questo
ragazzo ch’era stato il suo preferito egli si doleva di aver ceduto una
volta tanto alle preghiere di sua moglie, la quale, impuntatasi
nell’idea che Federico fosse di salute cagionevole, aveva, tra gemiti
e singhiozzi, scongiurato il marito di non fargli abbracciar la carriera
militare come i fratelli, e strappatagliene un giorno a malincuore la
promessa, se l’era fatta rinnovare solennemente al letto di morte.
Federico era quindi rimasto in casa, aveva frequentate le scuole
pubbliche, ed era giunto senza gloria fino all’università. Non gli
mancava nè cuore nè ingegno, ma aveva uno spirito indisciplinato,
ripugnante a studi regolari, turbato piuttosto da vaghe inquietudini
d’artista. Onde nel bel mezzo del corso di legge gli era saltato il
ghiribizzo di darsi alla pittura, con grande sdegno del colonnello, il
quale nè amava l’arte, nè credeva a questa vocazione improvvisa.
N’eran seguite scene violente, per merito delle quali Federico aveva
finito col non studiar nè pittura nè legge e col menare una vita oziosa
e dissipata. Allora il padre gli aveva posto un dilemma. O mettersi in
grado di prendere la laurea entro un anno, o partire subito per
Londra, ove un antico compagno di cospirazioni del colonnello,
arricchitosi nel commercio, impiegava volentieri dei giovani italiani
per mandarli, dopo un tirocinio più o meno lungo, presso le sue case
filiali di San Francisco o di Sidney. Federico che della laurea non
voleva saperne accettò la seconda proposta; meglio far il minatore in
California o il pastore in Australia che incretinirsi su una scranna di
giudice o assottigliare il cervello nei cavilli avvocateschi.
E partì con una cert’aria spavalda che il colonnello, cattivo psicologo,
attribuì a perversità d’animo, mentre Federico, dal canto suo,
risentiva profondamente l’affettata indifferenza del padre. Come
avviene quando c’è un equivoco che non si chiarisce subito, la
freddezza reciproca andò a mano a mano crescendo; padre e
figliuolo si accusavano in silenzio di poco cuore e non si
scambiavano che lettere brevi, fredde e insignificanti.
Frattanto entrò in casa la Bice portando nella dimora solitaria un
nuovo alito di giovinezza, togliendole quell’aspetto triste e desolato
ch’essa aveva nelle prime settimane dell’assenza di Federico. Senza
volerlo, senza saperlo, la fanciulla nuoceva al cugino. Una frase
dello zio lo mise in guardia. — Non parliamo di quello scapato. —
egli le disse. — Ora sei tu che ne tieni il posto.
Tenere il posto di Federico? No, ciò non poteva, non doveva essere.
Ed ella dichiarò allo zio che prima che accadesse una cosa simile
sarebbe tornata in collegio, sicura di farvisi accettare dalla direttrice
come assistente.
Alla lunga si calmò, ma fermando il proposito di esercitar tutta la
propria influenza per sopire quel dissidio domestico. Pur non tardò
ad accorgersi che l’impresa era ardua ed esigeva infinite cautele.
Non le fu difficile mettersi in relazione con Federico, avendola lo zio
stesso incaricata talora di scrivergli in vece sua. E Federico le
rispose in principio diffidente e guardingo, poi, via via, più sciolto ed
espansivo. A lei rivelava la tristezza del suo esilio, l’acuta nostalgia
da cui era sovente assalito, la sua ripugnanza ad allontanarsi ancora
di più dall’Italia, la sua sfiducia assoluta di far buona prova nella
mercatura. Ma soprattutto le discorreva dell’arte, ch’egli aveva
ripreso ad amar con passione, che coltivava in segreto, e nella quale
avrebbe potuto forse non esser degli ultimi se gli fosse stato
permesso di dedicarvisi intero.
Il colonnello Bedeschi aveva tempra di despota, non d’inquisitore, e
avrebbe stimato inferiore alla sua dignità lo spiar le corrispondenze
della nipote. Delle lettere ch’ella riceveva da Federico egli sapeva
quel tanto che a lei piaceva di dirgliene, ed è naturale ch’ella gliene
presentasse un’edizione riveduta e corretta. Accennava alla
condotta regolare del giovine, al desiderio ch’egli manifestava di
riacquistare l’affetto e la stima del padre.... soggiungendo
timidamente che a parer suo non c’era ragione di tenerlo più oltre in
castigo a Londra, e meno che mai di spedirlo in capo al mondo.
La prima volta che la Bice toccò questo tasto, Bedeschi montò su
tutte le furie. — O ch’ella pretendeva di dargli lezioni? Ella, una
bambina, con quell’esperienza che aveva? Badasse ai casi suoi e
non s’impicciasse di ciò che non la riguardava. Se Federico le
scriveva delle sciocchezze, padrone; e padrona lei di rispondergliene
altrettante, ma non venisse a far la saccente. Aveva capito?
La fanciulla non si smarrì d’animo per questo rabbuffo nè perdette di
vista la sua meta. A ogni occasione opportuna ella tornava alla
carica, sopportando in santa pace le sfuriate dello zio, il quale, in
cuor suo, non si rammaricava troppo ch’ella difendesse il cugino. Ma
il colonnello aveva riputazione d’uomo forte, d’uomo inflessibile, e
certe riputazioni sono come un patrimonio da conservare. Bedeschi
non voleva che si dicesse ch’egli s’infemminiva cogli anni. Accadeva
poi un fatto curioso. Quantunque egli non osasse confessarlo a sè
stesso, la Bice gli diventava più cara per la sua generosità nel
prender le parti di Federico, e appunto col diventargli più cara gli
rendeva meno sensibile la mancanza del figlio.
S’era sbagliata strada. La Bice lo riconobbe e mutò tattica. Da due o
tre mesi ella non parlava di Federico che quand’era strettamente
necessario il parlarne, pareva rassegnata non solo alla relegazione
del cugino a Londra, ma anche alla sua partenza per Sidney o San
Francisco.
— È frivola e obliosa come tutte le donne, — pensava il vecchio
soldato. — Que’ suoi grandi ardori battaglieri sono sbolliti.
E non le sapeva grado della sua docilità. Era meno sicuro di aver
ragione dacchè nessuno gli dava torto.
Ella intanto ne pesava le parole, ne scrutava i silenzi, i gesti,
l’espressione della fisonomia, arrischiando di tratto in tratto con finta
ingenuità una frase, una domanda, come un generale che spinge
innanzi i suoi esploratori per esaminare il terreno.
— Se Federico deve lasciar l’Europa, — ella disse una mattina, —
suppongo che verrà prima a salutarci.
Bedeschi levò il capo con un movimento brusco. — Non so....
Forse.... Vedremo.... — E sentendo lo sguardo della nipote fisso
sopra di lui, si alzò da sedere e uscì dalla stanza.
La corrispondenza fra i due cugini durava non interrotta e non
vigilata. Federico tradiva spesso la sua impazienza, accusava la
Bice di non spiegar sufficiente energia per agevolargli il ritorno in
patria, dichiarava che assolutamente a Londra non ci poteva stare e
che avrebbe finito col fare un colpo di testa.... Poi, nella medesima
lettera, chiedeva scusa della sua petulanza e prometteva di seguire
a occhi chiusi i consigli della sua savia cuginetta, ch’egli si ricordava
in vestito da collegiale e che aveva giudizio da vendere a lui e a
molti altri meglio di lui.
E la savia cuginetta gli aveva scritto un giorno con gravità di esperta
diplomatica: — Un colpo di testa può anche esser necessario, ma
bisogna saper scegliere il momento di farlo. Il momento lo sceglierò
io.

II.

— Eccomi, — disse la Bice comparendo nel salotto ove si trovava lo


zio.
Egli gettò via il giornale L’Esercito che stava leggendo e si preparò a
darle una risciacquata di capo pel suo lungo ritardo. Ma l’aspetto
singolare in cui ella gli si presentava gli strappò invece un sorriso dal
labbro. E disse soltanto: — Finalmente!... E in quale arnese!
La ragazza aveva un lungo grembiale bianco che le scendeva dalle
ascelle ai piedi, le maniche del vestito rimboccate fino ai gomiti, le
mani e i polsi impiastricciati di farina, e teneva appunto le mani
aperte e le braccia larghe, discoste dai fianchi, per non insudiciarsi di
più. Aveva un po’ di farina anche sul viso e nei capelli.
— Eh, non ho terminato che adesso — ella rispose. — Sono in
tenuta di fatica.
— Vada a mutarsi dunque.... presto.
— Vado.... ma che cosa voleva, zio, che mi ha fatto chiamare?
È vero. Che cosa voleva? Non se lo rammentava neppur lui.... Ah sì,
voleva rimproverarla. E riprese: — Perder la giornata per fare un
dolce. Vergogna!
— Fare e rifare, caro zio.... Senza dubbio, la ricetta era sbagliata....
Si figuri che se non ci mettevo un bicchier di latte di più veniva fuori
qualcosa di duro come una palla di cannone....
— E ce l’hai aggiunto di tuo capo?
— Già.... La cuoca non vuole responsabilità. È un’impertinente. Sa
quel che ha detto? “Mi perdoni, ma io non intendo immischiarmi ne’
suoi pasticci.„
— Ha ragione.... Ma non forzerai neanche me a mangiarlo il tuo
pasticcio.
— Oh lo assaggerà almeno.... per poter suggerirmi le correzioni da
farsi domani.
— Domani?
— Ma scusi, non eravamo d’accordo? Quella d’oggi è una prova....
Domani poi che ci sono i veterani a pranzo.
La Bice chiamava così tre ufficiali in pensione, antichi commilitoni
dello zio, il quale li invitava a desinare un paio di volte all’anno.
— Per i veterani — interruppe il colonnello — manderemo a
prendere dall’offelliere un dolce che non sia duro come una palla di
cannone.
La ragazza fece un segno di protesta.
— E a proposito — ripigliò Bedeschi — che ghiribizzo è stato quello
di voler che invitassi i miei amici per domani e non per oggi?
— Volere? — disse la Bice con accento sommesso. — Ho pregato....
Mi pareva che la vigilia di Natale fosse meglio passarla in famiglia.
La fronte del colonnello si annuvolò. — Famiglia numerosa in verità
— egli borbottò fra i denti.
— Ma! — sospirò la Bice.
— Vatti a vestire, va, — soggiunse lo zio.
Ella non si moveva.
— Che c’è adesso?
— Nulla.... Pensavo.
— A che cosa?
— Pensavo a tanti anni fa.... l’anno prima ch’io andassi in collegio,
quando il Natale si festeggiò qui tutti uniti.... Che tavola allegra!
C’erano il mio babbo e la mia mamma, c’era la zia, e Vittorio e
Augusto, venuti in vacanza per una quindicina di giorni, e
Federico.... Noi due eravamo i più giovani.... Egli faceva mille
biricchinate e mi legò con la treccia alla spalliera della seggiola.... Oh
mi par ieri.... E ora gli uni son morti, gli altri dispersi pel mondo.
Si voltò commossa, con le pupille umide.
Lo zio, infastidito, le diede sulla voce. — Per carità, non mi far
piagnistei. I morti lasciamoli in pace, e quanto a quelli che sono
dispersi, due calcolo che siano con noi; il terzo, il tuo carissimo
Federico, è meglio dimenticarlo com’egli dimentica.
Ella fu in procinto di mettergli la mano sulla bocca per farlo tacere.
Ma si ricordò ch’era tutta infarinata e si trattenne in tempo,
sorridendo in mezzo alle lacrime: — Non le dica neanche per
ischerzo queste cose. Se Federico non ha ancora scritto, questo non
significa che abbia dimenticato.... Giurerei che la lettera è in viaggio.
Il colonnello fece una spallucciata. — Del resto, peggio per lui. A me
non importa proprio niente.
Balzò in piedi e ripetè alla nipote: — Vatti a vestire. A meno che oggi
non si debba rinunziare al pranzo....
La ragazza guardò l’orologio. — Pel pranzo ci vorrà un’oretta.... o
un’oretta e un quarto.... secondo il punto in cui sarà il dolce.
— Insomma, Bice, — saltò su lo zio aggrottando le ciglia, — ogni bel
gioco dura poco.... Vada e torni vestita entro venti minuti, e quando
torna, a qualunque punto sia il dolce, disponga perchè portino subito
in tavola.... Marsch.
— Oh, — esclamò il colonnello appena rimasto solo. — È
indispensabile di por ordine a questa faccenda. Colei con le sue
smorfiette ottiene sempre quello che vuole.
Il peggio si è ch’ella lo faceva diventar patetico, sentimentale, lui, il
colonnello Bedeschi! Non aveva dovuto rasciugarsi gli occhi, quella
sera stessa, nel guardare le fotografie di famiglia? Non era stato lì lì
per commoversi quando la Bice aveva evocato la memoria di quel
Natale lontano? Non si crucciava fuori di luogo e di modo perchè
quel caposcarico del figliuolo minore tardava a mandare gli auguri
per le feste? Non soffriva all’assenza di questo ragazzo più assai
che non volesse ammettere di soffrire? Non vedeva con un certo
sgomento avvicinarsi il tempo nel quale Federico avrebbe dovuto
andare di là dall’oceano? Non c’erano dei momenti in cui gli sarebbe
venuta una gran tentazione di richiamarlo?
No, così non poteva durare. Il colonnello aveva bisogno di ricuperar
la stoica impassibilità d’una volta, anche a costo di allontanar da sè
la nipote. Ell’aveva diciott’anni, era piacente, graziosa, possedeva un
quarantamila lire di suo, altre ventimila gliene avrebbe date lui, non
doveva esser difficile di trovarle un marito.... Trovarle marito, e poi
rimaner solo, con Battista, l’ordinanza, e coi veterani per commensali
nelle grandi solennità.... Che bella prospettiva! Tanto bella che il
colonnello, nell’eccesso della gioia, diede sulla tavola un pugno così
formidabile da far quasi cadere il lume. Indi se la prese col giornale
L’Esercito che gli parve indegno di avere nemmeno un associato;
stracciò in due pezzi il numero che aveva fra le mani e ne fece due
pallottole che scagliò a due angoli della stanza. Dopo le quali gesta
tirò fuori il suo cronometro per vedere se fossero trascorsi i venti
minuti ch’egli aveva assegnati alla Bice per la sua toilette.
— Venti minuti giusti, nè uno di più nè uno di meno, — disse la
giovinetta entrando proprio in quel punto. Ella vestiva un abito di
lana celeste con guarnizioni di peluche, portava al collo un filo di
corallo, e buccole pur di corallo agli orecchi. Nei folti e lucidi capelli
aveva intrecciato un nastrino di velluto rosso che ne faceva meglio
spiccare il colore castano scuro e dava risalto ai suoi occhi bruni e
vivaci. Del resto non aveva lineamenti regolarissimi, nè poteva dirsi
bella nello stretto senso della parola, ma la persona agile e svelta e
l’espressione della fisonomia dolce ed arguta ad un tempo la
rendevano preferibile a molte vantate bellezze.
— Sfido un’altra a far così presto, — ella continuò avanzandosi
verso lo zio, che, suo malgrado, era rimasto colpito dalla geniale
apparizione. Ma egli era armato contro le seduzioni e rispose in
tuono burbero: — Bene, bene.... E che necessità c’era di mettersi in
fronzoli?... Tutte civette, le donne....
— Oh zio, mi son messa il vestito buono e i coralli che mi ha
regalato lei il mese passato.... Dovevo lasciarli sempre chiusi in
cassetto?
— Non dico questo.... Se ci fosse qualcheduno a pranzo.... Domani,
per esempio....
— Oh, pegli estranei.... Se però esige che vada a mutarmi di
nuovo...?
— Sì, per non finirla più.... Hai dato gli ordini in cucina?
— No, veramente.... Volevo darli adesso....
Bedeschi mise un’esclamazione poco parlamentare e tirò con
violenza il campanello.
Si presentò Battista.
— Il pranzo è pronto? — chiese il colonnello con voce tuonante.
Il servo guardò la signorina.
— Non guardate la signorina, guardate me, e rispondete.
— Ma, — balbettò Battista. — Dev’esser pronto tutto.... tranne il
dolce.... che la cuoca dice che non sarà pronto mai....
— Gelosia di mestiere, — rimbeccò la Bice.
— Se tutto è pronto, tranne quello che non sarà pronto mai, —
ripigliò il colonnello, — scodellate la minestra immediatamente.
Battista voleva soggiungere qualcosa, ma la padroncina con un
gesto di rassegnazione lo pregò di tacere.
Di lì a poco, nel salotto da pranzo bene riscaldato ed illuminato, zio e
nipote sedevano a tavola l’uno di fronte all’altra e parevano entrambi
in poco felici disposizioni d’umore. Lo zio trovava da ridire su tutte le
pietanze, la nipote, d’ordinario chiacchierina e vivace, s’era
ammutolita ad un tratto, e in preda a una singolare inquietudine
s’agitava sulla sedia, tendeva l’orecchio ai più lievi rumori, e alzava
ogni tanto gli occhi verso la mostra d’un orologio.
La Bice arrossì come uno scolaro colto in fallo, e disse: —
Osservavo ch’è molto tardi.
— Bella scoperta! Di chi la colpa?
In quel momento si sentì una scampanellata alla porta di strada; la
Bice balzò fuori della stanza, e Battista, che quella sera serviva
peggio del solito, rovesciò una bottiglia di vino sulla tovaglia.
— Imbecille! — urlò il padrone. E avrebbe aggiunto chi sa quali altri
epiteti se non fosse stata la curiosità di saper chi era venuto.
— Andate di là, — egli ordinò al domestico che non se lo fece
ripetere due volte, — e tornate subito a dirmi chi è.
Ma Battista non tornò subito. Tornò invece la Bice con una strana
espressione nella fisonomia, e si fermò sulla soglia.
— Ebbene? Che cos’è successo? Siamo in un ospedale di pazzi? —
chiese il colonnello.
— Oh zio, — rispose la ragazza. — Se mi fa quei visacci non ho
coraggio....
— Finiamola.... Chi è venuto?
— È venuto.... un forastiero....
— Un forastiero.... Chi?....
La Bice esitava.
— Chi, in nome del cielo?
— Oh sa, faccio come nelle commedie, io.... Avanti, Federico.
E tirò a sè il battente dell’uscio, dietro a cui si trovava Federico in
persona.
Si ha un bell’esser corazzati contro le debolezze umane, si ha un bel
voler foggiarsi sul tipo inflessibile di Bruto primo e di Manlio, allorchè
un figliuolo che si credeva di non rivedere per un gran pezzo vi
compare dinanzi all’improvviso, e con uno sguardo più eloquente
d’ogni parola vi chiede perdono de’ suoi trascorsi e ridomanda la sua
parte di affetto e il suo posto al focolare domestico, è impossibile
non cedere al bisogno di spalancargli le braccia. Il tempo delle
riflessioni verrà, verrà forse il tempo di pentirsi dell’aver ceduto a
questo primo movimento; intanto il cuore, sia pur di sorpresa, riporta
una vittoria che non è mai senza conseguenze per l’avvenire.
Tutto ciò accadde al colonnello Bedeschi, il quale faceva inutili sforzi
per nascondere la propria emozione, e girando intorno a questo figlio
piovutogli dall’Inghilterra e divorandolo con gli occhi, tradiva la sua
sollecitudine con una sequela di domande: — Non sei mica malato,
eh? — Non hai mica patito freddo per viaggio? — Sei stanco? — Hai
ancora da pranzare?
E mentre Federico ch’era un florido giovinetto sui ventitrè anni gli
rispondeva che stava benissimo, che s’era rifocillato a Verona, ma
che nondimeno avrebbe preso volentieri una tazza di brodo, il
colonnello scoteva il braccio della Bice.... — Via, perchè non ti
muovi?... E dov’è quello stupido di Battista? O non c’è del brodo
caldo in cucina?
— Ce n’è, ce n’è.... Ecco Battista con la zuppiera. È provvisto alla
cena, alla camera, a tutto....
— Tu sapevi dunque?
— Naturalmente.
— E anche la servitù?
Battista evitava lo sguardo del padrone. Federico non alzava il naso
dal piatto. La Bice sola, imperterrita, affrontava il fuoco.
— Anche la servitù. Da questa mattina.... L’avevo detto io.
— Ma perchè tanti sotterfugi?
— Ecco.... — principiò Federico al quale sembrava poco
cavalleresco il non accorrere in aiuto a sua cugina.
Ma il padre lo interruppe. — Tu bada a mangiare. La Bice ha la
lingua sciolta.
— Oh, — ripigliò questa, — mi spiccio in due parole. Se Federico
scriveva a lei che non ne poteva più delle nebbie di Londra, che
provava un bisogno imperioso di riscaldarsi al sole d’Italia, di
rivedere la casa, la famiglia, insomma se le chiedeva il permesso, lei
non glielo avrebbe dato, e allora come si faceva a disubbidire?...
Invece di scrivere a lei, Federico scriveva a me.... Non era un
segreto la nostra corrispondenza....
— No, certo; però non mi sarei immaginato che ne usaste per
cospirare. Alle corte, quel permesso che forse non avrei dato io, l’hai
dato tu.
— Al momento opportuno ho incoraggiato Federico a perorar la sua
causa in persona. Lo assicuravo che suo padre era un uomo severo,
ma un cuore come ve ne son pochi....
— Basta, basta, — disse il colonnello. — È inutile dorare la pillola. —
Indi rivolgendosi al figliuolo: — E il mio amico Giraldi, il tuo
principale, è anche lui della congiura? Son sei o sette mesi che non
mi manda una riga.
Federico estrasse del taccuino una lettera e la porse al padre.
— Oh! — disse questo. — Son due lettere, una dentro dell’altra.
— Leggi e vedrai.
Giraldi scriveva ad Annibale Bedeschi lodandosi della condotta e
della intelligenza di Federico, ma soggiungendo che non gli pareva
uomo nato pel commercio, e che, secondo lui, era molto meglio
lasciargli studiar l’arte per la quale mostrava disposizioni singolari. In
prova di che inchiudeva un biglietto del celebre Whitty, uno dei primi
pittori di Londra, che aveva visto i disegni del giovine e ne traeva i
più lieti pronostici per l’avvenire.
Il colonnello diede un’occhiata al biglietto in questione. — È in
inglese! — egli esclamò. — O che che cosa devo capirci io?
Federico si offerse di tradurlo. Ma la Bice propose di rimetter le
spiegazioni al domani. Erano tutti stanchi, e Federico in particolare
cascava dal sonno.
— È vero, — assentì il colonnello. — Federico dovrebbe andarsene
a letto.
A questo punto la Bice si picchiò la fronte con la mano. — E il mio
dolce? Battista, fate il piacere di domandarne conto alla cuoca.
Il cugino mise un’esclamazione ammirativa. — Anche di pasticceria
te ne intendi?
— Un poco.
— Uhm! — fece il colonnello.
Battista rientrò in salotto con aria contrita, e depose davanti al
padrone un piatto che conteneva un oggetto informe.
— È questo il tuo dolce? — chiese ironicamente lo zio dopo alcuni
vani tentativi di fenderne la crosta col coltello.
La Bice, mortificata, non riconosceva più l’opera sua. — Così me lo
hanno ridotto?
Federico non potè trattenere una sonora risata.
— Hai torto di ridere, — disse la giovinetta. — Quella, vedi, è tutta
malizia della cuoca, invidiosa de’ miei trionfi. Un’altra volta....
— Non c’è altra volta che tenga, — protestò il colonnello. — Basta
una, ce n’è d’avanzo.
La Bice si strinse nelle spalle. — Il mio dolce avrà servito a ogni
modo a far ritardare il pranzo. Se la corsa fosse arrivata in orario,
Federico avrebbe desinato con noi, anzichè trovarci alle frutta.... E
adesso....
— O che c’è ancora?... Non volevi che tuo cugino andasse a
riposarsi?
— Sì, ma poichè rimane un sorso di vino nei bicchieri faccio un
brindisi ai due assenti Vittorio e Augusto.
— Con tutto il cuore, — risposero a una voce padre e figliuolo.
— Agli assenti e al reduce, — ella soggiunse.
— Sia pure.... Anche al reduce, — ripetè il colonnello avvicinando il
suo bicchiere a quello di Federico. — Abbiamo però sempre dei
conti da regolare.

· · · · · · · · · · · · · · · ·

Quando Federico si mosse per salire nella sua camera, la Bice lo


accompagnò fino sul pianerottolo. Egli non aveva parole abbastanza
per ringraziarla, per esaltare il suo spirito, la sua bravura.
— Non facciamoci illusioni — ella disse. — Non cantiamo vittoria
troppo presto.
— Con te, mia cara, si vinceranno tutte le battaglie, — replicò il
cugino. E dopo una breve pausa, abbassando la voce e
avvolgendola d’uno sguardo ch’esprimeva il più sincero entusiasmo:
— Sai che ho fatto un’altra grande scoperta?
La Bice abbassò involontariamente gli occhi. — Quale?
— Che sei diventata proprio bella.... ma proprio.... non è già un
complimento.
— Pazzo che sei! — disse la ragazza imporporandosi in viso. —
Buona notte, buona notte. — El o piantò lì col lume in mano,
incantato a guardarla.

· · · · · · · · · · · · · · · ·

Come si accomodassero le faccende il dì appresso, che


conseguenze avesse nell’avvenire la grande scoperta di Federico, a
che risultato approdasse la visibile simpatia de’ due cugini, son tutte
cose che non si possono saper subito.... Al Natale prossimo.... forse.
LA ZIA TERESA.

Quella sera, quando s’udì la scampanellata del postino, in casa


dell’avvocato Ettore Gualtieri avevano appena finito di desinare. La
cameriera entrò in salotto portando due giornali pel padrone e una
grossa lettera coperta di francobolli per la signorina Amelia.
La signorina Amelia, una leggiadra giovinetta di forse diciott’anni,
divenne rossa ed esclamò: — È della zia Teresa.
E agitando la lettera con aria trionfale soggiunse: — Scommetto che
qui c’è la fotografia. Era tempo.
Con un oh, oh di curiosità tutti quanti si strinsero intorno all’Amelia.
Erano in quattro, l’avvocato Ettore e la signora Luisa sua moglie, la
Carolina, una ragazza in quell’età critica nella quale è arrischiato
ogni pronostico sulla bellezza femminile, e Amedeo, un fanciullo
sgarbato come sogliono essere i maschi dagli otto ai quindici anni.
— Bada che voglio i francobolli — gridò appunto Amedeo con la sua
voce di pentola fessa.
— E io il monogramma — disse la Carolina.
L’Amelia fece un gesto d’impazienza. — Dio, che noiosi! Avrete i
francobolli, avrete il monogramma, ma non istatemi addosso così.
L’avvocato allontanò col braccio i due importuni e diede alla sua
figliuola maggiore un temperino perch’ell’aprisse la busta senza
stracciarla.
Bisogna notare che, dal signor Ettore in fuori, nessuno dei presenti
aveva conosciuto questa zia Teresa. In primo luogo, dei giovani ella
non era zia ma prozia; aveva cioè sposato molto tempo addietro uno
zio dell’avvocato, un Gualtieri anch’esso, dimorante a Nuova York fin
dal 1849 e arricchitosi colà negli affari. Al momento del matrimonio il
signor Temistocle (l’uomo si chiamava così) aveva
quarantacinqu’anni ed ella ne aveva venti, nè alcuno credette ch’ella
lo prendesse per inclinazione. Comunque sia, questo signore, tanto
più vecchio della moglie, s’era conservato vispo ed arzillo e veniva
ogni due anni in Italia; ella invece tra per le cure da prestarsi al
padre che l’aveva seguita in America, tra per gli acciacchi di cui si
lagnava, non aveva più ripassato l’Oceano.
Erano queste le ragioni ch’ell’adduceva scrivendo all’Amelia, ma il
marito, ne’ suoi viaggi in Europa, affermava che la ragione vera era
la pigrizia, era la paura del mare.
— La salute — egli soleva ripetere — l’ha buonissima, e ingrassa di
giorno in giorno. Anzi, inter nos, questo potrebb’essere un altro dei
motivi pei quali le ripugna di tornare dove c’è tanta gente che se la
ricorda giovine e bella.... Donne, sempre donne.... Già, ne avete la
prova.... ha perfino scrupolo di farsi fare il ritratto per mandarvelo.
Alla lunga, come si vede, lo scrupolo ella lo aveva vinto, e la lettera
giunta quella sera all’Amelia conteneva realmente la sospirata
fotografia.
Bisogna convenire che la prima impressione fu tale da richiamare
alla memoria le parole poco galanti del signor Temistocle. La zia
Teresa aveva l’aspetto d’una donna attempata, più florida del
necessario, senza studio d’eleganza nel vestito ch’era liscio, d’una
sola tinta scura e chiuso fino al collo. Nei capelli spartiti sulle tempie
e ravviati dietro alle orecchie era appuntato un velo nero; la mano
sinistra non si vedeva, la destra, appoggiata sulla spalliera della
poltrona, si protendeva troppo innanzi e appariva di proporzioni
esagerate. Gli occhi ch’erano stati bellissimi, che si capiva dover
essere belli tuttora, erano guastati dalla fotografia mancante
assolutamente di nitidezza e di rilievo.
Vi fu un breve silenzio durante il quale un osservatore attento
avrebbe potuto notare nel volto della signora Luisa un risolino di
trionfo che contrastava con l’aria malsoddisfatta degli altri e
specialmente dell’avvocato. — È questa la zia Teresa? — parevano
domandare i ragazzi. E l’avvocato: — È proprio lei?
Amedeo ruppe il ghiaccio. — Sapete a chi somiglia?... Alla signora
Venosti.
Il babbo gli slanciò uno sguardo fulmineo. — Sciocco!
Anche l’Amelia protestò con gran vivacità.
Ma la signora Luisa venne in aiuto del figliuolo. — Amedeo non ha
tutti i torti. La ricorda....
— No, mamma, no, — rimbeccò l’Amelia che non poteva tollerare
questi paragoni tra una zia dilettissima e una conoscente ridicola da
lei messa in canzonatura infinite volte.
— Si fa presto a dir no, — insistè la signora Luisa. — È un fatto.
Il signor Ettore perdette la pazienza. — Insomma vorrei sapere dove
la si trova questa famosa rassomiglianza....
— Nella bocca, per esempio....
— Oh santo cielo. Nella bocca?... Se la Venosti non ha quasi più
denti....
— E in questo ritratto, con tua licenza, la bocca è chiusa e non
possiamo sapere se i denti ci siano o non ci siano. E poi il taglio
della bocca non ha nulla a che fare coi denti....
— Via, mamma, — ripigliò l’Amelia, — la Venosti ha più di
sessant’anni.
— E credi forse che tua zia sia una bambina?... I suoi cinquanta
deve bene averli....
L’avvocato fece un energico segno negativo col capo. — Nemmen
per sogno.
— O quanti allora?
— È un conto semplice. È nata nel 1843 e siamo nel 1888.
— Da ottantotto a levar quarantatrè rimangono quarantacinque, —
esclamò Amedeo per mostrar la sua perizia nell’aritmetica.
— Precisamente quarantacinque, — ripetè il signor Ettore.
— Più quelli della balia, — soggiunse sghignazzando la moglie.
— È una bella ostinazione, — replicò l’avvocato. — Vuoi un’altra
prova? La Teresa si sposò nell’agosto 1863, venticinque anni fa, nè
più nè meno.... E aveva compiuto i vent’anni in quel mese stesso.
— Sarà, — disse la signora Luisa con quella riluttanza che hanno le
donne a darsi per vinte, soprattutto in certe questioni. — A ogni
modo se non sono che quarantacinqu’anni non gliene faccio le mie
congratulazioni.... Ne mostra molti di più.
Frattanto l’Amelia, che aveva scorso rapidamente la lettera della zia,
ne lesse una mezza pagina ad alta voce: “È la prima volta che vado
dal fotografo dacchè ho lasciato l’Europa. E ci andai per contentarti.
Desidero in compenso di aver le fotografie rinnovate di tutti voi altri.
Ho la tua, quelle di tua sorella e di tuo fratello, ma le ultime
rimontano al 1885 e ritengo che ne avrete di più recenti. Non ho poi
quelle della tua mamma e del tuo babbo, e mi sarebbe così caro di
averle.... Il tuo babbo mi trova molto cambiata, non è vero? Eh, il
tempo passa per tutti, e per noi donne passa più presto che pegli
uomini.„
— Bisogna tornar da Vianelli, — disse l’Amelia ripiegando il foglio.
— Torniamoci addirittura domani. — propose la Carolina.
— Oh, — replicò la madre, — ci andrete voi. Io farò tirar qualche
altra copia dei ritratti del 1885, quelli che ci siam fatti, il babbo ed io,
pochi mesi dopo dei vostri.
E rivolgendosi al marito: — Mi sembra che anche tu potresti
spicciarti allo stesso modo.
L’avvocato sorrise. — No, io voglio esser sincero. Non voglio farmi
passar per più giovine di quello che sono. È giusto che la zia Teresa
trovi cambiato me come io trovo cambiata lei.
Questa dichiarazione in cui c’era una punta d’ironia crebbe le
disposizioni irascibili della signora Luisa. — Io non sono tenuta ad
aver tanti scrupoli, ella disse con piglio acre. — La zia Teresa non mi
ha conosciuta e non può quindi trovarmi cambiata. Già non crederei
d’esser così cambiata in tre anni.
La signora Luisa, sebben rasentasse la quarantina, aveva ancora le
sue pretese.
Il marito non le diede la soddisfazione di rilevare le sue parole, ma si
mise a sgridare Amedeo che, uscito un momento dalla stanza, vi
rientrava con malagrazia rovesciando una seggiola sul suo
passaggio.
Avvezzo ai rimproveri paterni, il ragazzo si limitò a rialzare la
seggiola e posò un grosso album sulla tavola. — Me li lasci levare
questi francobolli? — egli chiese all’Amelia stendendo la mano verso
la busta.
— Aspetta un momento.... Non scappano mica....
— Giacchè ho qui l’album....
La Carolina ricordò timidamente che doveva avere il monogramma.
— Lo so, lo so, l’ho già sentito.... Son gusti incomprensibili.... Ne
avete non so quanti di questi francobolli di Nuova York, di questi
monogrammi della zia Teresa.... Giurerei che non avete altro nei
vostri splendidissimi album....
— Oh sì! Io ho quarantotto francobolli tutti diversi....
— E io ventitrè monogrammi.... tutti magnifici....
— Figuriamoci....
Mentre i figliuoli si bisticciavano, la signora Luisa s’era rimessa a
guardare la fotografia, e, siccome ell’era di quelle che nelle dispute
non la finirebbero mai, riattaccò il discorso al punto di prima.
— Che col tempo le linee della fisonomia e della persona si
modifichino, è troppo naturale — ella disse riconsegnando il ritratto
all’Amelia; — ma altro è modificarsi, altro è trasformarsi.... E a me
non entrerà in mente che questa donna sia stata bella.
Gualtieri si strinse nelle spalle. — Ce ne furono poche di belle come
lei.
— Basta sentir lo zio Temistocle, — notò l’Amelia.
— Oh, lui, s’intende, — rispose la madre. — Se l’ha sposata senza
un soldo è segno che a lui pareva bella....
— Pareva?... Era bella, era bella.... con o senza il tuo permesso, —
ribattè infastidito l’avvocato.
— Non riscaldarti il sangue.... È evidente che tutti e due, zio e
nipote, la trovavate la Venere dei Medici.... Però una Venere che
preferì chi poteva vestirla di porpora e di velluto.
La signora Luisa non aveva ancora terminata la frase imprudente
ch’ell’era già pentita di essersela lasciata sfuggire. Ma ormai era
troppo tardi.
Una fiamma passò negli occhi di Ettore Gualtieri. Pur si contenne e
con una voce che l’emozione rendeva più penetrante, — Luisa, —
egli disse, — a te non è lecito ignorare per qual ragione la Teresa
Rosnati abbia sposato mio zio che aveva più del doppio della sua
età e ch’ella non amava. Che se tu lo hai dimenticato, stimo
opportuno di ricordartelo, anche perchè i nostri figliuoli, qui presenti,
non giudichino male una zia che non hanno mai vista, che forse non
vedranno mai, ma dalla quale non hanno ricevuto che gentilezze....
Oh non è una storia lunga.... La Teresa era le mille miglia lontana
dall’idea di quel matrimonio quando suo padre, rovinato da cattive
speculazioni, si trovò in procinto di fallire. Il suo maggior creditore
era Temistocle Gualtieri, stabilito da un pezzo in America, ma giunto
allora in Europa per uno de’ suoi viaggi d’affari. Corso a Venezia per
questo minacciato fallimento, egli vide la ragazza, se ne invaghì e ne
chiese la mano. Pur d’ottenerla, non solo egli accondiscendeva ad
annullare il debito che il Rosnati aveva verso di lui e lo aiutava a
liquidar onorevolmente la casa, ma gli offriva un posto lucroso nel
suo banco a Nuova York. D’altra parte, egli diceva chiaro e tondo

You might also like