Professional Documents
Culture Documents
CLassroom Assesment For Teachers - Raymond H. White
CLassroom Assesment For Teachers - Raymond H. White
Witte
TM
CLASSROOM
ASSESSMENT
FOR TEACHERS
RAYMOND H. WITTE
Miami University
Published by McGraw-Hill, an imprint of The McGraw-Hill Companies, Inc., 1221 Avenue of the Americas, New York, NY
10020. Copyright © 2012. All rights reserved. No part of this publication may be reproduced or distributed in any form or by
any means, or stored in a database or retrieval system, without the prior written consent of The McGraw-Hill Companies, Inc.,
including, but not limited to, in any network or other electronic storage or transmission, or broadcast for distance learning.
Some ancillaries, including electronic and print components, may not be available to customers outside the United States.
1 2 3 4 5 6 7 8 9 0 DOC/DOC 0 9 8 7 6 5 4 3 2 1
ISBN: 978-0-07-337870-1
MHID: 0-07-337870-4
All credits appearing on page or at the end of the book are considered to be an extension of the copyright page.
Witte, Raymond H.
Classroom assessment for teachers / Raymond H. Witte.
p. cm.
Includes bibliographical references and index.
ISBN-13: 978-0-07-337870-1
ISBN-10: 0-07-337870-4
1. Educational accountability—United States. 2. Teacher effectiveness—United States. 3. Academic achievement—United
States. I. Title.
LB3051.W49994 2012
371.1020973—dc22
2010053786
www.mhhe.com
This textbook is dedicated to the memories of Dr. Jack E. Weih, my father-in-law, and
Dr. James Sheehan, my colleague and friend. Both of these men made the world a better
place with their kindness, humor, and genuine joy of living.
iii
Dr. Raymond Witte received his Ph.D. from the University of Kentucky in 1991 and
has been teaching at Miami University since 1993. During his tenure at Miami, he has
taught a wide range of courses, including the assessment and intervention courses in both
the undergraduate teacher education program and the graduate school psychology pro-
gram. His areas of research interest and publication include assessment, school transition,
mental health, and diversity. He has authored over 30 articles and chapters in various jour-
nals, professional newsletters, and resource sites. He also directs and manages the Class-
room Assessment Resource Center (http://web.me.com/dr.raymondwitte/Classroom
Assessment/Welcome.html), which is a noncommercial Web site for teachers that is
devoted to providing professional information and resources about classroom assessment.
Dr. Witte has extensive P–12 school experience as a school psychologist, as a preschool/
extended kindergarten program director, and as an assessment coordinator for a public
school district. Dr. Witte is currently working on an assessment textbook designed for
special education teachers.
iv
Preface xix
CHAPTER 1
Teaching, Learning, and Assessment 1
Introduction 2
The Instructional Importance of Classroom Assessment 2
Assessment as Part of the Instructional Process 3
Assessment: Serving an Educational Purpose 4
Assessment Reality 4
Introductory Statement of Purpose from No Child Left Behind Act 5
Chapter Activity 6
Classroom Assessment in the Lives of Students 6
Assessment Defined 7
The Instruction, Learning, and Assessment Model 8
Where Are My Students? 9
Where Do My Students Need to Be? 10
How Do My Students Get There? 11
Instruction-Learning-Assessment Alignment 12
Assessment in Action: The Impact of Instruction-Learning-Assessment Alignment 13
Professional and Ethical Assessment Practices 14
Definition of Ethics 15
Ethics in the Classroom 15
National Education Association Code of Ethics 16
Focus on Assessment: Ethics in Assessment 19
Chapter Reflection 20
Chapter Review 20
Conclusion 20
Response to Key Focus Question 21
v
Assessment Check 21
Resources 22
References 23
CHAPTER 2
Elements of the Instruction-Learning-Assessment
Model 25
Introduction 26
Chapter Activity 26
Classroom Teaching and Assessment 27
Identifying Learning Targets 27
Pre-Assessment 30
Focus on Assessment: Use of Consensus-Grams 31
Assessment in Action: Effective Use of Pre-Assessment 32
Preview of Instructional Material 33
Instructional Approach and Techniques 34
Formative Assessment 35
Learning Feedback 37
Self-Assessment 38
Focus on Assessment: The Teacher as a Self-Assessment Model 39
Summative Assessment 40
Chapter Reflection 42
Chapter Review 43
Conclusion 43
Response to Key Focus Question 44
Assessment Check 45
Resources 45
References 45
CHAPTER 3
Academic Standards and Evidence of Student
Learning 48
Introduction 49
What Do Your Students Need to Know? 49
Chapter Activity 49
Know What Your Students Need to Know 54
Learning Targets, Objectives, and Goals 54
Ohio Science Academic Content Standards (Approved in 2002) 55
Hierarchical Arrangement of Instructional Planning 57
Demonstration of Learning: Bloom' s Taxonomy 58
Structure of the Knowledge Dimension of the Revised Taxonomy 61
Structure of the Cognitive Process Dimension of the Revised Taxonomy 62
Planning Your Lessons: Guiding Questions 64
Evidence of Student Learning 65
Gathering Useful Evidence 67
Informal Observation 67
Formal Observation 69
Group Observation 69
Focus on Assessment: The Utility of Observational Information 70
Types of Observations in the Classroom 71
Narrative 72
Interval and Event Recording 73
Checklists and Ratings 74
Questioning 76
Assessment in Action: Questions as Assessment Devices 78
Interviewing 79
Student Products 79
Chapter Reflection 80
Chapter Review 81
Conclusion 81
Response to Key Focus Question 81
Assessment Check 81
Resources 82
References 82
CHAPTER 4
Validity, Reliability, and Avoiding Assessment
Bias 84
Introduction 85
Validity: Does It Measure What It Is Supposed To? 85
Chapter Activity 85
Content-Related, Criterion-Related, and Construct-Related Evidence of
Validity 87
Assessment in Action: Construct Validity in the Classroom 90
Validity for Teachers 92
Reliability for Teachers 92
Response Requirements and Scoring Guidelines for an Essay Question 93
Reliability: The Same Every Time 94
Stability/Test-Retest, Alternate Form/Parallel Forms, and Internal Consistency/Split
Halves or Odd-Even 96
Focus on Assessment: Speed and Power Test Issues 101
Inter-rater Reliability 101
Avoiding Bias 105
Potential Test Item Bias Influences Examined by the West Virginia Department of Educa-
tion Assessment Office 107
Chapter Reflection 109
Chapter Review 110
Conclusion 110
Response to Key Focus Question 110
Assessment Check 111
Resources 111
References 111
CHAPTER 5
Formative Assessment: Guiding Student
Learning 113
Introduction 114
Formative Assessment Defined 114
Chapter Activity 115
CHAPTER 6
Performance Assessment 136
Introduction 137
What Is Performance Assessment? 137
Features of a Performance Event 138
The Focus of Performance Assessments 139
Chapter Activity 142
Focus on Student Production 142
Assessment in Action: Performance Assessment and a Driver’s License 143
Different Needs, Different Measures 144
Classroom Portfolios 145
Project-Based Learning 148
Focus on Assessment: Performance Activity within a Health Portfolio 149
Rubrics 151
Building an Effective Rubric 153
Let’s Bake, I Mean Make, a Rubric 155
Checklists 159
Make Sure Your Students Know the Performance Event Well 161
Chapter Reflection 161
Chapter Review 162
Conclusion 162
Response to Key Focus Question 162
Assessment Check 163
Resources 163
References 164
CHAPTER 7
Self-Assessment and Progress Monitoring 166
Introduction 167
Chapter Activity 168
Self-Assessment: Connection with the Teaching and Learning
Process 169
Elements of the Self-Assessment Process 170
Self-Assessment Techniques 171
Self-Assessment Review of a Writing Portfolio 176
Steps in the Process 178
A Progress Review Script 179
The Student as Progress Manager 180
Assessment in Action: Self-Assessment and Student Feedback 181
Focus on Assessment: Self-Assessment in a Third-Grade Classroom 183
Self-Assessment Begins with the Teacher 186
Self-Assessment Questions for Teachers 186
Chapter Reflection 187
Chapter Review 187
Conclusion 187
Response to Key Focus Question 188
Assessment Check 188
Resources 188
References 189
CHAPTER 8
Summative Assessment: Confirming Student
Achievement 191
Introduction 192
Chapter Activity 192
Summative Assessment Construction 193
Focus on Assessment: The Unpopular Truth 194
Assessment Construction Procedures 195
General Test Construction Rules 197
Assessment in Action: Verbal Examinations and Confirmation of Learning 199
Selected and Constructed Response Items 200
Selected Response: Multiple Choice 201
Multiple-Choice Test Items: Dos and Don’ts 202
Selected Response: True-False 203
True-False Test Items: Dos and Don’ts 204
Selected Response: Matching 204
Matching Test Items: Dos and Don’ts 205
Constructed Response: Fill-in-Blank and Listing Questions 206
Constructed Response: Short Answer 207
Short Answer and Completion Test Items: Dos and Don’ts 208
Constructed Response: Essay 208
Essay Test Items: Dos and Don’ts 210
Table of Specifications and Your Test 211
Constructed Responses Beyond the Traditional Test 213
Written Construction Performances 213
Visual Construction Performances 214
Verbal Construction Performances 215
Fine/Gross Motor Construction Performances 216
Reviewing Performances with Students 216
Chapter Reflection 217
Chapter Review 218
Conclusion 218
Response to Key Focus Question 218
Assessment Check 218
Resources 219
References 219
CHAPTER 9
Grading and Student Evaluation 221
Introduction 222
Chapter Activity 222
Grading: A Professional Challenge 223
Grading: Purpose Is Everything 224
Grading: A Part of the Educational System 224
Assessment in Action: Grading Is More Than the Numbers 226
Grading: Three Central Issues for Teachers 227
Evaluating Student Work 227
The Composition of Grades 232
Product, Process, and Progress 234
Guiding Principles 235
Focus on Assessment: Electronic Grade Books 237
Professional Consensus 239
Comparison Types 240
Norm-Referenced Grading 240
Criterion-Referenced Grading 241
Aptitude Comparison or Self-Referenced Grading 242
Grading Students with Special Needs 243
The Learning Race 244
Uniform Grading Standard for All Students 245
Grading Based on Effort 246
Grading Based on Progress 247
Grading Adaptations 247
Grading Practices and Realities 248
Chapter Reflection 250
Chapter Review 250
Conclusion 250
Response to Key Focus Question 250
Assessment Check 251
Resources 251
References 251
CHAPTER 10
Statistics in the Classroom 254
Introduction 255
Why Do I Need to Use Statistics? 255
Chapter Activity 255
Start with Data 256
Distributions 257
Focus on Assessment: Benefiting from Student Evidence 258
Central Tendency: Mean, Median, and Mode 261
Variability or Dispersion from the Mean 263
Normal Distribution 264
Standard Scores 265
Examining Student Performance: Correlation Coefficients 268
Standard Error of Measurement 270
Looking within a Testing Performance 271
Determining the Difficulty Level of Test Items 272
Item Discrimination 274
Assessment in Action: Using Statistics in Reviewing Student Performances 275
Student Responses to Test Items 276
Chapter Reflection 276
Chapter Review 276
Conclusion 276
Response to Key Focus Question 276
Assessment Check 277
Resources 278
Answers to Assessment in Action Activity 278
References 279
CHAPTER 11
Standardized Tests and High-Stakes
Testing 280
Introduction 281
The No Child Left Behind Act 281
Assessment in Action: Confession of a Veteran Teacher 282
CHAPTER 12
Assessment of Students with Identified Learning
Needs 309
Introduction 310
Special Education Legislation 311
Individualized Education Plan 311
Chapter Activity 312
Special Education Disability Categories 313
Autism 314
Deaf/Blindness 315
Developmental Delay 315
Emotional Disturbance 315
Hearing Impairment 316
Mental Retardation (Intellectual or Cognitive Disability) 317
Multiple Disabilities 318
CHAPTER 13
Classroom Assessment and Response to
Intervention (RTI) 339
Introduction 340
Chapter Activity 341
How Response to Intervention Works in the Classroom 342
Tier 1 344
Tier 2 346
Tier 3 347
Assessment in Action: RTI in the Classroom 347
CHAPTER 14
Affective Assessment 364
Introduction 365
Know the Limits and Boundaries of Affective Assessment 365
Cautions with Affective Assessment 366
Selection of Affective Assessment Areas 367
Student Attitudes 367
Student Interests 369
Personal Beliefs and Values 370
Personal Perceptions and Viewpoints 371
Chapter Activity 371
Assessment in Action: Affective Assessment in the Classroom 372
Generating Affective Assessments 372
Constructing a Likert Self-Report Measure 374
Construction Steps for a Likert Self-Report Measure 374
Topic or Domain Selection 375
Construction of the Items 375
External Review 376
Field Testing 376
Administering the Inventory 377
Scoring and Analyzing Responses 377
CHAPTER 15
Getting Started on the Assessment Path 389
Introduction 390
Chapter Activity 390
Your Own Classroom Assessment System 391
Focus on Assessment: Take a Look Around You 391
Classroom Assessment Framework Exercise Part 1 395
Determining What Your Assessment System Will Do 395
Classroom Assessment Framework Exercise Part 2 396
Determining What Assessment Measures Will Be Used in Your Classroom 396
Classroom Assessment Framework Exercise Part 3 399
Aligning Your Classroom Assessments with Instructional Learning Impact 399
Classroom Assessment Framework Exercise Part 4 400
Collecting Confirming Evidence of Student Learning 400
Classroom Assessment Framework Exercise Part 5 401
Making Informed and Effective Decisions 401
N ow that confirmation of student progress and learning plays a key role in how
educators and schools are reviewed and evaluated, assessment in all its forms
has become the critical discussion point in education. The field of education
recognizes the importance of accountability, and with No Child Left Behind (NCLB),
state standards, and other accreditation and review boards, the documented advancement
of student learning and achievement is necessary.
Professional and state standards require teachers to be effective assessors in their class-
rooms; this text helps teacher candidates meet those required standards of practice. In
particular, the instruction and practice activities of the text provide pre-service teachers
with opportunities to develop an essential knowledge and skill set that will allow them
to effectively assess, monitor, and confi rm student learning progress in their classroom.
The format of the text and its activities provide the opportunity for teacher candidates
to learn why assessment is done, how it is integrated into instructional delivery, and the
central role it plays in the progress monitoring and review of student performance in the
classroom.
TEXT COVERAGE
A range of topics are covered within the chapters so that each teacher obtains a solid
assessment knowledge and skill foundation. Essential components such as formative
assessment, summative assessment, and self-assessment are covered, along with additional
topics such as performance and affective assessments, construction of test items for class-
room tests, high-stakes testing, interpretation of test scores, validity and reliability issues
with classroom assessments, applied statistical procedures, grading, and the utilization of
student evidence and data in the educational decision-making process. In addition, chap-
ter activities and assessment issues are provided that are designed to engage the learner
in meaningful self-reflection of their current assessment perspectives and future assess-
ment practices. A continuous emphasis throughout this textbook is the importance and
direct linkage of assessment process and content with actual instructional practice in the
classroom. The conceptual and functional reality of classroom assessment as it relates
to instructional delivery is also examined. A pragmatic and interconnected instruction-
learning-assessment model for teachers to follow is provided, along with the important
procedures used in assessing student progress and the documentation of P–12 learning
impact. This is done to demonstrate the essential relationship that exists between assess-
ment and instruction.
xix
The need of P–12 students to learn and demonstrate self-assessment in regards to their
learning goals and objectives is emphasized. Learner self-assessment represents one of
the growth areas in assessment and is potentially the most important skill domain that
any student will learn. Given its instructional value, an entire chapter is devoted to this
subject. Assessment of learner progress at all levels in the classroom, along with specific
content on assessment as it relates to special needs learners, is addressed. Due to greater
inclusionary practices and the academic standards that are held for all learners, regular
education teachers need to possess greater knowledge and skill levels in evaluating the
progress of all students in the classroom, including those with special learning needs.
In addition, as public mandates for standardized achievement test results grow, teachers
must be able to explain test fi ndings to parents, children, and other interested parties, as
well as be able to use test data to help guide instruction in the classroom.
This textbook explains, in a direct and meaningful way, what classroom assessment
is and why it is essential for teachers to be able to measure the progress of their students.
Making content, concepts, procedures, and applications understandable and useful to stu-
dents is essential, and classroom assessment is the vehicle by which educators can confi rm
the acquisition of learning outcomes for their students.
FEATURES
The following features appear throughout this text.
Chapter Key Focus—Each chapter of the text is centered on a fundamental
assessment question that appears at the beginning of each chapter.
Chapter Objectives—Specific learning objectives appear at the beginning of each
chapter, stating what will be achieved upon successful completion of the chapter.
Key Terminology—Key terms appear in boldface throughout the chapters, and
are defi ned in the comprehensive glossary at the end of the text.
Chapter Activity—A specific learning activity that connects to the content is
presented within each chapter.
Assessment in Action—An assessment example that may include authentic expe-
riences or responses by field practitioners, teacher opinions, or samples of classroom
assessments appears within each chapter.
Focus on Assessment—Within each chapter, a relevant example, key issue, or
authentic experience related to the chapter topic is provided for the reader.
Chapter Reflection—Chapter reflection questions require students to reflect
back on their initial responses to the Chapter Activity presented at the beginning
of each chapter.
Conclusion—Each chapter ends with a summary of the main points presented.
Response to Key Focus Question—At the end of every chapter, a concise
response to the chapter’s key focus question is provided.
ACKNOWLEDGMENTS
I would like to express my sincere appreciation to the following individuals for their
valuable and constructive help with this textbook: Dr. Christine Messina, Dr. Susan
Mosley-Howard, Dr. Aimin Wang, Dr. Tom Southern, Dr. Jane Cole, Ms. Teresa
McGowan, Ms. Ellen Hill, Ms. Denise Krallman, and of course my editorial colleagues
Alexis Breen Ferraro, Jennifer Bonnar, and Sarah Kiefer. I would also like to thank the
countless teachers who have helped with this book. It is a pleasure to work with each and
every one of you.
TEACHING, LEARNING,
AND ASSESSMENT
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
Why does classroom assessment need to be an integral part of your instructional
process?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Understand the instructional importance of classroom assessment.
♦ Describe how classroom assessment directly enhances instruction and learning
outcomes for students.
♦ Define classroom assessment and the specific actions and functions that take
place when implemented.
♦ Identify the basic qualities and components of effective classroom assessment.
♦ Answer the three questions of the instruction-learning-assessment model as part
of any instructional activity or lesson.
♦ Acknowledge the concept of instruction-
learning-assessment alignment and its essen-
tial role in the evaluation of student progress.
♦ Distinguish the ethical principles and issues
that are connected to assessment procedures
and practices.
♦ Adhere to professional ethical principles and
standards when assessing student progress and
achievement.
© Creatas/PunchStock
Introduction
A strong subject or content knowledge and skill base has long been identified as an essen-
tial characteristic of a professional educator. Along with other recognized competencies
such as planning and delivering effective instruction, understanding and respecting the
diversity of students, professional collaboration, being an active member of the learning
community, etc., classroom assessment has now been added to the list of professional
characteristics required of teachers. But why does a teacher need to be competent and
skilled in the assessment of student learning?
© Creatas/PunchStock
ASSESSMENT: SERVING AN
EDUCATIONAL PURPOSE
Teachers assess student learning for many reasons, including:
♦ To know where students are in regards to their knowledge and skills before instruc-
tion begins.
♦ To know what knowledge and skills are acquired, and when, by students during
the lesson or learning activity.
♦ To document and confi rm student progress.
♦ To make appropriate instructional adjustments or modifications, if needed, during
a lesson.
♦ To inform each learner of his or her progress based on the stated learning objectives
and expected outcomes of the lesson.
♦ To enhance the effectiveness of teaching and the learning decisions that are made
in the classroom.
♦ To review and evaluate the quality of student-generated work and products.
We assess because we need to know information about students. We need to know
many things, and assessment is the vehicle through which desired information can be
obtained and used. As a professional educator, you are in the business of making impor-
tant educational decisions about students and the education they receive, and you will
be held accountable on a daily basis for those decisions. Having an assessment plan and
system within the classroom will help you make good decisions for and about students,
and that’s what teachers need to be able to do.
Assessment Reality
Until recently, the topic of classroom assessment did not generate high interest or value
for many educators. Assessment was seen by many as involving only standardized or
group achievement testing. It was done because it was connected to state or federal
mandates, and it was often considered “foreign” and outside the boundaries of the
classroom except for the traditional end-of-the-chapter test and/or provided grade.
All of that has now changed. Since accountability of student performance plays a key
role in how educators and schools are reviewed and evaluated, assessment has become
a critical discussion point in schools and districts across the nation. Moreover, with
the impact of No Child Left Behind, the documentation of actual academic progress
of students, particularly those at-risk academically, has taken center stage, and so has
classroom assessment (see the box “Introductory Statement of Purpose from No Child
Left Behind”).
Federal legislation (Public Law 107–110) known as the No Child Left Behind Act
(NCLB) was signed into law by President George W. Bush in January of 2002. NCLB
requires states (that are receiving federal funds) to utilize a comprehensive academic assessment sys-
tem, which documents the progress and educational outcomes of all students, including those that are
low-achieving, economically disadvantaged, or identified with special learning needs. In particular,
this law mandates that all teachers must be highly qualified (i.e., certificated or licensed),
that “scientifically based research” be used and followed in the schools, that students
have greater public school choice when schools do not demonstrate appropriate academic
progress and/or improvement, and that all students’ academic progress be measured and
followed on a yearly basis. All states at a minimum are required to provide standardized
testing in reading and math in grades 3 through 8, and at least once in high school. By
the 2007–2008 school year, it was established that testing in the area of science had to be
formally conducted at least once in grades 3–5, 6–9, and 10–11.
In addition, adequate yearly progress (AYP), as defined by scoring at the proficient or
better levels on state-required achievement tests, must be demonstrated by all students. In fact,
by 2014, all students must be at or above the proficient level in content skill areas (read-
ing, math, science, etc.); schools that do not meet this benchmark will be identified and
could possibly receive sanctions if progress is not demonstrated. Since these test scores
hold such importance, daily instructional practice is directed toward the preparation and
successful completion of these tests. To a great extent, these mandated tests drive the
instruction in classrooms across the country. Consequently, teachers need to be fully
aware and engaged with this assessment reality so that the best educational decisions can
and are made for students.
The continued importance of assessment is reflected in A Blueprint for Reform: The
Reauthorization of the Elementary and Secondary Education Act, released by the U.S. Depart-
ment of Education in March 2010. With the focus on college and career preparation and
under the heading for better assessments, it was stated with presidential approval that:
We will support the development and use of a new generation of assessments that
are aligned with college- and career-ready standards, to better determine whether
students have acquired the skills they need for success. New assessment systems will
better capture higher-order skills, provide more accurate measures of student growth,
and better inform classroom instruction to respond to academic needs (U.S. Depart-
ment of Education, 2010, pp. 3–4).
As a professional field, P–12 education clearly recognizes the importance of account-
ability in accordance with federal mandates, state standards, and other accreditation
and review boards, along with the necessary documentation of the change in students’
knowledge, achievement, and skills. Consequently, teachers must be able to accurately
and efficiently measure the progress of their students even as it relates to daily instruc-
tional practice. Yet even if outside pressures for verification and documentation didn’t
exist, the field of education would still be addressing this desired outcome (with perhaps
less anxiety and discontent). Why? By the nature of being an educator, every teacher
wants and needs to know if their lessons and teaching are truly effective, and that their
students are learning what they need to learn. The selection and use of appropriate assess-
ment measures and procedures can help answer that very question, and answering that
question is critical, both for the teacher and the students that he or she teaches.
CHAPTER ACTIVITY
If one of your classmates, a parent, or even a student asked you what classroom assessment
is, what would you tell them? How would you defi ne or describe it? Granted that you are
just beginning to learn about this subject, it is still important to identify at this point in
your professional preparation what you know and believe assessment to be about. So take a
moment and write down your defi nition of classroom assessment.
As you construct your defi nition, be sure to include the specific activities and proce-
dures that make up this process. Also, provide a brief description of your view of the rela-
tionship between teaching and assessment. Again, write it down and if possible discuss your
responses with those of your classmates. Whatever level of review is possible (individual
or group), examine how teaching and assessment are connected and in what ways they are
directly linked.
CLASSROOM ASSESSMENT
IN THE LIVES OF STUDENTS
Based on research spanning several decades (Black & Wiliam, 2009; Bloom, 1976, 1984;
Bloom, Hastings, & Madaus, 1971; Boston, 2002; Chappuis, 2005; Fuchs, Deno, & Mir-
kin, 1984; Fuchs & Fuchs, 1986; Guskey, 2003; Hattie & Timperley, 2007; McTighe
& O’Connor, 2005; Stiggins & DuFour, 2009), student achievement has been found to
improve, in some cases dramatically, when students are actively engaged in their learn-
ing and receive feedback on their performance during an instructional event. According
to the Assessment Reform Group (1999), based on the extensive assessment and learning
review work of Black and Wiliam (1998a), students (particularly low-achieving students)
can achieve at high levels if five instructional/assessment practices are followed, including:
When these practices are implemented, students show greater learning and achieve-
ment gains, and teachers that take advantage of these practices are likely to have students
who evidence greater academic success in the classroom (Boston, 2002; Chappuis, 2005;
McTighe & O’Connor, 2005; Stiggins & DuFour, 2009). This is about as close to a learn-
ing guarantee in teaching as you can get. So now that you know assessment contributes
directly to the academic achievement of students, it is time to examine how the assess-
ment process can operate and function in the classroom setting.
ASSESSMENT DEFINED
In the previous section, each reader was asked to provide a definition of classroom assess-
ment. When students in my classes are asked to provide their defi nitions, a wide range of
Bananastock/AGE Fotostock
descriptions and viewpoints are provided, along with some personal feelings mixed in as
well (e.g., “I hated being graded so I don’t like doing it to my students.”). A fairly consis-
tent but erroneous fi nding is that assessment is often viewed as synonymous with grad-
ing and that it serves the same purpose. Not surprisingly, pre-service teachers, as well
as those already in the field, possess a variety of viewpoints, experiences, and personal
expectations about assessment and its potential role in their classroom and professional
career.
Central to the purpose of classroom assessment is that meaningful evidence of student
learning must be obtained. Consistent with Black & Wiliam (1998b), the following defi-
nition of classroom assessment is provided: classroom assessment is viewed as an active
process of systematically collecting and analyzing student learning evidence in order to make effective
educational decisions, which are designed to promote student learning. Four essential actions are
required within this defi nition of classroom assessment:
1. Classroom assessment requires a consistent and ongoing process of student learning
evidence collection. This may include the collection of different types of informa-
tion over extended periods of time based on a particular need.
2. The examination and review of that evidence is necessary and expected so that it
can be determined whether effective learning is taking place and students are pro-
gressing because of it.
3. Important instructional decisions and adjustments need to be based on the critical
review of the collected information.
4. Instructional decisions need to be made with the intent of promoting learning for
all students in the classroom. This is the purpose and a big “why” of the assess-
ment process: to make decisions that are in the best interests of students and their
learning.
Some readers may be surprised with this defi nition and the actions that need to take
place. Assessment is not focused on, or restricted to, giving and/or grading tests. It’s really
about collecting needed information to make good instructional and learning decisions.
It’s important to understand that testing and evaluation is an important process in educa-
tion as well as in most other human endeavors. Testing must be used correctly and for its
intended purposes; if that is done, then true benefits can be obtained. However, testing
as a formal evaluation process is only a part of a much larger process recognized as class-
room assessment.
Assessment Teaching
Preview and Pre-Assess Where are
my students?
Formative Assessment
and Learning Feedback
How do my
students get there?
Instruction-
Student (e.g., cooperative
Assess progress Learning-
Self- learning activities,
“midstream” Assessment
Assessment modeling,
Alignment
inquiry-based
learning)
Formative Assessment
and Learning Feedback
pragmatic pedagogical approach shown in Figure 1.1, there are three modified assessment
questions that need to be asked relative to your students and the instruction they receive:
1. Where are my students?
2. Where do my students need to be?
3. How do my students get there?
These basic but essential questions are connected to core assessment processes within this
teaching model that involve each of the instructional questions.
(i.e., formal evaluation of progress and/or performance). The primary purpose is to assess the
progress of the students so that they can be informed of what they still need to learn if
they are to reach the intended learning targets of the lesson that is being taught (Boston,
2002). Think of this process as an educational map where the teacher as well as the stu-
dents must be able to locate themselves on the learning path, ascertain whether or not
they are headed in the right direction, and determine what knowledge and skills still
need to be acquired before the intended destination is reached.
traditional paper-and-pencil tests (e.g., math facts for addition, subtraction, multiplica-
tion, and division) are also typically utilized and completed. Several different measures
can and should be used in order to collect evidence of learner progress.
What is interesting and important to recognize and understand is that all of the listed
assessments have as much relevance and utility for the fi rst or second graders at the ele-
mentary level as for the seventh or tenth graders at the middle and high school levels. The
sources of student evidence and how it is collected may vary, but the need does not. In
addition, the use of multiple sources of confirming data allows for a comprehensive and
balanced approach in confi rming student progress. What is clear from a measurement
perspective is that the reliance on one primary source of data as evidence for student
progress is not effective or defensible from an assessment standpoint.
Therefore, it is important to view assessment measures (e.g., portfolios, tests, journals,
checklists, rubrics) as not age, grade, or content specific but as individual “tools” that are
needed based on a particular assessment job or need. So if it is important to document
the specific skill levels of a student or a group of students, direct observation and verifica-
tion through a skill checklist will likely be needed. On the other hand, if the acquisition
and retention of a specific knowledge base or set must be confi rmed, a traditional test
would be effective in assessing the acquired knowledge of the students, especially if lim-
ited time constraints exist. What is important is to make sure that, as the instructor, you
have many assessment “tools” that can address any particular need or issue that may arise
within the classroom.
INSTRUCTION-LEARNING-
ASSESSMENT ALIGNMENT
A fundamental requirement with the instruction-learning-assessment alignment model
is that instruction in the classroom must align with intended standards and performance
expectations. Standards are written statements, usually at the state or national level, that
specify what students are expected to know and be able to do based on preset perfor-
mance or competency levels within certain content and knowledge areas. The classroom
assessment must be designed so that the students’ progress in accomplishing the learning
outcomes that are connected to those standards is accurately measured. Assessment is most
effective and useful when it matches with the instructional content that is taught in the classroom,
thereby providing instruction-learning-assessment alignment (Beck, 2007; Cohen &
Hyman, 1991). For example, if a teacher is following the state mandated academic con-
tent standards that are approved and required for a specified grade level, then the stu-
dents’ progress relative to that academic content (e.g., math, science, history, English)
should be assessed and evaluated, since students are ultimately held accountable for learn-
ing and obtaining competency with that identified material.
Unfortunately, assessment targets can sometimes become lost and other skills or areas
are taught that are not part of the intended standards. If this occurs, instruction-learning-
assessment alignment can go off-target and something other than the intended standards
may get measured and evaluated. It is crucial to maintain a clear focus on the desired
ASSESSMENT IN ACTION
goals for the students and to possess valid and reliable assessment measures so that mean-
ingful data on actual student progress can be obtained. Teachers can and often go beyond
the required educational outcomes in lesson plans, and that is completely acceptable. The
assessment system, however, should be “laser accurate” when it comes to evaluating stu-
dent accomplishments relative to all required learning outcomes.
As Earl (2003) has identified, “there is no single activity called classroom assessment”
(p. 3). Classroom assessment is made up of many parts that constitute a teacher-directed
process that strategically collects student data that is used to help direct and enhance the
instructional process. Yet as important as it is to know what classroom assessment is, it
is just as important to know what classroom assessment is not. The comparisons listed
in Table 1.1 provide a review of the core elements that are recognized as essential to an
effective assessment system compared to one that is not (Guskey, 2003; Marzano, 2006;
McTighe & O’Connor, 2005).
have a strong ethical responsibility to make sure the results they obtain are as accurate
and reliable as possible and that the procedures and measures used in obtaining those
results are genuine for their intended purpose and use.
Definition of Ethics
The terms “ethical standards” or “ethical decisions” are often used in professional discus-
sions, but what actually constitutes a person’s professional ethics is not always constant
and consistent across disciplines and/or professional organizations. If asked to define or
describe one’s ethics, many, if not most, people indicate that ethics direct one’s professional
conduct and actions, which are based on a set of recognized standards of practice that are established
and followed by a member of a certain profession. Professional standards are in place and fol-
lowed in order to provide the necessary safeguards for the individuals or clients (i.e.,
students) that are directly impacted by the actions and decisions that are provided by cer-
tain individuals (i.e., doctors, psychologists, educators). Within the educational profes-
sion, professional competencies are emphasized and expertise in general areas of practice
or knowledge are usually expected (e.g., learner characteristics, instruction, assessment)
along with the desired and/or expected professional behaviors that are connected with
those areas.
In the case of educators, since students are in the direct care and supervision of teach-
ers, professional care must be provided as part of any action or decision that impacts
a student. Life-changing or altering decisions (e.g., selection into advanced placement
courses, receiving scholarships, passing or failing classes, qualifying for special education
services) can be made or influenced by teachers and for that reason ethical guidance and
direction is always expected.
The National Education Association (NEA) code of ethics, including its Preamble
and Principles I (Commitment to the Student) and II (Commitment to the Profession),
is provided in the box on pages 16–17. The NEA has over 3 million members and is the
nation’s largest education association.
Preamble
The educator, believing in the worth and dignity of each human being, recognizes the
supreme importance of the pursuit of truth, devotion to excellence, and the nurture of the
democratic principles. Essential to these goals is the protection of freedom to learn and to
teach and the guarantee of equal educational opportunity for all. The educator accepts the
responsibility to adhere to the highest ethical standards.
The educator recognizes the magnitude of the responsibility inherent in the teaching
process. The desire for the respect and confidence of one’s colleagues, of students, of parents,
and of the members of the community provides the incentive to attain and maintain the
highest possible degree of ethical conduct. The Code of Ethics of the Education Profession
indicates the aspiration of all educators and provides standards by which to judge conduct.
The remedies specified by the NEA and/or its affi liates for the violation of any provision
of this Code shall be exclusive and no such provision shall be enforceable in any form other
than the one specifically designated by the NEA or its affi liates.
Principle I
Commitment to the Student
The educator strives to help each student realize his or her potential as a worthy and effec-
tive member of society. The educator therefore works to stimulate the spirit of inquiry, the
acquisition of knowledge and understanding, and the thoughtful formulation of worthy
goals.
In fulfi llment of the obligation to the student, the educator—
1. Shall not unreasonably restrain the student from independent action in the pursuit of
learning.
2. Shall not unreasonably deny the student’s access to varying points of view.
3. Shall not deliberately suppress or distort subject matter relevant to the student’s
progress.
4. Shall make reasonable effort to protect the student from conditions harmful to learn-
ing or to health and safety.
5. Shall not intentionally expose the student to embarrassment or disparagement.
6. Shall not on the basis of race, color, creed, sex, national origin, marital status, politi-
cal or religious beliefs, family, social or cultural background, or sexual orientation,
unfairly—
a. Exclude any student from participation in any program
b. Deny benefits to any student
c. Grant any advantage to any student
7. Shall not use professional relationships with students for private advantage.
8. Shall not disclose information about students obtained in the course of professional
service unless disclosure serves a compelling professional purpose or is required by
law.
Principle II
Commitment to the Profession
The education profession is vested by the public with a trust and responsibility requiring the
highest ideals of professional service.
In the belief that the quality of the services of the education profession directly influ-
ences the nation and its citizens, the educator shall exert every effort to raise professional
standards, to promote a climate that encourages the exercise of professional judgment, to
achieve conditions that attract persons worthy of the trust to careers in education, and to
assist in preventing the practice of the profession by unqualified persons.
In fulfi llment of the obligation to the profession, the educator—
1. Shall not in an application for a professional position deliberately make a false state-
ment or fail to disclose a material fact related to competency and qualifications.
2. Shall not misrepresent his/her professional qualifications.
3. Shall not assist any entry into the profession of a person known to be unqualified in
respect to character, education, or other relevant attribute.
4. Shall not knowingly make a false statement concerning the qualifications of a candi-
date for a professional position.
5. Shall not assist a noneducator in the unauthorized practice of teaching.
6. Shall not disclose information about colleagues obtained in the course of professional
service unless disclosure serves a compelling professional purpose or is required by
law.
7. Shall not knowingly make false or malicious statements about a colleague.
8. Shall not accept any gratuity, gift, or favor that might impair or appear to influence
professional decisions or action.
FOCUS ON ASSESSMENT
ETHICS IN ASSESSMENT
Examine the following ethics pledge and imagine that you have been asked to review and
sign it.
ETHICS PLEDGE
I, (sign your name here), hereby agree to accept and abide by the code of ethics established within my
profession as a teacher throughout my entire professional career. I understand that the primary purpose
of these ethical principles and guidelines is to protect the students I teach and not to safeguard myself or
my practice as an educator. I also understand that the advocacy for the rights and educational needs of
children may create potential conflicts with the school system that employs me. Consequently, I must
adhere to my code of ethics even if doing so adversely impacts my personal status in that system. I also
accept the responsibility of remaining current on new developments and professional responsibilities so
that I can work proactively on enhancing school policies and procedures and ultimately on serving the
interests and needs of all children in the classroom.
If you were asked to sign this ethics pledge, or one like it, would you? Why or why not? If
possible, sit down with a small group of your peers and colleagues and discuss the potential
challenges to adhering to an ethics pledge like the one just presented. Discuss the role eth-
ics will play in your professional career as a teacher.
Bananastock/AGE Fotostock
CHAPTER REFLECTION
Reexamine the defi nition of classroom assessment that you completed earlier in the
chapter. Along with your defi nition, you may be envisioning how, as a process, assess-
ment might take place in your future classroom. You also completed a brief description
of the relationship between teaching and assessment in that chapter activity. At this time,
would you make any further changes or modifications to your defi nition? If so, what
would those changes be and why would you make them? Compare your new defi nition
with your original and examine what has remained and what has changed. Are you sur-
prised with your fi ndings? Why or why not?
In generating your own defi nition of assessment, you begin to make it real and per-
sonal for yourself as a future educator. Don’t be surprised if that defi nition continues
to change; that’s not unusual. What’s most important (and underlies the instructional
purpose for completing this activity) is to help make assessment a relevant and meaning-
ful issue for you at this time in your professional preparation. As the assessment journey
continues, you will hopefully begin to see more critical connections between teaching
and assessment, which are examined in more detail in the next chapter.
CHAPTER REVIEW
Conclusion
As Black and Wiliam (1998b) have stated, “Learning is driven by what teachers and pupils
do in classrooms” (p. 139). Because of that reality, what becomes critical is the purpose-
ful and accurate measurement of the learning of students. Through a strategic approach
of addressing the instructional questions of where students are, where they need to be,
and confirming their educational destinations through the collection of data or learn-
ing evidence, teachers can follow and support instructional plans with confirmation of
P–12 learning impact. In this way, teachers can consistently evaluate the knowledge and
skill levels of their students before, as well as while, a lesson is being delivered, identify
the end point and outcomes of the instructional process, and determine the instructional
pathways that will need to be taken, along with monitoring the learning status of the
students during the process.
Assessment exists as the essential complement to teaching: the fl ip side of the instruc-
tional coin. With an effective classroom assessment system in place, a valid demonstra-
tion of student learning and progress connected to classroom instruction and experience
can be confi rmed. In addition, if the classroom assessment system is aligned with the
intended academic content standards, then direct evidence that students have acquired
expected knowledge and skills mandated by district, state, or national standards can be
provided.
Assessment is now an expected skill for teachers, and competency must be demon-
strated. It is a skill that is expected to grow and develop over one’s professional career
and one that must be guided by the highest standards and ethics of the profession. On
a personal level, accurate assessment of learning in the classroom provides the affi rma-
tion that a teacher’s professional effort and time helped to generate positive, and poten-
tially life-enhancing, accomplishments for his or her students. Along with this comes
an undeniable sense of pride and satisfaction that students have truly benefited from
the lessons and the learning experiences they have been provided.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
RESOURCES
The Standards for Teacher Competence in Educational Assessment of Students at www
.unl.edu/buros/bimm/html/article3.html demonstrate the profession’s recognition for
teacher assessment competency.
Currently, the book Standards for Educational and Psychological Testing (American Educational
Research Association, American Psychological Association, & National Council on
Measurement in Education, 1999) provides the most up-to-date information on the
standards that need to be recognized and followed whenever testing and assessment
procedures are implemented (e.g., test construction, fairness in testing, the use of test
findings, educational testing and assessment).
The National Board for Professional Teaching Standards (NBPTS) at www.nbpts.org pro-
vides detailed information on the five core standards (including assessment and monitor-
ing student learning) that must be met in order to receive the distinction of a National
Board Certified Teacher.
The Council of Chief State School Officers (CCSSO) at www.ccsso.org is a national orga-
nization consisting of state and federal school officials that examine and review major
educational issues that affect K–12 education. Connected to that site is information on
the Interstate New Teacher Assessment and Support Consortium (INTASC). INTASC
is a consortium of state education agencies and national educational organizations that
directly influences the preparation and professional development of teachers. Along
with content standards, INTASC has created 10 core professional standards, and Stan-
dard 8 deals exclusively with assessment.
REFERENCES
American Educational Research Association, American Psychological Association, &
National Council on Measurement in Education. (1999). Standards for educational and
psychological testing (3rd ed.). Washington, DC: Author.
Assessment Reform Group. (1999). Assessment for learning: Beyond the black box. Cambridge,
UK: University of Cambridge School of Education.
Beck, M. D. (2007). Alignment as a psychometric issue. Applied Measurement in Education,
20(1), 127–135.
Black, P., & Wiliam, D. (1998a). Assessment and classroom learning. Assessment in Educa-
tion, 5(1), 7–74.
Black, P., & Wiliam, D. (1998b). Inside the black box: Raising standards through class-
room assessment. Phi Delta Kappan, 80(2), 139–148.
Black, P., & Wiliam, D. (2009). Developing the theory of formative assessment. Educa-
tional Assessment, Evaluation, and Accountability, 21, 5–31.
Bloom, B. S. (1976). Human characteristics and school learning. New York: McGraw-Hill.
Bloom, B. S. (1984). The 2 sigma problem: The search for methods of group instruction
as effective as one-to-one tutoring. Educational Researcher, 13, 4–16.
Bloom, B. S., Hastings, J. T., & Madaus, G. F. (1971). Handbook on formative and summative
evaluation of student learning. New York: McGraw-Hill.
Brookhart, S., Moss, C., & Long, B. (2008). Formative assessment that empowers. Educa-
tional Leadership, 66(3), 52–57.
Boston, C. (2002). The concept of formative assessment. Practical Assessment, Research &
Evaluation, 8(9). Retrieved from http://pareonline.net/getvn.asp?v=8&n=9
Chappuis, J. (2005). Helping students understand assessment. Educational Leadership, 63(3),
39–43.
Chappuis, S., & Chappuis, J. (2008). The best value in formative assessment. Educational
Leadership, 65(4), 14–19.
Cohen, S. A., & Hyman, J. S. (1991). Can fantasies become facts? Educational Measurement:
Issues and Practice, 10(1), 20–23.
Earl, L. (2003). Assessment as learning: Using classroom assessment to maximize student learning.
Thousand Oaks, CA: Corwin Press, Inc.
Fuchs, L. S., Deno, S. L., & Mirkin, P. K. (1984). The effects of frequent curriculum-
based measurement and evaluation on pedagogy, student achievement, and student
awareness of learning. American Educational Research Journal, 21, 449–460.
Fuchs, L., & Fuchs, D. (1986). Effects of systematic formative evaluation: A meta-analysis.
Exceptional Children, 53, 199–208.
Guskey, T. R. (2003). How classroom assessments improve learning. Educational Leader-
ship, 60(5), 7–11.
Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research,
77(1), 81–112.
Marzano, R. J. (2006). Classroom assessment and grading that work. Alexandria, VA: ASCD.
McTighe, J., & O’Connor, K. (2005). Seven practices for effective learning. Educational
Leadership, 63(3), 10–17.
National Education Association (NEA). (1975). Code of ethics. Retrieved from http://
www.nea.org/home/30442.htm
No Child Left Behind Act of 2001 (ESEA), Pub. L. No. 107-110 (2002).
Stiggins, R., Arter, J., Chappuis, J., & Chappuis, S. (2004). Classroom assessment for student
learning: Doing it right—using it well. Portland, OR: Assessment Training Institute, Inc.
Stiggins, R., & DuFour, R. (2009). Maximizing the power of formative assessments. Phi
Delta Kappan, 90(9), 640–644.
U.S. Department of Education. (2010). A blueprint for reform: The reauthorization of the
Elementary and Secondary Education Act. Retrieved from http://www2.ed.gov/
policy/elsec/leg/blueprint/blueprint.pdf
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How do the instruction-learning-assessment components fit together during a
classroom activity or lesson?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Identify the key elements of the instruction-learning-assessment model.
♦ Explain the components of the instruction-learning-assessment model.
♦ Justify the necessity of sharing intended learning targets with students.
♦ Acknowledge the purpose and value of pre-assessment.
♦ Organize a pre-assessment activity.
♦ Identify the purpose of an instructional
preview.
♦ Acknowledge the requirement of research sup-
port for instructional approaches or techniques
used in the classroom.
♦ Understand the importance of feedback and its
impact in promoting effective learning.
♦ Differentiate the purposes of formative assess-
ment, summative assessment, and self-
assessment and how evidence from these
approaches can be used to promote student
learning.
Bananastock/AGE Fotostock
25
Introduction
The relationship and the components that link teaching, learning, and assessment
together are examined in this chapter. In addition, as a teacher-in-training, it is impor-
tant to know the assessment expectations that you will be held accountable to and more
importantly how these assessment competencies will impact your role as a teacher. To
help start this investigation, please complete the following activity.
CHAPTER ACTIVITY
Many, if not most, states have specific assessment standards and competencies that teachers
are expected to meet and demonstrate as part of their teaching practice. Can you identify
and list the assessment standards and skills that teachers in your state are expected to dem-
onstrate as part of their teaching? If you can, that’s great, since most pre-service teachers
are not typically exposed to that knowledge set early in their program of study. You are just
beginning to develop your assessment foundation; however, familiarity with the assessment
standards that you will be expected to demonstrate and will possibly be evaluated against
is very important.
When conducting a search on assessment standards, one of the fi rst places you may want
to visit is your state education Web site to see if specific teacher assessment standards exist
and where that information can be obtained. In addition, university faculty and teachers
in the field are excellent resources, as they can provide useful information and insight into
the assessment demands of a teacher in the classroom. As you conduct your online search
and participate in discussions with professional colleagues, ask yourself if you possess the
knowledge and skills that are necessary to demonstrate at least a beginning level (i.e., a level
of understanding and skill that is expected for current practitioners) of proficiency relative
to the listed competencies.
Based on the following assessment competencies (selected from the Ohio, Hawaii, and
Wisconsin educator assessment standards), rate your current assessment knowledge and
skills on a scale of 1 (extremely weak) to 5 (very strong) relative to these statements:
a. Knowledge of various assessment types, their appropriate uses, and the data they can
generate. (1 2 3 4 5)
b. Ability to identify, develop, and utilize a variety of assessments (e.g., formative, sum-
mative, self ). (1 2 3 4 5)
c. Ability to use data to review and monitor student progress and learning, and to plan or
differentiate instruction based on collected data. (1 2 3 4 5)
d. Ability to collaborate with and effectively communicate academic progress to students,
parents, and colleagues. (1 2 3 4 5)
e. Ability to involve students in self-assessment and goal setting in relation to their learn-
ing goals. (1 2 3 4 5)
Now examine your ratings for these items. Don’t worry if you couldn’t give yourself 4s
and 5s across these listed competencies. For now it’s just important to fi nd out where you
see yourself relative to these skill areas. Please record your answers, as your ratings for these
items will be discussed at the end of the chapter.
CLASSROOM TEACHING
AND ASSESSMENT
Pre-service teachers possess the unique advantage of being able to construct and refi ne
their classroom assessment system before they begin their teaching career. Clearly, much
can be learned, developed, and formulated before your system becomes operational in
a real classroom setting. Unfortunately, a real drawback of any theoretical system is the
limited opportunity to try out and conduct actual assessment practices in a classroom
setting. The fi rst chance to try out an assessment system typically occurs during student
teaching, and that is still a limited, full-time experience.
For practitioners in the field, the use and integration of an effective classroom assess-
ment system typically requires changes and adjustments to existing classroom routines
and instructional procedures. As pointed out in Assessment for Learning by Black, Har-
rison, Lee, Marshall, and Wiliam (2003), considerable time and effort may be necessary
for teachers when incorporating effective assessment practices into the classroom. But it
can and should be viewed as an organized redistribution of effort (not adding more to
an existing work foundation) that is developed over time in a careful, methodic manner.
As Black et al. point out, “the message is not about working harder, but about working
smarter” (p. 2). But before you can get somewhere, you need to know where you are, and
that surely applies to the practice of classroom assessment.
Whenever you focus on teaching you must also focus on assessment, because they are
directly connected (Earl, 2003). Assessment as a process serves as an important decision-
making tool that teachers use to enhance the learning progress of their students. Since
assessment must function as a seamless part of the teaching process, several intercon-
nected parts or components should be present and operational during any formal lesson
plan or planned learning activity in order to maximize instructional impact (Figure 2.1).
Let’s examine each of these instructional components.
Identified
Learning
Targets
Summative Pre-Assessment
Assessment of Students
Instruction-
Learning-
Assessment
Self- Model Preview of
Assessment Instruction
Instructional
Formative
Approach and
Assessment
Techniques
toward acquiring the expected grade-level standards and proficiencies appropriate for
the students you are teaching. Typically, a district’s curriculum guide is designed to
help with this process, along with state supports and resources. In addition, many states
(e.g., Hawaii, Alaska, Indiana, Idaho, Georgia, Oregon, Louisiana, and Ohio) have cre-
ated exceptional instructional management Web sites that help teachers create lesson
plans that are matched up with the appropriate grade-level learning goals. In particu-
lar, the “Instructional Management System” section on the Ohio Department of Edu-
cation’s Web site (ims.ode.state.oh.us) provides valuable instructional resources that can
be used to develop lessons specific to content and grade level that are directly aligned
with the recognized academic content standards for that state. Customized lesson plans
and potential assessment items for the classroom are generated within this system. The
Ohio Resource Center (ohiorc.org) also provides instructional assistance and support in
regards to grade-mandated standards and competencies. Many states provide these kinds
of resources, usually in the form of Web pages or Internet links.
Bananastock/AGE Fotostock
Along with curriculum alignment, learners need to know and understand the
expected performance standards that are connected with the lesson and its material. This
is where identifying the knowledge and cognitive processing levels, according to Bloom’s
modified taxonomy (Anderson & Krathwohl, 2001), is most helpful for both teacher and
students. The level of understanding and performance needs to be clearly articulated
by the instructor and understood by the intended learners. For example, if the higher
order processing skills of analysis and evaluation are expected, then those skills need to
be taught and practiced within the lesson in order for the students to acquire those very
skills. Yet if knowledge and comprehension skills are emphasized in the classroom, then
students will most likely acquire and use those processing levels in their work. Remem-
ber that what is taught, and at the skill level at which it is taught, is what will likely be
demonstrated by those who receive the teaching.
Also, classroom assessments must directly align with the intended learning targets of
the lesson. For example, if a teacher presents a lesson on the significance of the rights that
are protected under the First Amendment, and the instruction is formatted toward that
goal but at the end of the lesson a test is given on the three branches of government and
their responsibilities, a clear mismatch and instruction-learning-assessment misalignment
exists.
Pre-Assessment
Once learning targets are known and shared with the intended learners, pre-assessment
prior to the teaching of a lesson is needed for the diagnostic purpose of fi nding out
what students know and don’t know about the planned learning activity (Oberg, 2009).
Two basic areas typically reviewed in pre-assessment include (1) assessment of prerequi-
site skills, and (2) assessment of the information and skills to be taught. Both areas are
important. Assessment of prerequisite skills, however, is not always necessary given that
information from previous assessments or progress reviews can confi rm the existence of
certain knowledge and comprehension as well as skill levels.
Pre-assessment is usually designed to focus on determining how well-informed stu-
dents are about a new topic (e.g., the causes and impact of the American Civil War) or the
strength of a skill area (e.g., multiplying two-digit numbers with and without regroup-
ing). For example, for a third-grade English lesson designed to teach skills and strategies
for writing persuasive paragraphs, an effective pre-assessment activity may involve ask-
ing students to write a paragraph expressing and defending a point of view relative to
an article they have just read. With this kind of activity, the teacher can determine the
students’ writing skills and their performance levels before the lesson is started. With the
use of learner or student pre-assessment data, a lesson can proceed as originally planned
or it can be appropriately modified based on the collected fi ndings.
Important pre-assessment considerations should include the realization that all of the
intended learners need to be pre-assessed in order to get an accurate assessment of the
entire classroom. Getting a “representative sample” where the selection of a few students
reflects the entire class, even if that is possible, is likely to be difficult and inaccurate.
Therefore, involving the entire class in the pre-assessment activity or review provides the
most valuable and accurate performance picture. Also, the pre-assessment activity should
gauge the students’ current skills and knowledge relative to the learning targets and out-
comes of the intended lesson or activity that is to be provided.
Pre-assessment typically requires a short period of time and can be easily embedded
within the normal instructional process. In some cases, it can occur before the actual
instructional unit is even started. For example, if the pre-assessment is focused on deter-
mining whether the students have the prerequisite skills necessary for the current lesson,
a review of the last test performance and/or current work products that involved these
skills should provide that needed information. All that is required is checking previous
work, prior to the lesson, to determine that they have the necessary skills and under-
standing to benefit from the lesson that is to be provided.
When the pre-assessment involves reviewing how much students know relative to the
information and skills that are to be covered with the current lesson or learning activity,
then a checklist can be completed to determine whether or not they are demonstrating
these skills in their current work. The checklist review can take place a day or two before
the new lesson is to be provided. If new content is being introduced, then a pre-test of
the new material, in the form of a short paper-and-pencil performance, computer-based
review program, or small group oral question-and-response approach, could be admin-
istered. Obviously, these activities would need to be administered before the new lesson,
since the acquired information is critical in determining whether the lesson proceeds as
originally designed or whether it should be modified.
Pre-assessments can be packaged in different ways. For instance, they can take the
form of pre-tests that review specific content and material or they can be checklist-
oriented approaches that list the necessary steps for certain operations, such as a suc-
cessful computer search or the solution of an algebraic equation. Whatever method is
considered, the pre-assessment should confirm that the intended learners have the prior
knowledge and the prerequisite skills to complete the work (Sousa, 2006) in the intended
FOCUS ON ASSESSMENT
USE OF CONSENSUS-GRAMS
If time is limited or committed to other lesson components, an attempt to acquire “per-
ceptional” information from students can be made through the use of a consensus-gram. A
consensus-gram usually takes the form of a chart turned away from the class that contains a
low to high scale for students to anonymously mark to indicate their perceived understand-
ing, skills, and/or performance level relative to a statement or specific question asked by
the instructor. This will not reflect a direct skill or knowledge review, but it can provide
a general measure of the students’ perceptions and/or beliefs regarding their skill level and
comfort with the topic that is to be presented.
For example, the students could be asked to rate their knowledge from 1 (very low) to 5
(very high) based on the following question: “List and explain the steps that occur during
the process of photosynthesis.” The responses would then be reviewed to see, in general,
how confident the students are in their knowledge of the steps that occur during photo-
synthesis. This information can help give a teacher a better understanding of the learning
status of his or her audience before actual content or instruction is provided.
Overall, a consensus-gram is:
• A quick and relatively nonstressful self-appraisal activity for students.
• Useful in generating students’ perceptions of the skill/knowledge sets that are to be
used and/or developed during the instructional process.
• A tool that provides information that can be used to inform and direct instruction.
A simple show of hands (or a thumbs-up/thumbs-down) might also work, but realize
that there tends to be a “false positive” or a “fake good” effect with that practice. In order
to avoid social embarrassment and ridicule, students will often indicate that they know or
can do something when in reality they can’t do it at all or as well as indicated by the show
of hands.
lesson, as well as verify that the material in the soon-to-be-delivered lesson has not been
previously learned or mastered.
Beyond the mechanics of providing an accurate pre-assessment lies the importance
of understanding what students understand about the subject area before the lesson is
started and the new content is examined ( Jensen, 2005). An important consideration
with any preview is that teachers understand the pre-existing knowledge base of the
intended learners. As reported by the National Research Council (2000), “teachers need
to pay attention to the incomplete understandings, the false beliefs, and the naïve ren-
ditions of concepts that learners bring with them to a given subject” (p. 10). Teachers
must deal with this reality, because “If students’ initial ideas and beliefs are ignored, the
understandings that they develop can be very different from what the teacher intends”
(National Research Council, 2000, p. 10). It is imperative for teachers to know what
their students know and how well they understand it before embarking on new material
or understanding.
ASSESSMENT IN ACTION
© image100 Ltd
A carefully constructed preview can provide learners with a valuable setup and frame-
work that can be used to structure all subsequent lessons and activities. It sets the tone
and provides the educational expectation of where everyone should end up when the
instruction is completed. It is basic human nature to want to know where you are going;
students at every grade and age level are no different. A preview provides the impor-
tant road signs that need to exist throughout the educational journey, as well as at the
intended fi nal destination.
Formative Assessment
Within the field of assessment, no area has received more recent attention or contains
greater potential educational opportunity for teachers and students than formative assess-
ment (Brookhart, Moss, & Long, 2008; Fisher, Grant, Frey, & Johnson, 2008; Heritage,
2007; Stiggins & DuFour, 2009; Wiliam, 2006). Formative assessment, or “assessment
for learning,” serves an essential and direct purpose: to provide the learner, as well as the instructor,
with useful feedback regarding his or her present performance in order to improve or enhance that per-
formance in the future. As Black and Wiliam (1998) have indicated, assessment recognized
as formative serves this purpose only when student learning evidence is used to directly
influence the teaching process. Students are active participants in the classroom setting,
but it is the teacher who facilitates the learning process by providing helpful observations,
quick checks, or other active feedback, corrective or otherwise, to the student during a
lesson or activity. This type of assessment is not evaluative by design, so a “no cost” effect
is associated with the performances or products that are generated by students. Conse-
quently, learning and performance mistakes, if made, do not alter or directly impact a
student’s formal status, grades, or general standing within a class.
Research has documented the substantial impact of formative assessment on increasing
academic achievement, particularly for low-performing students (Black, Harrison, Lee,
Marshall, & Wiliam, 2003; Black & Wiliam, 1998; Fuchs & Fuchs, 1986). The instruc-
tional framework of the corrective feedback aspect of the formative process connects to
the theoretical work and earlier research of Bloom and his “mastery learning” approach
(Bloom, 1971, 1974). Consistent with the instructional model was the use of formative
learning assessments that were designed to give students individualized feedback on their
learning, including areas of accomplishment as well as issues that required further work
and practice (Bloom, 1976). In this way, minor or temporary learning problems that
exist are addressed and corrected, thereby minimizing the likelihood of the development
of significant long-term learning issues. As reported by Guskey (1997), Bloom’s mas-
tery approach centers on two primary components: (1) learning performance feedback
(including corrective and enrichment support), and (2) instructional alignment. Students
need to know what they have acquired and can complete independently, as well as what
they haven’t learned.
According to Bloom, specific performance feedback is only part of the needed
response. A corrective process and response is also necessary. For identified problem
areas (e.g., skill deficits, weak or minimal skill progression, inconsistent performance
levels), remediation strategies and procedures need to be followed in order for students
to continue to make sustained progress. Students need to be informed and shown that
they have not acquired the desired learning targets. When this happens, their instruction
should be modified to incorporate needed intervention strategies that will help acquire
the intended learning objectives and outcomes, recognizing that a variety of strategies
SOURCE: Guskey, T. R. (2005, April). Formative classroom assessment and Benjamin S. Bloom: Theory, research,
and implications. Paper presented at the meeting of the American Educational Research Association, Montreal,
Canada.
and techniques may be necessary given the wide range of student needs. This reflects the
“re-teaching” or continued instruction that should be a part of any formative assessment
process (Heritage, 2007).
Guskey (2005) stipulates that Bloom’s notion of instructional alignment consists of
three interconnected foundations (see Table 2.1). First, clear and identifiable learning
goals or standards need to be identified. This notion corresponds directly to “where stu-
dents need to be” in the instruction-learning-assessment model. A clear endpoint with
skill and performance expectations should exist, and this “learning destination” must
be communicated to the learners before the instructional lesson is ever started. As you
and your students follow your educational map, this is the place you want them to end
up when the journey is completed. This would be followed by appropriate and effective
instruction that should generate informed and competent learners relative to the pro-
vided lesson plan material.
In combination with the instruction comes the feedback and corrective instruction
that is necessary in order for the learners to maintain the proper instructional pathway
and the intended level of progress. As Guskey (2005) has stated:
Providing feedback, correctives, and enrichments; and ensuring instructional
alignment takes relatively little time and effort, especially if tasks are shared col-
laboratively among teaching colleagues. Still, results show that the systematic use
of these elements helps many more students learn well, significantly reduces varia-
tion in student learning outcomes, and closes gaps in the achievement of different
groups of students (p. 9).
Formative assessment is focused on giving systematic confi rming or corrective
instructional feedback to students along the instructional pathway. In this way, student
learning should not veer off the intended pathway, at least not without the awareness
of the teacher, and if misdirection does occur, corrective action can immediately take
place. Throughout the process, students remain informed of their relative progress and
the learning challenges and goals that remain as part of the provided lesson plan or learn-
ing activity. Formative assessment in conjunction with learning feedback allows a teacher
(along with his or her students) to confi rm, refi ne, and direct self-correction procedures,
if correction is necessary. Through this process, students are more likely to achieve genu-
ine competency or “mastery” of the knowledge and skills of the presented lesson.
Learning Feedback
A major factor that contributes to successful and effective learning is feedback. Critical
elements of effective teacher feedback that have been identified include the following
(Hattie & Timperley, 2007; Van Houten, 1980):
♦ identifying and addressing a specific aspect of a performance;
♦ precise performance feedback;
♦ immediate feedback as opposed to delayed feedback;
♦ feedback frequency;
♦ positive feedback regardless of performance;
♦ correct modeling of the desired response; and
♦ differential feedback with an emphasis on improvement.
Collectively, these practices are typically recognized by educators as basic and founda-
tional to student learning success. So how can these practices be used to directly influ-
ence the feedback process in the classroom? What should the feedback that is provided in
your classroom look like? In general, feedback needs to be specific (as opposed to general)
in scope to the individual learner’s performance and to the task that is to be completed.
It is viewed as more effective if provided as soon after the event is completed as possible,
and preferably more than once over the course of the activity. More feedback is prefer-
able to less, and receiving feedback over multiple occasions is more effective. In addition,
positive statements and actions are much more powerful and reinforcing than those that
are negative or punitive in nature.
Based on a comprehensive review of the literature, particularly in feedback systems
for school improvement, Visscher and Coe (2002) have determined that the research on
academic feedback is still developing and requires continued investigation. Yet despite
the lack of detailed information on the complex factors that influence the effectiveness
of feedback in the classroom environment, a few tentative findings have been identified
that are connected with the beneficial effects of feedback on school performance. They
include the following:
♦ Feedback should focus specifically on the obtained and desired performance as it
relates to the learning task. Moving beyond the task itself to focus on self-esteem
issues or additional skills can potentially lessen the impact of the classroom feed-
back process relative to the desired learning task and target.
♦ Ensuring perceived competency in the learner, along with providing and selecting
useful intervention suggestions (e.g., greater effort, more review, different learning
strategies), is important for learner success.
♦ Individual performance comparisons should be followed, as opposed to compari-
sons to others, and feedback should be designed to develop specific goals that are
needed to meet the desired performance level.
♦ Feedback should be directed toward completing the primary task. Although valu-
able and necessary, completing secondary aspects of the task should not overshadow
the need to address the central aspects of the task.
Bananastock/PictureQuest
♦ Feedback should be useful, accurate, and fair. It must have perceived value for the
learner. If it does not, then the feedback will not be relevant or desired, and an
intended change in behavior is unlikely.
As a matter of actual classroom practice, many teachers differentiate their feedback, usu-
ally informally, into the areas of descriptive feedback (informing the learner of what was
done correctly as well as incorrectly); corrective feedback (providing specific help and
clarity on amending incorrect work); and evaluative feedback (how strong the student’s
work or performance was based on the requirements of the task or project). Providing
different levels of information and feedback to students is essential if student progress is to
be maintained. Learning feedback is really about giving useful information to the learner
in such a manner that it is positively received and acted upon. This requires knowing
what to do, knowing what to say, and knowing how to say or present it so it is accepted.
As a teacher, it is important to recognize the immense value that feedback has on the
learning process.
Self-Assessment
Self-assessment or “assessment as learning,” according to Earl (2003), is an expansion area
in the field of assessment that holds great learning potential for students as well as teach-
ers. Within the educational arena, self-assessment as a process involves the internal review
of an individual’s progress along with the recognition of required adjustments and modifi cations,
based on that review, that need to be made to reach an intended learning target. With this form of
assessment, students learn to self-monitor their own learning progress and to recognize
needed learning “corrections” (e.g., to re-examine certain terms or follow a set of pro-
cedural steps in proper order) along the way so that they can attain the intended learning
outcomes of the lesson or activity. In order to develop this skill, however, conditions and
opportunities must be provided in the classroom where students can learn and practice
how to evaluate and monitor their own learning progress.
An internal review process requires the active instruction of procedures and tech-
niques that the learners are to demonstrate independently. For example, activities such as
implementing periodic progress checks, work completion reviews, and additional repeti-
tions or practices can all be included within a student’s self-assessment program. But pro-
viding instruction, giving opportunities, and even modeling procedures are often needed
for self-assessment development. If teachers follow a self-assessment process, then this
will not only enhance their instruction but will also help their students acquire the skill
in question.
Educators are in the best position to review their own teaching and its impact on
students. A teacher is the fi rst one to know what worked and what didn’t when a lesson
is provided or when an instructional activity is completed. As part of the self-review
process, teachers can systematically determine what should be kept and what should
be eliminated based on the desired effect of the learning activity. When instructional
adjustments are made, classroom feedback from students, along with their acquired and
FOCUS ON ASSESSMENT
demonstrated skill attainment, can provide confi rmation as to whether or not the desired
instructional and learning effects were obtained.
Self-assessment is a process that requires effort, practice, and critical review on the
learner’s part, and it needs to be a community event. It should be “public” because one
of the primary ways we learn is by watching and modeling the behaviors of others, and
students defi nitely watch their teachers (Bandura, 1969, 1973, 1977; Bandura & Walters,
1963). By modeling self-assessment as a productive and useful learning tool, teachers can
provide direct instruction to their students as well as provide practical experience with
the process.
Beyond the benefit of teaching students to be more effective learners in the classroom,
self-assessment is a life skill that can serve learners in many aspects. Given the complexity
and challenges of today’s world, we want young people to make “good decisions” and the
“right choices.” In order to do this, they need to possess an effective review process, and
self-assessment can help with that mission.
Summative Assessment
Summative assessment, or “assessment of learning,” is the confirmation of accumulated
learning through a formal evaluation process and represents the formal and traditional measure and
evaluation of a student’s accumulated learning. Within the classroom, summative assessment
involves the formal measurement and evaluation, often in the form of testing, of stu-
dent learning at some predetermined point. Chapter tests or fi nal exams serve as key
classroom examples of this kind of measurement, and they are designed to evaluate an
individual’s acquired learning. Most tests that are given in school are used to fulfi ll a
summative function.
Anyone who has gone through a formal education system has experienced the effects
of summative assessment. It gets the most attention from both the field of education and
the general public. Information from summative assessment is used for various official
school purposes such as grading, report cards, promoting students to the next grade, or
selecting students for advanced study courses. This form of assessment is important and
clearly essential, given its basic function of measuring learning gains of students; how-
ever, because of the continuous push for accountability and ever-higher test scores, the
importance of summative assessment often overshadows other forms of assessment.
Results from standardized achievement tests are summative in design and are consid-
ered and often used as official indications of a student’s skill level and learning progress
in certain content areas such as reading, math, and science over a specified period of time
(usually an academic year). Review of academic progress often involves state and national
comparison levels. International comparisons are also available with some standardized
achievement tests. In addition to wanting to know if students can perform specific tasks
and demonstrate certain skills, we, as a demanding public, also want to know how well
the students we teach perform when compared to their peers, whether they are in the
next town, state, or country, and this desire will likely never go away. Like it or not,
many envision the educational process as a race and want to know how the children are
performing relative to one another in that race.
Achievement tests that are completed in school are often viewed as “high stakes” tests
due in large part to NCLB. By law, school districts are held directly accountable for
the performance and progress of all students. If reasonable progress (determined by each
state) is not evidenced, possible fi nancial restrictions and other sanctions can be adminis-
tered. What is unique about NCLB is that all children must demonstrate academic prog-
ress, including students with special needs and English language learner (ELL) students,
as well as other students who traditionally score poorly on standardized tests. School
districts have a vested interest in making sure all students are performing and progressing
appropriately in the classroom in order to have the greatest opportunity to perform well
on these measures. And that is as it should be, although this places intensive demands on
each and every school, especially those with limited resources and support (e.g., instruc-
tional materials, classroom aides) to meet this mandate.
Summative assessment serves as the formal evaluation of student learning progress at
designated instructional points. Due to this important function, this form of assessment
will not diminish now or in the future. Therefore, it is essential that valid and effective
measures be constructed and utilized by teachers. It is imperative that any constructed
summative assessment link directly to the learning objectives or outcomes that are iden-
tified with the provided lessons or learning projects that are completed as part of an
instructional unit.
For students to genuinely learn and benefit from instruction, formative assessment,
summative assessment, and self-assessment all need to be part of the instructional pro-
cess. Although configured for different purposes, these assessment practices collectively
provide the opportunity for academic success to be maximized for all learners in the
classroom (Table 2.2). All have value and purpose and need to be considered and utilized
with any classroom assessment system.
CHAPTER REFLECTION
At the beginning of the chapter, you were asked to complete a self-rating on a composite
list of professional educator assessment skills. Now take a moment to review your ratings.
The five assessment competencies represent a broad range of expected skills. In particu-
lar, the fi rst competency (a) deals with general knowledge of assessment (which usually
has the highest rating), while the second (b) and third (c) competencies deal with the
ability to construct and implement assessments in the classroom and revise instruction
based on the data collected from those assessments. The second and third competencies
constitute the real “doing” part of assessment, and pre-service teachers, as well as practi-
tioners in the field, typically demonstrate less confidence with these skills; consequently,
their ratings are usually lower than their ratings on the fi rst competency.
The fourth competency (d) deals with the capacity to work with and communicate
assessment results effectively to others. Teachers often see themselves as possessing a real
strength in this area, due to the interpersonal aspect of this skill and its implementation.
Competency number five (e) represents a greater understanding of assessment, since this
involves “teaching” students how to self-assess their progress and set academic goals. This
is more than just knowing, it’s about teaching actual assessment concepts and content to
students in the classroom. This is generally perceived as more difficult, and many teachers
provide a lower rating, in fact maybe their lowest rating, with this competency relative
to the others listed.
These assessment competencies are representative of the professional skills that are
typically expected of teachers practicing in today’s classrooms across the nation. As is
evident from this short list, the expectation of multiple skills and the ability to use those
skills to monitor and improve instruction, as well as the direct instruction of this content
to students, exists for every teacher, regardless of content specialization and program
training. This is the reality and the challenge that faces each educator in the classroom.
When this form is used with teachers (both pre-service and practitioners), many, if
not most, generate ratings of 2s and 3s at best across the listed competencies. These
results are not surprising for pre-service teachers, as they are just beginning to develop
their assessment knowledge base and have had limited opportunities to develop and
practice their assessment skills, particularly in actual classroom settings and under real
classroom conditions. Because of this, as a group the ratings are usually low. However,
this pattern can also be found with practitioners in the fi eld, as some teachers receive
little formal instruction or training in the area of classroom assessment. What they have
learned about assessment they have learned in the fi eld as part of on-the-job training.
But it is important to note that what is learned on the job doesn’t automatically make it
“the right stuff.”
If you generated mostly 3s and 4s, some comfort or even perceived strengths may exist
within the skills that received those ratings. If 5s are listed across the board, then it’s pos-
sible you will gain little from the information in this book. This kind of rating profi le,
however, is unusual and does not represent the majority of pre-service or current teach-
ers. The important message here is that regardless of your particular ratings, it’s impor-
tant to know where you see yourself regarding the assessment skills and competencies
that will certainly be expected from you as a teacher.
If your ratings are low or lower than you would like, don’t worry. This is not a sign
that you will be a bad teacher. On the contrary, this should be viewed as a good teacher
becoming an even better one, since gaining new information will extend this evolving
professional skill to its highest level. This becomes an area of development, something
that requires opportunity, time, practice, and effort in order to develop into a strong,
useful skill.
CHAPTER REVIEW
Conclusion
Teaching, learning, and assessment are directly connected. This interconnected process
contains several key parts and components that need to work together in order to pro-
duce effective learning results. As you continue to examine and integrate this model into
your future teaching design, keep these components and their essential qualities in mind:
Learning Targets
♦ Clear, precise learning outcomes of what the students are expected to do must
exist.
♦ Targets need to be directly connected to the provided instruction and curriculum
(i.e., the specific subject content and skills that are covered in lessons).
♦ All assessment needs to be aligned with the learning outcomes or targets.
♦ Learners need to be informed of the learning targets prior to instruction.
Pre-Assessment
♦ All students should be involved in pre-assessment activities.
♦ Pre-assessment activities should measure intended learning targets.
♦ Pre-assessment data should be reviewed and instruction modified, if necessary,
based on that review.
Preview of Instruction
♦ Learners should be informed of what they will be learning and what they will be
expected to know and be able to do.
♦ The importance and relevancy of what is to be learned should be explained.
Selection of Instructional Techniques
♦ Evidence-based methods and techniques should be considered and selected.
♦ Plans for differentiated instruction need to exist.
Feedback
♦ When using feedback in the classroom, be as specific to the task and the student
performance as possible.
♦ Try to provide feedback as soon and as frequently as possible.
♦ Positive feedback and encouragement are much more effective and desired than
negative or punishing responses.
Formative Assessment
♦ This tool reflects an active feedback collection process used to make necessary
instructional decisions and to help address individual student learning needs and
issues.
♦ This is “learning at no cost,” as mistakes that are made are corrective, not evalua-
tive, in design, so there is no impact on a student’s grade.
♦ Formative assessment should provide helpful information and learning feedback to
every student along the instructional pathway.
Student Self-Assessment
♦ Students need to be taught how to assess their own learning progress and to make
essential adjustments based on information they collect.
♦ Students should be able to model self-assessment approaches by directly watching
and interacting with their teachers.
♦ Self-assessment is a lifelong skill that can serve students in every aspect of their lives.
Summative Assessment
♦ Summative assessment is a needed indicator of acquired learning at a particular
instructional point and is most often completed at the end of a lesson or activity.
♦ The assessment needs to link up directly with the intended learning objectives or
outcomes that are identified with the lesson or learning project.
♦ This assessment is evaluative in nature, where the quality of the learning and the
demonstration of the acquired skills is examined and reviewed.
♦ Constructing valid and effective summative assessments (e.g., quizzes, rubrics, or
exams) is a very important component in the overall classroom assessment process.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How do you see the various components of the instruction-learning-assessment
model operating together as a connected process?
2. What is gained (in regards to student learning) when teaching and assessment are
integrated together?
3. How can these instruction-learning-assessment elements fit within your current
lesson plan design?
RESOURCES
Many states possess specific assessment standards for teachers. Prospective teachers should
visit the Web site for the department of education of the state they intend to work in
so that they are aware of the assessment expectations for teachers within that state.
Teacher assessment standards and expectations are particularly comprehensive on the
following state sites: Ohio (esb.ode.state.oh.us), Hawaii (www.htsb.org), and Wiscon-
sin (dpi.wi.gov/tepdl/standards.html).
The Instructional Management System section of ims.ode.state.oh.us/ode/ims/lessons
serves as an excellent resource for teachers in Ohio in developing instructional lesson
plans that are integrated with effective assessment techniques and measures. Many
states have similar instructional resources.
Regarding NCLB, two major Web sites that provide current information, evidence,
and research on evidence-supported instruction include the What Works Clearing-
house (ies.ed.gov/ncee/wwc) and The Promising Practices Network (www.promising
practices.net).
The New Zealand Te Kete Ipurangi Web site at www.tki.org.nz/r/assessment provides
useful information on assessment and feedback and how these factors contribute to
successful learning.
REFERENCES
Anderson, L. W., & Krathwohl, D. R. (2001). A taxonomy for learning, teaching, and assess-
ing: A revision of Bloom’s taxonomy of educational objectives. New York: Longman.
Bandura, A. (1969). Principles of behavior modifi cation. New York: Holt, Rinehart & Winston.
Bandura, A. (1973). Aggression: A social learning analysis. Englewood Cliffs, NJ: Prentice
Hall.
Bandura, A. (1977). Social learning theory. Englewood Cliffs, NJ: Prentice Hall.
Bandura, A., & Walters, R. H. (1963). Social learning and personality development. New York:
Holt, Rinehart & Winston.
Black, P., Harrison, C., Lee, C., Marshall, B., & Wiliam, D. (2003). Assessment for learning:
Putting it into practice. Maidenhead, UK: Open University Press.
Black, P., & Wiliam, D. (1998). Inside the black box: Raising standards through classroom
assessment. Phi Delta Kappan, 80(2), 139–148.
Bloom, B. S. (1971). Mastery learning. In J. H. Block (Ed.), Mastery learning: Theory and practice
(pp. 47–63). New York: Holt, Rinehart & Winston.
Bloom, B. S. (1974). An introduction to mastery learning theory. In J. H. Block (Ed.),
Schools, society and mastery learning (pp. 3–14). New York: Holt, Rinehart & Winston.
Bloom, B. S. (1976). Human characteristics and school learning. New York: McGraw-Hill.
Brookhart, S., Moss, C., & Long, B. (2008). Formative assessment that empowers. Educa-
tional Leadership, 66(3), 52–57.
De Graaff, S., Bosman, A., Hasselman, F., & Verhoeven, L. (2009). Benefits of sys-
tematic phonics instruction. Scientifi c Studies of Reading, 13(4), 318–333. doi:
10.1080/10888430903001308
Earl, L. (2003). Assessment as learning: Using classroom assessment to maximize student learning.
Thousand Oaks, CA: Corwin Press, Inc.
Fisher, D., Grant, M., Frey, N., & Johnson, C. (2008). Taking formative assessment
schoolwide. Educational Leadership, 65(4), 64–68.
Fuchs, L., & Fuchs, D. (1986). Effects of systematic formative evaluation: A meta-analysis.
Exceptional Children, 53, 199–208.
Guskey, T. R. (1997). Implementing mastery learning (2nd ed.). Belmont, CA: Wadsworth.
Guskey, T. R. (2005, April). Formative classroom assessment and Benjamin S. Bloom: Theory,
research, and implications. Paper presented at the meeting of the American Educational
Research Association, Montreal, Canada.
Hattie, J., & Timperley, H. (2007). The power of feedback, Review of Educational Research,
77(1), 81–112.
Heritage, M. (2007). Formative assessment: What do teachers need to know and do? Phi
Delta Kappan, 89(2), 140–145.
Jensen, E. (2005). Teaching with the brain in mind (2nd ed.). Alexandria, VA: Association for
Supervision and Curriculum Development.
Lawrence-Brown, D. (2004). Differentiated instruction: Inclusive strategies for standards-
based learning that benefit the whole class. American Secondary Education, 32(3), 34–62.
McTighe, J., & O’Connor, K. (2005). Seven practices for effective learning. Educational
Leadership, 63(3), 10–17.
National Research Council. (2000). How people learn: Brain, mind, experience, and school.
Washington, DC: National Academic Press.p
Oberg, C. (2009). Guiding classroom instruction through performance assessment. Jour-
nal of Case Studies in Accreditation and Assessment, 1, 1–11. Retrieved from http://aabri
.com/jcsaa.html
Sousa, D. A. (2006). How the brain learns (3rd ed.). Thousand Oaks, CA: Corwin Press.
Stiggins, R., & DuFour, R. (2009). Maximizing the power of formative assessments. Phi
Delta Kappan, 90(9), 640–644.
Tomlinson, C. A. (2001). How to differentiate instruction in mixed-ability classrooms. Alexan-
dria, VA: Association for Supervision and Curriculum Development.
Tomlinson, C. A., Brighton, C., Hertberg, H., Callahan, C. M., Moon, T. R., Brimijoin,
K., Conover, L. A., & Reynolds, T. (2003). Differentiating instruction in response to
student readiness, interest, and learning profi le in academically diverse classrooms: A
review of literature. Journal for the Education of the Gifted, 27, 119–145.
Van Houten, R. (1980). Learning through feedback: A systematic approach for improving academic
performance. New York: Human Sciences Press.
Visscher, A. J., & Coe, R. (Eds.). (2002). School improvement through performance feedback.
Lisse, The Netherlands: Swets & Zeitlinger Publishers.
Wiliam, D. (2006). Formative assessment: Getting the focus right. Educational Assessment,
11, 283–289.
Yeynes, W. (2008). A meta-analysis of the relationship between phonics instruction and
minority elementary school student academic achievement. Education & Urban Society,
40(2), 151–166.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
What do my students need to learn in the classroom and how do I use student-
generated evidence to confirm that learning?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Identify the need to connect district and/or state curriculum requirements to
daily lesson plans.
♦ Understand the importance of connecting classroom assessment procedures to
identified learning goals and outcomes in the classroom.
♦ Utilize Bloom’s taxonomy table when integrating learning objectives into a
lesson plan.
♦ Explain the importance of empirical evidence in documenting the learning
progress of students in the classroom.
♦ Implement observation methods to record and examine
student performance in the classroom.
♦ Utilize questioning and interviewing as assessment
tools to confirm student understanding and progress.
♦ Select student products that provide useful evidence of
obtained learning outcomes.
♦ Utilize a variety of data collection measures and
methods for the documentation of student learning
© Getty Images outcomes in the classroom.
48
Introduction
“What do students need to learn and be able to do in the classroom?” is a fundamental
question in education and it is as relevant today as it was 50 or even a 100 years ago.
Not surprisingly, the answer to that question has changed many times based on what
has been determined to be the necessary knowledge and academic skill base for young
people. Even today, as the needs of society continue to change (e.g., technology, global
communication), our educational system must be able to recognize and adjust to those
evolving needs by providing meaningful learning experiences and opportunities that
prepare lifelong learners for the demands of an ever-changing world (Breivik & Senn,
1994; Longworth, 2003; Senge, Cambron-McCabe, Lucas, Smith, Dutton, & Kleiner,
2000).
CHAPTER ACTIVITY
On a piece of paper, list the 10 most important things that you believe you learned in
school. The list can include actual skills (operating a computer, reading, doing mathemati-
cal calculations, etc.), a specific knowledge base (foreign language, health, specialty areas
such as digital technology, etc.) or general life skills (interpersonal dynamics, functional
independence, social awareness, etc.), or any other acquisition that comes to mind. Once
the list is complete, ref lect on how these skills were acquired. Did they start in preschool
(e.g., reading) and develop throughout your educational career, or was it more of a later
post-secondary learning experience (e.g., public speaking)? Also, look at each of your list-
ings and determine why they are so important to know or be able to do. If possible, share
your list with your classmates or friends for comparison purposes and then answer the fol-
lowing questions:
1. Do you believe having these skills and possessing this knowledge makes you a produc-
tive and competitive member in today’s society? Why or why not?
2. Do you believe these skills and knowledge sets can be traced back to particular learn-
ing goals and outcomes that existed within the curriculums you were exposed to as a
student? Or were they obtained merely by educational chance or opportunity because
of the schools you attended?
3. Compared to your parents, do you believe you have been taught similar or different
skill and knowledge sets, and why do you think that is the case?
about mathematical processing? What level of computation and analysis is needed for
today’s high school graduates? What about technology? What about general problem-
solving skills, personal wellness, or interpersonal-based goals such as cooperation, col-
laboration, or critical thinking? It is likely that many of us would agree that foundation
life skills, perhaps even the ones just listed, are important and essential (Marzano, 2006).
Moreover, what about specific knowledge or content bases (e.g., science, social studies)
and the skills connected to these areas?
Without question, teachers identify personal learning goals for their students within
the academic realm (e.g., be able to read and comprehend material at or above grade
level, develop an enjoyment or even a love of learning, pass the AP exam in calculus)
although behavior-based goals (e.g., improved attendance, positive peer interaction) are
not unusual. Beyond the personal goals that teachers may hold for their students, each
school district maintains the responsibility to generate and follow a curriculum guide
that sets forth the expected academic goals and objectives, across subjects and within each
grade level, that all students within grades P–12 are expected to acquire and demonstrate
as part of their school performance.
What skills were essential for these students to be competitive learners in the early 20th century?
© Dynamic Graphics/JupiterImages
A curriculum guide reflects a school district’s documented plan of the goals and learning
outcomes that are expected to be acquired by students as they move from grade to grade. It essen-
tially serves as the educational list of learning requirements that students are expected to
meet before they graduate. For instance, in the sixth grade in the Lakota School District,
an expected learning outcome for students is to be able to independently write a letter,
complete with a stated purpose and function, using standard business letter format (Fig-
ure 3.1). Beyond the obvious importance of this information for the sixth-grade teacher
(where this skill must be taught and mastered), work on the prerequisite skills necessary
to complete this task must be done and completed in the earlier grades. Moreover, in the
later grades, teachers will be utilizing and building on this expected student-acquired
skill and competency to connect with more involved and detailed projects and products.
For a school district, a curriculum guide truly represents the master educational schedule
with the expected arrival times (i.e., grade or grade bands) and skill “destinations” that
are anticipated and required if students are ultimately to graduate with all their neces-
sary competencies. It also serves as an educational contract, in that the responsibility for
creating learning opportunities to acquire those skills, experiences, and knowledge sets
rests with the schools and the instruction that is provided throughout all the classrooms
in the district.
Curriculum guides are direct ref lections of mandated changes and requirements at the
state and/or national levels regarding expected student proficiency and skill development
(see the box “Ohio Science Academic Content Standards” on page 55). Included within
the curriculum guides are embedded standards, which are broad statements of academic goals
and competencies that students are expected to acquire, and benchmarks, or identified learn-
ing goals that are expected to be accomplished by a specified time, usually set at certain grade lev-
els. These learning goals are written and supported by professionals in academic content
areas (e.g., science, math, language arts) and endorsed through professional associations
that eventually appear as specific knowledge and skill expectations through identified
learning targets and outcomes for students in your classroom.
State-mandated academic standards require school districts to address specific learning
goals, usually at specified grade levels or grade bands (e.g., K–2, 3–5), and in so doing
try to ensure comparable learning outcomes for all students within a state. Therefore,
student learning goals and outcomes identified in school curriculum guides need to be
connected to the standards and benchmarks that are endorsed by state departments of
education. This connection is critical, as each statewide testing program, which is man-
dated in NCLB, will ref lect those intended learning goals. In order for students to dem-
onstrate strong test performances and results, each school district must make sure that
daily instruction ref lects the state standards, since those tests are specifically designed to
address and measure those learning outcomes and skills.
This is at the heart of the debate around high-stakes testing and accountability in P–12
education today. Most educators, along with the general public, are not against account-
ability. It is a necessary attribute of any formal profession and it is an essential quality for
a profession that is involved in the documentation of human learning. The controversial
part is determining what sources of student learning evidence should be used to measure
and make up that accountability.
• write persuasive essays that establish a clear reports to present information that supports a
position and include organized and relevant clear position with organized and relevant evi-
information to support ideas dence about the topic or research question
• write legibly in both formal and informal • demonstrate active listening strategies (e.g.,
writing asking focused questions, responding to cues,
• utilize spelling strategies to spell words correctly making visual contact)
in daily writing
• summarize the main idea and draw conclusions
• use correct capitalization and punctuation when from oral presentations and visual media
writing
• interpret the speaker’s purpose in presentations
• identify and use knowledge of all eight parts of
and visual media (e.g., to inform, to entertain,
speech (e.g., noun, pronoun, verb, adverb, adjec-
to persuade)
tive, conjunction, preposition, interjection) to
improve writing • identify the persuasive techniques (e.g., band-
• use knowledge of subject–verb agreement with wagon, testimonial, glittering generalities,
collective nouns, indefinite pronouns, com- emotional word repetition, and bait and switch)
pound subjects, and prepositional phrases to used in presentations and media messages
improve writing • demonstrate an understanding of the rules of
• generate a topic (assigned or personal inter- the English language, select language, and adjust
est) and open-ended questions for research and speaking content appropriate to purpose and
develop a plan for gathering information audience
• identify appropriate sources, and gather relevant • deliver formal and informal descriptive presen-
information from multiple sources (e.g., school tations that convey relevant information and
library catalogs, online databases, electronic descriptive details
resources, and Internet-based resources)
• deliver informational and persuasive presenta-
• compare and contrast important findings and
select sources to support central ideas, concepts, tions (e.g., expository, research) that: support
and themes the main idea or position with relevant facts,
details, examples, quotations, statistics, stories,
• use quotations to support ideas
and anecdotes; organize information, including
• use an appropriate source of documentation, a clear introduction, body, and conclusion and
with teacher assistance, to acknowledge sources follow common organizational structures when
(e.g., bibliography, works cited) appropriate (e.g., cause–effect, compare–
• use a variety of communication techniques contrast); and draw from several sources and
including oral, visual, written, or multimedia identify sources used
As part of effective assessment practice, there should be multiple sources of student per-
formance, but often the single measuring stick that the states and the federal government
rely on are mandated group achievement tests. The reasoning behind their use is that if
these mandated tests are well constructed and centered on identified academic standards,
consequently the results from these tests should provide useful and accurate data on how
well students are actually acquiring and learning those identified competencies.
Grade 2
Physical Sciences
Forces and Motion
1. Explore how things make sounds (e.g., rubber bands, tuning fork, and strings).
2. Explore and describe sounds (e.g., high, low, soft, loud) produced by vibrating objects.
3. Explore with fl ashlights and shadows that light travels in a straight line until it strikes
an object.
Grade 1
Explore the effects some objects have on others even when the two objects might
not touch (e.g., magnets).
Investigate a variety of ways to make things move and what causes them to change
speed, direction, and/or stop.
Grade 2
Explore how things make sounds (e.g., rubber bands, tuning fork, and strings).
Explore with flashlights and shadows that light travels in a straight line until it strikes
an object.
and appreciation of others, acceptance and tolerance, and social justice, as well as other
important learner goals. Learning targets can be identified that go beyond specific aca-
demic issues.
PhotoDisc/Getty Images
The term learning outcome, or what a student should know and be able to do because of
an instructional experience, is often used interchangeably with learning target; however, an
outcome is usually viewed as a desired effect whereas a learning target is an anticipated
aim or mark that is to be acquired. For example, in completing this chapter, an antici-
pated learning outcome for each reader would be that they recognize how various kinds
of student evidence can be used to confirm learning change and progress. Yet within this
chapter, learning targets would focus on specific and recognized acquisitions such as the
selection and usage of an observation technique to collect student data, or the completion
of a taxonomy table that lists the knowledge and cognitive processing levels of a lesson’s
objectives. The desired learning outcome, if obtained, comes about only if learning tar-
gets are met, and that is what you want your students to hit.
However, before a student reaches the level of an identified learning target, broader
guiding instructional elements are involved in the hierarchical arrangement of instruc-
tional curriculum (see the box “Hierarchical Arrangement of Instructional Planning”).
For instance, students’ progress in the classroom is examined relative to the educational
goals that are identified. An educational goal is recognized as a broad statement of what
students are expected to be able to do. Goals can be written so that they are general or very
specific, but they usually require an extended period of time to acquire as they are rarely
accomplished within a short period of instructional time. Some goals may require several
weeks or perhaps even months before they are accomplished.
Educational objectives are based on the broader educational goals but are more nar-
row statements of expected student performance that are to be achieved upon completion of an entire
instructional unit of study. Also, just like goals, objectives can be written so that they are
general and allow for considerable instructional variation, or they may be very specific
and particular relative to a desired performance. No matter how they are constructed,
objectives must clearly state in an action verb format (e.g., students will list, describe,
produce, analyze) what students must be able to do and accomplish as part of the instruc-
tion, and evidence needs to be collected to show that indeed the students have acquired
the knowledge, skills, and expectations of the objectives.
Consider the following educational objectives:
♦ Based on an unknown data set, students will identify the three measures of central
tendency (mean, median, and mode) and the variance with 100 percent accuracy.
♦ On a map of the human brain, students will locate and identify the lobes of the
cerebral cortex along with their primary, secondary, and tertiary cortical functions.
These objectives possess clear action verbs (i.e., identify, locate) so it is apparent to
the students what specific actions and skills are needed in order to meet these objec-
tives. In addition, a criterion and accuracy expectation of 100 percent is presented in
the first objective. The conditions, or how the learning performance will be conducted
(i.e., unknown data set and a map of the human brain), are also provided, which helps
the learners to know and practice in advance of the classroom performance of these
objectives.
Clearly stated and specific learning targets and outcomes need to exist within les-
sons. This “essential clarity” corresponds directly to the end question of the instruction-
learning-assessment model (“Where do my students need to be?”). Classroom assessment
connects with learning targets through the instructional questions that are generated in
your instructional planning. Assessment as a process requires collecting needed evidence
that is to be used to address and answer questions involving student progress. This is vital
because if relevant evidence is not collected, as an instructor you have no way of con-
firming the ongoing progress of your students.
HIERARCHICAL ARRANGEMENT
OF INSTRUCTIONAL PLANNING
Goals
(General learning aspirations for students)
Objectives
(Statements of what students should be able to successfully complete)
Learning Targets
(Accomplishments a student seeks and attempts to acquire)
DEMONSTRATION OF LEARNING:
BLOOM’S TAXONOMY
Bloom’s original taxonomy of the cognitive domain includes six major categories,
including (1) knowledge, (2) comprehension, (3) application, (4) analysis, (5) synthe-
sis, and (6) evaluation. The cognitive categories are placed in ascending order based on
growing complexity and abstraction (Krathwohl, 2002). Knowledge is viewed as the foun-
dational processing level involving simple recall and memorization of information (e.g.,
the student identifies the parts and structures of the mosquito larva; the teacher plans and
organizes the science content for the lesson). Comprehension ref lects basic understanding,
awareness, and use of information (e.g., the student identifies and executes the proper
sequence of steps in operating a compound microscope; the teacher follows the pre-
scribed delivery and implementation steps of the lesson plan). Application involves the use
of new and/or existing knowledge in addressing and finding solutions to various prob-
lems or questions (e.g., the student examines and illustrates the functions of the body
structures on the presented slides; the teacher uses observation techniques and checklists
to document proper microscope operation and student learning).
Analysis requires the learner to break an idea or concept into its various parts and
thereby expose the relationship of those parts to the collective or unified whole or sys-
tem (e.g., the student identifies observed structures and organizes them into specific
functional categories such as respiration, elimination, or general defense; the teacher
gathers pre-assessment information to determine prior knowledge/skills of students
before the lesson). Synthesis requires the learner to integrate related pieces in order to
form a complete pattern or idea similar to the experience of matching and combining
pieces of a puzzle (e.g., the student connects all structures and functions of body parts in
order to describe the initial water experience of larva; the teacher collects and organizes
student learning evidence in order to confirm and document the learning status of each
student).
Finally, evaluation involves the critical review and examination of some product, idea,
or procedure, complete with identified assets and limitations (e.g., the student interprets
the functional integrity of the mosquito larva as a living organism within the food chain;
the teacher evaluates the learning progress of the students and determines whether they
have met the desired learning outcomes of the lesson). All of the aforementioned cogni-
tive skills and corresponding learner actions are necessary and important to master if
students are to become effective learners (Table 3.1).
In addition to the cognitive dimensions, an affective domain (Krathwohl, Bloom, &
Masia, 1973) was also generated with the categories hierarchically arranged in the fol-
lowing order:
1. Receiving phenomena (i.e., basic awareness and attention to the environment and
events occurring within it);
2. Responding to phenomena (i.e., focused attention and active engagement in an
event or experience);
3. Valuing (i.e., attaching worth or value to something with potentially strong com-
mitment to a set of beliefs and/or principles);
4. Organization (i.e., the continual examination of a life view through the integration
of different values, experiences, and knowledge/skill sets); and
5. Internalizing values (i.e., the creation of a value system that consistently directs
one’s behavior and decisions).
These skill areas require educators to recognize the importance of consciousness and
sensational awareness and the role that they play in the overall learning experience. All
too often we focus only on the cognitive dimension and forget that the learner’s feelings,
attitudes, and personal perceptions directly impact any learning event or activity that
is created. Specific skills were not produced or arranged by Bloom for the psychomo-
tor domain. However, other theorists have generated models that are recognized when
focusing on the development of physical coordination, behavior, and complex motor
skills (Harrow, 1972; Simpson, 1972).
Since many learning tasks and projects require work that involves more than one skill
category, all identified skill levels need to be recognized and targeted, as well as assessed,
if mastery of those intended skills is to be demonstrated. As a general rule, knowledge
and general comprehension is expected for every instructional piece or content area that
Note. “Learner Actions Associated with Cognitive Processes.” Provided by A. T. Wyatt, 2008. Reprinted with permission of
the author.
is covered, for without this acquisition, more enhanced skills such as evaluation are not
possible. This explains to some extent why the lowest processing levels are the most typi-
cally identified in educational objectives (Krathwohl, 2002). But if more developed skills
are to be acquired, they need to be part of the active instructional process; otherwise they
remain unidentified and are not intentionally taught, and more importantly not prac-
ticed and learned by students that need to acquire them.
Building upon Bloom’s original work, Anderson and Krathwohl (2001) revised
Bloom’s taxonomy by incorporating the cognitive process across four knowledge dimen-
sions. The four knowledge dimensions (factual knowledge, conceptual knowledge, pro-
cedural knowledge, and meta-cognitive knowledge) indicate what kind of knowledge
set is to be learned in an objective or goal (e.g., the scientific method as procedural
knowledge, the central and peripheral nervous systems as conceptual knowledge) and the
six dimensions of the cognitive process (remember, understand, apply, analyze, evaluate,
and create) ref lect what the learner is supposed to be able to do with that knowledge
(see the boxes “Structure of the Knowledge Dimension of the Revised Taxonomy” and
“Structure of the Cognitive Process Dimension of the Revised Taxonomy”). This system
allows teachers and other professional educators to clearly identify the knowledge and
skill areas that are to be developed within the specific learning objectives and goals listed
in a lesson plan. This new taxonomy generates a useful two-dimensional table known as
the taxonomy table, where the specific skill performance(s) within an instructional objective can
be plotted relative to both the knowledge and cognitive process dimensions.
Structure of the Knowledge Dimension of the Revised Taxonomy. From Anderson, Lorin W., David
R. Krathwohl. A Taxonomy for Learning, Teaching, and Assessing: A Revision of Bloom’s Taxonomy of Edu-
cational Objectives. Published by Allyn and Bacon/Merrill Education, Boston, MA. Copyright 2001 by
Pearson Education. Reprinted by permission of the publisher.
Bloom’s revised taxonomy provides an essential framework that can be used in design-
ing and delivering effective instruction in the classroom. With the revised taxonomy,
different levels of knowledge are recognized and can be focused on relative to a hierarchy
of cognitive skills and levels. It allows a teacher to directly target specific knowledge
elements and processing qualities within a predetermined coverage domain or area in a
lesson.
For example, if you are providing a lesson on the human brain and objective 1 of that
lesson includes the identification and function of A. R. Luria’s three cortical units of the
brain (Luria, 1966; 1973), conceptual knowledge is being utilized (i.e., knowledge of
brain structures and how they function) along with factual knowledge (i.e., knowledge of
terminology and details of the brain). In Table 3.2, 1s have been placed in accordance with
the knowledge and cognitive process being tapped within that objective. For objective 2,
however, students must review the cortical damage of a head trauma victim and provide
© David Ashley/Corbis
an intervention/recovery plan based on the findings of that case. Different skills are being
asked for with objective 2. In particular, conceptual as well as procedural knowledge sets
(e.g., knowing how brain structures and neural systems work together) are involved with
this task along with higher processing skills such as evaluation and creation (e.g., generat-
ing and producing a rehabilitation plan for this patient based on the unique qualities of
the trauma). Number 2s in the table represent the skills involved with this objective. Both
objectives are important, yet they involve the development and use of different skills, and
this coverage is important to document before the lesson is ever provided.
With this approach, any instructional objective, goal, or standard can be reviewed and
examined relative to the knowledge base and cognitive processing skills that are needed
and required for the identified learning outcome. The use of the taxonomy table provides
an excellent way of predetermining what areas of knowledge will be learned, along with
the corresponding cognitive skills that will be utilized and developed. This serves as a
coverage map and shows very clearly what knowledge and skill areas are being empha-
sized within a specified lesson.
♦ What evidence is needed to confirm that the students acquired the desired outcomes
from the lesson? It was not long ago that some teachers held the professional mind-
set that if a lesson was taught, then that was enough. No evidence of actual student
learning was necessary or even expected. The contention that “I taught the mate-
rial, therefore you have learned it” permeated their instructional delivery. Due to
recent changes and expectations in the classroom, that notion has basically faded
away. Evidence of academic improvement and progress or behavior change must
now exist. In the field of education today, and consistent with the intent of NCLB,
direct evidence of student learning impact must be demonstrated. Beyond the rec-
ognition that evidence needs to be collected and used comes the critical question
of determining what kinds of data need to be collected. This is where each educa-
tor decides on how the different forms of assessment, pre-assessment, formative
assessment, summative assessment, and self-assessment are to be used and integrated
within the instructional plan and in the classroom setting.
♦ Does instruction-learning-assessment alignment exist within the lesson? Directly
assessing the learning progress of the students on the presented material and content
is a necessity. Regarding the solar system lesson mentioned in the first question, the
knowledge of the planets of the solar system, characteristics, and orbit patterns, as
well as additional planetary information, is the focus of that instruction. However,
if students are assessed not on their knowledge of that information, but rather on
the Greek and Roman gods’ inf luence on planetary names or the constellations of
the zodiac, then an instruction-assessment misalignment exists. Therefore, a basic
alignment question that a teacher should always ask is, “Do I have tangible evi-
dence that my students have acquired the desired learning goals and desired out-
comes of the presented lesson?” If that can be answered in the affirmative, a teacher
can be confident that instructional success and impact has been obtained with the
lesson.
Are there additional questions that should be added to this list? If so, what questions
would you add and why? As stated earlier, assessment is really about what is done for, and
not to, your students. What is most important is the information that is collected and
how it is used. That is the essence of this process. As a teacher you want to make sure
that the information you are collecting allows you to make necessary and effective edu-
cational decisions for the students in your classroom.
it is to be collected must exist. The right information or data has to be found relative to
the question that is being asked. It’s not just about collecting information, but about col-
lecting the right information.
The terms “information” and “data” are used interchangeably both within this text
and by most educators. Evidence is comparable to these terms, yet it does exceed them
in meaning. Evidence implies correctness or rightness to what is collected; it is proof that helps
confirm the finding. Educators are no different from other professionals in their need to find
confirmatory evidence of work effectiveness and impact. Teachers need to know that
students have learned the desired goals of the lesson; in essence, teachers seek out and
collect confirmatory evidence that desired progress toward (perhaps even mastery of )
intended outcomes has been met.
When collecting evidence of student learning, certain assumptions are made that must
be recognized. We like to think certain things exist of their own accord. For example,
few people have difficulty accepting and acknowledging the concept of achievement. In
truth, achievement is an explanatory concept that exists by inference. Let’s consider the
following example. When a child takes a test and performs well on that test, we say that
he or she has achieved something, has evidenced academic success, or has learned certain
content or skills. In reality, a child has taken a test, and we infer from that test perfor-
mance certain things such as achievement, mastery of material, and academic excellence.
We never experience or encounter in an empirical or directly observable sense concepts
such as achievement, motivation, aptitude, or other hypothetical constructs. What we are
able to experience (and measure with considerable precision) and document are behaviors
and performances (e.g., completing addition problems, writing a paragraph, completing a
literature search) that are believed to represent those constructs.
Academic achievement is confirmed to exist through collected evidence such as skill
observations, test performance, problems that are correctly solved, or by homework that
is completed. Discrete behaviors and skills that are directly observable and measurable are referred
to as empirical evidence, and this evidence is used in the confirmation of the concept.
It is important that we as educators are aware of these assumptions and recognize the
limits of the data and information that we collect and attribute to various concepts. Now,
recognizing these limitations doesn’t necessarily weaken the constructs. In fact, it could
be argued that these concepts are actually strengthened because they force us to carefully
select our confirming data sources, and also require us to rely on multiple sources of sup-
portive information to justify those confirmations.
Empirical evidence has direct implications for how educators document student learn-
ing and progress in the classroom. For classroom assessment, empirical evidence is rec-
ognized as evidence that is directly observable and measurable and that can be reported
in clear and meaningful ways. Momentary impressions, estimations of progress, or other
unsystematic approaches that are based on subjective and inconsistent determinations do
not meet the standard of evidence confirmation. Given higher accountability standards
in education, it is essential that teachers, as well as students themselves, be able to produce
direct evidence for learning and mastery of material, and obtaining information from
more than one source is needed in order to triangulate or “fix” onto a student’s learning
progress. As Butler and McMunn (2006) indicate, there is no set number of measures that
must be completed in this process. However, relying on only one measure of progress
provides limited information and places great pressure on the accuracy of that measure
and its findings. In the case of confirming evidence, more is always better.
Informal Observation
Informal observation, a relatively unstructured approach to the collection of information, some-
times referred to as anecdotal observation, is one of the most common ways that teachers
collect information on students (Figure 3.2). As teachers move about the room and check
on students and their work, they are continuously adding information to the active mem-
ory files that they keep on their students. During these “spontaneous reviews,” the collec-
tion of formal data (e.g., percentage of on-task time, behavior frequencies) is unlikely, but
behavior occurrences, general impressions, and basic information may still be acquired.
Informal observations often occur at random times without preset notions of what to
look for and why. The primary advantage with this approach is that the teacher is not
predisposed to look for something in particular. Consequently, behaviors or problems that
do emerge during the observation may represent true concerns or issues for that student,
especially if a pattern can be demonstrated (e.g., observed on multiple occasions and across
academic settings). This provides for a technically unbiased viewing, and that is important
because all too often we can convince ourselves that a problem or issue exists (i.e., shop-
ping for evidence to support an unsubstantiated problem) when in reality it may not.
OBSERVATION RECORD
Time: 10:30–11:15
The children had just finished their group activity and were
getting in line. A child (Sarah) got in front of the line and tried to
lead the group out of the classroom. Her teacher (Mrs. Smith)
told her that it was not her turn to lead the group and Sarah
started to cry. She then pushed the girl next to her and started
walking toward the end of the line. Mrs. Smith stopped Sarah and
talked with her and the entire class about the importance of not
hitting or pushing others. She also went over the rules of how
someone becomes a line leader. The students then went down to
the restrooms. They then returned and were allowed to play at a
center of their choice. At the math center, a boy (David) was
matching up numbers with flash cards that had the correct set of
items on the cards. He was able to match the numbers 1–10 with
the correct flashcards. He then left the center to get his teacher so
she could see what he had done.
A major limitation with this approach is that because the collection process can be
open and nonspecific, few if any particular details or issues may emerge from the obser-
vation. Also, the recording process can get cumbersome, as information may or may not
be recorded during an informal observation. If notes are taken in the process, the scripted
notes can be recorded and collected without any focus or direction. Also, considerable
time demands are often required. Since time is a valuable commodity in the classroom,
this type of observation must be carefully considered. Collected pieces of information
can get jumbled and easily rearranged, especially if a consistent recording process is not
followed. In addition, one of the biggest limitations with the informal approach is the
lack of comparability of the results from one observation to the next (e.g., a specific
behavior or event is not consistently reviewed and recorded). The ability to extend the
collected findings to other settings and situations is limited.
Formal Observation
In order to generate more meaningful and stronger empirical-based data that has greater
accuracy, permanency, and functional utility, more structured formal observation (a
planned and structured review usually focused on a specific issue or condition) recordings, com-
plete with data collection and recording requirements, are needed. However, it must also
be recognized that even with highly systematic procedures, an absolute and complete
recording of a student’s behavior can never really be accomplished during one observa-
tion, and that is why several observations need to take place over a period of time before
a “representative sample of the behaviors” begins to emerge (Sattler, 2002).
Formal observations, and the data that are generated from them, are considered most
effective when intended behavioral targets have been clearly identified and defined. This
allows the observer to focus on an identified behavior while avoiding nonessential behav-
iors during the recording period. For example, if during in-class work time a teacher
notices that several students are experiencing what appears to be confusion, frustration,
and general disengagement with their math worksheets, a focused observation process
on just the math skill being practiced and demonstrated in the exercise (e.g., two- and
three-digit division problems) may be in order. In this way, the skill acquisition process
of the students could be examined and reviewed in greater detail in order to determine
if a skill competency issue is connected to the student difficulties. If competency can be
ruled out, then other possible contributing factors (e.g., number of problems to complete,
amount of time given to complete items) can and should be examined and reviewed.
As Sattler (2002) has stated, “Observational methods are particularly useful for
(a) studying behaviors that have a relatively high frequency, (b) assessing global behav-
iors such as aggression, social withdrawal, sociability, and attention, and (c) evaluating
a child’s progress over time” (p. 86). Since teachers deal with academic and behavioral
problems that demonstrate high or consistent rates of occurrence, formal observation
formats can be helpful for documentation purposes as well as for collecting data on how
effective changes in instructional practices are on the identified concerns.
Group Observation
Observations are often focused on collecting information on an individual, but group
observations, designed to collect information on multiple participants, can be easily completed
and can be extremely helpful. In fact, there are some situations, including group projects
and team activities, where observation of a group of students is desired and necessary.
Consider the following observation that involves a group project in a ninth-grade social
studies class. Students are required to work in small groups of five or fewer students on
FOCUS ON ASSESSMENT
• Some combination of the listed possibilities is likely, as most school and learning issues
are complex and can involve a number of contributing and overlapping factors and
inf luences.
• Other possibilities?
Because of the observational information, Johnny’s teacher is more informed as to what
is happening and can begin to examine the potential issues that are surrounding and con-
tributing to this issue. She is in a much better position to make changes to see what might
work to improve Johnny’s compliance with the activity. For example, having him start his
math work earlier in the day may provide him with the additional time that may be needed
to complete this task. Also, determining if he is actually bored with the practice items on
the worksheet must be examined and resolved. Simply moving him closer to the teacher’s
desk may increase his on-task behavior and completion of the math sheets. Starting to col-
lect information on Johnny’s ability to focus and sustain his attention may also be neces-
sary, especially if this pattern is evidenced in other activities during the day.
Clearly, there are many possible reasons for Johnny’s current behavior, and with useful
information such as observational data, effective learning strategies and interventions can
be implemented. What becomes important in utilizing observations is clearly identifying
what questions need to be answered and then selecting the recording methods that will
provide the most relevant information to help answer those questions.
a Civil War project in the classroom. The teacher wants to document how well the stu-
dents work together on this project. Therefore, the focus and purpose of this observation
is on student collaboration and cooperation. The observation checklist lists the desired
collaboration behaviors (i.e., active discussion with members of the group, taking turns
to address the group, listening to others, sharing responsibilities and duties for the group)
that should be demonstrated in this exercise. The skills are listed on the form, as are the
names of the students within each group (Figure 3.3). Moving from group to group, with
comparable observation time for each group, the teacher records with Xs what behav-
iors are demonstrated and by whom. A section at the bottom of the form also allows the
teacher to record information and other details associated with the observation (e.g., time
of day, noted distractions, energy level of the group, task demands).
TYPES OF OBSERVATIONS
IN THE CLASSROOM
When observing and recording student behavior, several methods can be used, but with
each method there are recognized advantages as well as limitations that exist as information
is collected. General descriptions of the primary observation approaches used with students
in classrooms are provided in the following sections, along with relevant examples.
GROUP OBSERVATION
Additional Notes:
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
Narrative
A narrative record involves the chronological recording of an event as it is taking place. Usually
set within an identified setting or school event, such as recess or fifth-period English
class, and for a specified amount of time, information can be recorded in general terms
with limited details or in great detail relative to specific conditions or specific target
behaviors (e.g., quality of classroom work, student-to-student interaction). The observer,
like a photographer who has telephoto capabilities with his camera, has considerable
f lexibility to focus on specific issues as they evidence themselves during the observation.
As an evidence collection method, the narrative observation gives a teacher a simple
way to record the behaviors and learning events of students in the classroom or in other
educational settings. It can provide a snapshot, albeit a limited one, of what is currently
happening for students. Its greatest value lies in providing a foundation for more specific
and focused evidence collection. It can identify important observation targets or issues
that may need to be examined further. Since the narrative approach does not follow
uniform and systematic data collection operations, major evidence limitations, including
generalization of findings, comparability of findings across observers, and the limited
Interval 1 2 3 4 5 6 7 8
Target + + + - + + + +
Peer - - - - - - - -
Name: Jim
Grade: 2
Task: Addition of two-digit numbers with and without carry-over
Addition roblems
P Completed 5 5 3 4 3 20
outbursts in class, the number of times a student is out of his seat), along with other
important factors, can be documented with this approach.
A frequency recording, or how often a behavior or specific event occurs within the observa-
tion period, can be generated with this approach, along with the ability to generate a rate
of behavior measure. This involves a simple calculation of dividing the number of occur-
rences of the behavior by the time of the observation period. In Figure 3.5, the record on
Jim demonstrates the quantitative data that can be generated from a simple event record-
ing. In this particular case, Jim completed all 20 addition problems within the provided
time, answered 16 correctly, demonstrating 80 percent accuracy, and completed on aver-
age about four problems per minute.
Name: ___________________________
Date: ____________________________
Yes No
Was any lens touched? ___ ___
Was lens tissue, a lens cloth, or a lens pen used
in cleaning a lens? ___ ___
Were all parts of the microscope known and
correctly identified? ___ ___
Were slides correctly prepared with cover slips? ___ ___
Were slides correctly placed on the stage with
stage clips? ___ ___
Did the student start with the low power
objective first? ___ ___
Was the proper light control set? ___ ___
Was proper focusing demonstrated? ___ ___
Does the student understand the opposite rule
for moving slides? ___ ___
Did the student demonstrate appropriate care
and safety around the microscope during the lab? ___ ___
After the lab did the student return all materials
to their correct places and demonstrate proper
clean-up procedures? ___ ___
♦ Not Acceptable: The applicant does not meet the knowledge, skill, and perfor-
mance requirements. The student is unable to perform at a satisfactory level based
on current classroom information.
So what makes this rating scale imprecise and therefore ineffective? The qualitative
descriptors that are used are only as good as the text clarity and detail provided to delin-
eate one level from the next. Rating descriptors such as “outstanding” or “acceptable” by
themselves are not sufficient for review purposes. That is why the text used to describe
the categorical adjectives must be clear and specific in order to ensure rating consistency
and accuracy, and that is not the case with this scale.
For example, to earn an “acceptable” rating, a student must meet most of the knowl-
edge, skill, and performance requirements. But what does “most” involve? What per-
centage or majority of requirements must be met to in order to receive this rating? Also,
the knowledge, skill, and performance requirements connected to this rating must be
identified. In addition, the rating stipulates that the student must be able to perform at a
satisfactory level based on current classroom information. Therefore, “satisfactory” must
be clearly defined. Overall, the level of clarity in the descriptor narratives must be such
that anyone reviewing a student work product would be able to unmistakably identify
whether it meets that rating or not. In addition, the student must also clearly understand
what is required and expected as part of any provided ratings.
In order to make this rating more precise, “most” could be replaced with “80 percent”
or “four out of the five” knowledge, skill, and performance requirements. The listed
or expected requirements would need to be precisely stated, such as the following first
grade requirements:
1. Identifies all letters of the alphabet by sight
2. Identifies all alphabet sounds
3. Produces all alphabet sounds on command
4. Identifies numerals 1–10 by sight
5. Performs correct number-object associations for numbers 1–5
Finally, “satisfactory” could be defined as the student independently and accurately
completing a teacher-initiated task at least two out of three times. Based on these sug-
gested changes for the “acceptable” rating, how could you clarify the “excellent” and
“not acceptable” ratings? More information about rating scales and rubrics and how they
can used in the classroom is provided in Chapter 6.
QUESTIONING
Direct questioning involves the collection of evidence through the verbal responses of students; it
is a discussion or dialogue with an intended purpose. Educators use this technique throughout
the school day. It serves as a quick and authentic way to confirm students’ understanding
of presented material, use of procedures, and skill execution, as well as a way to confirm
progress. Used as part of or as a follow-up to a lesson or class activity, teacher-directed
questioning can generate valuable learning feedback that can result in skill correction
while the learning event is actually taking place.
These moment-driven questions can be connected to reviews, skill checks, or other
activities that are conducted during the normal course of the class day. Typically individu-
alized to the situation and student’s work, well-constructed and delivered questions can
serve to confirm that the student truly understands the task demands through his or her
verbal responses. Often the spontaneous questions that are generated have their origin in
the confirming questions (who, what, where, when, and why), since these are natural ques-
tions to ask when new information or skills are acquired or connected to a pre-existing
knowledge or skill base. In addition to asking individualized questions, the integration of
standard questions that are asked of all students is important. Asking key questions ensures
that the students as a whole understand and have processed the intended material of the les-
son. This confirmation is crucial for current and future extensions of that learning.
Due to the need to confirm retention of information and/or general understanding
of material, teacher-generated questions directed toward students often focus on the
foundational processing skills of Bloom’s revised taxonomy. Clearly, students need to be
able to remember and understand information and possess a strong foundation of gen-
eral knowledge. But most would argue that there is an even greater need for asking
questions that require students to directly apply their knowledge and/or skills, or that
pose situations requiring students to critically analyze informational sources and examine
the underlying relationships that may exist. The practice of using information is impor-
tant, and this “mental practice” only happens when educational circumstances are cre-
ated (e.g., higher-level questions and requests are made of students) in the classroom that
require students to think and respond with the skills and at the level necessary to answer
the questions. This is particularly the case when the cognitive skills of evaluation and
creation are involved. In order for students to develop strong evaluation and construction
skills, they must be placed in situations where they are required to use those skills.
Beyond the questions that are crafted at the moment, predetermined “probing ques-
tions” that have a specific instructional purpose can be constructed so that they tap into
specific processing skills, along with different areas of knowledge as described in Bloom’s
revised taxonomy. However, an important aspect must be acknowledged when asking
questions of students. Students naturally provide responses that answer only the questions
that are asked. For instance, if asked basic knowledge or comprehension questions, those
are the types of responses that are typically generated by students. However, if asked
higher-level processing questions, those types of responses must be generated by the stu-
dents. What is asked for in the questions is what is generally received in the answers that
are provided, and as the instructor you have complete control over the level and type of
questions that are asked of your students.
For example, if new information is being examined and the students have no back-
ground knowledge about the subject area, then a basic knowledge and comprehension
questioning approach would certainly seem justified, especially at the beginning of the
lesson. As additional information and understanding is acquired, a focus on higher-level
questions may be desired. In all likelihood, questions that ref lect all levels, from the most
basic to the highest, should be involved at different times if an instructor really wants to
know the extent of the students’ knowledge, skills, and usage of the material that they are
ASSESSMENT IN ACTION
and inaccurate feedback and will not ref lect the learning impact on the classroom as a
whole. As with any data collection method, this approach must be used carefully and in
combination with other data sources, and be focused around key instructional questions
and outcomes that are connected to the provided lesson or learning activity.
Interviewing
As part of the traditional interview, the classic image of two people engaged in a focused
discussion comes to mind. An interview, by design, is a directed conversation where specific
information is obtained from the interviewee by way of direct questions from the interviewer. Teach-
ers rarely engage in formal interviews due to the lack of extended time that is required to
complete an interview. However, during the course of the school day, teachers are some-
times able to visit with students for a limited period of time to review progress or address
other school issues. Through informal mini-interviews or sessions, teachers are often able
to collect information that can then be used to directly assist those students.
The mini-interview, a time-restricted personal meeting centered on a specific issue, is typi-
cally associated with the mental health field and is used by some medical schools in the
selection of physician candidates and other health field professionals (Eva, Reiter, Rosen-
feld, & Norman, 2004a, 2004b), but it has direct implications and usage in the classroom.
For example, during a mini-interview, questions that are particular to a specific child
can be asked; in the process, in-depth and valuable information can be obtained. In
addition, standard questions that are asked of every child can also be completed. In this
way, general feedback from the class can be obtained along with specific and individual
information from each child. This one-on-one time that is generated as part of “review
sessions” during class time can also help to form the student-teacher bond and strengthen
important interpersonal connections that will be needed over the course of the academic
year. Individual time with students is generally time well spent. Unfortunately, it is dif-
ficult to come by given all of the daily demands in the classroom.
STUDENT PRODUCTS
Data collected through observations can be used to document learning progress. In addi-
tion, specific responses to instructor-generated questions can also serve as confirmation
of what students know as well as what they do not know. Still, what students generate
as a matter of their classroom work is the gold standard of student learning evidence. As
educators, we rely on student production as a critical indicator of their progress and abil-
ity to take what they have learned and generalize it to other activities, including real-life
situations outside the classroom. Student products meet the requirements of empirical evi-
dence, as they have recognizable beginning and end points; generate tangible, concrete constructions
that can be examined; reflect the direct work and effort of the individual (under most conditions); and
meet the accountability mandate of “show me.”
Student products can and do represent multiple skills and talents and involve different
output responses such as verbal performances (e.g., speeches, debates, group discussions,
stage productions) and written expression performances (e.g., short stories, essays, let-
ters, journal entries, portfolios, research papers), as well as traditional test performances
(e.g., completing various test items including multiple-choice questions, matching, fill-
in-blanks, essays). The use and collection of a variety of student products is important
because it allows for a more accurate and detailed assessment and evaluation of a student’s
progress and growth, especially with skills that are expected to develop over time. In
regards to instruction, multiple student products and performances create opportunities
for greater and more focused intake, development, and refinement of new content, along
with hierarchical skill development.
The typical range of learning activities and instructional assignments require the
combination and integration of several learning modalities (e.g., visual, auditory, motor
skills), but most student products or performances are identified with a primary modality
output. For example, a speech is normally considered a verbal product, and a visual prod-
uct is the construction of a map or a drawing. In this process, students are constructing a
response or product, so the term constructed response is used to describe work and a per-
sonal product that is generated solely by each student in response to a given question or instructional
request. Writing a paper falls in the constructed response category, since each student must
construct an original text to meet the requirements of the assignment.
A constructed response is typically associated with paper-and-pencil tests and with
specific test items that require a student to create novel responses to those items. Well-
recognized constructed response forms include traditional essay and short answer ques-
tions that are often included in classroom tests and exams. There are many other types
of activities that require constructed responses and products from students, including
journals, presentations and performances, research projects, experiments, and portfolios.
The other type of student product is referred to as selected response, and this format is
used predominantly in classroom tests or exams. With selected response items, a student
is required to select the correct answer or response from a set of potential answers that are provided.
This approach is used with many standard test items, including true/false, matching, and
multiple-choice items.
In the classroom setting, students are typically asked to complete a variety of learn-
ing products. Student products, along with observational data and verbal reports, are the
main forms of evidence that teachers use to document student learning. In addition to
documenting knowledge and skill gains, these types of evidence can be used to judge or
evaluate the quality and extent of that learning. The following chapters include specific
information on how different types of evidence can be embedded and collected within
classroom assessment processes (i.e., formative, summative, and self-assessment), along
with the formal evaluation of student work through grading practices.
CHAPTER REFLECTION
At the beginning of this chapter, you constructed your “top ten list” of the most impor-
tant things learned in school. As you look over that list again, answer this final question:
What skills and knowledge sets do the students in your current or future classroom need in order
to be productive and competitive contributors in today’s society as well as the world of tomorrow?
Keep in mind that these students will be competing on an international level and will be
expected to possess a “world class” education.
CHAPTER REVIEW
Conclusion
Educators are obligated to make sure that their students learn the material and skills that
they will be held accountable to regardless of their level of instruction. That is why lesson
goals need to match up with the intended district and/or state academic standards that
students are required to demonstrate and master at their respective grade levels. Included
with standards-driven instruction is the important process of determining and clearly
identifying learning targets and outcomes that are connected to those specific curricu-
lum and academic content standards. This includes a clear identification of what students
are supposed to be able to do, and must include the processing levels and cognitive skills
that are connected with those outcomes. The teacher, as he or she directs the classroom
activities, must keep the intended learning targets in sight, as must the students who are
required to obtain them; both must know what is to be accomplished. In the pursuit of
student learning targets and outcomes in the classroom, empirical evidence needs to be
collected in order to confirm whether these desired targets and outcomes are reached.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. Describe the importance of teaching to recognized learning targets and outcomes
that are based on the academic standards that exist within a school district’s cur-
riculum guide. Why does evidence to document student learning progress need to
be collected?
2. Does Bloom’s revised taxonomy provide a helpful structure and framework for the
development of your lessons and what you want and expect from your students?
Why or why not?
3. Do you feel comfortable recognizing and potentially using the different methods
identified to collect student learning evidence that are presented in this chapter?
Why or why not?
RESOURCES
Education World at www.education-world.com/standards/national/index provides an
extensive listing of national standards across major content areas including math, Eng-
lish, and social studies. Content and subject standards for each state are also provided
on this site.
For more reading on observation and different observation formats, this chapter pro-
vides an excellent overview on the subject: Hintze, J. M., Volpe, R. J., & Shapiro, E.
S. (2008). Best practices in the systematic direct observation of student behavior. In
A. Thomas & J. Grimes (Eds.), Best practices in school psychology V (pp. 319–335).
Bethesda, MD: The National Association of School Psychologists.
REFERENCES
Anderson, L. W., & Krathwohl, D. R. (2001). A taxonomy for learning, teaching, and assess-
ing: A revision of Bloom’s taxonomy of educational objectives. New York: Longman.
Bloom, B. S., Englehart, M. D., Furst, G. J., Hill, W. H., & Krathwohl, D. R. (1956).
Taxonomy of educational objectives: Handbook I: The cognitive domain. New York: David
McKay Company, Inc.
Bloom, B. S., Hastings, J. T., & Madaus, G. F. (1971). Handbook on formative and summative
evaluation of student learning. New York: McGraw-Hill.
Breivik, P. S., & Senn, J. A. (1994). Information literacy: Educating children for the 21st century.
New York: Scholastic Inc.
Butler, S. M., & McMunn, N. D. (2006). A teacher’s guide to classroom assessment: Under-
standing and using assessment to improve student learning. San Francisco: Jossey-Bass.
Eva, K. W., Reiter, H. I., Rosenfeld, J., & Norman, G. R. (2004a). An admissions OSCE:
The multiple mini-interview. Medical Education, 38, 314–326.
Eva, K. W., Reiter, H. I., Rosenfeld, J., & Norman, G. R. (2004b). The relationship between
interviewer characteristics and ratings assigned during a multiple mini-interview. Aca-
demic Medicine, 79, 602–608.
Harrow, A. (1972). A taxonomy of psychomotor domain: A guide for developing behavioral objec-
tives. New York: David McKay.
Krathwohl, D. R. (2002). A revision of Bloom’s taxonomy: An overview. Theory into
Practice, 41(4), 212–218.
Krathwohl, D. R., Bloom, B. S., & Masia, B. B. (1973). Taxonomy of educational objectives:
The classifi cation of educational goals: Handbook II: Affective domain. New York: David
McKay Company, Inc.
Longworth, N. (2003). Lifelong learning in action: Transforming education in the 21st century.
London: Kogan Page.
Luria, A. R. (1966). Higher cortical functions in man. New York: Basic Books.
Luria, A. R. (1973). The working brain. New York: Basic Books.
Marzano, R. J. (2006). Classroom assessment and grading that work. Alexandria, VA: ASCD.
Sattler, J. M. (2002). Assessment of children: Behavioral and clinical applications (4th ed.). San
Diego, CA: Jerome M. Sattler, Publisher, Inc.
Senge, P., Cambron-McCabe, N., Lucas, T., Smith, B., Dutton, J., & Kleiner, A. (2000).
Schools that learn: A fi fth discipline fieldbook for educators, parents, and everyone who cares about
education. New York: Doubleday.
Simpson, E. J. (1972). The classifi cation of educational objectives in the psychomotor domain.
Washington, DC: Gryphon House.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How do you construct classroom assessments that accurately and consistently mea-
sure student performance?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Define validity and its essential role in the
assessment process.
♦ Describe the different aspects of validity and their
importance in constructing accurate classroom
assessment procedures and measures.
♦ Define reliability and its essential role in the
assessment process.
♦ Explain the different forms of reliability and
their importance in developing consistent class-
room assessment procedures and measures.
♦ Understand the critical relationship between
validity and reliability.
♦ Identify assessment bias and its detrimental
effects on the collection of student evidence.
♦ Describe how to minimize assessment bias when
collecting student evidence through classroom
Bananastock/AGE Fotostock assessments.
84
Introduction
Student evidence can be collected through an array of measures. Whatever the method or
technique, no matter its form or function, it must accurately measure what it is designed
and intended to measure and it must be consistent in the data that it provides. Validity and
reliability are essential qualities and components of any measurement activity, whether it
involves a fourth grade in-class writing project or district-wide testing. Due to the wide-
ranging impact of these factors, all educators need to understand the importance and
value of validity and reliability and their inf luence on any assessment process, wherever
it takes place. In addition, avoiding assessment bias is critical if a teacher really wants to
collect accurate and useful information and evidence on student learning progress.
CHAPTER ACTIVITY
At this very moment, scan your immediate surroundings. This could take place in your
room, the library, at the kitchen table, or a thousand other places. Regardless of your loca-
tion, identify five objects that are located close to you. Once those objects have been found,
determine if these objects are valid (they do what they were designed to do) and if they are
reliable (they perform their function in a consistent and uniform way).
In most instances it is likely that all of your chosen objects will meet these criteria,
unless they are broken or in need of repair, and there is a very important reason for that.
We depend on products, machines, various operations, and other human-engineered events
to meet those requirements. If they did not, then their utility and effectiveness would be
compromised and nothing of value would be obtained. Meeting these two conditions are
critical in the area of assessment, as well. A measure, procedure, test, demonstration, or any
other classroom condition or event must be valid and accurate for its intended purpose,
and it must also be reliable so that the same or comparable measurements are obtained each
time the procedure is implemented.
the Standards for Educational and Psychological Testing (1999), validity is now viewed
as a unitary entity, and evidence is collected to support the accuracy and authenticity of
test scores and their interpretation. In particular, five categories of validity confirmation
exist and include (1) evidence based on test content, (2) evidence based on response pro-
cesses, (3) evidence based on internal structure, (4) evidence based on relations to other
variables, and (5) evidence based on consequences of testing.
This is a positive move for the measurement field, and the consideration of validity as
a unitary concept certainly improves the psychometric properties of tests and other eval-
uation procedures. Yet not all of the aforementioned validity confirmation areas relate
directly to teachers and their classroom issues, so a modified validity review is in order.
Therefore, with a focus on classroom instruction and validity issues related directly to
standardized tests, a review of validity evidence (ref lecting the established elements of
content, criterion, and construct-related validity) connected to specific functions and
purposes is provided in order to connect it directly to the needs and uses of the teacher.
For classroom assessment purposes, validity involves accurately measuring what a test or
procedure has been designed for and is intended to measure so that the results can be accurately inter-
preted and utilized (Reynolds, Livingston, & Willson, 2006). For example, it is assumed
that a math test can and does measure a student’s math achievement. If relevant test items
and questions are selected and these items correspond to and evaluate identified math
skills, the math results should be interpreted as accurate indicators of the students’ math
learning and progress. As mentioned in the last chapter, it is important to recognize the
fact that we are actually inferring achievement by way of performance on the provided
test items. We can’t directly see a student’s achievement, but we can see his or her written
computation responses to questions that we then infer ref lect the student’s achievement
and learning progress. So as a teacher you infer student progress based on assessment
results and make decisions based on that information, and that is why it is critical that
assessment and procedures used in the classroom be valid. If valid results are obtained,
then effective instructional and educational decisions can be made. That is ultimately
what assessment is all about: providing educators and their students with valid data that is
then analyzed in order to make good educational decisions.
Validity:
The degree to which a test
or instrument measures what
it is intended to measure
questions, or other classroom tasks to ensure that they accurately represent the material
and ref lect the skills that are expected to be covered and acquired as part of an instruc-
tional plan. And the key is to make sure that the instructional material that is covered
aligns with and is appropriately represented and examined within the selected class-
room assessment. A test or other evaluation measure should be carefully constructed so
that the desired knowledge and skills that the students are expected to acquire are accu-
rately evaluated. That is why teachers need to develop a table of specifications whenever
they construct a formal summative test.
A table of specifications is essentially a graphic framework that identifies the test items
or other measures (e.g., rubric) that are constructed to tap the listed objectives and outcomes of the
lesson. The value of this practice is that it provides a visual display and confirmation
of the assessment alignment between what has been taught and what is being formally
evaluated. Both general content, ref lected in the identified questions or items, and the
required cognitive processing skills necessary to answer those questions are identified.
A table of specifications for an algebra test (pre-test and post-test) for seventh graders
appears in Table 4.1. It provides a visual map of the test items (listed as Question 1, 2, 3,
etc.) and the knowledge and cognitive skill domains that are being covered. The example
in Table 4.1 demonstrates a consistent coverage of material and skill levels for both the
pre- and post-tests, and the teacher has a clear idea of what objectives are being measured
and the cognitive levels and skills that are required to correctly answer those test items.
Criterion-related evidence of validity examines the relationship between scores on a
particular test with some external criterion. In the field of education, you might see this used
in the prediction of potential college achievement success based on the performance on
an early entrance exam like the American College Test (ACT) or the SAT Reasoning
Test. Predictive and concurrent and validity evidence represent the two basic approaches
to demonstrate this validity.
With the predictive approach, a test is given followed by a specified interval of time,
and then the criterion or performance is evaluated. For example, it is now common for
preschool programs, as well as in kindergarten, to have students complete a screening or
entry test, and the results from those tests are often used to predict a student’s potential
academic success if enrolled. When the completion of the test and the criterion evaluation occur
at the same approximate time, then we have concurrent criterion-related validity. The perfor-
mances from the two measures are compared and a correlation coefficient, or more accu-
rately, a validity coefficient, is generated to measure the degree to which one measure
estimates the performance on some other criterion measure.
Predictive validity evidence involves the capacity to determine how well someone will
perform a future task based on a previous measured performance. This kind of evidence holds
particular importance in education as test score performances, especially in the areas of
achievement or general aptitude, are sometimes used in the selection process for admis-
sion into certain courses, programs, or other educational experiences. If a test or measure
demonstrates strong evidence of predictive validity—that is, if it possesses a strong valid-
ity coefficient—then it can be used to support and validate educational selections.
For example, a math entry exam demonstrates a high predictive validity coefficient
(r = .85) with later academic success in an advanced statistics class. Knowing students’
scores on this test can help identify the students that will likely do well and others who
TABLE OF
BLOOM' S TAXONOMY
wit78704_ch04_084-112.indd 89
SPECIFICATIONS
Pre-assessment areas Remember Understand Apply Analyze Evaluate Create
Justify a general rule for a pattern Question 2 (FK) Question 4 (CK) Question 7 (PK)
or a function by using physical
materials, visual representations,
words, tables, or graphs.
Use variables as unknown Question 6 Question 3 (CK)
quantities in general rules when (FK) Question 5 (PK)
describing patterns and other Question 7 (CK)
relationships.
Model problems with Question 1 (FK)
physical materials and visual Question 2 (FK)
representations, and use models,
graphs, and tables to draw
conclusions and make predictions.
VALIDITY, RELIABILITY, AND AVOIDING ASSESSMENT BIAS
FK—Factual Knowledge
89
CK—Conceptual Knowledge
PK—Procedural Knowledge
2/11/11 10:37 AM
90 CHAPTER 4
may likely require more help and assistance in the class. Of course it is important to be
aware that correlation evidence is not causative, so anticipated outcomes are possible
but certainly not absolute. Many factors contribute to academic success, and the entry
test that is examining current skill level is an important factor, but it still is only one of
many that will affect the eventual outcome of a student’s performance in the statistics
class.
Construct-related evidence of validity involves the collection of confirmatory evidence
that demonstrates the existence of a recognized concept or construct that a test purports to measure
(Worthen, Borg, & White, 1993). As was mentioned previously, hypothetical constructs
are created to help provide understanding and recognition of complex human phenom-
enon. For example, depression is a psychopathological condition, a clinical diagnosis, and
a hypothetical construct. If you mention depression in a conversation, almost everyone
has some idea of what you are referring to, yet it doesn’t exist as a physical entity. At best
it is manifested as a chemical imbalance within the brain, but even today that is diffi-
cult to identify and confirm relative to the standards of empirical evidence. In fact, the
ASSESSMENT IN ACTION
Just as car shoppers soon realize that all cars are not the same (e.g., different styles, func-
tions, amenities, cost, maintenance requirements), educators need to possess and use that
same review “lens” when examining tests or instruments and the constructs they claim to
measure. As a professional educator, it is important to make sure the constructs are justified
(i.e., that they are clearly defined and supported by corroborating evidence and/or support
from other relevant studies and findings) and can stand up to your personal review. Con-
sider the following questions as part of the review process:
• Does the test or instrument define the construct in a clear and meaningful way?
• Does the test (or other assessment procedure) material really measure what it says it
does? Is the supporting evidence that is provided convincing?
• Does the construct (and its provided definition) match with your expectations as well
as those of your colleagues in the school or district?
• Are there any inconsistencies in what is stated as being measured and what is actually
being measured?
Make sure the answers to these questions are satisfactory before you buy into the notion
of any construct and how it is measured. Be a cautious and appropriately demanding edu-
cational consumer when it comes to any measure or procedure that involves your students,
particularly when they, and their skills, are being formally evaluated. It is your assessment
right and responsibility to ensure that adequate information and evidence exists that con-
firms the construct’s validity as well as the measure’s utility in the classroom.
inconsistent scores or ratings will eliminate the possibility of getting any meaningful
results from which to make appropriate inferences (i.e., based on the collected findings,
the students have or have not learned the material and desired skills from the lessons). In
order to ensure high reliability, teachers need to implement basic yet important regula-
tory procedures during their assessments. The following are essential reliability mainte-
nance procedures that need to be adhered to.
♦ Proper review of material prior to an examination provides a uniform and consis-
tent preparation for the students.
♦ Directions for any assessment activity must be clear and understood by all partici-
pating students.
♦ Consistent administration procedures must be followed at all times during all
assessment activities.
♦ Test items or tasks need to be accurately and uniformly presented and evaluated
(requiring objective scoring methods).
♦ Maintaining uniform and appropriate completion time helps to ensure equity and
fairness for all those involved in an assessment activity.
♦ Elimination of distractions and interruptions during an assessment is essential.
All of the listed procedures are important; however, the consistency in scoring items or
student performances generally draws the most attention, and that is why objective scor-
ing methods must exist. What must be avoided is assigning different scores or ratings
when an item is scored. This is relatively easy to accomplish with selected response items
(e.g., picking the right answer from a set of responses as is done with multiple-choice
items) but can be more difficult for constructed response items like essay questions. That
is why grading responses to essay questions must be completed prior to the administra-
tion of a test, and must evidence clear scoring guidelines based on the information that is
provided, or not provided, in the student response (see the box “Response Requirements
and Scoring Guidelines for an Essay Question”).
Also, in regards to the internal consistency of test items within a classroom test, teach-
ers need to be observant and recognize general performance patterns for individual stu-
dents as well as the collective group. In particular, a student’s performance on items that
measure a specific content area or skill base should be consistent, given that the items
are measuring the same or similar knowledge base or skill. For instance, if the student’s
learning and preparation has been strong within certain areas or skills, then it would be
expected that all items tapping that area or skill would be passed or that the performance
would be strong overall. On the other hand, if the learning and preparation were poor,
the opposite pattern would be anticipated.
Regarding the impact of reliability, a teacher must be able to demonstrate the sys-
tematic execution of all assessment administrations in the classroom so that the same
consistent measurement can be expected each time a test or other assessment activity is
conducted. But remember that reliability doesn’t just involve the measure, since what
the student does, or doesn’t do, can affect the consistency of the test results. That is why
school letters go out to parents about making sure their child gets enough rest, eats well,
and is ready to do their best on whatever tests are being administered by the district.
The effort and motivation of the students can directly affect the results of any exam or
performance. If a student gives his or her best effort on one day and “blows off ” a test
the next, then unstable and inconsistent test results are generated. Getting the best effort
from every student is just as important in obtaining reliable results as ensuring that a
test or activity is procedurally consistent in its administration and internally sound in its
construction.
Reliability is a measurement necessity that must exist for any assessment process. But
what exactly is reliability? As Gronlund (1985) stated, “reliability refers to the consistency
of assessment results” (p. 212). In regards to classroom assessment, reliability involves the
obtainment of consistent findings where the same or similar results are acquired each
time a measurement takes place. When the findings from an instrument or procedure has
high reliability, you have confidence that the information you have received would be
the same or similar to information if a measurement was taken at a different time or over
a series of several times. Of course certain procedures demand a high reliability factor.
Take, for instance, the importance of obtaining consistent readings from a blood pressure
gauge. Uniform pressure readings over multiple administrations must exist if it is to have
meaningful diagnostic medical value. The same standard applies to a weight scale. Stable
and consistent readings each time the device is used is expected and necessary if it is to be
considered reliable and valuable to the user.
Reliability is an essential element in the measurement process, but how is it obtained?
First of all, a high level or degree of reliability is desired and sought after, since reliability
is not an all-or-nothing condition (i.e., you either have it completely or you don’t). Reli-
ability exists as a matter of degree and magnitude, and the further and higher you move
along the reliability continuum, the more certainty will exist with your findings and sub-
sequent decisions. Therefore, in order to acquire a high level of reliability, actions must
be taken to obtain it, such as taking steps to minimize error within measurement instru-
ments as well as the assessment procedures that are used. Every measurement instrument
and procedure contains some amount of error, so measurement error cannot be entirely
eliminated. However, by taking steps to minimize procedural error (e.g., giving clear
administrative directions and procedures, ensuring appropriate item coverage and dif-
ficulty level, determining appropriate number of test items), the integrity and stability of
the measurement instrument and the process can be established and hopefully maintained,
thereby ensuring consistent and reliable findings during each and every administration.
Now you might say to yourself, “Why should I use or develop some test or procedure
that has error associated with it?” That is a very good question to ask, by the way. As
Worthen, Borg, and White (1993) state, “Educational measures do not need to be per-
fect to be useful” (p. 139). The key is to create measures and procedures that generate
comparable results each time they are implemented. As a teacher, you assume a measure
(e.g., a classroom test) that is given to a student will generate reliable and consistent scores
or findings whenever that test is given. And as an educator you desire this quality in
all school procedures, especially when examining the progress of students. If scores or
ratings that are generated do not possess high reliability, then those scores have limited
value for you and your students, and the collected confirming evidence of student learn-
ing is suspect.
For example, if a student receives a score of 100 (ref lecting 100 percent accuracy) on
a certain test and then takes the same test (containing the same test items) later that same
day and receives a score of 70 (ref lecting only 70 percent accuracy), then a high level of
inconsistency with these scores is demonstrated. The reliability associated with this test
is very suspect, since highly discrepant performances on the same test material over a
very short testing interval have been evidenced. Since it is highly unlikely that this stu-
dent lost significant information and skills between the two completions of this test, the
disparate results indicate that the test is not a useful and reliable measure of how well the
student comprehended and mastered the covered material.
Several contributing factors or conditions could be at work, thereby interfering
with and minimizing the stability of the results. For instance, given such a huge dif-
ference between the two performances, perhaps the print quality and clarity of the test
items differed between the two tests, providing a high degree of reading inaccuracy. If
machine-scoring sheets were used, maybe response items were filled in incorrectly or the
answer sheets were misread and incorrectly scored. Student performance factors could
have contributed to these findings. Perhaps the student suddenly experienced a migraine
headache, which could have caused nausea problems during the second testing, thereby
explaining the lower test performance. Since these and other situations are possible, they
would need to be considered and independently ruled out. In this way, the stability and
associated confidence of the reliability of your test can be maintained. If procedural
integrity is not pursued, reliability is suspect with these test scores, or any other test
scores, ratings, etc., and confidence regarding the stability of a student’s performance can
never really be obtained.
As busy people, we expect our watches or cell phones to be reliable and consistently
function no matter the time or place. We need to contact individuals at various times of
the day as well as get to class on time. Unfortunately, if your watch runs fast during one
hour and slows during another, or stops altogether, then your work is compromised and
your life quickly starts to resemble chaos theory. The same applies to the measurement of
student learning and achievement. As educators, we cannot function if faced with the real-
ity that a student’s science project could be rated as advanced in the morning and deficient
in the afternoon. We expect findings to be consistent regardless of when a measurement
or evaluation is conducted. This is critical, because important school decisions are based
on information or data that are automatically assumed to be reliable and stable over time.
In regards to educational measures and tests, three major types of reliability (i.e.,
stability/test-retest, alternate form, and internal consistency) are considered essential
(Gronlund, 1985; Worthen, White, Fan, & Sudweeks, 1999). Reliability plays a direct
role in regards to the usefulness of information that a teacher collects on his or her stu-
dents in the classroom, and as such it is essential for teachers to understand how it oper-
ates and inf luences the stability of educational results (Figure 4.2). As these issues are
reviewed, however, it is important to recognize that even when these reliabilities are
demonstrated with a test or project, that still does not automatically ensure a high and
continued consistency status. In order to obtain high reliability in the classroom, it must
be constantly sought after and confirmed through vigilance and effort.
Reliability:
The consistency of the
scores obtained from a
test or measure
Alternate Form:
Internal Consistency:
Stability or Test-Retest: Sometimes referred to as
Involves the completion
Involves the same form parallel and involves
of one test or instrument
of a test or instrument two different versions of
at one time
a test or instrument
A key reliability issue deals with the consistency and stability of results that are gen-
erated from a measure or procedure. This is referred to as stability reliability, and it
involves the consistency and stability of a measure’s results over repeated administrations. This type
of reliability is usually referred to as test-retest reliability. The common way that the
relationship between two administrations is examined is through a correlation coef-
ficient, a number that is used to represent the strength and direction of a linear relationship between
two events, and in this particular case it functions as a reliability coefficient. This coef-
ficient can range from +1.0 to –1.0, and represents the strength of similarity between the two scores
or ratings. A correlation of +1.0 represents a perfect relationship between the measures,
0 represents no relationship at all between the measures, and –1.0 represents a perfect
inverse relationship between the measures.
As an educator, you need to know that the measures you use and complete in your
classroom would yield the same or comparable findings and results. Therefore, you
Type of
reliability Participants Material Administration
Stability
(Test-Retest) same same test different times (twice)
Alternate Form
(Parallel Forms) same different but same time
comparable tests
Internal Consistency
(Split Halves/Odd- same different items within same time
Even) same test (same
construct)
should be able to ask, “Would my students get the same or similar scores and ratings each
time their tests or projects are evaluated?” and be able to answer that affirmatively with
confidence, and only a measure with internal stability can provide that. This assurance
is important, since with all measurement procedures there are limitations that must be
recognized and understood.
For example, as the amount of time between administrations increases, the reliabil-
ity of the scores typically decreases, especially considering qualities like achievement,
motivation, and anxiety, that by their nature change over time. As more time elapses
between administrations, the potential effect of other factors on the measurement process
will likely increase and affect the reliability of the findings. Also, a carryover effect from
taking the same test twice can inf luence the internal stability of a test or procedure. As
Reynolds, Livingston, and Willson (2006) point out, “When evaluating the stability of
test scores, one should consider the length of the test-retest interval in the context of the
characteristics being measured and how the scores are to be used” (p. 92).
As a teacher you are probably saying to yourself, “When in the world would I give
the same test to the same group of students?” Once again this is a reasonable question to
be asking. In most situations you will not have the need. This type of reliability is com-
monly completed with standardized tests and batteries in order to provide measurement
confidence with the test forms that are produced. However, there are potential classroom
implications and uses. Since lessons are becoming more standards-focused, the tests and
performances that are conducted to measure learning progress relative to those standards
are now very important.
The need to check and confirm, or reconfirm, student progress and skill retention
has dramatically increased, so repeated testing of similar material is not unheard of
in schools. In fact, it is becoming a more common practice, as teachers want to make
sure their students have mastered content that is evaluated as part of district- or state-
mandated achievement testing. For example, in Mrs. Clark’s third-grade classroom, her
students have taken the same 50-item computation test involving single-digit multi-
plication problems on two separate occasions (with a two-week interval between the
administrations). She wants to know if her students have memorized and maintained
their mental math facts in regards to single-digit multiplication. Here are the findings
from those administrations:
First administration Second administration
Carol 49 50
Susie 48 48
Marji 48 47
Ray 47 48
Sarah 46 48
Riley 45 47
Joy 44 46
Alice 42 43
Sally 40 42
Jack 39 41
Mean 44.8 46
Test-Retest Correlation: .961
What does the data tell you about the performances of these students?
♦ Collectively, these students demonstrate strong math computation skills based on
both administrations, with an overall group performance increase during the sec-
ond testing.
♦ The students have retained their relative “rank” or position based on the second
test administration.
♦ As a group, it appears that they have been able to maintain and in some cases fur-
ther refine their math skills based on the most recent performances.
♦ An exceptionally strong reliability coefficient was demonstrated between the two
test administrations.
♦ The two-week interval between administrations did not appear to adversely affect
the student performances. (There is no recognized uniform or correct time interval
as this always depends on the skills or behaviors being reviewed.)
♦ Other findings?
In addition, test-retest reliability information is typically reported as part of formal
achievement tests and batteries; consequently, it is important to know and understand
what reliability is claimed with tests. For instance, if a test-retest reliability coefficient
of .92 is reported for an advanced placement science test that is being used in your class,
school, or district, as an educator you can hold confidence in the internal stability of this
placement test and in the consistency of the results that would be generated from this
measure.
An additional type of reliability generated for some standardized achievement tests is
alternate form reliability. This type of reliability occurs when two parallel or equivalent forms
of a test are developed. Reliability is obtained when both test forms are administered and
a correlation between the two sets of scores is computed. Having two comparable test
forms is quite valuable and necessary, especially for high-stakes tests. Additional work
and effort, however, is needed in order to develop two tests with strong alternate form
reliability. Both forms must be comparable to each other in several key areas, such as
material coverage and content, item difficulty, length, and general administration.
Teachers sometimes report that they do not have the time to develop comparable
test forms. Yet in many instances, equivalent test forms can be a valuable commod-
ity for the classroom teacher. For example, having alternate test forms is very useful
when summative assessments with large groups of students are involved. Also, different
forms can be administered to different classes, thereby lessening potential cheating and
other performance inf lation issues. This can also lessen the advantage for students who
take these exams later on in the day or week, depending on the testing schedule. Even
if alternate test forms are not emphasized in your classroom, familiarity with alter-
nate form reliability and its function is important. Remember, parents and the general
public will assume you know and understand most everything about tests, what they
measure, how accurate and reliable they are, and so on, and being a knowledgeable
consumer is very important.
Internal consistency reliability is focused on examining how consistently items within
a test correspond to each other and uniformly measure a concept or attribute (e.g., multiplication
skills, knowledge of the Civil War). By design, questions and items that make up a test
should be correlated if they are attempting to measure the same thing, and this form of
reliability attempts to quantify the internal consistency of the actual test items. There
are two basic approaches that are used to determine internal consistency: split halves and
odd-even.
The split halves reliability approach involves separating a single test into two equal halves
(e.g., first half as compared to the second half ) and generating a correlation between the scores
of each half-test. The Spearman-Brown formula, a formula designed to examine reliability
based on test length, is typically used to generate the reliability coefficient for items that are
viewed as dichotomous (i.e., only two possible options), which generate either a correct
or incorrect outcome. However, this only works if the difficulty level of the test items is
equally distributed throughout the test. When tests are constructed, especially classroom
tests, the easier items are sometimes placed at the beginning of a test, with the harder
items at the end of the test. If the split halves approach is used under these conditions,
an inaccurate reliability coefficient would be generated. In order to contend with this
possibility, the odd-even reliability approach can be used. With this approach, all even-
numbered test items make up one half of the test and the odd-numbered items make up the other
half. This way a more even balance of the difficulty level of the items is obtained, so a
more accurate reliability estimate can be determined.
Since only one test administration is necessary to generate an internal reliability coef-
ficient, and basic consistency information is provided, this method is very useful and
popular in determining the reliability of classroom tests (Kubiszyn & Borich, 2007). It
is important, however, to acknowledge the assumption that we make here, which is that
both halves of the test are equivalent. In reality, this may not be the case. Also, this
approach is not effective with tests that are considered to be speed tests. A speed test
contains a large number of items and questions (usually selected response items) with the expecta-
tion that few students will answer all the questions within the time limits of the test, compared
to power tests, which typically contain a combination of selected and constructed response items
that are specifically designed and administered to provide adequate time so that most, if not all, stu-
dents can complete the test. With speed tests, inf lated reliability estimates are usually found.
However, if a power test approach is taken, then split half reliability estimates tend to be
effective and accurate.
Another approach that is used to estimate the reliability of test scores from a single test
involves the use of the Kuder-Richardson formulas (known as formulas KR-20 and 21),
in which all test items are compared with each other instead of with different halves of
the test. As reported in Worthen, Borg, and White (1993), these formulas are designed to
provide an “estimate of what the average reliability would be if all possible ways of split-
ting the test into halves were used” (p. 154).
For test items that are not dichotomously scored, Cronbach (1951) developed the
coefficient alpha, which is more commonly referred to as Cronbach’s alpha, to measure
the internal consistency of a single test or measure. This approach is consistent with
the Kuder-Richardson formulas, but allows for a greater variety of test item formats,
including essay questions and items that have partial-credit scoring (Worthen, Borg, &
White, 1993). Given that this approach is computationally complex, classroom teachers
are usually not in the position (and perhaps do not possess the desire) to complete this
FOCUS ON ASSESSMENT
Inter-rater Reliability
Another important reliability measure for teachers to know about is inter-rater reliability.
Inter-rater reliability examines the consistency that is demonstrated between two or more rat-
ers that independently score a test or project. When a test is scored, a correlation is calculated
based on the scores of the raters. Generating a correlation, however, may not work well
within the boundaries of the classroom. Therefore, consistent with this approach, an
BananaStock/JupiterImages
solid grammar, pacing, and appropriate choice of words and expression. The stu-
dent provides a clear perspective and analysis of the poem.
Rating of 2—A general or generic analysis is provided. The logic in the author’s
thinking is clear, but the analysis lacks specific focus. A lack of clarity of the
author’s perspective is evident. Grammar is usually correct but may lack style and
expression.
Rating of 1—No recognizable analysis and focus on the work is apparent. Minimal
interpretation and review is provided, and grammar limitations, including errors
in spelling, noun–verb agreement, and awkward phrasing, are evident. The writ-
ing style is fragmented, inconsistent, and generates little to no f low or personal
expression.
Based on the provided ratings for this assignment, examine this sample of student work.
What rating would you give it and why?
The poem “When We Two Parted” by Lord Byron describes the loss of a love and
the intense feelings associated with that loss. In this work, he mourns the break-up of
a romantic relationship and focuses on his personal feelings and mood at the moment
when both lovers are forever parted. The author’s emotional viewpoint and tone is evi-
dent by the use of phrases such as “in silence and tears,” “half-broken-hearted,” and
“truly that hour foretold sorrow.” A personal self-indulgent approach is taken in the
poem as opposed to an emphasis on the love or romantic bond that existed. The secret
despair, sense of rejection, and pervasive sorrow are emotional elements there are found
throughout the entire poem.
A personal coldness and emotional numbness is also described. He states that, “the
dew of the morning sunk chill on my brow—it felt like the warning of what I feel
now.” He cares but he doesn’t want to care any longer as he has been hurt and has
sustained a wound that he will never recover from. At the conclusion of the poem, he
grieves silently and assumes that if he ever sees his beloved again, “after long years,”
he would be met “with silence and tears.” Clearly, Lord Byron emphasizes the mel-
ancholy and deep despair of a love that was never to last. But his emotional reaction
appears reasonable, controlled, and never seems over the edge or irrational.
Lord Byron uses simple but effective wording to portray a strong sense of melan-
choly to the poem. The words he uses are clear and focused with obvious meaning and
emotion attached to them. Because Lord Byron never reveals his beloved’s name, nor
gives any details or explanations for the break-up, an element of curiosity and wonder
exists. The reader is left to make assumptions and interpretations on their own which
adds to the poem’s appeal and mystery.
Both Mrs. Hamilton and Ms. Pham rated this particular piece of student work as a 2.
They agreed that they liked the expression in the writing, but that a general, not specific,
analysis of the poem was provided. Ideas and sentences were not arranged in a clear order
and style. It was presented as a collection of impressions without organization or a clear
progression from beginning to end. Both agreed that with some work, this piece could
be rated a 3.
1 2 3
Ratings of Judge 2 ___________________________
(Ms. Pham)
3 0 2 7
2 0 8 3
1 5 1 0
Bananastock/AGE Fotostock
results obtained from classroom measures and procedures, then consistent information
and findings can be expected and used with confidence in educational decisions and
actions.
AVOIDING BIAS
As valuable as validity and reliability are, so too must teachers be aware of bias, the subjec-
tive preference or penalization of some individuals over others, and how it can adversely inf lu-
ence student learning evidence. Now at first glance, you might be surprised to find bias
being discussed in an assessment textbook. When bias is mentioned, most people imme-
diately think of a subjective viewpoint or prejudiced act that is committed against some
individual, group, or race. In regards to assessment, we need to look at a specific condi-
tion referred to as assessment bias. As Popham (2008) has reported, “Assessment bias
refers to qualities of an assessment instrument that offend or unfairly penalize a group of
students because of students’ gender, race, ethnicity, socioeconomic status, religion, or
other group defining characteristics” (p. 73).
Assessment bias is directly connected to the materials and measures that teachers may
use in their classrooms. In particular, biased instruments and procedures create a disad-
vantaged learning situation for a specific student or students that can consequently limit
their academic progress, while at the same time creating a potential learning advantage
for others. It is important to note from the start that as professional educators, teachers
don’t go out and try to develop biased assessments in their classrooms. Genuine attempts
to offend or penalize certain groups of students aren’t part of their pedagogy and cer-
tainly aren’t encouraged in their training programs and in the schools where they teach.
It is more often the case that teachers are unaware of how utilized procedures or con-
tent may cause certain students distress and limit their progress compared to others in
the classroom. A lack of knowledge is certainly not an excuse, and that is why this topic
is being examined and reviewed. As a teacher candidate, valuable information can be
obtained with this topic so you can avoid assessment bias in your classroom.
Bias is most typically found in material that is used to test or evaluate students’ knowl-
edge or skills. Assessment bias can also be demonstrated in actual assessment procedures.
For example, when administering certain district math tests, it is common to require that
students keep their own calculators at home and use calculators provided by the school
as part of the testing. This is done to avoid cheating by bringing in more powerful cal-
culators, and it provides for uniform technology support for all the students. In general,
this is a good procedure. However, if we find that a certain group of students are totally
unfamiliar with this type of calculator, then an assessment bias is present for that group.
Also, if laptop computers are used as part of a test and only a certain kind of computer is
provided to the students, then students who are familiar with that type of computer may
have a distinct advantage over those who do not. This constitutes bias within the assess-
ment process itself, and though rare, it can occur. When it does, it provides for an inac-
curate measurement of student learning and accomplishment.
Bias is more often found with material that is used for tests or other evaluative mea-
sures. This typically involves individual tasks and items that are constructed as part of
the measures. For instance, if sport examples are used in a test, these items may be biased
in favor of students who are athletes and biased against those who are not. The student
athletes may not have a greater skill base, but because they have a greater understanding
of the context and detail of sporting situations, their chances of being successful with
these items are enhanced. Also, if music-generated items are used, or a music background
provides an advantage, then students who are band members are likely to be advantaged,
while their nonmusical classmates are not. These events ref lect an unfair penalty because
they advantage one group while at the same time disadvantaging other groups. Student
performance must be determined by the ability, effort, and preparation of each student
without any outside advantage or disadvantage. When this is accomplished, as a teacher
you can have confidence that students have truly earned their marks and that fairness has
been maintained.
Items can also be offensive by portraying individuals within a group in unattractive,
demeaning, and insulting ways. It wasn’t very long ago that on test items, physicians
were always portrayed as men and nurses were always women. That gender stereotype
has been basically and appropriately eliminated. New possibilities, however, are always
emerging. For instance, recently a political cartoon was reviewed as part of a government
class. The cartoon displayed a young skateboarder with tattoos smoking a cigarette and
hanging out on a street corner. Under the cartoon was the line “The future of America.”
How do you think the students in the class who skateboard or have tattoos are impacted
by that assignment? What stereotypes and misconceptions are being generated by this
cartoon? Even if the intent of the instructor is to have students evaluate existing stereo-
types, some students are immediately placed in a biased and potentially confrontational
situation, and are viewed differently from students who do not possess these character-
istics. This is the kind of material that needs to be reviewed, and possibly eliminated, if
bias is to be avoided in the classroom.
As a classroom teacher you typically don’t have the luxury of having a group of inde-
pendent reviewers examine your assessment procedures and material to determine if
students could be offended or biased. This is usually reserved for more comprehensive
district- or state-generated assessment and testing (see the box “Potential Test Item Bias
Inf luences Examined by the West Virginia Department of Education Assessment Office”).
So what is a teacher supposed to do about assessment bias? You do your best to be
aware of assessment bias and you make sure that it does not become a systematic part of
your teaching and assessment. Biased results invalidate any findings that you have, and
that must be avoided at all costs. In developing your sensitivity to this topic, you may
become aware of things that you do or ideas that you hold that need to be examined
and possibly changed. As teachers we never want to acknowledge that what we do in
the classroom or any other aspect of our teaching is biased against any student or group
of students. But the possibility exists, and for that reason we, as professional educators,
need to recognize that reality and implement review procedures on a consistent basis to
safeguard against assessment bias.
Consider the following suggestions as you develop your own set of guidelines to min-
imize and/or eliminate assessment bias in your classroom:
♦ Don' t assume common prior knowledge and/or experiences exist for your students.
Since students come in with varied backgrounds, interests, learning opportunities,
and so on, their “inconsistency” of experience needs to be recognized, particularly
when testing and assessment activities are involved. When constructing test items
or other assessments, try to create items, questions, or tasks that all the students
have directly experienced, preferably those that they have all shared under your
supervision within the classroom. For example, consider the following essay test
item dealing with cardinal points or directions:
During a spring kayaking trip to Alaska, two campers run into a blizzard two
days into their trip and are forced to travel by land to try to get back to their
base camp. They need to identify basic directions (north, south, east, and west)
but are without the aid of a compass or GPS and are out of radio contact. How
can they go about determining direction, and what materials, if any, will they
need? Answer: During the day, the shadow-stick method with the sun can be
used, and during the evening, locating the North Star might work.
This question is a good example of a biased test item, since it is set up as an
outdoor camping question that contains certain references some students may not
be familiar with, especially those who may come from urban or suburban settings
(e.g., students have to know what a kayak and base camp are). Students who are
knowledgeable about the outdoors would have an undue advantage over students
who do not have that kind of experience. A more unbiased approach would be
to use a shared common experience among the students to set up the test item or
activity. For example, consider this replacement item:
In class last week, we investigated the concept of cardinal points and learned
how to determine these directions without the aid of modern instruments. Now
imagine that you are in the soccer field behind the school and you need to
determine all the primary directions (north, south, east, and west). Without the
aid of a compass, GPS, or other instruments, how can you go about determining
direction, and what materials, if any, will you need?
This is a better test item because it sets or primes the students, as the first sen-
tence serves as a learning reminder about the topic. Also, the directional task is
completed in a recognizable setting. This item accomplishes the same measurement
end as the prior item and minimizes potential assessment bias.
♦ Don' t assume directions are clear and understood by all your students. To maximize
the directions of a test or activity, make sure clear and basic words are used that all
your students can understand. The “readability” or reading level required to follow
and comprehend the directions should never exceed the grade level of the students
being tested; in fact, the directions should be below grade level so the weakest stu-
dents can read and understand the directions. Also, stay with a set script or template
with your directions, and use this script during in-class activities whenever pos-
sible, so that your students are familiar with and understand what you want them to
do during an assessment event.
♦ Don' t surprise your students with test items or tasks they haven' t seen before a test or
activity is administered. It is important for your students to have as much experience
as possible with test questions (as they are asked, complete with skill level expecta-
tions) that they will see on a test prior to that event. This is essential, because as a
teacher you want to really find out what they know and understand, and familiar-
ity with the test items and tasks will maximize that effort. Otherwise, you end up
measuring how well students can figure out the directions and the unique aspects of
the items, and the measure of the desired knowledge or skill becomes a secondary
issue. The best approach to take is to use the same assessment items and procedures
with classroom work and learning activities that you use in more formal assessments
or evaluations with your students. Don’t ask your students to complete questions
that they haven’t seen before or have not practiced in the classroom (e.g., extended
response items), since this will impact the accuracy of the results and interfere with
obtaining a true measure of what your students have learned.
♦ Make sure the grading process and evaluation of student work is clearly explained
and understood by your students. In addition to following good assessment practices
when constructing and implementing tests, it is important for your students to be
fully informed of how their work and learning will be evaluated. Once again,
there should be no surprises, and the system should be clearly explained. A fair and
equitable grading system for all your students needs to be in place, a topic that will
be more fully discussed in Chapter 8.
CHAPTER REFLECTION
At the beginning of the chapter, you were asked to identify five common objects and
evaluate the importance that both validity and reliability play in the functional utility of
those objects. Validity and reliability are essential conditions for any procedure or measure
and that certainly applies to all assessment procedures utilized in the classroom. Continue
the validity/reliability search the next time you are in a classroom, even if it is part of
another class or a personal visit. During that visit, identify the assessment practices in that
classroom and try to determine if those measures or procedures meet the necessary valid-
ity and reliability conditions. As you now know, having one doesn’t necessarily mean the
other will be present. Unfortunately, you can find situations where a test or procedure has
validity but not reliability, or more tragically, where it’s reliable but not valid.
CHAPTER REVIEW
Conclusion
Both validity and reliability must exist for any measurement process, whether it occurs
inside or outside the classroom, and teachers need to be knowledgeable about these con-
cepts. Moreover, deliberate efforts to develop and maintain high levels of validity and
reliability within all classroom assessments need to be made because without these essen-
tials, assessments may generate questionable and misleading data, from which inaccurate
inferences can be made, followed by truly damaging decisions.
As the teacher and assessment professional in the classroom, you need to make every
effort to ensure that any measures utilized (for which multiple sources of data should
exist) truly measure what they are supposed to in order to use the results to make accu-
rate inferences about students and their learning. Without strong indices of validity and
reliability, the measurement process and classroom assessments are essentially worthless.
On the other hand, when practices and procedures that maximize these qualities are fol-
lowed, teachers can reap tremendous long-term instructional benefits and can be confi-
dent in the accuracy of the student evidence that is collected. That is why teachers need
to make validity and reliability awareness part of their standard routine and practice.
Only an introduction to the concepts and impact of validity and reliability on class-
room assessment have been provided here, and in no way should this review be seen as
exhaustive coverage of these topics. Readers are encouraged to seek out other resources,
some of which will be listed in the Resources section.
The issue of assessment bias was also examined in this chapter. Teachers need to be
sensitive to and identify procedures or content that might offend or disadvantage students
simply because they are a member of a particular group. Bias is something that must be
looked for specifically; otherwise it may go unnoticed. Most importantly, bias invalidates
any assessment results that are obtained, resulting in the collection of inaccurate data.
experiences, and so on needs to take place in order to ensure that individuals or groups of
individuals are not unfairly penalized during any assessment process.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. What does obtaining stable and consistent results from tests and other classroom
performances provide a teacher with? Describe the important relationship between
reliability and validity.
2. Identify and explain the different forms of validity evidence and their importance
in obtaining accurate assessment findings. In order to make correct inferences and
learning decisions, what must exist for you as a teacher?
3. What actions are you going to take to address assessment bias in your future class-
room? How is becoming aware of and sensitive to assessment bias the first step?
RESOURCES
The Social Research Methods Web site at www.socialresearchmethods.net provides a
number of resources and links that teachers will fi nd useful regarding basic statistical
issues and terms including reliability and validity.
For a complete review and explanation of basic validity and reliability concepts, the Sage
publication entitled Reliability and Validity Assessment (1979) by Edward Carmines and
Richard Zeller should be considered.
For a comprehensive review of assessment bias with culturally and linguistically diverse
groups, read the 1994 article by Robert Hernandez in The Journal of Educational Issues
of Language Minority Students, “Reducing Bias in the Assessment of Culturally and Lin-
guistically Diverse Populations” (Volume 14, pp. 269–300).
REFERENCES
Carmines, E. G., & Zeller, R. A. (1979). Reliability and validity assessment. Newbury Park,
CA: Sage Publications.
Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika,
16, 297–334.
Gronlund, N. E. (1985). Measurement and evaluation in teaching (5th ed.). New York: Mac-
millan Publishing Company.
Kubiszyn, T., & Borich, G. (2007). Educational testing and measurement: Classroom application
and practice (8th ed.). Hoboken, NJ: John Wiley & Sons.
Popham, W. J. (2008). Classroom assessment: What teachers need to know (5th ed.). Boston:
Allyn & Bacon.
Reynolds, C. R., Livingston, R. B., & Willson, V. (2006). Measurement and assessment in
education. Boston: Pearson Education, Inc.
Standards for Educational and Psychological Testing. (1999). Washington, DC: American
Educational Research Association.
Worthen, B. R., Borg, W. R., & White, K. R. (1993). Measurement and evaluation in the
schools. New York: Longman Publishing Group.
Worthen, B. R., White, K. R., Fan, X., & Sudweeks, R. R. (1999). Measurement and
assessment in schools (2nd ed.). New York: Longman Publishing Group.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How do I promote student learning during an instructional activity or lesson?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Define formative assessment and
differentiate it from summative
assessment.
♦ Distinguish the purpose of formative
assessment practices for students in
the classroom.
♦ Utilize different approaches and
measures in generating formative
feedback to learners.
♦ Develop an effective and usable
formative assessment approach for
learning events in the classroom.
♦ Acknowledge the importance of
integrating formative assessment
into your instruction.
♦ Identify the role formative assess-
ment should play in teacher review
and evaluation of instruction.
© Roy McMahon/Corbis
113
Introduction
The formal recognition and evaluation of a student’s work and progress is not designed to
be an isolated educational event. It only has real value and instructional meaning when
it occurs in combination with systematic formative assessment. Summative assessment is
connected to the instructional plan of where you want your students to end up regard-
ing their educational destination, but formative assessment is a critical component in the
instructional process of getting them to that place; in essence, it promotes student prog-
ress and provides instructional direction as the learning experience unfolds.
Formative assessment results provide a teacher with essential information on what his
or her students know and what they do not know, so that necessary instructional adjust-
ments can be made (e.g., information or skills can be retaught) based on that informa-
tion. As Wiggins (1998) has posited, “imagine an educative assessment system that is
designed to improve, not just audit, student performance” (p. 1). It helps to provide that
all-important mirror image and progress ref lection of student learning status at different
points during the learning process. Consequently, formative assessment must be viewed
as a necessary component of the “real time” part of teaching, where students receive crit-
ical review, confirmation, and correction as they complete their work and make desired
progress in the classroom.
CHAPTER ACTIVITY
Formative assessment provides the essential feedback and information that is needed in
order to successfully learn and complete a task. Remember back to when you first learned
how to ride a bicycle. Every attempt at riding was a critical learning experience, and as the
rider you knew almost immediately, through muscle and balance feedback, how success-
ful your efforts were in trying to maneuver and master this fascinating machine. Assessing
learning as it is occurring is an essential need for most activities, especially those that take
place in the classroom.
Now respond as if your current instructor has provided you with the following ques-
tions to answer:
1. Define formative assessment and identify its purpose for any learning event or activity.
2. Who should be the primary beneficiaries of formative assessment?
3. What will formative assessment “look like” in your classroom?
Through the lens of the “instructor,” the generated responses to these questions can
help indicate how well individual students, as well as the class as a whole, understand the
material. And from the perspective of the “student,” this information provides essential
feedback on key concepts that you understand or still need to work on depending on your
performance on these particular items. By the way, how complete were your responses?
Did you answer all of the questions? If you did, that’s great, but it’s unlikely given that you
only just started reviewing this chapter. If you didn’t, you know what is coming and what
to focus on as you read the chapter. What is so helpful for a learner regarding formative
assessment is that vital performance feedback is provided without a grade or mark. Students
can learn from their responses (including mistakes) without a formal cost or evaluation
placed on their performance.
Nichols, Meyers, & Burling, 2009). Once a student’s learning progress and educational
status is determined, instruction can be maintained and support and instructional clarifi-
cation can be provided, if needed, to help keep the learner on the intended academic path
(Boston, 2002).
The quality and effectiveness of the information that is gained from a teacher’s forma-
tive assessments is invaluable and has tremendous impact on the current and cumulative
learning effects for the student who receives it. It can be the difference between a student
understanding a presented concept or not. It’s the difference between acquiring a skill
and the accurate future demonstration of that skill or not being able to do so. Many con-
tend, and justifiably so, that meaningful formative assessment and feedback can be the
educational deal breaker for any student. As Stiggins (2005) has pointed out:
The students’ mission is no longer merely to beat other students in the achievement
race. At least part of their goal must be to become competent. Teachers must believe
that all students can achieve a certain level of academic success, must bring all of their
students to believe this of themselves, must accommodate the fact that students learn
at different rates by making use of differentiated instruction, and must guide all stu-
dents toward the attainment of standards (p. 326).
The issue of differentiated instruction is relevant for all students, not just those with
special needs, and formative assessment is essential in providing learning feedback
information as various approaches and modified instruction techniques are considered
for students who struggle to learn and acquire expected learning outcomes. Formative
assessment and differentiated instruction are essential instructional partners, and as such
need to be viewed as inseparable. The topic of differentiated instruction is covered in
more detail in Chapter 12.
FOCUS ON ASSESSMENT
THE DECISION
Integrating formative assessment within the fabric of your daily teaching regime requires
a certain amount of time and effort. Now you may ask “do I have to do this?” and the
answer is “no,” since as the teacher you retain direct control of the instructional process in
the classroom. But if you ask “do I really need to do this?” then the answer is absolutely
“yes,” since educational feedback is essential to your students, who must master the cov-
ered material. Moreover, knowing what your students have and have not acquired during
the learning event is critical to you as the instructor of the lesson. As an example, consider
this scenario: Would you send your students off on a 20-mile hike through an unmarked
section of forest armed with only a basic map and compass without providing them with
effective formative feedback on how to use a compass and read a map prior to the event?
Obviously you wouldn’t, and the instructional challenges in your classroom, while perhaps
not survival-oriented, are no different.
nucleus
nucleolus
ribosome
cytoplasm mitochondrion
FIGURE 5.1 ANIMAL CELL DIAGRAM IN SIXTH-GRADE
SCIENCE CLASS
centrioles
Golgi
apparatus
smooth Golgi
endoplasmic vesicles
reticulum ribosome
mitochondrion
FIGURE 5.2 TEACHER COPY OF ANIMAL CELL
With this particular exercise, and based on the formative findings, this teacher found
that the majority of the students could accurately identify the listed structures in the cell.
Just a few students required more individual work and review. Most generated accurate
diagrams; however, the explanations of the structures and how they worked together
varied tremendously within the class. Some students clearly understood all the structures
and their functions, while others were still unclear or confused. This formative exercise
provided important learning feedback, and based on that feedback a review was planned
the next day on the cell structures and their interconnected functions.
The observation-based review just described is an effective assessment technique
for this teacher with this activity. There are many ways to check and monitor student
learning and progress, including observation, practice work activities, checklists, dem-
onstrations, and performances. To further extend the formative assessment effect of this
particular exercise, student self-assessment procedures could also be embedded in it. For
instance, after the students complete the cell diagram and label all the parts, they could
be given a copy of the teacher’s cell picture to compare to their own.
To enhance this learning activity, the additional cell structures found on the teacher’s
copy could then be identified as part of a “challenge” follow-up search once the pri-
mary structures have been identified. Additional information sources such as compre-
hensive textbooks, Internet resources, or biological reference books could be used as part
of the follow-up work. Work dyads could also be set up so that students could review
their work and responses in small peer work groups. One or all of these accuracy review
strategies could be utilized. In the process, a better product is constructed and students
are engaged in autonomous, self-directed review practices. In addition to more accurate
work production, the students learn to review and evaluate their own learning progress,
which is a critical life skill that serves students well beyond their school career. More
information on self-assessment is provided in Chapter 7.
Determining Importance
4 Identifies most (greater than 90 percent), if not all, presented key concepts, ideas,
Exceeding and/or themes in the story. Explains in significant detail how they are important
Standard to overall text meaning.
3 Identifies an expected (80–90 percent) number of concepts, ideas, and/or
Meeting themes in the story. Explains with adequate detail how they are important to
Standard overall text meaning.
2 Identifies a minimal (less than 80 percent) number of all the presented key
Approaching concepts, ideas, and/or themes in the story. Explains with minimal detail how
Standard they are important to overall text meaning.
1 Guesses randomly and inaccurately attempts to identify important concepts,
Not Meeting ideas, or themes. No explanation of text meaning is provided.
Standard
ASSESSMENT IN ACTION
the teacher) decides how the assessment is to be implemented based on its intended pur-
pose. A test could just as easily function as a formative assessment, and in many instruc-
tional situations it is used for that very purpose. A test that measures learning progress
over a specified amount of material doesn’t have to possess a formal evaluation function;
it can serve as practice and/or as a preliminary review for a later test that does function as
an official summative measurement.
A variety of evidence collection techniques can serve a formative function if the need
exists. It is a matter of carefully sorting out the best techniques to use in the documenta-
tion and monitoring the achievement progression of your students. The selection of a
measure is important because it serves as the primary gauge of the learning accuracy of
the students. In addition, it needs to fit well within the routine and rhythm of the lesson
and the instructional style of the teacher. The more seamless it appears, the better, as the
assessment should exist as a natural part of the teaching process. Since teachers need to
follow the progress of their students over the course of a lesson, more than one collection
period and measure is desirable before students are formally evaluated at the end of an
instructional unit. These review points can also be designed so that they utilize minimal
classroom instruction time.
wit78704_ch05_113-135.indd 124
TABLE 5.1 APPROACHES USED TO ASSESS DESIRED LEARNING TARGETS
Constructed Selected
student student
Target Observation Checklist Questioning Interview Rubrics evidence evidence
Awareness and Not designed to Not designed to Good approach Can confi rm Not designed Good approach Good approach
understanding confi rm con- confi rm content for confi rming information for (or efficient to confi rm to confi rm
of information tent retention: retention: can comprehension as retention, but in) the review understanding of understanding
and content can document confi rm specific well as depth and time consuming of this kind of information and of informa-
knowledge specific actions actions taken that breath of knowl- and limited to a skill target strong confidence tion; efficient
or responses that may be based edge base few individuals with evidence and effective
may be based on acquired at a time due to required way to confi rm
on acquired knowledge self-generated acquisition of
knowledge responses large amounts of
content
Application Can confi rm Good approach Good approach Can confi rm Good approach Good approach Good and effi-
of informa- specific actions for documenting for determin- applied knowl- to rate the for examining a cient approach
tion and/or followed along the steps taken ing the extent edge and proce- learner’s student’s concep- for confi rming
procedural with the cor- and executed of the applied dures; typically demonstration tualization and student’s under-
understanding rectness and/ during an activity knowledge and limited by time and utilization implementation standing and gen-
or effectiveness or event procedures requirements of the knowl- of knowledge and eral application
of demonstrated and classroom edge and/or skill sets; requires of knowledge
actions resources procedures time and effort to and/or imple-
obtain results mentation of set
procedures
CHAPTER 5
4/12/11 5:20 PM
Constructed Selected
student student
Target Observation Checklist Questioning Interview Rubrics evidence evidence
wit78704_ch05_113-135.indd 125
Reasoning Not designed to Can be con- Can provide Good approach Good approach If items or Limited in deal-
and evalu- examine, review, structed so that useful evidence for obtaining for determin- requested prod- ing with skills
ation skills or confi rm ele- specific compo- and insight into detailed informa- ing the quality ucts are well of this kind and
with authentic ments in the nents of a process the components tion from the evidenced in designed, useful level
data, cases, or reasoning or are acknowledged and the thought learner; requires the demonstra- information and
scenarios evaluation pro- and confi rmed; process from a high time tion of these clarity can be
cess; can confi rm quality ratings the learners’ requirement skills and provided; high
overt behaviors can be added to perspective that usually does their usage level skills (e.g.,
and responses this method not exist in the with authen- writing, oratory)
inferred to be classroom tic events and may be required
involved in the situations with these
demonstrated demonstrations
process
Generation of Can confi rm the Can be used in Can be used to Comparable Good approach Good approach Extremely lim-
original prod- existence of the the confi rmation obtain a more effect as described for assessing the to confi rm ited in provid-
ucts or work product or work of meeting listed detailed under- in the question- attributes of the students’ ideas, ing any useful
and its intended criteria associated standing of the ing method, but product; quali- unique interpre- information with
FORMATIVE ASSESSMENT: GUIDING STUDENT LEARNING
effect, along with the piece of internal creative even more so tative distinc- tations, and gen- this skill target
with qualitative work process, along given the capac- tions allow for erative processes;
anecdotes with the specific ity to extract more precise most effective
details and ele- more information performance when written
ments unique to feedback expression is
each piece involved
125
4/12/11 5:07 PM
126 CHAPTER 5
anticipated learning targets. Specific instructional elements can be singled out (e.g.,
parts of the cell) if needed; however, information on whether the students have
acquired the final learning outcomes (e.g., understand the internal mechanisms and
physiology of the human cell) is most important.
2. Is the formative measure strategically placed in the teaching process to provide maximum ben-
efit to both the instructor and the learners? For instance, make sure that if only one
formative assessment is planned, it occurs no further than halfway through the
instructional sequence or learning activity. This should provide enough time to
make instructional adjustments if necessary, and also provide recovery time for stu-
dents who have veered off (or never got on) the intended learning pathway.
3. Are your students familiar and comfortable with the measurement forms and items that appear
in the formative measures? Do they know the purpose of these measures and that
there is a “no-cost” tag attached to these exercises? Full disclosure of the formative
assessment process and its function with the students in the classroom is imperative.
This will help to alleviate any stress or perceived anxiety that might be connected
with this assessment, especially if it possesses a “test-like” form or delivery.
time (e.g., independent seatwork, homework completion) in order to confirm that the
student acquired the desired skill.
The checklist information would be handed back to the students for their review,
along with any individual comments. Once again, formative supports and assessments
need to be designed and embedded within the instructional process, and can include the
following:
♦ the modeling, by the teacher, of the actual process of locating subject topics or
terms in a book;
♦ the verbal description and elaboration of the process that is provided along with the
demonstration;
♦ recording of procedural accuracy of the students via direct observation and check-
list documentation.
This approach could work for any procedural activity, and it provides written docu-
mentation on the performance status of each individual student. With this event, a clear
“show me” assessment approach is followed and is appropriate given the task and skill
target involved. Also, the generalization and continuation of this skill can and should be
checked at various times in order to confirm skill maintenance.
Considering a variety of assessment measures and determining those that fit best
within your instructional methods is important, since the assessment measures need to
be a natural part of the instructional process yet still provide essential learning status
information. As the learning progress data of the students is reviewed, they are separated
into two general groups: the group of students that is understanding and gaining mastery
of the information and skills that are being taught, and the group of students that is not.
This confirmatory information is important for both groups but especially for the latter,
as necessary adjustments and instructional modifications will be needed in order to get
them back on the expected learning pathway. The fact that some of your students may
not respond well to the academic instruction that is provided signals that alternative or
modified instructional methods are needed in order to reach those students.
Using formative assessment procedures to set up in a lesson doesn’t automatically
ensure that student learning will improve. Simply going to school doesn’t make you
successful, it’s what you do and learn in school and the application of your acquired
knowledge and skill sets to your future endeavors that define your success. Students
only benefit when the information that is collected from formative assessments is used to
enhance personal understanding and growth. It has to be meaningful and relevant to the
individual learner, and it has to be applied within the immediate instructional process.
FOCUS ON ASSESSMENT
(e.g., normal distribution and its characteristics, measures of central tendency, variance,
standard scores) and procedures. A summative test covering all of the presented mate-
rial, along with the interpretation of a case study, will be used at the end of the unit to
formally evaluate student accomplishment. During this unit, you want to monitor the
learning progress to ensure that your students are acquiring and demonstrating all the
knowledge and skills that are connected with the lesson goals. Terms, procedures, calcu-
lation, and interpretation of these operations are to be covered through lecture, discus-
sion, and small group learning sessions.
At this point, address and answer the following questions.
♦ Given the identified learning outcomes, what formative assessment techniques
would you consider? Why?
♦ How would you embed these activities within the daily instructional routine?
♦ Once you have identified the techniques and activities, discuss your selections with
your classmates. Are your reasons and justifications for their usage similar to your
colleagues’? Why or why not?
♦ Are the formative assessment activities strategically placed throughout the instruc-
tional process?
♦ As the teacher, would you be getting the student progress information and per-
formance feedback early enough to make appropriate instructional adjustments if
needed?
The selected assessment methods should depend on the kind of evidence that is
needed. For example, the students must be able to recognize and differentiate specific
statistical terms and operations. Therefore, different approaches to measure those skills
and provide evidence on the students’ proficiency relative to those skills (e.g., small work
groups responsible for the solution of a statistical problem, crossword puzzle of statisti-
cal terms) could be used. In addition, the rationale for the use of those measures is also
important. Just like students, educators need to be accountable for what they do in the
classroom, and that includes the use of assessment methods.
Now let’s move on and compare selections. With this exercise, I selected three differ-
ent formative assessment measures to be incorporated within the unit. First, a formative
or “practice” exam containing the terms and procedures to be completed is constructed.
This is to be administered halfway through the unit to make sure the students are acquir-
ing the basic understanding of the material. Since this is a formative exam and no grade
is given for this test, the students can check their answers and review steps and computa-
tions once they complete the test to determine which, if any, concepts are still unclear
and need additional work. Also, selected parts of this exam or the entire test can be taken
more than once if desired.
Second, two case studies that are comparable to the one provided in the summative
test are examined in order to help the students understand the various scores and their
meaning, along with the instructional implications of the presented test performances. In
essence, these cases serve as practice and review for the case study that is included in the
final exam at the end of the unit.
Also, exit slips are completed by the students at the end of every class period. Stu-
dents anonymously respond to listed questions and complete them before they exit the
classroom. The collective findings are examined and used as review for the next day. For
instance, the following question is listed on the exit slip: “What is still confusing to you
about what was covered in today’s class?” The student responses to that question would
be collected, and the listed concerns, if any, would be discussed at the beginning of the
next class period. This review process serves as a good way of recycling the previous con-
tent as well as checking for understanding before moving on to new material. In addi-
tion, selective questioning and review could be done with students who are observed to
be struggling with the material during class time.
Documentation of progress can be provided with these measures. With the formative
exam, accuracy percentages can be recorded by both the teacher and the student. In addi-
tion, error analysis (i.e., identifying correct and incorrect student responses) can help deter-
mine what specific areas or concepts the student is having difficulty mastering. For the
case studies, accuracy percentages can be calculated for the listed questions. More qualita-
tive information is usually obtained from exit slip responses, but even they can be modified
to get ratings or response percentages if so desired. Also, these three formative measures
can be easily constructed and embedded within the instructional framework of the lesson.
Now review and answer these questions.
♦ How similar were your selected methods to the ones just described?
♦ Do you agree that these are appropriate formative assessments for this unit? Why or
why not?
♦ What changes, if any, would you make to the selected list?
FORMATIVE ASSESSMENT
AND THE TEACHER
The ultimate purpose of formative assessment is to provide the learner with informa-
tion that directly helps improve his or her learning in the classroom. But there is an
equally important consequence with the use of formative assessment. As students gain
understanding and improve their skills relative to the instructional tasks in the lesson,
so too should a teacher acquire better understanding of the instructional effectiveness of
his or her teaching and its delivery. It is important for the teacher not to ignore the for-
mative feedback that is automatically provided as part of the instructional experience.
Don’t forget that you as the teacher are learning right alongside your students, and that
you can use your experiences to modify your own instruction based on what student
evidence tells you. Formative assessment lets you know whether or not your students
understand the presented material. If they do, then you should continue. If they don’t,
then re-teaching is needed before further progress can be obtained. Given that immedi-
ate student performance feedback is available, the following questions are natural ones
to ask.
♦ Was the lesson effective? Can the students demonstrate the intended outcomes?
♦ Did the students understand and connect the ideas and skills that were presented?
Do all students need further instruction with this material, or just a few? Should
the material be reviewed in a large group, a small group, or individually?
♦ Could the content be presented in a different way to be more effective? If yes, how?
How do teachers find the answers to these questions? Since students are not likely
to comment on how well or poorly a subject or concept is covered or explained (“Mrs.
Davies, you did a bang-up job of covering the difference between direct and indirect
objects today!”), every teacher needs to be able to use evidence collected from formative
assessment to gauge the progress of their students, as well as the instructional effective-
ness of the lesson that has just been provided to them. As experienced teachers would say,
“You need to be able to read the signs.” Just make sure those signs have student data and
evidence connected to them. Remember that how well your students do is dependent to
a great extent on your instructional performances, so these need to be done well. That is
why effective teachers use student performance information to guide their instructional
practice (Heritage, 2007; Stiggins & DuFour, 2009). The most important information
you receive is from your students, but it is only useful if it is recognized and integrated
into your teaching.
CHAPTER REFLECTION
Whether you are riding a bike or completing a lesson in class, information gained from
formative assessment is essential both for the teacher who orchestrates the learning expe-
rience and the students who are held responsible for learning the material and skills that
are covered as part of the instruction. What did your results from the formative assess-
ment questions that were completed at the beginning of this chapter tell you as a prospec-
tive teacher? What role will formative assessment play in your classroom? Information
from a formative assessment should provide clarity on a student’s progress and his or her
continuing needs while the learning process is unfolding, and that is why it needs to be a
part of the instructional process.
CHAPTER REVIEW
Conclusion
Formative assessment, which is under the instructional control of every teacher, is
designed to examine and improve the learning progress of students as they proceed
through material provided in their lessons. This assessment is focused on the learning
process as it is occurring, and the resulting impact on each learner. As Blaz (2008) states,
“Formative assessment should help determine what the students have mastered, what
they still need, and what needs to happen next” (p. 29). Moreover, the collected student
information directly inf luences the adjustments, if needed, to your instruction, along
with the consideration of other instructional techniques. What makes this assessment
form so essential is that the corrective information holds a “no risk” status for the learner.
Benefits are obtained in the acquisition of greater knowledge or a new skill without hav-
ing a formal evaluation attached to one’s performance.
Formative assessment helps to confirm that your students are making progress as well
as to maintain that progress toward their expected learning goals. Stiggins (2005) has
reported that
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How important is formative assessment in the planning of your classroom routine?
What functions will it serve?
2. How will you integrate formative assessments within your daily and weekly lessons?
3. How will you select the approaches needed to function as formative assessment
measures in your classroom? Are multiple measures considered and incorporated
so that more than one source of student learning and feedback exists?
RESOURCES
The Formative Assessment for Students and Teachers (FAST) at www.ccsso.org is an
extremely valuable resource initiated by and maintained through the Council of Chief
State School Officers (CCSSO). The FAST mission is to encourage discussion about
formative assessment, examine its use in impacting student learning, and support high-
quality teaching and learning in the classroom. Additional informational resources are
also provided and include related topics such as school accountability, English language
learners, and math and science assessment initiatives, as well as special education sup-
ports. Typing in “FAST” in the search bar on the CCSSO site will take you to this
information. Additional information can be obtained by typing in “assessment and
standards.”
The Educator’s Reference Desk at ericir.syr.edu contains thousands of documents and
references that can be examined on a wide variety of educational issues, including
formative assessment resources.
The Exemplars Web site at www.exemplars.com/index.php contains useful information
about formative assessment and how it can be integrated within the classroom. Infor-
mation and examples of rubrics are also provided.
REFERENCES
Ainsworth, L. (2007). Common formative assessments: The centerpiece of an integrated
standards-based assessment system. In D. Reeves (Ed.), Ahead of the curve: The power of
assessment to transform teaching and learning (pp. 79–101). Bloomington, IN: Solution Tree
Press.
Black, P., Harrison, C., Lee, C., Marshall, B., & Wiliam, D. (2003). Assessment for learning:
Putting it into practice. Maidenhead, UK: Open University Press.
Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education,
5(1), 7–74.
Blaz, D. (2008). Differentiated assessment for English language arts. Larchmont, New York: Eye
on Education.
Bloom, B. S. (1969). Some theoretical issues relating to educational evaluation. In R. W.
Tyler (Ed.), Educational evaluation: New roles, new means (National Society for the Study of
Education Yearbook, Vol. 68, Part 2, pp. 26–50). Chicago: University of Chicago Press.
Bloom, B. S. (1971). Mastery learning. In J. H. Block (Ed.), Mastery learning: Theory and
practice (pp. 47–63). New York: Holt, Rinehart & Winston.
Bloom, B. S. (1974). An introduction to mastery learning theory. In J. H. Block (Ed.),
Schools, society and mastery learning (pp. 3–14). New York: Holt, Rinehart & Winston.
Boston, C. (2002). The concept of formative assessment. Practical Assessment, Research &
Evaluation, 8(9). Retrieved from http://pareonline.net/getvn.asp?v=8&n=9
Chappuis, S., & Chappuis, J. (2008). The best value in formative assessment. Educational
Leadership, 65(4), 14–19.
Earl, L. M. (2003). Assessment as learning: Using classroom assessment to maximize student learn-
ing. Thousand Oaks, CA: Corwin Press, Inc.
Fuchs, L., & Fuchs, D. (1986). Effects of systematic formative evaluation: A meta-analysis.
Exceptional Children, 53, 199–208.
Heritage, M. (2007). Formative assessment: What do teachers need to know and do? Phi
Delta Kappan, 89(2), 140–145.
Herman, J. L., Aschbacher, P. R., & Winters, L. (1992). A practical guide to alternative assess-
ment. Alexandria, VA: Association for Supervision and Curriculum Development.
Nichols, P. D., Meyers, J. L., & Burling, K. S. (2009). A framework for evaluating and plan-
ning assessments intended to improve student achievement. Educational Measurement:
Issues and Practice, 28(3), 14–23.
Stiggins, R. (2005). From formative assessment to assessment for (FOR) learning: A path
to success in standards-based schools. Phi Delta Kappan, 87(4), 324–328.
Stiggins, R., & DuFour, R. (2009). Maximizing the power of formative assessments. Phi
Delta Kappan, 90(9), 640–644.
Wiggins, G. (1998). Educative assessment: Designing assessments to inform and improve student
performance. San Francisco: Jossey-Bass.
Wiliam, D. (2006). Formative assessment: Getting the focus right. Educational Assessment,
11, 283–289.
PERFORMANCE ASSESSMENT
KEY FOCUS
The central focus of this chapter is to examine the fundamental question: How
can performance assessment activities be effectively utilized within the classroom
setting?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Define a performance event and identify its two major parts.
♦ Distinguish the primary features of a performance event.
♦ Select an appropriate assessment method for a chosen performance event.
♦ Differentiate the basic forms and functions of portfolios.
♦ Explain the advantages and disadvantages of
using a portfolio in the classroom.
♦ Understand and follow the basic steps in con-
structing a portfolio.
♦ Comprehend the aim of project-based learning
(PBL) activities for a classroom.
♦ Describe the characteristics and purpose of a
rubric.
♦ Differentiate a holistic rubric compared to an
analytical rubric.
♦ Complete the needed steps in constructing a
rubric for a classroom performance event.
♦ Identify the importance of checklists in confirm-
ing student behavior and progress.
© Tim Pannell/Corbis
136
Introduction
Both inside and outside the classroom, students are expected to possess and demonstrate
integrated skill sets and behaviors, and performance events are a natural way of assessing
multiple learning dimensions (Cohen & Spenciner, 2003; Moss, 1992; Wiggins, 1989).
Through learning demonstrations and authentic work projects, students are given the
opportunity to perform a particular activity with the expectation that it be will done
accurately and independently with minimal outside guidance. These kinds of learn-
ing verifications meet the confirmation that is ref lected in Missouri’s state nickname of
“Show Me,” and are viewed as highly desirable by those connected to the learning pro-
cess: teachers, students, parents, and the general public, which includes future employers.
Ongoing professional interest, measurement support and research evidence, and other
advances have helped to continue the use and refinement of this assessment approach in
the classroom (Bond, 1995; Brennan & Johnson, 1995; Green, 1995; Wiggins, 1992).
WHAT IS PERFORMANCE
ASSESSMENT?
A performance assessment requires a learner to complete a performance event. A perfor-
mance event involves the presentation of some specific individual and/or group-based performance
that typically involves an external reviewer, usually the teacher or some other appointed official, who
observes and critically evaluates the performance based on a pre-set list of criteria. In addition to
the actual performance, the skills or competencies required to generate the specific work
product (e.g., the creation of a Web site, defending a position during a debate) are often
reviewed and evaluated as part of the process. There are two major parts that compose a
performance assessment. First, there is the actual performance-focused task, which can
be referred to as the performance event, and the evaluation of the task, which deals with
the assessment of the student’s or group’s performance relative to that event (Stiggins,
Arter, Chappuis, & Chappuis, 2004).
The tennis toss activity that was discussed in Chapter 5 can be considered a small per-
formance assessment event. The overall acquisition of the tennis serve, however, would
serve as a better example of a comprehensive performance event. As would be expected,
performance assessments often include several elements or components, given the com-
plex nature of the product or performance that is being produced by the participant
(Linn, Baker, & Dunbar, 1991; Wiggins, 1993, 1998).
For example, at sports development camps, it is now standard practice for junior play-
ers to receive a DVD and video analysis of their skill demonstrations and a review of
techniques, along with practice and game play review. In addition, skill checklists based
on various aspects of the game, such as technical position execution, agility, strategy,
physical and mental conditioning, and rules of the game, along with performance rat-
ings and individualized feedback response boxes, are often provided as part of the review.
Multiple formative assessments are used in order to generate multiple sources of feedback,
and this is a very effective approach for both short- and long-term learning conditions.
For the short term, it maximizes the time element by providing considerable information
to examine and review. With long-term conditions, developmental progress can be more
fully examined, progress monitoring can be documented, and a substantial record can be
provided.
If a teacher is looking to demonstrate standards-related progress and skill changes of
his or her students over an extended time period, performance assessments would be a
logical and appropriate assessment choice (Bhola, Impara, & Buckendahl, 2003; Moon &
Callahan, 2001). However, time and effort is required on the teacher’s part in the devel-
opment and implementation of performance assessments. With that prerequisite under-
stood, the amount and level of information that is provided to both the teacher and the
student certainly makes this a desirable assessment consideration.
Most work that is generated in the classroom involves a construction of some kind
(e.g., generating an answer to a question usually requires a writing performance). Never-
theless, performance events and assessments are by design focused on substantial criteria
and typically require distinguished and refined responses from students. There are sev-
eral critical learning characteristics or qualities associated with performance assessments
(Linn & Miller, 2005). For example, performance assessments require a student to per-
form, sometimes in front of others, and participate in an identified activity or task such as
a recital or the creation of a product such as a pen-and-ink portrait.
Within a performance event, a demonstration of high-level skills would be expected,
along with an extensive understanding of the knowledge connected with this event. This
level of understanding is often expected to be at such a level that the student or group can
independently perform and complete the activity. Moreover, students can obtain a level
of competency and relative mastery that enables them to instruct and teach the knowl-
edge and skills to other learners upon request. Also, the integration of several skills and/
or knowledge sets is often necessary in order to complete the performance event (see the
box “Features of a Performance Event”). As mentioned earlier, these events are authentic
in that they are designed to duplicate real-life events and required skills for work or other
functional life purposes.
• Higher level cognitive processing skills are typically required in the completion of the
performance event.
• Multiple skill integration involving different knowledge, skill, and ability sets are
evidenced within a performance event.
• Performance events reflect real world skills, experiences, and the capacity to do and
complete a task.
• Performance events are typically reviewed through an objective, systematic scoring
of the quality of the student work based on listed standards or proficiencies.
© Image Source
The reviewer involved in a performance assessment must be fully informed and well
trained when evaluating performance events. Clear, specific standards or performance
criteria need to be identified and consistently followed when reviewing student work,
and they must also be known and made available to the participants involved in the per-
formance assessment. So the event, with its knowledge and skill requirements, along with
the formal and systematic evaluation of the students’ performances relative to that event,
needs to exist as part of the performance assessment. For example, examine the perfor-
mance rubric in Table 6.1, used in the audition process for students interested in joining
a music performance ensemble. This measure serves as an initial review of the student’s
performance and is also used two or three times during the grading period to review the
quality of the student’s individual work (and progress) at those specific review times.
5 4 3 2 1
wit78704_ch06_136-165.indd 140
Tone Quality Tone quality is clear Tone quality is clear Tone quality suffers Tone quality is Tone quality is
and appropriate to and appropriate to minor mistakes often unclear. The unclear throughout.
the instrument at the instrument most at times. Proper instrument does The instrument
all times. Constant of the time. Air air support and not sound as it never sounds clear
appropriate air support and speed speed only occur should. Proper air and appropriate. Air
support and speed. occurs often. occasionally. support and speed support is lacking
only occurs for brief throughout entire
moments. excerpt.
Intonation Candidate performs Candidate encounters Candidate encounters Candidate encounters Candidate performs
excerpt with no 1–2 notes that are out 3–4 notes that are numerous intonation entire excerpt
intonation issues. of tune (sharp or f lat) out of tune (sharp or issues. Tuning of with inappropriate
All notes are in tune within the key of the f lat) within the key instrument may intonation. Tuning
within the realm of excerpt and intended of the excerpt and have been executed of instrument clearly
the key signature. pitch. intended pitch. inaccurately. did not occur.
Rhythm Candidate performs Candidate Candidate Candidate Candidate
with no rhythmic misinterprets 1 misinterprets 2 misinterprets 3 misinterprets 4+
errors. All rhythms rhythm within the rhythms within the rhythms within the rhythms within the
are executed passage. passage. passage. passage.
successfully.
Dynamics Candidate Candidate Candidate takes Candidate adheres Candidate performs
successfully adheres successfully adheres account of dynamic to some, but not all entire excerpt at
to all dynamics to dynamic markings markings but their dynamic markings. one dynamic level.
within the excerpt within the excerpt interpretation does Interpretation of No adherence to
as indicated within as indicated within not ref lect what is on dynamics does not or interpretation of
the music, including the music. However, the page. ref lect what is on the indicated dynamics
ideas such as candidate misses page. exists.
crescendos and sFp, crescendos, etc.
etc.
Note Accuracy Candidate performs Candidate performs Candidate performs Candidate performs Candidate performs
excerpt with excerpt with excerpt with excerpt with 5+ pitch with numerous pitch
no note errors. 1–2 pitch errors. 3–4 pitch errors. errors. Candidate errors resulting
Performer adheres Performer adheres Candidate adheres adheres to key from misinterpreted
to key signature and to key signature and to key signature but signature but ignores key signature
CHAPTER 6
4/12/11 5:09 PM
5 4 3 2 1
Articulation The articulation 1 articulation error 2–3 articulation 4+ articulation errors All articulation
interpretation of occurs in context errors occur in occur in context of indications are
wit78704_ch06_136-165.indd 141
the performer is of the excerpt, but context of the the excerpt. The ignored. All notes
f lawless and ref lects candidate otherwise excerpt. The general general interpretation are performed slurred
everything indicated executes articulations interpretation of of style of the excerpt or tongued. Style
on the page. effectively. style of the excerpt is is misinterpreted. of piece is ignored
slightly off. and indications such
as “marcato” are
ignored.
PERFORMANCE ASSESSMENT
Tempo The tempo is The tempo is The tempo is off The tempo is off by The tempo is
appropriate given the appropriate given by 10–20 clicks and 20–40 clicks and completely off from
marking in the piece the marking in the does not ref lect the does not ref lect the the appropriate
and style of the piece. piece and style of the style of the piece. style of the piece. tempo and does not
Performer adheres to piece. Performer does Performer does Student ignores ref lect the style of the
all tempo changes. not execute tempo not execute tempo tempo changes. piece. No adherence
changes effectively. changes effectively. to tempo changes
exists.
Continuity Candidate performs Candidate stops and Candidate stops and Candidate Student is unable to
the excerpt with restarts once in order starts 2–3 times in consistently stops finish the excerpt.
no stutters or stops. to regroup. order to regroup. and starts in order to
Beginning to end in regroup but makes
one take. it to the end of the
piece.
Posture/Embouchure Candidate is seated Candidate sits Candidate sits Candidate’s Candidate never
Formation with perfect posture down properly but down properly initial posture is sits with proper
from beginning to posture is affected but immediately inappropriate for the posture. Consistent
end. Embouchure by playing at times. shifts upon playing. given instrument. slouching and
formation is Embouchure Embouchure Embouchure laziness observed.
appropriate to the formation is formation is formation is passable. Embouchure is
instrument. appropriate to the acceptable. inappropriate for the
instrument. instrument.
Scale Scale performed Scale performed with Student misses the Student misses two Student misinterprets
f lawlessly. Pitches one missed note but same note on the notes up and down entire key signature
and tone quality recovery and tone way up and down the the scale. Tone of scale. Tone quality
suffer no mistakes. quality are observed. scale. Tone quality is quality may begin to may begin to suffer.
passable. suffer.
141
4/12/11 5:09 PM
142 CHAPTER 6
CHAPTER ACTIVITY
Take a moment and try to identify the performance events that you complete during the
normal course of a day. These events are often tied to essential needs and purposes, yet they
become so automatic we often take them for granted and forget their significance. They
can include a wide range of events, including cooking breakfast, balancing a checkbook,
driving to school or work, completing a computer search, and helping run your school’s
soccer or cheerleader practice, as well as a hundred more activities. The fact is that each of
our days is made up of performance events (many of which may be interconnected with
each other), and the success of our day and our immediate and long-term future is often
tied directly to how well they are completed.
As part of this activity, identify two performance events that you have to address and
complete as part of your daily routine. Based on those experiences, answer the following
questions.
1. Are the selected performance events essential to the daily goals and outcomes that you
(or someone else) have identified? Why or why not?
2. How would you rate your performance on each of them? Were they done well? How
do you know? Were there natural consequences and/or external feedback?
3. Are internal and/or external expectations associated with these performances?
Hopefully this exercise has pointed out how prevalent and relevant performance events
are within one’s daily experience, and as such represent the essential “doing” part of our
work and life.
move students beyond the foundation levels of cognitive processing (e.g., understand-
ing, comprehension) to higher processing skills. This provides an opportunity of going
beyond merely writing about a subject or topic and actually doing it. For example, per-
forming experiments during a science lab ref lects an “active doing” process, where stu-
dents are required to merge information and skills together in order to understand a
chemical process, identify specific chemical compounds and the elements that they are
made of, or perhaps identify a new pollutant that has entered a ground water system.
to extend what is known within a certain research field could all constitute meaningful
performance events for students in the classroom. With the research opportunities that
the Internet now provides, extensive literature searches, current research reviews, and
professional contacts are almost unlimited for students as they complete various perfor-
mance assessments within the classroom.
The nature of a performance event and the assessment of a student’s progress relative
to that event are to create a situation where a student is required to integrate his or her
knowledge and skills and demonstrate that understanding in some form of a demon-
stration or learning product. With performance assessments there is typically a deliber-
ate focus on the creation of original individual work that meets or exceeds specified
criteria. What is desired is to generate a student product that stands as clear evidence
that the required academic and learning outcomes and objectives identified as part of
the classroom lesson have been met. The performance event and the learning outcome,
as identified through the expected academic standards, must match up. Instruction-
learning-assessment alignment is as necessary here as it is for any assessment activity.
ASSESSMENT IN ACTION
Written performances:
♦ short stories
♦ essays
♦ professional/personal letters
♦ journal/lab entries
♦ portfolios
♦ research reports
Motor performances:
♦ dance recitals
♦ music recitals
♦ marching band presentations
♦ sporting events
Visual performances:
♦ studio art
♦ graphic design
♦ digital construction
♦ photography
Many if not all of these performances ref lect the integration of multiple sensory and
motor skills, but one emphasized modality tends to stand out with each particular per-
formance. In addition, this list is by no means exhaustive, but it does clearly demonstrate
that there are multiple ways to demonstrate a student’s acquired learning and skill devel-
opment, and many of these by design can be utilized as performance events.
One of the most common student products that is connected with performance assess-
ments is the portfolio (Arter & Spandel, 1992; Butler & McMunn, 2006). A portfolio can
be specific to a content area (e.g., writing), a certain medium (e.g., pastel or watercolor), or a certain
professional focus (e.g., photographic documentary), and it reflects an integration and development of
skills and/or experiences over an extended period of time (e.g., a student teaching portfolio). The list
of possible performance events is limited only by the imagination and determination of
the teacher in identifying those learning events. Since these events are “real world” in
their direct application and usage, they provide valuable data on learner progress to both
the teacher and the students themselves.
Classroom Portfolios
A classroom portfolio involves the collection and personal review of pieces of work,
sometimes referred to as artifacts, that demonstrate the growth and development of spe-
cific knowledge or skill sets. Traditionally, portfolios have been recognized as the pri-
mary work evidence in certain professions, including art, architecture, and photography.
However, this method has been assimilated into the field of education, and many teach-
ers find this approach particularly effective at documenting the process and progress of
student learning as it occurs over an instructional period (Butler & McMunn, 2006).
Different types of portfolios are designed to fulfill specific functions. For instance, the
developmental or process-oriented portfolio contains artifacts or student work produc-
tions that demonstrate, over time, the development of the student’s skills as evidenced by
the tasks that are completed in the portfolio. For example, for a research portfolio project,
this would include the construction of the outline of the research paper, the first draft
of the literature review and prior research, a reworked draft with teacher feedback, con-
struction of the second draft of the research literature, collection of data and/or responses
from the subjects interviewed as part of the study, analysis of the data/responses, etc. The
portfolio can be theme- or skill-focused, and the emphasis is on the continuous refine-
ment and improvement of the student’s work product throughout the process (Valencia &
Calfee, 1991). This type of portfolio is common in elementary settings, as teachers desire
evidence that their students are acquiring and improving certain foundation knowledge
and skill sets (e.g., a writing or math portfolio).
A product or best-pieces portfolio follows a showcase format, where the student’s best
quality pieces are demonstrated, and usually several different types or kinds of pieces are
collected (e.g., different types or styles of writing pieces, a variety of art pieces using dif-
ferent mediums such as charcoal, ink, or oil). The finished products, and not the process
that generated these products, are the focus of this type of portfolio. The selection of the
best pieces, usually based on a set of specific criteria, can be left up to the student, or it
can be a joint decision made with the teacher. Pre-selected standards for each artifact
are needed in order to facilitate the selection of the individual pieces. This approach is
common at the middle or secondary school level, given the high skill levels that typically
exist for more mature learners and students. Whatever method is selected, a portfolio
assessment can provide rich and detailed information about your students and ref lect
many desirable learning qualities. For instance, portfolios
♦ require and encourage self-assessment skills of the students as they complete the
required assignments;
♦ involve the creation and use of performance criteria and rubrics in the evaluation of
the quality of the work that is generated;
♦ can be connected to and reviewed by a wider public audience (e.g., portfolios can
be exhibited and/or judged as part of a community event, school- or district-wide);
♦ can be utilized at any grade level, although this format may be difficult to imple-
ment with the youngest of learners.
Due to the authentic and active construction nature of portfolios, instructional ben-
efits are recognized with their use. Yet as with any assessment method, there are limi-
tations, and the portfolio is no exception. In particular, some of the most recognized
benefits and shortcomings of this method are the following:
Advantages/strengths of portfolios:
♦ allow for the examination of the unique work of each student
♦ can involve one or multiple content areas (e.g., reading, writing, science) and the
integration of skills across knowledge and/or discipline areas
♦ directly involve each student in the review of his or her progress and in the plan-
ning of future learning behaviors
♦ require considerable student–teacher communication, interaction, and collabora-
tion on a consistent basis
♦ encourage student ownership and responsibility of the portfolio and its successful
completion
♦ focus is on continued improvement and the effort to achieve or even exceed per-
formance standards or outcomes
Disadvantages/weaknesses of portfolios:
♦ extensive time requirements
♦ if additional assessment measures (tests, homework, other projects, demonstrations)
must be completed in the classroom, even less time is available for portfolios
♦ clear portfolio goals, performance criteria for work products or artifacts, adherence
to timelines, and so on must be established before a portfolio is initiated
♦ the progress monitoring caseload can be overwhelming if not carefully managed
♦ scheduled individual meeting times with students are necessary, which further
limit general classroom instruction time
♦ due to the systematic collection and review of student work, teacher commitment
to the successful implementation of this assessment method is essential
As with any assessment procedure, it is important for students to know and under-
stand what the instructional purpose and demands are for an activity, and that certainly
applies to a portfolio. Sharing the expected learning outcomes prior to and through-
out the learning experience is essential, especially as a portfolio project may involve an
extended period of time, and students may need to be reminded of their expected goals.
In the collection phase, a student is required to collect most, if not all, of their prod-
ucts and constructions. Documentation of work in all its forms (e.g., class notes, first
drafts, journal recordings, diagrams, timelines) is necessary as the process unfolds and the
student gains a better understanding of how all these pieces of work start to form a pat-
tern for the learning process.
In the selection phase, depending on the type of portfolio, artifacts are identified that
will be used to judge the quality and extent of the knowledge/skill accomplishments that
are rendered in the collected works. This can vary from a set list of required items (e.g.,
first draft, second draft, peer and teacher review, and final copy) that everyone submits,
along with items chosen individually by the student. The role and selection power of the
student is clearly controlled by the teacher and the guidelines that are reviewed prior to
the implementation of the portfolio assessment. For example, a project rubric that lists
pre-selected criteria along with the ratings associated with various quality/performance
levels can be used independently by the student and/or jointly with the teacher to iden-
tify the status of the product, in order to determine whether it is ready to be included in
the final review process of the work in the portfolio.
Personal review and ref lection involves the direct evaluation of a student’s work by
the actual student or in collaboration with the teacher. This process involves direct self-
assessment, where the student, as objectively as possible, critiques his or her performance,
determines what has been achieved, and more importantly, what still needs to be com-
pleted or refined. Portfolios carry the potential for high-level self-assessment of a student’s
progress, and provide clarity to an instructor as to whether continued or modified instruc-
tion is needed in order to maintain and enhance student knowledge and skill improve-
ment. At this stage, students are expected to look into the “learning mirror” and provide
the most honest and direct accounting of their learning and their work as they can.
Finally, confirming the learning connections and outcomes is the critical endpoint
that must be provided in a portfolio assessment. Students should know and be able to
answer the question “What has been gained?” relative to their own learning, and the
teacher should have ample student-generated evidence among the collected artifacts to
confirm the answers that are provided. Whether it is to document student progression
along a skill continuum, provide evidence of actual skill proficiency, or identify one’s
functional status or level (e.g., emerging, proficient, advanced) within an instructional
module or unit, the purpose of the portfolio assessment must be clear and understood by
all participants.
Connections to larger learning issues and/or problems in a community or even the
larger world may be forged through this kind of learning and assessment activity. The
potential for extensive learning acquisition and integration is very high with a portfolio
assessment, but at the same time it requires a well-managed and guided instructional pro-
cess. In addition, it is important to make sure students are ready, capable, and inspired to
complete a portfolio project, since considerable time, effort, and resources are required in
order to complete an activity of this kind.
Project-Based Learning
Performance-based events and assessments are also usually connected to project-based
learning (PBL) activities. As an instructional approach, PBL attempts to address and solve
real-world issues through the use of active problem solving, which often incorporates
performance-based learning. This process requires critical student review (e.g., problem
FOCUS ON ASSESSMENT
800
700
600
500
400
300
200
100
0
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29
Consecutive Days
FIGURE 6.1 STUDENT CALORIE INTAKE CHART
identification, general problem solving, decision making, and overall investigative analy-
sis) and results in a publicly presented product or performance.
Students, working individually or in small research groups, serve as the primary inves-
tigators in identifying and addressing a relevant topic of concern or interest. Teachers
serve as facilitators, working to help refine the investigative questions and assisting in the
review process, as well as assessing student progress throughout the completion of the
project. The use of high-level thinking skills, along with an emphasis on creativity and
independent, autonomous functioning, are expected with PBL. The use of self-assessment
and peer reviews are also desired in this process. Projects are typically multidisciplinary in
scope and often rely on technology as part of the investigative process or when conducting
a research review.
For example, at a local high school, the final exam for approximately 80 physics and
engineering students was to design and construct a sustainable filtering system for fresh
drinking water to be used as part of a health clinic in a remote district of Tanzania, East
Africa. The submitted projects were reviewed by doctors, nurses, and engineers from the
nonprofit organization responsible for building the clinic and providing health services
to the people of the villages in that area. Student teams (ranging from two to four stu-
dents) presented their collected research, designs and models, and construction materials,
along with the results of their preliminary findings on how effective the systems were in
removing impurities from dirty water. Some designs were successful and some were not,
but the actual project engineer acknowledged that several interesting ideas were provided
by the students, many of which had not been considered by the organization prior to this
activity. This high-level “real-world” water filtration project is characteristic of the top-
ics that are addressed with PBL.
PBL as an instructional process has been connected to a wide range of grades and
domains (Foulger & Jimenez-Silva, 2007; Mitchell, Foulger, Wetzel, & Rathkey, 2009;
Polman, 2000) and provides genuine opportunities for students to engage in active learn-
ing and meaningful dialogue as they refine their problem-solving skills. As active seekers
of knowledge, they are responsible for the accuracy of the data they collect and the find-
ings that are generated from that data. But as with any instructional method, there are
issues that must be recognized with PBL.
For instance, there are no universally recognized standards for what constitutes a
“project.” In addition, considerable time and effort are needed for PBL to be correctly
implemented, and it must be carefully orchestrated and managed as a learning experi-
ence. A project-oriented learning event needs to have clear outcomes that involve the
examination of defined content and expected skills. Moreover, performance assessments
must be carefully designed and selected in order to collect needed student learning evi-
dence. Highly motivated, skilled, and task-oriented students are certainly required for
PBL activities.
With the creation of performance events comes the requirement of determining the
means by which they will be reviewed and evaluated. A student’s performance is usually
examined relative to a standardized set of steps or completed products through criteria
that are designed specifically for the review of that performance or product. Creating
the desired performance event is important. However, generating a clear assessment and
review of the performance event is just as important if accurate information on student
progress is to be obtained. One of the most prevalent and popular measurement tech-
niques used in the review of performance events involves the use of rubrics.
RUBRICS
Rubrics, as defined by Arter and McTighe (2001), are “scoring tools containing cri-
teria and a performance scale that allows us to define and describe the most important
components that comprise complex performances and products” (p. 8). The selected cri-
teria clarify the qualities that the work must possess in order to receive a certain score
or rating. In addition, if shared before or during the construction of a project, rubrics
provide each student with a clear and unmistakable image of the quality of the work
that is required. Rubrics are particularly useful in reviewing student work within the
classroom setting. With stated ratings, criteria, and work expectations, a rubric provides
every learner with a clear performance standard prior to the completion of a project. In
addition, once the work is started, it can provide useful feedback to a student throughout
the completion of the project.
Rubrics can be designed and constructed to serve as a general review for several work
products or activities; consequently, these serve as general rubrics. Conversely, rubrics
can also be created so that they are very selective or task specific and used for only one
particular activity or project. A teacher is likely to need both kinds of rubrics.
For instance, the advantage with general rubrics is that they can be used with differ-
ent activities and across a variety of settings and events. A specific rubric doesn’t have to
be developed for each separate activity. If the science teachers at your school had decided
that the use of the scientific method throughout the school year is a priority, then the use
of a general rubric could be used to evaluate students’ progress in following the scientific
method across a number of different science projects. General rubrics are particularly
economical in regards to development and construction work. A general rubric also pro-
vides assessment clarity and uniformity to the students in regards to the necessary execu-
tion of the skills or criteria that make up the rubric.
On the other hand, there are instances when only a focused rubric will work, given
the instructional need. When that is the case, a task-specific rubric must be designed to
meet that assessment need and provide detailed information to the learners on what is
required for the successful completion of the assignment.
No matter the breadth or focus of a designed measure, within each listed rubric, per-
formance criteria are identified so that different point values can be attributed to particu-
lar levels of performance. Rubrics are also designed to be either holistic or analytical in
their function. A holistic rubric involves a scoring mechanism where a single score or rating is
used to represent a student’s entire work (Table 6.2). The teacher or rater examines the entire
process or product, including all elements of that work, and then selects a single score or
rating that best represents the entire project (McMillan, 2007; Nitko, 2001). On the other
hand, an analytical rubric is designed to rate or score each identified criterion of the required
project (Table 6.3). These individual scores can also be combined to generate a total or
overall score (McMillan, 2007; Nitko, 2001). For instance, if there are four specific cri-
teria associated with an assignment, then four scores or ratings would be generated, with
each one rated based on its respective set of criteria, as well as a total composite score.
The analytical approach can provide more detailed “diagnostic” feedback to students
regarding their performance and the relative strengths and weaknesses of their work, as
opposed to the holistic approach.
Both holistic and analytical rubrics are useful and should be considered when assess-
ing classroom projects and performances. Due to how they are structured, they possess
particular strengths as well as limitations. As the instructor you need to decide which
approach best fits your assessment needs in regards to the intended purpose of the rubric,
and that should be accomplished before the project is initiated. Many factors (e.g., the
instructional nature of the activity, classroom resources, time restrictions) can inf luence
which rubric approach is chosen.
In general, a holistic rubric provides an overall indicator of performance with minimal
feedback. In contrast, an analytical rubric is designed to provide considerable feedback in
particular skill areas, depending on how the rubric is constructed. In fact, a criterion pro-
file can be generated that allows a student to examine their areas of relative strength as
well as those that require further refinement. However, this wealth of information comes
at a cost, and that cost is in the form of time. Analytical rubrics are typically more time
consuming in regards to construction than the holistic rubrics. In addition, considerably
more time is needed to review projects that are evaluated using an analytical rubric. Both
rubric types serve useful purposes, and it is up to you as the learning planner to decide
which works best with a given instructional project.
like and what are the construction steps? Before you invest time and effort into designing
a rubric, make sure you are clear about its purpose. Make sure it will provide you with
the information you truly need. Assuming that a rubric fits your assessment needs, it then
must possess certain qualities, such as:
♦ instruction-learning-assessment alignment;
♦ clear representation of criteria and performance levels across that criteria;
♦ distinct point values associated with each skill/performance level;
♦ meaningful performance descriptors.
Royalty-Free/Corbis
Now examine the criteria that you have generated with those that are already pro-
vided. Which ones match up best with the lesson outcomes? The criteria that are going
to be evaluated within the rubric have to be selected. Therefore, let’s agree that five cri-
teria are essential for the rubric and this project and that they must connect directly with
the objectives of the lessons. One of the five will be selected to serve as a working exam-
ple as the rubric is constructed. When identifying criteria for your rubrics, it is important
to match them directly to the expected standards and learning outcomes connected to
the instruction. Using standards-based wording and clarity helps ensure the instruction-
learning-assessment alignment exists.
Selected Criteria
1. Food preparation
2. ?
3. ?
4. ?
5. ?
Food preparation was selected since it relates directly to the first lesson objective of the
cooking unit, and there would be no disagreement that it should exist as an expected cri-
terion for the rubric. Once a criterion is selected, the performance levels must be estab-
lished, along with the key descriptors and identifiers associated with each performance
level. It is important to make sure that the performance levels are distinct from each
other so that clear response differences are evident. This is essential for rating purposes as
well as for justifying the point values that exist for each of the levels. The goal is to cre-
ate a rating mechanism wherein each student’s work is accurately identified according to
the listed performance level and descriptors within that respective level. If done well, it
will enhance the rating reliability of the instructor as he or she reviews each project and
accurately rates a student’s performance based on the listed criteria.
The number of performance levels can vary among different rubrics. However, it is
generally best to keep the levels to a minimum and not to exceed seven, as the ability to
differentiate between levels becomes more difficult as the number of levels increases. On
the other hand, a minimum of at least three levels should be considered when construct-
ing a rubric. The selection of the number of performance levels is completely under the
control of the educator who is creating the rubric. Based on what has been presented so
far, how many performance levels do you think should exist in the baking rubric? As you
provide your answer, make sure you provide a rationale for your decision.
Given the straightforward nature of the material and the cooking task to be demon-
strated, as well as the desire to ensure high accuracy in rating the cooking performances,
a three-performance-level system seems appropriate. Sometimes simple and direct is
the best method, and that approach will serve us well with this task. Assuming that
there is agreement that a three-performance-level system is best for the cooking rubric
that is being developed, the following labels will be used to identify each of the three
Level 2—Acceptable (only four descriptors are met, three of which must include clean
cooking space, fresh ingredients, and washed and cleaned ingredients)
• entire cooking space clean
• fresh ingredients (no food used beyond expiration date)
• all necessary ingredients washed and/or cleaned
• one additional criteria
Level 1—Unacceptable (only one or two of the required Level 2 descriptors are met)
Given that food preparation involves several discrete procedures, critical elements of
that practice have to be identified. As the instructor, you would determine which spe-
cific elements or descriptors (sometimes referred to as anchors) are required to be dem-
onstrated for each respective level. A list of required elements was constructed for Levels
2 and 3. Another approach could have been to collapse these items together and provide
more detailed narrative text (including a representative performance used to illustrate
each required criteria), but since a simple and straightforward approach was taken regard-
ing this assignment, the listing of essential actions and events seems to be most appropri-
ate with this rubric.
Now examine the four remaining criteria and determine how they would be pre-
sented along a three-performance-level system. Once the criteria and their respective
performance levels have been established and qualified with the descriptors, seeking out
professional feedback on the rubric from your colleagues is an important and necessary
step before it is used in reviewing each student’s work on their pot roast project. Also,
if possible, conducting a small pilot or initial test of the rubric would be valuable. As an
assessment measure, this rubric would require the observation and review of the entire
cooking process, from the pre-baking stage to the final process of tasting the roast.
Since this rubric contains performance levels across five identified criteria, an analytic
rubric exists and rating feedback can be provided across those five areas. In addition,
specific point values need to be associated with each criterion and performance level. If
all five criteria are seen as possessing equal value relative to the cooking process, then it
is reasonable that all of them should be seen as having equal importance, and no criterion
should be weighted more (or given more point value) than any other. Regarding the
performance levels, different point values need to exist for each level. Since three levels
exist with the cooking rubric, a simple matching approach of assigning points according
to level could be followed (e.g., Level 3 is worth 3 points, Level 2 is worth 2 points, and
Level 1 is worth 1 point). A different point scaling could also be used. For example, Level
3 is worth 5 points, Level 2 is worth 3 points, and Level 1 is worth 1 point. In this pro-
cess it is important to (1) determine the relative importance of each performance level,
(2) provide good logic for implementing that point system, and (3) make sure that the
point system is understood by the learners completing the project or activity. In regards
to assigning point values to different levels, whole numbers and not fractions should be
used. For measurement purposes, the use of an ordinal scale (e.g., recognized ratings
and rankings of 1, 2, or 3 and the points that correspond to those ratings only) does not
account for fractional usage and it would severely limit the rating format of the rubric, as
well as confound the findings and the potential decisions that would be generated from
the obtained ratings. It is always best to use the rating scale of a rubric as it is designed to
be used and not to modify it based on personal preference or convenience.
You have now gone through the essential process of developing a rubric. How does it
feel? Still need practice? Of course you do; everyone does. You will want to repeat this
process with lessons and tasks that are specific to your content area and interests. How-
ever, with repeated practice, your knowledge and experience base will grow along with
your general sense of confidence, and that’s a very good beginning.
CHECKLISTS
As useful and popular as rubrics have become, it’s important to understand that a rubric
is not always necessary with every classroom performance activity. Sometimes other
measures, such as a checklist, can better serve your assessment needs. Checklists involve
the listing of essential components of a behavior or procedure, and are not as robust in
regards to the collection of assessment data as the rubric. Checklists are typically used
when a teacher needs to review student work or when learning needs to be confirmed.
They serve an important but basic function of recording whether or not certain behaviors
are exhibited. Therefore, if during the instructional process you need to check whether
or not certain behaviors or conditions are present or a set of skills has been demonstrated,
a checklist is well suited for that assessment purpose.
When completing a checklist, the teacher provides a mark or check from a list of items
and indicates whether a certain behavior is demonstrated as part of the performance. A
simple recording process is typically followed, and no qualitative variations are noted
behaviors are noted on the checklist. In addition, checklists can also be connected to
specific observations or used to check whether specific steps have been followed during
a specific activity (e.g., a math equation or a science experiment). In general, checklists
provide a focused structure for the content that needs to be reviewed within an activity.
By the nature of its design, a checklist provides an array of separate tasks, steps, or
behaviors that are recognized as necessary by the creator of the list. Because in most
instances the author of the checklist is the teacher, checklists become important listings
of skills that must be acquired and demonstrated by students. They become recognizable
and tangible goals of immediate instructional targets because of their construction. Stu-
dents are able to direct their attention to acquiring the listed skills or behaviors. Remem-
ber, if as an instructor you can provide a list of the things you want your students to
know and be able to do, then do it. This is one of the most direct ways of showing your
students what they must know and be able to do during and after a lesson is covered.
CHAPTER REFLECTION
In the Chapter Activity at the beginning of this chapter, you were asked to identify two
performance events that you had to complete during the normal course of the day. A
rating of the execution of those performance events was also requested. As you ref lect on
your selected acts and their completion, what, if anything, is needed for you to maintain
that performance, and what else must be done in order for you to improve your func-
tional level relative to those performance events?
During the course of the day in the “real world,” most of us rarely get individualized
feedback or specific ratings in regards to various performance activities that we complete.
What we do receive are natural consequences based on the decisions or actions that are
taken. For example, if we conduct an effective and accurate computer search, we find
the information we are seeking. If we don’t, we do not find what we are looking for,
and frustration is the only thing obtained in the search. In other words, we pay the price,
and sometimes a heavy one, for not meeting and successfully addressing the challenges of
various performance events in our lives.
School and classroom environments are unique in that they can (and should) provide
learning opportunities for students where useful feedback on skills and performances
can be provided without immediate real-world consequences. Having students complete
performance events, including procedures that evaluate how well those events are com-
pleted, simply makes good instructional sense. Whether students are giving a presenta-
tion in class, presenting a portfolio of work, or following the required steps in solving a
problem, these are natural and expected learning skills that students must practice and
perform well both in school and eventually in their chosen area of work. Consequently,
as educators we must be able to accurately assess the learning progress connected with
these kinds of learning events.
CHAPTER REVIEW
Conclusion
Performance assessments are centered on student-generated work and products. This
work is typically presented to others and is evaluated according to predetermined stan-
dards or performance criteria that are known to the students before the project or prod-
uct is even started. The products or performances that are reviewed often take different
forms, such as oral presentations (e.g., speeches, debates, panel discussions, and mock
trials), skill demonstrations (e.g., recitals, athletic events, concerts, or gallery showings),
or classroom-based work (e.g., science projects/demonstrations, experiments, or inde-
pendent research projects).
Moreover, the work products or performances that are generated are often complex;
they can involve an array of skills and require extensive practice, as well as a high invest-
ment of effort and time on the part of the participant. Performance assessments in the
classroom are important and extremely useful for teachers and their students, since the
performance events demand the integration of both knowledge and skill requirements as
part of an actual or authentic demonstration.
teacher, the value and recognized need for assessing the level and quality of student work
as part of performance events must first be acknowledged. By design, performance events
often require sophisticated and integrated knowledge and skill sets (providing for the
assessment of higher-level skills and products), which typically involve considerable time,
effort, practice, and classroom resources. Therefore, “less is really more” when it comes
to performance assessment, so it is important for every teacher to carefully design the per-
formance assessments that will most effectively demonstrate the intended knowledge and
skill development and acquisition of their students. Also, given the potential educational
magnitude and impact of these authentic skill demonstrations, a performance assessment
can and perhaps should be considered an “assessment showcase” event, with additional
assessment activities supporting and/or supplementing this primary experience.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How can performance events/assessments be integrated within the instructional
format of your lessons in the classroom? What function should these student learn-
ing progress assessments serve?
2. How will you balance the time, effort, and resources that will be necessary to
provide performance assessments?
3. Why is instruction-learning-assessment alignment just as critical for performance
assessments as it is for other forms of assessment? What measurement techniques
will you use with your performance assessments? Why?
RESOURCES
Performance Assessment Links in Science (PALS) at www.pals.sri.com is a resource bank
of science performance assessment tasks identified with grade-level standards as well as
tasks associated with grade level bands.
Lesson Plan Central at lessonplancentral.com provides a wide array of useful rubric
resources, as well as other performance-related assessments and materials. Type
“rubrics” in the “Search Lesson Plans” field.
An extensive resource on rubric development can be found at www.rubrician.com. Sub-
mitted rubrics from teachers across all subject areas can be examined and reviewed.
RubiStar, at rubistar.4teachers.org, is a rubric-generating site that allows teachers to con-
struct their own rubrics across all basic content and specialty areas.
General assessment information, along with a selection of useful checklists for project-
based learning projects, can be found at pblchecklist.4teachers.org.
An extremely useful Web site for generating a portfolio assessment in a foreign language
classroom is provided by the National Capital Language Resource Center (NCLRC)
at www.nclrc.org. Information on basic components (e.g., validity, reliability, portfolio
contents, rubrics) for creating a portfolio assessment is provided. This site is relevant for
any content area, not just foreign language.
The Buck Institute for Education at www.bie.org is a comprehensive site providing
project-based learning resources for educators. Resources include a Tools section that
includes videos, toolkits, teacher feedback, related links, and so on, as well as research
and professional development opportunities (e.g., workshops on PBL).
Project Based Learning at pbl-online.org provides several project-based learning resources,
including a video library, research, Web links, and templates and information on how
to develop standards-focused projects.
REFERENCES
Arter, J., & McTighe, J. (2001). Scoring rubrics in the classroom: Using performance criteria for
assessing and improving student performance. Thousand Oaks, CA: Corwin Press, Inc.
Arter, J., & Spandel, V. (1992). Using portfolios of student work in instruction and assess-
ment. Educational Measurement: Issues and Practice, 11, 36–44.
Bhola, D. S., Impara, J. C., & Buckendahl, C. W. (2003). Aligning tests with states’ con-
tent standards: Methods and issues. Educational Measurement: Issues and Practice, 22(3),
21–29.
Bond, L. A. (1995). Unintended consequences of performance assessment: Issues of bias
and fairness. Educational Measurement: Issues and Practice, 14(4), 21–24.
Brennan, R. L., & Johnson, E. G. (1995). Generalizability of performance assessments.
Educational Measurement: Issues and Practice, 14(4), 9–12.
Butler, S. M., & McMunn, N. D. (2006). A teacher’s guide to classroom assessment: Under-
standing and using assessment to improve student learning. San Francisco: Jossey-Bass.
Cohen, L. G., & Spenciner, L. J. (2003). Assessment of children and youth with special needs
(2nd ed.). Boston: Allyn and Bacon.
Foulger, T. S., & Jimenez-Silva, M. (2007). Enhancing the writing development of Eng-
lish learners: Teacher perceptions of common technology in project-based learning.
Journal of Research on Childhood Education, 22(2), 109–124.
Green, B. F. (1995). Comparability of scores from performance assessments. Educational
Measurement: Issues and Practice, 14(4), 13–15.
Linn, R. L., Baker, E. L., & Dunbar, S. B. (1991). Complex, performance-based assess-
ment: Expectations and validation criteria. Educational Researcher, 20, 15–21.
Linn, R. L., & Miller, M. D. (2005). Measurement and evaluation in teaching (9th ed.). Upper
Saddle River, NJ: Merrill/Prentice Hall.
McMillan, J. H. (2007). Classroom assessment: Principles and practices for effective standards-
based instruction (4th ed.). Boston: Pearson.
Mitchell, S., Foulger, T. S., Wetzel, K., & Rathkey, C. (2009). The negotiated project
approach: Project-based learning without leaving the standards behind. Early Childhood
Education Journal, 36(4), 339–346.
Moon, T. R., & Callahan, C. M. (2001). Classroom performance assessment: What should
it look like in a standards-based classroom? NASSP Bulletin, 85(622), 48–58.
Moss, P. A. (1992). Shifting conceptions of validity in educational measurement: Implica-
tions for performance assessment. Review of Educational Research, 62(3), 229–258.
Nitko, A. J. (2001). Educational assessment of students (3rd ed.). Upper Saddle River, NJ:
Merrill.
Polman, J. L. (2000). Designing project-based science: Connecting learners through guided inquiry.
New York: Teachers College Press.
Stiggins, R. J., Arter, J. A., Chappuis, J., & Chappuis, S. (2004). Classroom assessment for
student learning: Doing it right—using it well. Portland, OR: Assessment Training Insti-
tute, Inc.
Valencia, S. W., & Calfee, R. (1991). The development and use of literacy portfolios for
students, classes, and teachers. Applied Measurement in Education, 4, 333–345.
Wiggins, G. (1989). A true test: Toward more authentic and equitable assessment. Phi
Delta Kappan, 70, 703–713.
Wiggins, G. (1992). Creating tests worth taking. Educational Leadership, 49(8), 26–33.
Wiggins, G. (1993). Assessing student performances. San Francisco: Jossey-Bass.
Wiggins, G. (1998). Educative assessment: Designing assessments to inform and improve student
performance. San Francisco: Jossey-Bass.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How can self-assessment practices be integrated into the daily instructional rou-
tine of the classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Define self-assessment and how it can serve your prospective students.
♦ Describe the educational purpose and function of self-assessment practices in the
classroom.
♦ Summarize self-assessment and
progress monitoring techniques
that students can utilize in the
classroom.
♦ Embed self-assessment activi-
ties into an instructional plan
or lesson.
♦ Acknowledge the importance of
each student serving as his or
her own progress manager.
♦ Construct a personal self-
assessment program that is
followed and reviewed after the
delivery of any instructional
Comstock/PictureQuest
lesson.
166
Bananastock/AGE Fotostock
Introduction
Self-assessment represents a relatively new skill acquisition that has demonstrated aca-
demic effectiveness and continued promise for students as well as teachers (Ross & Bruce,
2007; Ross, Hogaboam-Gray, & Rolheiser, 2002; Ross, Rolheiser, & Hogaboam-Gray,
1999). Stiggins, Arter, Chappuis, and Chappuis (2004) state that “Self-assessment is a
necessary part of learning, not an add-on that we do if we have the time or the ‘right’
students” (p. 44). As a process, self-assessment functions as an internal review system that
can include selecting and prioritizing individual learning goals or outcomes, monitoring
one’s progress toward those learning outcomes, and determining what individual adjust-
ments, if any, are needed throughout an instructional experience. As Earl (2003) has
stated, “Over time, students move forward in their learning when they can use personal
knowledge to construct meaning, have skills of self-monitoring to realize that they don’t
understand something, and have ways of deciding what to do next” (p. 25). She believes,
and appropriately so, that each individual student is the critical factor in any learning
event.
Earl further contends that:
Students, as active, engaged, and critical assessors, can make sense of information,
relate it to prior knowledge, and master the skills involved. This is the regula-
tory process in metacognition. It occurs when students personally monitor what
they are learning and use the feedback from this monitoring to make adjustments,
adaptations, and even major changes in what they understand. Assessment as
learning is the ultimate goal, where students are their own best assessors (p. 25).
Unfortunately, minimal instructional time and curriculum attention is directed to
self-assessment in today’s classrooms (Black & Wiliam, 1998). Yet at the same time, it
is common to find mission statements for schools or educational programs that include
descriptions of their students as “self-sustaining learners” or “lifelong learners.” If the
field of education and the general public truly want students to be lifelong learners, then
our children need to be taught, develop, and learn internal review and monitoring mech-
anisms that can be implemented during any learning experience. In addition to the joy
and excitement of learning new things, the necessity of being able to self-regulate and
monitor the effectiveness of one’s learning and overall progress is essential both inside
and outside the classroom.
The generation of a personal assessment and monitoring system doesn’t happen auto-
matically. It is important to recognize that self-assessment is a learned skill and behav-
ior set that most learners don’t acquire by chance (Boud, 2000; McDonald & Boud,
2003). Like other presented content and material, this skill must be taught, practiced,
and refined over time, and the primary setting where this instruction can take place is
the classroom setting. Only with time, effort, and repeated practice opportunities can
this assessment format become an effective learning tool and lifelong process for your
students.
CHAPTER ACTIVITY
Self-assessment starts with self-awareness. Students need to recognize that they are active
learners in the classroom and are capable of directing their own efforts toward any educa-
tional accomplishments. Part of that process is making sure students are aware of themselves
and their potential role in their own learning. Having students involved in the process of
generating their own personal scripts is a great way of getting them directly involved. For
example, a potential script for Mrs. Warner’s second-grade classroom could include the fol-
lowing statements:
• I am a student in Mrs. Warner’s second-grade class.
• I am learning how to multiply one- and two-digit numbers.
• I am learning the rules of multiplying and carrying numbers.
• I must practice in order to improve my calculations.
• I am responsible for my learning.
Imagine you are meeting with your students in your classroom. How would you present
this idea to your students? What statements would you want to include for the script with
your students and how would you encourage them to generate some of their own state-
ments for the script?
SELF-ASSESSMENT: CONNECTION
WITH THE TEACHING AND LEARNING
PROCESS
In Chapter 1, you were encouraged to ask three assessment questions when engaged in
providing an instructional activity or lesson:
1. Where are my students?
2. Where do my students need to be?
3. How do my students get there?
By virtue of addressing these questions, instructors are also forced to examine where
they are in that process. For example, starting with the first question, a teacher must
know where he or she is in regards to the preparation of the teaching event. Addressing
and answering these questions requires a teacher to enter into a process of self-assessment
as part of any instructional event. Consequently, self-assessment questions for the teacher
evolve into the following:
1. Where am I in preparation for the lesson?
2. Where do I want and/or need to direct my students?
3. How do I get my students there?
This ref lection is critical in the unveiling of the lesson that is provided, so that the learn-
ers acquire the outcomes that are expected to be obtained before the lesson is completed.
Moreover, as a teacher engages in self-assessment and progress review, it is important for
students to see that process demonstrated so that value can be identified with it (McDon-
ald & Boud, 2003). This is essential if students are to engage in self-assessment practice. If
they see it being utilized by significant models in their life (you will be one, so get used
to it), they will be more likely to acquire and consistently engage in self-assessment prac-
tices themselves (McDonald & Boud, 2003).
No matter the nature of the instructional journey, every learner wants and needs to
be informed of his or her educational destination. In addition, in order for anyone to be
able to self-assess learning progress, they must know and be able to identify the standard
against which that progress is being measured. Precise, understandable goals and out-
comes must be provided so that clarity can be maintained between the desired behavior
and the current or emerging behavior. This creates a needed reference point by which
individuals can then judge their work and their progress toward the ultimate product or
performance standards. That is why providing examples of excellent student work serves
as an important and authentic way to demonstrate the desired performance standard.
Regardless of the content or skills that are being taught, learners need to know exactly
where they are going in the lesson, and they need to know what they are supposed to be
able to do by the time the lesson is finished.
On the other end of the instructional process, learners must be able to determine their
relative status, including the strength of their prerequisite skills, prior to the initiation
of a learning event. Students benefit tremendously from a preview of what they will be
taught in an upcoming lesson. If clarity is provided on the expected learning outcomes,
students can better evaluate their readiness when starting a new lesson or unit (McTighe
& O’Connor, 2005). Through ref lective action they can ask the important question, “Do
I have the necessary skills and prior learning experience in order to get the most from
this instructional experience?” They can ask themselves if they are ready for this, and if
the answer is “no,” then a quick review on the student’s part can take place and instructor
assistance can be sought.
Beyond the recognition that self-assessment is a good thing for students, self-assessment
and progress monitoring is now a skill expectation for P–12 students. For example, as part
of the assessment standard for teachers in the state of Ohio, as well as many others, class-
room students are expected to be aware of and utilize appropriate self-assessment skills in
their own learning. In particular, Ohio teachers are required to meet seven educator stan-
dards that address the following domains (Ohio State Board of Education, 2007, p. 12):
SELF-ASSESSMENT TECHNIQUES
The value of self-assessment extends far beyond the educational arena. In particular, self-
assessment is an essential practice in the business/corporate leadership world, as companies
strive for continuous improvement to remain competitive in today’s demanding market-
place. As part of the self-assessment process, major businesses and corporations engage in
designed self-review steps, which typically involve activities that incorporate the following:
♦ setting specific targets for financial and quality improvement;
♦ measuring a company’s progress toward meeting identified targets;
♦ using information and data to provide institutional feedback and review;
♦ gauging progress against competition within an industry; and
♦ continuous review based on the market and new economic opportunities.
These steps are necessary and desirable within a market economy. Within education,
the ultimate goal of teaching is to produce lifelong, independent learners, and a major
part of that goal is to produce learners who have the capacity to assess their own progress
and make necessary learning adjustments. Therefore, in order for students to become
effective self-assessors, they must be given opportunities to learn and practice this skill,
and that falls within the domain of the classroom.
Self-assessment usually takes two general forms. The more common approach involves
the use of teacher-directed activities where students are required to check and exam-
ine their learning progress (McDonald & Boud, 2003). These activities are designed to
require students to review their work and determine what they have learned and what
areas are still unclear or confusing. The actual forms and activities will vary depending
on the developmental level of the students, but even the youngest students can begin to
review their own work (Towler & Broadfoot, 1992). Simple but useful self-assessment
activities can be embedded within the normal classroom routine. For example, consider
these easy-to-implement self-assessment activities for the classroom.
♦ Allow your students to review some of their assignments. They can score their own
work or the work of their classmates. Providing a checklist of the expected items,
answers, or required steps can help them recognize what is needed in order to per-
form at the highest levels (see the box “Math Scoring Checklist”).
♦ Provide scoring rubrics to the students as an easy way of having them understand
what criteria are expected as part of a project or product. Also, include them in the
selection of the criteria for the project to increase their understanding of the assign-
ment and their connection to it (see the box “Math Performance Task”).
Data Analysis
Name:
Date:
Score Description
4 All attempted estimations accurate. Uses appropriate mathematical operations
with no mistakes. Draws logical conclusions supported by graphs. All explana-
tions are consistent with provided data.
3 Makes good estimations. Uses appropriate mathematical operations with few
mistakes. Draws logical conclusions supported by graphs. Most explanations
are consistent with provided data.
2 Attempts estimations, although many inaccurate. Uses inappropriate math-
ematical operations and evidences consistent mistakes in work. Draws no logi-
cal conclusions supported by graphs. Offers explanations but not related to
provided data.
1 All attempted estimations inaccurate. Uses inappropriate mathematical opera-
tions. Draws no conclusions related to graph. Offers no explanations of thinking.
0 No response/task not attempted.
SOURCE: Math Performance Scoring Rubric. Provided by Dr. Craig Mertler, University of West
Georgia. Reprinted with permission.
♦ Provide examples of the products or performances the students are trying to com-
plete. Use past student work, preferably before the project is started, so that your
students can see how it met the expected criteria (Figure 7.1).
Performance Rubric
3 points: Strong collective use of line, shape, and space
must be evidenced in the construction. Work must
demonstrate the following qualities:
• Outline and contour of object(s) well defined
• Emotional response evoked in work
• Accurate shadow and texture demonstrated
• Obvious three-dimensional quality to the work
• Appropriate use and/or balance of positive and
negative space
• Creative/unique integration of design and form
2 points: Adequate use of line, shape, and space with
only four of the six qualities demonstrated in the work.
1 point: Marginal use of line, shape, and space with no
more than two of the six qualities demonstrated in the
work.
(continued)
General results/findings:
♦ Hold whole class or small group discussions, along with scheduled student–teacher
interviews where certain products (e.g., pieces of writing, articles read) can be dis-
cussed. It really depends on the teacher and how he or she wants to incorporate
self-assessment training into the instruction.
♦ Use classroom exit slips, administered at the end of a class period, where students
anonymously respond to specific teacher-generated questions (e.g., “What did you
learn today?” “What are you still confused about?”). This is done in order to make
the students consciously aware of their present learning status, take stock of current
progress, and determine what must be continued or changed, if anything, to reach
the expected instructional outcomes. Exit slips can be administered as frequently as
every day if so desired.
Learning or progress checklists, involving the required actions that need to be completed for
a project, can also serve as learning guides and confirmation of instructional targets. Listing
assignments that must be finished, writing terms that must be recognized and defined,
completing the prescribed sequence of steps that need to be followed in order to solve a
problem or equation, or even locating specific battlefields on a map can provide a “connect-
the-dots” self-assessment that helps a learner determine whether or not the presented
material is really understood. For example, the checklist in Figure 7.2 enables the learner
to document the completion of various assignments and the corresponding knowledge/
skill sets they are connected to, as well as record personal comments about those assign-
ments. Progress checklists become arrays of essential items that need to be confirmed as
acquired and mastered; they serve as necessary “mile markers” throughout the completion
of a project or exercise. Through their use, students can quickly determine if they are still
on the expected instructional road and are heading in the right direction.
Whenever progress checklists are used, a note of caution is needed. It is not uncommon
for students, especially when they are not closely monitored or are first learning to self-
review, to just check off items or steps without really assessing their work or progress. It is
important to guide your students to ensure that appropriate time and attention is given to
the self-review process. Just as students who are caught running in the halls are requested
to come back and walk, providing a procedural structure and format that monitors the
students, yet allows them to methodically review their level of work (what still needs to be
done, what can be improved, etc.), is essential as part of the instructional process.
Based on how items on a checklist are structured, a review of general or specific
knowledge or skill sets can be conducted. For instance, imagine that your students are
reading a story, and they need to check their comprehension at various points through-
out. A self-assessment checklist could be created that requires each reader to ref lect upon
his or her comprehension by reviewing and answering questions that examine whether
specific strategies are being used (e.g., reading the summary or conclusion of the story
before it starts, completing plot and character predictions, rereading sections, or high-
lighting and notating significant events). Providing written summaries after each reading
that report the accurate sequence of events, main characters and their actions, and what
will likely happen in the forthcoming chapters could also be a part of the checklist.
Along with a comprehension check, basic information on general interest, motivation,
and progress satisfaction, as well as short- and long-term goals, could also be collected.
The design of the self-assessment activity should be individualized to the needs of the
instructional experience and what the instructor wants the students to acquire as part
of the experience (see the box “Self-Assessment Review of a Writing Portfolio”). With
repeated exposure to these kinds of activities, students can begin to generate their own
self-assessment activities, complete with the questions they want and need to answer dur-
ing an instructional activity.
This kind of instructional approach provides a planned opportunity for students to
check their immediate progress, as well as to acquire and refine a review habit that is inter-
nalized into a personal review process and system. This kind of activity can help “prime
the pump” and get students into the habit of checking their progress as they complete an
instructional activity. It forces a learner to critically ref lect on his or her current perfor-
mance level and determine what learning adjustments need to be made. The information
and insight that is collected through the ratings is usually perceptual data unless actual skill
analysis or determination activities are included within the self-assessment activity. Nev-
ertheless, this information can help provide a useful check of where the learner believes
he or she is relative to the provided instructional material. Moreover, the confirmation of
that awareness is critical for the learner to acknowledge throughout the acquisition of any
new material.
SELF-ASSESSMENT REVIEW
OF A WRITING PORTFOLIO
Pathways Home
In recent years, portfolios of student performance and products have gained support from
educators, who view them as a way to collect evidence of a student’s learning. For many
educators, portfolios are an attractive alternative to more traditional assessment methods.
The following instructions enable students to transform their reading notebook into a
portfolio in an attempt to make literature more challenging, integrated, and meaningful for
students. It engages the student in the investigation of a real-world problem by gathering and
sharing information, communicating ideas and fi ndings, refi ning perceptions, and creating
artifacts.
Portfolio Power
Turn your notebook into a portfolio by:
1. Arranging all your works of writing from most to least effective, including all evi-
dence of the writing process behind each fi nal draft.
2. Reflecting on your two best works, and on a separate sheet(s) of paper for each work,
answering the following questions.
• What makes this your best (second best) work?
• How did you go about writing it?
• What problems did you encounter?
• How did you solve them?
• What goals did you set for yourself?
• How did you go about accomplishing them?
Place this evaluation of your process and product in front of each fi nal draft when
completed.
3. Answer these two questions on a single sheet(s) of paper at the front of your portfolio.
• What makes your most effective work different from your least effective work?
• What are your goals for future writing?
4. Including an illustrated Cover or Title Page and a Table of Contents at the beginning
of your portfolio.
Your self-evaluation of your two best works will be as important to your grade as the teach-
er’s assessment of your fi nal two drafts and their writing process. The teacher will also con-
sider the number of works in your portfolio, the range of styles, and your progress as a writer.
When students understand the criteria for good work before they begin a literacy activ-
ity, they are more likely to meet those criteria. The key to this understanding is to make the
criteria clear. As students evaluate their work, you may want them to set up their own crite-
ria for good work. Help them with the clarity of their criteria as they assess their own work.
Students’ observations and reflections can also provide valuable feedback for refi ning
strategies they use, thinking about their responses to fi nd out what they are really learning,
and to see if they are learning what you are teaching them.
SOURCE: info@ncrel.org. Copyright © North Central Regional Educational Laboratory. All rights
reserved. Reprinted with permission.
For some classes (e.g., communications, math) or school teams (e.g., debate, sports), a
performance review is likely integrated within the self-assessment practice of a student’s
performance. For instance, when preparing for an upcoming debate or public speech,
students can use a skills checklist along with their last videotaped speech. The video-
tape is reviewed by a student and examined based on a list of skills that must be dem-
onstrated during any speech. A comparable approach is followed with athletes or band
members as they review game films or taped musical performances. Self-assessment
activities require students to look at their performances and determine their functional
status at that time, as well as what they need to continue to improve upon. For that rea-
son alone, this approach has become very popular and effective for a variety of learning
situations.
The script can also consist of specific questions that you want your students to ask after
a lesson has been presented or a learning activity has been completed (see the box “Basic
Learning Ref lection Questions”). However, responding to ref lection questions does require
a certain amount of uninterrupted time so that the students can respond to these questions.
Any set of questions (usually two or more) can constitute a script, and questions within
a script can vary depending on the learning focus and situation. For instance, consider
these possible questions for your own script.
© GeoStock/Photodisc/Getty Images
Beyond the specific self-assessment measures that are directly connected to a lesson,
the development and use of a written or verbal review script, complete with progress
questions, is highly recommended. As valuable as a script can be for students, it is impor-
tant to realize that the most important goal is to provide planned time in order for them
to examine their progress and acquire the knowledge and skills that they are missing or
have incomplete mastery of (Chappuis & Chappuis, 2008).
question. This question functions much like a bulls-eye found on an archery or shooting
range, in that it helps identify whether or not the learning “target” was hit and how close
to center the attempt was. This question can be addressed via an entry or exit slip and
can then be followed up with a quick daily journal entry, an oral response in class, or a
skill-check exercise that provides more objective evidence that, indeed, the material has
been learned.
In order to facilitate the internal review process, in-class activities can be structured
so that self-assessment review is produced as part of classroom work or projects. This
could involve peer-review activities, homework activities, or simple progress reviews
(e.g., classroom exit slips) that are completed. For instance, with exit slips, which are
completed before the students leave class, responses to learning review questions can be
checked for general comprehension and perceived competency.
Within the self-assessment process, basic components must be recognized, including
personal awareness, instructional knowledge, and learner control. Students must be able
to recognize and accept the responsibility that they are in control of reviewing their own
work, marking their actual progress, and using appropriate learning strategies, as well as
estimating what still needs to be done to reach the identified goals. Each student needs
to know that he or she is involved in a self-directed experience that is requiring them to
review and monitor their own learning progress.
In addition to awareness is the need to possess knowledge of learning strategies. Stu-
dents need to understand a variety of learning strategies and be able to connect them
with the right learning tasks. For example, memorizing information may work when
confronted with the introductory lesson on parts of the human cell, but when complet-
ing math problems, knowing and following operational steps to solve computation-based
problems will be required. Students need to have as many learning tools as possible so that
they can apply them to the various learning needs experienced in the classroom setting.
ASSESSMENT IN ACTION
More importantly, the questions require students to complete a quick self-check of their
own understanding of the presented material. Many teachers also include a statement that
has the students rate or respond to how effective the day’s activities or lessons were. This
important addition on the exit slip represents a small but meaningful attempt on the teach-
er’s part to self-assess the instructional impact of the lesson that was provided at that time.
If this information is collected on a regular basis, then a consistent source of student feed-
back on instruction over an extended period of time can be obtained. Getting this kind of
information and feedback is clearly important, since it is good to know how the instruction
is progressing in the estimation of those who matter the most. We must not forget the fact
that teachers are “educational performance artists,” and because of that, it is important to
perform well in order to reach that critical student audience.
EXIT SLIP
2. What was the most useful information or skill that you got from
class today? ________________________________________
__________________________________________________
__________________________________________________
__________________________________________________
One of the most important aspects of self-assessment is the idea that students have
control over their learning process. Internally-driven statements such as “Do I under-
stand what I just read?” or “Does this information connect with what I learned earlier?”
serve as important verbal prompts in the learning review process. Personal control serves
an important “office manager” function for a learner, whereby the learning actions are
organized, reviewed, managed, and implemented.
Along with the process of self-assessment, students need to understand and demonstrate
basic skills in monitoring their progress so they can demonstrate their knowledge. For
example, one simple but effective way to show progress is by collecting relevant data (e.g.,
Spanish vocabulary words read) and then presenting that evidence (e.g., colored columns or
pie charts that represent the percentage of number of words correctly defined and/or enun-
ciated) so accuracy can be demonstrated during one or more learning checks. Videotaping
a demonstration of a skill or providing independent ratings on a specific piece of work over
time could also serve as useful progress evidence. With self-assessment comes the need to
monitor and record current as well as future performances. The collected information pro-
vides the basis for adjustments that may be necessary as part of the learning event.
FOCUS ON ASSESSMENT
Name: Date:
1. What did I learn during the class activity?
2. What did I make in the class activity?
3. What do I need to practice?
4. What do I need from my teacher?
The student responses are collected and reviewed by Mrs. Oliver during recess. These
questions are designed to provide feedback to the individual students and the teacher, as
well as to facilitate and further develop the students’ writing and self-expression skills. This
is a quick and useful procedure that is easily embedded within the instructional delivery of
the class activity. It also can become a memorized mental script, with classroom support,
that can guide a student as he or she completes other assignments.
After reviewing the responses, Mrs. Oliver uses the collected information to ask follow-
up questions before the next class activity, or to provide additional teaching on the subject
if the student responses suggest that is needed. She considers this assessment process essen-
tial to the class activity format.
After reviewing the items on the report card, what grades did you give yourself? Were
strong grades listed across the board or did it depend on the item? If several low grades
were generated, then what does that mean, if anything, for you? In particular, low grades
can serve as personal focus points where more attention and examination is required in
these important life domains.
As you have probably already determined, there is no official mental health report
card. Nevertheless, as reported in Witte (2006), “the important message here is that
some standard(s) of mental health review and status is needed”(p. 157). The focus of
self-assessment is for each individual learner to recognize and set learning goals and out-
comes, review his or her learning progress, and provide the necessary learning adjust-
ments that may be needed in order to obtain the desired outcomes.
The questions in the preceding box serve as examples and are provided to help edu-
cators start to develop their own self-assessment process. The self-ref lective process is a
genuine attempt to objectively and systematically review what has transpired and then
identify what has worked, as well as recognize what still requires additional effort and
practice. This is an essential skill for anyone who desires to reach the highest levels of his
or her profession, and that includes teaching.
After reading this chapter you may still be asking yourself if this self-assessment
instruction is really necessary. Quite honestly, it’s one of the most important things you
can do for yourself and the students you teach. Involving yourself in professional self-
ref lection and review by addressing questions like these can only enhance your pedagogy
and effectiveness with your students. More importantly, in modeling this process and
providing orchestrated learning opportunities for your students to learn and practice self-
assessment strategies, they can then acquire self-review skills that can be used regardless
of the learning situation, content, or grade level.
CHAPTER REFLECTION
In the Chapter Activity, awareness of self-assessment was examined along with the devel-
opment of a basic student review script. Now consider the same activity for yourself. As a
teacher, what general or standard statements should be part of your own personal review
script? Consider the self-assessment questions that were just discussed as potential review
points for your script. Also, how would you go about monitoring and evaluating your
self-assessment practices? Provide a list of activities you would follow to ensure that you
are engaged in instructional review and that collected evidence is used in the continuous
improvement of your teaching.
CHAPTER REVIEW
Conclusion
Self-assessment is a learned skill, acquired and refined by way of instructional opportuni-
ties, that requires the learner to maintain a self-monitoring review of his or her learning
progress, as well as determine what individual adjustments, if any, are needed throughout
the instructional experience. Progress monitoring is necessary because learners need to
possess accurate performance checks of their progress relative to the learning outcomes
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. Describe the role and function self-assessment will have in your classroom.
2. Explain how you would construct and insert self-assessment activities into your
daily or weekly lessons.
3. Explain how you would help your students develop and follow a self-assessment
script as they encounter new learning experiences in your classroom.
RESOURCES
An informative self-assessment reference is Dr. Lorna M. Earl’s book Assessment as Learn-
ing: Using Classroom Assessment to Maximize Student Learning (2003), published by Cor-
win Press, Inc. This is a book that every teacher should own or have access to as part
of his or her professional library.
The Iowa Department of Education Web site at www.iowa.gov/educate provides useful
information on many forms of assessment, and it also provides direction on how stu-
dents can use assessment as part of the learning process. A number of specific programs
within the state that integrate student assessment practices within the curriculum are
listed and described. Click on “Student Assessment.”
Learning Point Associates at learningpt.org, which is associated with the North Central
Regional Educational Laboratory, provides a wealth of information for teachers on
various topics, including legislation, current practice issues, and testing and assess-
ment topics, as well as more specific resources for classroom assessment. Typing key
descriptors (e.g., “self-assessment,” “classroom assessment”) into the search box gener-
ates extensive lists of documents and potential classroom applications.
Edutopia (What Works in Education—The George Lucas Educational Foundation) at
www.edutopia.org provides a variety of assessment resources (e.g., articles, videos,
professional blogs). Typing “self-assessment” into the search box provides an extensive
list of useful articles, blog topics, and teacher modules on this subject.
REFERENCES
Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Educa-
tion, 5(1), 7–74.
Boud, D. (2000). Sustainable assessment: Rethinking assessment for the learning society.
Studies in Continuing Education, 22(2), 151–167.
Chappuis, S., & Chappuis, J. (2008). The best value in formative assessment. Educational
Leadership, 65(4), 14–19.
Earl, L. (2003). Assessment as learning: Using classroom assessment to maximize student learning.
Thousand Oaks, CA: Corwin Press, Inc.
McDonald, B., & Boud, D. (2003). The impact of self-assessment on achievement: The
effects of self-assessment training on performance in external examinations. Assessment
in Education, 10(2), 209–220.
McTighe, J., & O’Connor, K. (2005). Seven practices for effective learning. Educational
Leadership, 63(3), 10–17.
Ohio State Board of Education. (2007). Standards for Ohio educators. Retrieved from http://
www.ode.state.oh.us/GD/Templates/Pages/ODE/ODEDetail.aspx?page=3&TopicR
elationID=1269&ContentID=8561&Content=90022
Pintrich, P. R., & De Groot, E. V. (1990). Motivational and self-regulated learning compo-
nents of classroom academic performance. Journal of Educational Psychology, 82(1), 33–40.
Ross, J. A., & Bruce, C. D. (2007). Teacher self-assessment: A mechanism for facilitating
professional growth. Teaching and Teacher Education, 23, 146–159.
Ross, J. A., Hogaboam-Gray, A., & Rolheiser, C. (2002). Student self-evaluation in grade
5–6 mathematics: Effects on problem solving achievement. Educational Assessment, 8(1),
43–58.
Ross, J. A., Rolheiser, C., & Hogaboam-Gray, A. (1999). Effects of self-evaluation train-
ing on narrative writing. Assessing Writing, 6(1), 107–132.
Stiggins, R. J., Arter, J. A., Chappuis, J., & Chappuis, S. (2004). Classroom assessment for
student learning: Doing it right—using it well. Portland, OR: Assessment Training Insti-
tute, Inc.
Towler, L., & Broadfoot, P. (1992). Self-assessment in the primary school. Educational
Review, 44(2), 137–151.
Witte, R. (2006). Mental health management for educators and the children they teach.
In A. Wang (Ed.), Selected topics on educational psychology and school psychology (pp. 115–
190). Dalian, China: Liaoning Normal University Press.
SUMMATIVE ASSESSMENT:
CONFIRMING STUDENT
ACHIEVEMENT
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How can I generate the most effective summative assessments in my classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Define summative assessment and its fundamental purpose for both students
and teachers.
♦ Acknowledge that how an assessment measure is used defines its purpose and
function in the classroom.
♦ Identify five guiding principles that need to be followed when constructing a
summative measure in the classroom.
♦ Distinguish appropriate test construction rules involved in test development,
implementation, and test completion review.
♦ Generate a summative test assessment incorporating different test item formats
designed to measure students' knowledge and cognitive processing skills.
♦ Apply effective assessment practices when constructing selected response and
constructed response items.
♦ Create and utilize a table of specifications to document content, knowledge, and
skill coverage when using a test as a summative assessment measure.
♦ Describe the use and benefit of student products as summative measures in
evaluating student learning and progress.
♦ Comprehend the value of reviewing summative assessment performances with
students.
191
Introduction
Summative assessment is the use of any measure whose primary role and purpose is to
confirm and formally recognize student learning accomplishments (Gronlund, 2006).
These assessments, whether they are performance checklists, classroom projects, or tests,
serve a critical role in the assessment of learning, where every student’s accumulated
progress is measured and documented. This is a necessary function that must be met.
Summative assessment should not overshadow the other assessment forms, but unfortu-
nately it often does. That typically happens due to an inf lated importance that is placed
on it by the user (e.g., educators, administrators, legislators, the general public). If done
well, the learning effects from all assessment avenues, including formative and self-
assessment, enable students to perform at the highest levels possible, which is hopefully
demonstrated on official summative measures. Due to several issues, including time limi-
tations, ease of administration and content coverage, and other related factors, tests and
comprehensive exams are often the preferred measure in the formal assessment of student
work in the classroom.
CHAPTER ACTIVITY
There are few situations associated with as much anxiety, dread, and unpleasantness as
taking a test. We all have personal feelings and memories associated with taking tests in
school. If you asked a random group of students at any grade level what their least favorite
thing to do in school is, the most likely answer would be “taking a test.”
The reality is that it doesn’t have to be that way. We as educators often maintain this
educational perception. There is nothing inherent in the act of responding to and complet-
ing a test that makes it universally unpleasant. The attitudes and expectations of educators
and students toward testing have more to do with how we prepare (or more often don’t
prepare), present, and implement a testing experience.
For this activity, a quick check of your testing memories is needed in order to identify
common elements or characteristics that made one testing experience “good” and another
one “bad.” What specific actions, procedures, or experiences produced a positive experi-
ence, compared to a negative, intolerable memory? Although the majority of your testing
experiences might be negative, you probably have a few positive testing memories. Write
down your recollections and examine the qualities that contributed to making it a positive
or negative experience. If possible, compare your responses with those of your classmates.
Are there common themes and experiences? If so, what are they? Do the collective findings
make sense? Are there any surprises? Why or why not?
As you review your list, examine the items for both positive and negative experiences
and determine which were directly controlled by the teacher (e.g., test review and practice)
and which were under your control as the student (e.g., amount of time spent studying for
the exam, personal attitude and motivation). In retrospect and now through the lens of a
professional educator, do you believe your teachers did an effective job of preparing you
and your classmates for the tests that you encountered? Why or why not? Also, as a student,
did you do an effective job of preparing for tests in the classroom?
SUMMATIVE ASSESSMENT
CONSTRUCTION
Every educator wants to be effective in his or her teaching, and that also applies to the
use of summative assessments. Unfortunately, there is not a quality meter that can be
scanned over a test or other summative measure to indicate whether it is good or average
or just plain bad. Following sound assessment practices, however, will increase the likeli-
hood that a measure will be able to do what it is intended to do, and that is to accurately
measure the learning progress and achievement of students based on the lesson outcomes
and instructional material that is covered.
As you begin to investigate the material in this chapter, it is important to recog-
nize and understand that technically any selected measure (e.g., rubric, project, portfolio,
experiment, test) can be used as a summative or formal evaluative measure of student
work. As mentioned previously, the nature and function of an assessment is determined
by its intended purpose. If a rubric is designed to review a final project and that project
serves as a formal evaluation of a student’s learning, then that rubric functions as a sum-
mative assessment.
On the other hand, if a test (which is usually viewed and used as a summative assess-
ment) is administered to provide formative feedback to students, then it is by design not
a summative measure, and the student responses are not evaluated or graded. An assess-
ment measure can serve a variety of purposes; it is dependent upon how the teacher iden-
tifies the purpose and function of the assessment. The rubric in Table 8.1 can certainly
function as a summative assessment; it just has to be used in that fashion by the teacher.
For example, after reading Animal Farm by George Orwell in class, each eighth-grade
student is required to produce several written reports, one of which is a three-page,
double-spaced report on the main characters in that book. The reports are judged on the
established rubric, and points are connected to each of the ratings (rating of 4 = 10 points,
rating of 3 = 8 points, rating of 2 = 6 points, and rating of 1 = 4 points). These points,
along with points accumulated from the other required reports, are added together to
produce a grade that is used to describe the quality of each student-produced work.
With the understanding that any assessment measure can serve different assessment
purposes and functions, tests, to a considerable degree, are used as formal summative
measures in most classroom settings. But what exactly is a classroom test and what is its
function? In regards to a working definition, a classroom test consists of a collection of
items that are designed to measure and evaluate a learner’s skills, knowledge, general performance,
or other capabilities as they relate to the content and lesson objectives that are covered as part of the
instructional process. A classroom test typically serves as a formal assessment of your stu-
dents, and this work is usually graded and evaluated through some standard procedures
within the school district. The selection of the test items and the creation of the test itself
are typically completed by the teacher; however, preconstructed subject and skill tests
and exams are also commercially available and used in the classroom. These tests may
include chapter reading tests as part of a reading program or tests provided with specific
textbooks that cover material within specific chapters.
Commercial package programs that accompany textbooks or certain academic series
programs (e.g., reading and math) can be very appealing and useful, as content and chap-
ter tests are often provided as part of the programs from the publisher. However, it is
extremely important to make sure the test items align with the content that is actually
taught and the instructional objectives that are covered. If that alignment doesn’t exist,
clearly students would not be tested on the material and skills that were provided in their
lessons in the classroom. The teacher is ultimately responsible for ensuring that instruction-
learning-assessment alignment exists no matter what form an assessment might take or
whether measurement material is personally constructed or borrowed from a commercial
provider.
Generating effective summative assessments is essential, and teachers are responsible
for the construction of assessments that will be used to directly measure the learning gains
of their students. As you begin to develop your knowledge and experience constructing
summative assessments, it is important to identify the rules, guidelines, and operating
procedures that you will follow. As you begin this professional journey, consider some
basic rules that will help guide the construction of your summative assessment foundation.
FOCUS ON ASSESSMENT
teaching–learning experience. Also, the measurement of student progress has been treated
as a secondary subject matter that teachers should probably know something about, but has
not been stressed as central to effective teaching. Some teachers even believe that assessment
of student progress actually interferes with the design and delivery of a lesson and its impact
on students. Others contend that assessment is not the responsibility of teachers. More rea-
sons could be provided, but the reality is that these perceptions are inaccurate and outdated
and hold no value for any teacher who is teaching in today’s classrooms.
The practical truth, supported by an ever-growing research base, is that assessment
maintains a critical role in the implementation of effective instruction and the documenta-
tion of student learning progress (Bangert-Drowns, Kulik, Kulik, & Morgan, 1991; Black
& Wiliam, 1998a; Bloom, 1984; Fuchs & Fuchs, 1986), particularly in the area of reading
(Fuchs, 1998; Fuchs, Deno, & Mirkin, 1984; Fuchs, Fuchs, Hamlett, & Ferguson, 1992;
Fuchs, Fuchs, Hamlett, & Stecker, 1991; Fuchs, Hamlett, Stecker, & Ferguson, 1988).
Without a well-designed evidence-based assessment system in the classroom, you have only
a teacher performing in front of a group; there is no verification that students are really
learning what they need to learn. Consequently, assessment mechanisms and procedures
must be in place if a systematic and reliable approach in the monitoring of student progress
is to be established.
As teachers, we want to believe that we are good at what we do and that students benefit
from their learning activities. However, the condition of being a good teacher and positive
student learning outcomes are not necessarily related. As Airasian and Russell (2008) point
out, there is a critical difference between good teaching and effective teaching. A teacher
can demonstrate essential qualities and practices (e.g., participates in instructional prepara-
tion, evidences a caring demeanor toward students, maintains strong organization skills,
knowledgeable in content area) that are recognized as good teaching. Those qualities are
desired, but that shouldn’t be the ultimate goal. The goal is to be an effective teacher, which
is someone who can demonstrate and document that students are truly acquiring the skills
that they are expected to learn (as evidenced through the collection of student data). Since
the confirmation of student learning is most often demonstrated through the use of summa-
tive activities and performances, every teacher must be skilled in the selection or construc-
tion of summative assessment measures.
© Getty Images
they are supposed to know and be able to demonstrate as part of their testing per-
formance. You need to make sure the items or tasks, whatever form they take
(e.g., essay, multiple choice, rubrics, checklist), match up with the intended objec-
tives of the lessons or chapters that are covered. Whether test items are generated
by you or through a commercial testing bank, you must ensure that the response
items, whether selected or constructed in design, clearly align with the instruc-
tional objectives of the material that has been covered.
2. The skill and performance levels (i.e., Bloom’s Revised Taxonomy) that will be
measured with the summative assessment need to be identified before the assess-
ment is administered. Both the knowledge levels and cognitive processes need to
be identified, and as the instructor you must ensure that the material has been
taught and practiced at these levels as part of the daily classroom instruction. For
example, if a test on some introductory material is constructed, it is likely that
the more basic skill levels (e.g., remembering, understanding, application) will be
emphasized in the test items, compared to higher-order cognitive processing skills,
and that may be desired and appropriate. However, higher-level skills can and
should be measured with some items, usually essay questions within a test format,
so that an appropriate skill performance balance is demanded from the students.
Also, students need to be familiar with and practice using the measurement forms
that appear in the summative measures (e.g., responding to essay questions prior to
a formal test).
♦ Avoid unfamiliar item formats that the learners have not seen or answered before.
Learners should be familiar with every type of item used within the test they are
given. In order to help gain test item familiarity, use similar test items in practice
tests, formative quizzes, or other instructional activities prior to the administration
of a summative test.
♦ After the instruction is completed, but prior to the administration of the test, gen-
erate a test review that provides an accurate and complete review of the test mate-
rial to ensure your students are fully prepared for the examination. The value of
the test (i.e., how many points it is worth) and the point values of the various test
items should be provided (e.g., 1 point for true-false items; 2 points for multiple-
choice items).
Test construction:
♦ All test items should be clear and concise and excess words or phrases avoided.
Make sure each test item question ref lects one complete idea and one response
from the student. Each test item must be fully understandable and independent
from every other test item.
♦ A variety of test item types (e.g., multiple choice, essay, matching, true-false) should
be considered and incorporated within a test, if at all possible. This helps eliminate
any construction bias in favor of one type of test response item. For example, if
most of the test items are essay questions, that would favor those students who have
confidence with this type of question or possess strong writing skills versus those
who do not.
♦ Clear and understandable directions need to be provided at the beginning of the
test.
♦ Questions should be constructed so they ref lect a singular statement or topic and
should not be designed to mislead the reader. Test questions need to accurately
determine if someone knows specific information or can perform a certain skill
set. Test questions should not be constructed with confounding words or misdi-
rections. Intentionally misleading or “trick” questions only serve to increase the
anxiety and insecurities of the students and serve no valid measurement purpose.
♦ If selected response items are used (e.g., multiple choice), make sure to provide
appropriate responses or alternatives for each question. For example, if a ques-
tion requires the students to identify the largest planet in the solar system, provide
names of real planets and do not provide responses that are not related to the sub-
ject matter.
♦ Provide the number of questions that can be completed based on the amount of
time that is available to the student. If a majority of selected response questions are
being used, less time would be required, compared to a test that possesses a consid-
erable number of essay questions. Typically, test items require at least one to two
minutes for students to read, understand, and formulate an answer.
♦ If constructed response items such as short answer or essay questions have higher point
values relative to other test items, indicate that value next to each of the questions.
ASSESSMENT IN ACTION
Test implementation:
♦ Administer the test within one period or session if at all possible, as this will help to
ensure fair and uniform examination conditions. If more time is required, it is best
if additional time is provided immediately after the class period and not on a differ-
ent day.
♦ As was already mentioned, provide adequate time to complete the test and be avail-
able to provide any clarification of an item that may be requested by a student.
♦ Maintain a quiet classroom environment that is free of distractions so all students
can focus their attention on the demands of the test.
♦ Complete an item analysis of the test, where each test item is examined relative
to the number of students who passed the item and the scoring or rating frequen-
cies that were evidenced for essay questions, and record your impression of stu-
dent responses and their general performance. Obtaining student impressions and
comments relative to test demands can be extremely valuable in the continued
improvement of a test.
♦ Learn from and improve upon test construction after each administration. Identify
what worked well and, more importantly, what did not. This review is invaluable
in the test construction process and for future examinations.
♦ Remember that well-stated test questions connected with clear directions contrib-
ute greatly to the validity of the test results and decisions that are made based on
those results.
Along with the good test construction practices listed above, educators must follow
sound test item construction practices. This involves the details of how to put together
effective test questions that require students to generate the desired knowledge or skills that
they are expected to learn as part of their instruction. Unfortunately, if tests are made up of
poorly constructed and unclear test items, students do not have a reasonable chance to show
what they have learned. In addition, teachers cannot make good educational decisions rela-
tive to their students because of ineffective or invalid test data. This condition must be
avoided; otherwise summative assessment is meaningless for both you and your students.
A variety of these measures are typically used based on the skills being developed as well
as the teacher’s knowledge, experience, and comfort with classroom assessment measures.
Student-generated products are highly desirable and necessary to document student
learning progress, but it is important to select the items that most effectively demonstrate
that they have acquired the knowledge and skills of that lesson. In essence, the teacher
needs to select the best assessment and evidence collecting measures for the job. In order
for teachers to do this, they must first become familiar with the various methods that are
used to document student work. Let’s start with well-recognized selected response items
that are commonly used in documenting learning progress.
Both approaches can be used to help measure your students’ knowledge and skills.
However, there are specific issues associated with multiple-choice items that must be
reviewed. To start with, the aforementioned questions contain only one correct answer
in the response sets, and most multiple-choice questions are designed that way. A best-
answer approach can also be used, where theoretically all the alternatives can be correct
to some extent but one answer is the strongest response based on the test question. For
example, consider the following item:
3. Which of the following is the best predictor of future student achievement in
college?
a. ACT score
b. SAT score
c. parental education level
d. past academic performance
e. extracurricular involvement
All of the listed answers are recognized factors that are connected to predicting school
achievement; however, the strongest predictor of current and future academic perfor-
mance is past academic performance.
Multiple-choice items are popular and effective, and for good reason. They are rela-
tively fast and easy to score, and a moderate number of items can potentially cover a large
amount of information and content. One of the greatest advantages is that these items
can be constructed to evaluate a wide range of cognitive processes (e.g., retention and
knowledge-based responses to questions that involve application and analysis skills). On
the down side, since the answer is provided in the set of options, it is possible for stu-
dents to guess, without any actual knowledge or acquired skill, and correctly identify the
answer from the listed responses.
♦ Make sure the answer responses make grammatical sense when paired with the
question stem so that all alternatives present logical responses.
♦ Avoid the use of the articles “a” or “an” at the end of the stem, as these can provide
unwarranted clues as to the correct answer based solely on grammatical accuracy.
♦ When constructing responses or alternatives, make sure all the responses are possi-
ble and are connected to the content. They need to be viewed as legitimate answers
to the question.
♦ Varying the word length of the responses is necessary, since in general longer
answers tend to be correct, and test-wise students often identify this trend. Pro-
viding correct answer responses that alternate between short, moderate, and long
word length can help minimize this tendency.
♦ General test construction wisdom dictates that certain alternative responses should
be carefully reviewed and in some cases avoided. For example, the response “all of
the above” is not an effective option, due to the fact that if more than one of the
presented responses can be determined to be true, the “all of the above” response
becomes the obvious answer. If present, this response option can artificially inf late
the passage rate of an item while having little to do with genuine knowledge on
the student’s part. On the other hand, the “none of the above” response can and
should be used when it is desirable in order to increase the difficulty level of the
item. This response requires the learner to examine and eliminate all remaining
item responses as being invalid, which may be necessary for the design of some
test items.
♦ Regarding the test item response pattern, it is important as the test constructor to
make sure that correct responses are randomly alternated in regards to the listed
position across the items that are used in a test. Each letter used in the response
should demonstrate approximately the same percentage of use across the entire test.
Also, never intentionally embed a specific pattern or try to spell out words with
the letters of the responses (e.g., 1. b, 2. a, 3. d). This provides potential clues to the
learner that should not exist.
♦ Make sure no item is dependent upon the answer of another item. Each item should
remain independent. The items can address a common topic or theme (e.g., the life
of Leonardo da Vinci), but the test items must remain distinct from each other.
not, or some other condition that can be judged as either correct or incorrect. True-false
items are easy to score, and due to their concise nature, a large number of items covering
a considerable amount of material can be embedded within most tests.
However, there are clear limitations with this test type. As a rule, true-false items
measure lower-level processing skills and are rarely constructed to extend beyond the
knowledge and factual recognition level. Little extensive information is required of the
learner beyond the general recognition that a statement is either true or false. Also, the role
of guessing is a huge confounding issue for these types of items. A respondent has a 50/50
chance of getting the item right merely by guessing the correct response.
Column A Column B
1. Alaska a. Albany
2. California b. Charleston
3. Kentucky c. Columbus
4. Minnesota d. Frankfort
5. New York e. Juneau
6. Ohio f. Knoxville
7. South Dakota g. Pierre
h. Sacramento
i. Springfield
j. St. Paul
be internally consistent. In particular, the primary item statements need to share a similar
attribute (e.g., all states), and all the provided answers must also be alike in regards to
some specific quality (e.g., all capitals).
Since matching items are easily embedded within a standard test design, they are typi-
cally used in classroom tests. In general, matching items are usually constructed to assess
basic knowledge or comprehension and by nature of their item structure cannot easily
measure a student’s higher reasoning or processing skills. In addition to being able to
cover a broad range of content (e.g., terms, definitions, significant events, related parts
or functions of a procedure or process), matching items are desired by teachers because of
their quick and easy scoring capabilities.
♦ Provide a limit of no more than 10 premises for each matching test exercise.
Beyond that number, the visual connections and alignments become confusing and
increase extraneous errors.
♦ It is recommended to list more responses than premises in all matching items. In
this way, the respondent must be able to accurately identify each match as opposed
to the easier task of aligning an equal number of premises and responses.
♦ When listing items within a column, use some standard ordering method (e.g., list
items alphabetically) so that a consistent listing method is provided.
♦ Both columns must fit onto a single page of a test. Actually, for all test items it is
essential that no question be split so that part of the item is on one page and the
other section is listed on the next page of the test.
response items are designed to set particular perimeters around a student’s response to a question,
particularly as it relates to the focus and potential length of response. Response restrictions can be
generated through the use of specific guiding words (e.g., list, describe, define) within
the question or statement, as well as limits set on the actual length of the written response
(e.g., a few sentences to several paragraphs). In addition, if an objective and uniform
scoring method is in place, these items can be scored with a high level of reliability and
consistency. On the other hand, since these items can be structured to examine certain
issues, areas, and learner outcomes, little if any response variation can be demonstrated,
and that is where extended response items can be utilized.
The extended response allows a respondent maximum flexibility in constructing a writ-
ten response to a question. Typically, few restrictions are included (e.g., page limit) and the
respondent can execute “free thinking” in the generation of his or her response. The
huge benefit of this question format is the opportunity to examine each student’s high-
est levels of thinking and organization. If constructed well, extended response items can
examine analysis, evaluation, and creation skills of the student that are rarely tapped with
other test item formats. However, due to the large time and effort required for this type
of item, only a limited number of these items are usually included in a test. Time limits
alone restrict their usage, so these questions must be carefully selected within a summa-
tive assessment.
Essay questions are popular with teachers and will remain that way. Important high-
level learning evidence can be gathered with these items. Yet they require more time to
complete than other test item types, and that must be recognized when completing a test.
Essay questions tend to ask for more in-depth responses on specific issues or situations
and do not survey a broad review of learning or material that is covered in the classroom.
Also, designing an objective and reliable scoring system for the written responses is
essential. For example, the expected answer to the question must be provided prior to the
administration of the test. In addition, the specific elements (e.g., essential steps or levels
and their descriptions) of the answer must be recognized, along with the credit that is
given to each of the components of the answer. This provides for a standardized review
and a consistent evaluation of each response.
♦ In order to ensure that your students are familiar with responding to essay questions
and the higher-level responses that can go with them, provide in-class opportuni-
ties where essay questions are practiced as part of the normal instructional process.
♦ When scoring essay questions, review all student responses to the same question
before moving on to other questions. Also, when scoring the answer do not attempt
to identify the respondent, as this will help to control for potential instructor bias
and enhance objectivity in the scoring process.
TABLE OF SPECIFICATIONS
AND YOUR TEST
As you may recall from Chapter 4, a table of specifications was provided for a seventh-
grade algebra quiz. A table of specifications for a fourth-grade art test is presented in
Table 8.3. A table of specifications allows a teacher to identify test items with the knowl-
edge and cognitive levels of Bloom’s Taxonomy and the skill outcomes of a lesson or
instructional activity. This “testing map” provides a teacher with complete coverage of
the content and provides a clear image of what is being covered at each respective skill
level. With this information, the coverage and focus of a test can be reviewed for its
appropriateness and utility in meeting your immediate summative assessment needs in
the classroom.
The generated table creates an important visual confirmation guide to the test items or
tasks that have been designed to evaluate specific knowledge and cognitive skills. These
test items become the direct links to the listed instructional outcomes, and ultimately to
the general learning standards for those students. For classroom assessment to work, there
must be a continuous connection among the assessment measure and/or tasks, the lesson
outcomes, and the eventual global learning standards.
This process starts with selecting and aligning questions or tasks with identified learn-
ing objectives. In Table 8.3, since the art teacher decided to use a test to serve as the
summative assessment, test items must be generated to measure distinct knowledge and
cognitive skills for each of the listed lesson outcomes.
For example, test item #4 was generated in order to obtain written evidence to con-
firm a student’s comprehension and differentiated understanding of two essential art con-
cepts: positive and negative space. On the other hand, test item #6 is only designed to
measure recall of basic line formations used in visual construction. Test item #11 requires
a student to understand the concept of depth and also recognize and employ techniques
that are needed to create it. A higher level of skill application is demonstrated in this item
as opposed to the previously listed test items. Most importantly, each one of these test
items provides an important confirmation of learning at a particular level of knowledge
and cognitive processing for each of the listed lesson outcomes.
So as you approach the development of a test in your classroom, be sure to complete
this table before the test is administered. In this way you can be assured of what areas and
skills are being covered. These coverage areas and skill acquisitions will likely change
from test to test. For example, if you are introducing a subject area for the first time,
TABLE OF
SPECIFICATIONS BLOOM' S TAXONOMY
Lesson outcomes Remember Understand Apply Analyze Evaluate Create
Lesson Outcome 1— Q #1 Q #2 Q #3 Q #4 and 5
Identify and use (FK) (FK) (FK) (CK)
techniques and various
materials and mediums
when creating original
art products.
Lesson Outcome 2— Q #6 Q #7 Q #8 Q #7 and 8 Q #9
Defi ne and identify (FK) (FK) (CK) (CK) (PK)
primary elements found
in a landscape (e.g.,
horizon line, depth of
field, foreground).
Lesson Outcome 3— Q #10 Q #11 Q #12 Q #13 Q #14
Identify and utilize the (FK) (CK) (PK) (CK) (CK)
different techniques
for creating depth on a
two-dimensional surface
(e.g., relative size,
overlapping).
higher-level processing skills such as evaluation and creation may not be developed or
expected, and the focus of establishing the basic skills of understanding, comprehen-
sion, and application may be more appropriate. Also, this table can serve as an excellent
assessment reference if questions ever arise as to the nature of your testing approach and
design.
Test Items for the Knowledge/Cognitive Levels for the Listed Outcomes in
the Fourth-Grade Art Test
Item #4. What is the difference between positive and negative space?
Item #6. An example of a type of line that we studied in class is called a(n):
a. outline
b. contour line
c. inline
d. applied line
Item #11. The illusion of depth can be achieved in a two-dimensional piece of work
by using what?
a. perspective
b. line
c. shading
d. shape
can also record their daily journal entries into a voice recorder, so journal information
does not have to be written. However, audio entries are usually translated into a written
record at some point in time.
A common assessment method that involves the collection of written evidence is the
portfolio. With a portfolio, student-generated work is systematically collected and reviewed over
a certain period of time (e.g., a single semester or the entire academic year); pieces of work are placed
in the portfolio as evidence of their accomplishments and progress. Individual subjects and skill
areas can be emphasized or can be integrated together; however, teachers often follow a
singular content area, as evidenced in a writing portfolio. There are different purposes
for collecting and reviewing portfolios. For example, in a developmental or growth
portfolio, the progress of the student and his or her skills (e.g., as reflected in a writing portfolio)
are emphasized. In a best pieces portfolio, the emphasis is on the best constructions of the vari-
ous assignments that are required over the instructional period. Professional portfolios found in
art, photography, music, and other professions typically represent best pieces portfolios.
Portfolios have traditionally followed the paper product route. However, the electronic
portfolio approach is becoming more popular with both teachers and students.
In general, high-quality student work is produced in the review process found in
portfolios. However, beyond its recognized effectiveness and general popularity, this
type of classroom assessment usually requires considerable resources to maintain.
Pyruvic acid
NAD+
NADH CO2
Acetyl-CoA
NADH NAD
FADH2
NAD+ 2CO2 NADH
FAD
ATP
ADP Succinyl-CoA
FIGURE 8.1 CONCEPT MAP OF CITRIC ACID
CYCLE
Reprinted with permission of Kevin Hartzog.
ways, and that is why they are so often found in books and other learning materials.
Graphs are particularly helpful in representing and clarifying complex relationships or
organizing and packaging large amounts of information into visually meaningful units.
All of the aforementioned visual tools can serve as effective indices of student learning.
They integrate well with the traditional written production methods that are found in
the classroom setting.
Liquidlibrary/JupiterImages
function, as credible evidence must exist that documents the completion of schoolwork
and the evaluation of that work in identified courses of study. Beyond the official function
of summative assessments, there is an equally important need to provide useful feedback to
students in regards to their performance.
Some teachers may be reluctant to share results or to take time out of class to
examine test results either for individuals or for the class as a group. However, while
recognizing that instructional time is always limited, reviewing assessment perfor-
mances—especially summative assessments—is not a valueless activity. In fact, sum-
mative assessment results need to be viewed as a potential instructional opportunity
(Biggs, 1998; Black & Wiliam, 1998b). The process of reviewing summative perfor-
mance results with students in order to improve and enhance their future learning,
referred to as formative summative assessment, has received empirical support (Drouin,
2010; Wininger, 2005). Moreover, this positive effect has also been demonstrated with
peer review (Drouin, 2010).
Going over a student’s test performance, and in particular recognizing mistakes and
clarifying misunderstandings, can help provide a meaningful base for the next instruc-
tional period and for future summative assessment activities. This approach of building
on and correcting earlier misunderstandings provides for a more successful summative
performance, especially in content areas that require mastery of cumulative information
or skills. The simple truth is that we learn from our mistakes, so time must be taken to
examine those mistakes to ensure that they are not repeated in the future. That is exactly
what happens when you review test results or other summative assessment performances
with your students.
CHAPTER REFLECTION
At the beginning of the chapter, you were asked about your memories of testing. In addi-
tion, you were asked to examine whether your teachers did an effective job of preparing
students for tests and other summative measures. Now that you will soon be teaching in
your own classroom, how will you prepare your students? Will the summative measures
you design be effective and accurate relative to their intended purpose?
At this time, or in the near future, locate a test or quiz that you have recently con-
structed. As objectively as possible, review and critique its structure, function, and
general effectiveness in measuring student learning. If you have not created a test or
quiz to date, review an actual test or quiz that you have recently taken in one of your
classes. Identify the strong and weak points of the test. Examine how the test items
are written. Are they clear, direct, and effective in measuring the intended knowl-
edge or skills areas? Can the items be directly connected to skill sets and a knowledge
base in the intended instructional goals for the students? As your test review skills
improve, that same critical review can be placed on the construction of your own tests,
which will only enhance the effectiveness of the summative assessment process in your
classroom.
CHAPTER REVIEW
Conclusion
Creating and utilizing effective summative assessment measures is critical for teachers in
obtaining necessary information regarding the learning progress of their students. Sum-
mative assessment evidence is needed to show, as part of a formal review process, that
students have indeed acquired the intended learning outcomes of the lesson. The most
recognized summative assessment measure used by teachers is the classroom test. Com-
mon test item types (including multiple choice, true-false, matching, short answer, and
essay) used in summative tests were reviewed, along with key implementation guidelines.
However, other forms of student-constructed work and products can also serve as use-
ful summative assessments. Many of these measures were brief ly reviewed. Summative
assessment is a valuable source of student evidence, and teachers need to develop a solid
knowledge base regarding effective summative measures.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. Describe the steps you would follow when constructing a summative measure
(like a test) in your classroom.
2. Identify construction practices that need to be followed when generating various test
items (e.g., multiple choice, true-false, matching, short answer, essay) for your tests.
3. How should a table of specifications be used when aligning test items with the
knowledge and cognitive skill levels they are designed to measure?
4. How do you intend to use summative results to enhance the current and future
academic performances of your students?
RESOURCES
A useful reference for constructing effective classroom tests is W. James Popham’s 2003
book Test Better, Teach Better: The Instructional Role of Assessment, published by the Asso-
ciation for Supervision and Curriculum Development (ASCD). It provides informa-
tion on the importance of linking testing and teaching, along with procedures to fol-
low to ensure that meaningful classroom tests are constructed.
An interactive learning module entitled “Planning, Preparing, and Administering Class-
room Tests” that was prepared as part of the Alabama Professional Development
Modules is highly recommended for review. This module can be found at web.utk
.edu/~mccay/apdm/plan/plan_b.htm.
Identified as a self-study or part of a continuing professional development unit, the Class-
room Assessment at fcit.usf.edu/assessment/index.html provides basic test construction
suggestions and ideas.
REFERENCES
Airasian, P. W., & Russell, M. K. (2008). Classroom assessment: Concepts and applications (6th
ed.). Boston: McGraw-Hill.
Bangert-Drowns, R. L., Kulik, C. C., Kulik, J. A., & Morgan, M. T. (1991). The instructional
effect of feedback on test-like events. Review of Educational Research, 61(2), 213–238.
Biggs, J. (1998). Assessment and classroom learning: A role for summative assessment?
Assessment in Education: Principles, Policy & Practice, 5, 103–110.
Black, P., & Wiliam, D. (1998a). Assessment and classroom learning. Educational Assess-
ment: Principles, Policy and Practice, 5(1), 7–74.
Black, P., & Wiliam, D. (1998b). Inside the black box: Raising standards through class-
room assessment. Phi Delta Kappan, 80, 139–148.
Bloom, B. (1984). The search for methods of group instruction as effective as one-to-one
tutoring. Educational Leadership, 41(8), 4–17.
Drouin, M. A. (2010). Group-based formative summative assessment relates to improved
student performance and satisfaction. Teaching of Psychology, 37, 114–118.
Fuchs, L. S. (1998). Computer applications to address implementation difficulties associ-
ated with curriculum-based measurement. In M. R. Shinn (Ed.), Advanced applications
of curriculum-based measurement (pp. 89–112). New York: Guilford Press.
Fuchs, L. S., Deno, S. L., & Mirkin, P. (1984). The effects of frequent curriculum-based
measurement and evaluation on pedagogy, student achievement, and student awareness
of learning. American Educational Research Journal, 21, 449–460.
Fuchs, L. S., & Fuchs, D. (1986). Effects of systematic formative evaluation: A meta-
analysis. Exceptional Children, 53(3), 199–208.
Fuchs, L. S., Fuchs, D., Hamlett, C. L., & Ferguson, C. (1992). Effects of expert system
consultation within curriculum-based measurement using a reading maze task. Excep-
tional Children, 58(5), 436–450.
Fuchs, L. S., Fuchs, D., Hamlett, C. L., & Stecker, P. M. (1991). Effects of curriculum-
based measurement and consultation on teaching planning and student achievement in
mathematics operations. American Educational Research Journal, 28, 617–641.
Fuchs, L. S., Hamlett, C. L., Stecker, P. M., & Ferguson, C. (1988). Conducting curriculum-
based measurement with computerized data collection: Effects of efficiency and teacher
satisfaction. Journal of Special Education Technology, 9(2), 73–86.
Gronlund, N. E. (2006). Assessment of student achievement (8th ed.). Boston: Pearson Educa-
tion, Inc.
Popham, W. J. (2008). Classroom assessment: What teachers need to know (5th ed.). Boston:
Allyn and Bacon.
Wininger, S. R. (2005). Using your tests to teach: Formative summative assessment.
Teaching of Psychology, 32, 164–166.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
What is the most effective and equitable grading system I can use when evaluating
student progress and accomplishment in my classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Distinguish between the terms “grading” and “evaluation.”
♦ Identify the various purposes grades are used for in education today.
♦ Identify common grading methods, including letter, numerical, pass-fail,
standards-based, and narrative grading.
♦ Acknowledge the importance of achieve-
ment measures and other factors that
can constitute a student’s actual grade.
♦ Discriminate between the grading
criteria components of product, process,
and progress.
♦ Compare and contrast common grade
comparison models, including norm-
referenced, criterion-referenced, and
aptitude comparison or self-referenced
models.
♦ Distinguish among grading options for
students with special learning needs.
221
© Royalty-Free/CORBIS
Introduction
As student products are reviewed, the need for evaluation and grading of one’s accom-
plishments is required for official school purposes. Grading remains a consistent and con-
troversial issue in the field of education. Grading and evaluating student work is one of
the most significant and important duties that a teacher must execute, as its impact can
have significant long-term implications for students. Always remember that grading is
personal, so let’s start the investigation of this topic by answering the following questions:
♦ What are your most vivid memories of the grades you received in school?
♦ In regards to grades, do you in general have more positive or negative recollections?
♦ What specific emotions would emerge for you when your report cards went home?
If you were like most students, receiving grades was a strange combination of dread,
excitement, panic, relief, and perhaps a few other feelings all wrapped together. From
an academic as well as a psychological standpoint, grades have an impact on the lives of
people, and for some a lifelong effect can be reported. For that reason alone, it is impera-
tive that educators be as informed about grading and evaluation as possible. Everyone
has been affected by grades in some way, and that’s why such strong emotions are often
attached to those memories. Your students will be no different.
CHAPTER ACTIVITY
Before the issues surrounding grading and evaluation are reviewed, review your own grad-
ing experiences. In this way, the potential impact of those memories on your current views
and biases regarding this topic may be more clearly understood. This brief cathartic exer-
cise is only designed to increase your sensitivity to this issue and to engage you in personal
ref lections about it; grading epiphanies are highly unlikely, at least at this point in the
chapter.
Based on your own personal school experiences (elementary, high school, college, or all
combined), construct your personal responses to the following questions.
1. As a student, did you like getting grades on your work (e.g., tests, projects)?
2. What kinds of grades did you get? How did they affect you?
3. Did you think the grades you received were fair and accurate?
4. What did your grades tell you about your performances?
5. Did you receive any additional information (e.g., written comments) from your teach-
ers beyond a recorded grade? If so, was the information valuable?
6. Do you believe that the grades you received truly ref lected your understanding and
skill level of the material that was being evaluated?
7. In your opinion, what functions, if any, do grades perform for a student?
If possible, share your responses with your classmates. Compare the similarities and
differences of the experiences and determine if any core themes emerge. What do those
themes tell you about the grading process and its impact on your educational experience?
GRADING: A PROFESSIONAL
CHALLENGE
If you ask teachers to list their least favorable professional duties, grading always seems
to show up. This activity is viewed as one of the most unwanted aspects of the job, and
there are a lot of reasons for this viewpoint. To start with, the process of grading requires
constant attention and effort on the part of a teacher. Tests, quizzes, projects, and so on
have to be developed and scores must be recorded and tracked for each individual student
over the course of a semester or year. Just maintaining the accurate documentation of the
scores and entries within a grade book can be daunting. In addition, because the specific
value of a student’s work or performance is often determined, the stress and emotional
energy that is associated with this responsibility can take a heavy toll on any educator.
Also, the reality is that teachers are placed in the unenviable position of teaching
and supporting students in their learning while at the same time judging their work
and progress. Since grades constitute permanent performance records, those who are
affected, including students, their parents, and even teachers, can experience high anxi-
ety and stress. As is evident from these listed experiences alone, grading is a genuine and
ongoing professional challenge for educators. Also, given that academic standards and
instructional targets exist, the review and evaluation of student work and performance
relative to those expected levels is absolutely necessary. Consequently, the need for grad-
ing and evaluation will not diminish for teachers.
As part of this process, it is important to define and differentiate terms that are often
used when evaluating student work or performances. First of all, “grading” and “eval-
uation” of student work are sometimes used interchangeably in the classroom setting.
Grading is the formal evaluation of student work, which involves determining the value and qual-
ity of that work, based on some set of standards or performance criteria, and then providing some
formal recognition of that value to the learner, usually in the form of a grade, number, or mark
(Guskey & Bailey, 2001; Nitko & Brookhart, 2007). Evaluation is a necessary element of
any grading or review system. Evaluation involves the actual process of judging and determin-
ing the quality of work of a student (Popham, 2008). Determining whether some product or
performance is excellent or outstanding versus something that is satisfactory involves the
reviewer, in most cases the teacher, making that judgment. As already mentioned, this
judgment should not be arbitrary but based on some recognized set of criteria.
Within the classroom, grading typically ref lects teacher-directed procedures, where
the teacher completes a consistent and organized review of each student’s performance
and renders a critical analysis of a student’s achievement based on designated criteria. The
grading of student work is often connected to summative assessment measures that stu-
dents complete as part of their instruction. Consequently, these assessment products are
viewed as quite valuable in documenting student progress or competency.
Assessment and grading or evaluation are not synonymous and should not be viewed
as such. In particular, assessment, as emphasized in this text and defined in the first chap-
ter, involves the active and ongoing collection of student learning evidence that is used
to make effective instructional decisions. Assessment is the comprehensive and dynamic
process of reviewing and enhancing student learning, whereas grading and evaluation
serve a more circumscribed, but nonetheless important, role in judging and determining
the quality and worth of students’ products or performances in the classroom.
with percentage grading (a scale ranging from 0 to 100 that indicated the percentage of
material that the student had learned) considered a popular approach among educators in
the early part of the 20th century.
That system was later surpassed by the more popular categorical grade systems that
involved a three-point scale (excellent, average, and poor) and the contemporary five-
point letter grade system (excellent—A, good—B, average—C, poor—D, failing—F),
which most of the general public is familiar with today (Butler & McMunn, 2006). Con-
cerns over the inconsistency of grades created interest in a pass-fail system; however, that
was short-lived, as the standard letter grade system became strongly entrenched within
education in the later decades of the 20th century and has continued to the present day.
Consistent with the changes in grading systems over the years have been the con-
cerns and controversies surrounding the use of grades in the formal evaluation of student
learning (Guskey, 1994; Marzano, 2000). Interestingly, most of the concern has not been
centered on the importance of the formal evaluation of student work and progress. It
has centered more on the grading process; to be more precise, the inconsistency of grad-
ing practices and the resulting problems of ineffective student evaluation. Grading prob-
lems that center on the inconsistency in measurement processes and practices have been
known for decades, yet they remain unresolved even in today’s classrooms (Brookhart,
1991; Frary, Cross, & Weber, 1993; McMillan, Myran, & Workman, 2002). However, as
the accountability movement has spread throughout the educational field, there has been
renewed interest in addressing grading issues and the impact teacher evaluation has on
student learning and achievement. Given that grades are to remain a part of the formal
educational experience, it is essential that all educators become as knowledgeable as pos-
sible on this subject. That means that this chapter should only serve as an introduction to
a very complex and ever-evolving set of professional issues.
To formally start this introduction, the function of a grade is examined. By definition,
a grade constituted by a letter, symbol, or other performance marker is used to represent
© Royalty-Free/CORBIS
ASSESSMENT IN ACTION
the quality of a student’s work or performance. In addition, the term “mark” is generally
used to describe the collection or summation of a series of grades over a specified time
period, such as a quarter or semester. However, a grade or mark by its construction and
Letter Grades
The use of letter grades is the most common and well-recognized grading method, with the set
of capital letters A, B, C, D, and F serving as the primary markers. These letters are ordered
Numerical Grades
Another common grade reporting method is numerical grades, which involve the use of
number ranges or bands (with 100 set as the ceiling of the range) that are typically paired with letter
grade equivalents. For example, the common 10-point grading scale is shown in Table 9.1.
In addition, the eight-point scale (93–100 is an A, 85–92 is a B, 77–84 is a C, 70–76 is
a D, and 69 and below is an F) and nine-point scale (92–100 is an A, 83–91 is a B, 74–82
is a C, 65–73 is a D, and 64 and below is an F) are also used by school districts across the
country. Consistent with the letter grade system, the numerical grade system is familiar
to the general public, as it has a long history of use in schools and its popularity is second
only to letter grades. In addition, an overall composite number can be generated, usually
through the process of averaging all collected scores to obtain a final summative score,
which provides a summary mark of a student’s work over an extended period of time.
This number scale allows for more precision and discrimination of a student’s perfor-
mance compared to letter grades.
However, consistent with the letter grade system, the numerical grade system still
requires collapsing a huge amount of student information into a single number or score.
As with letter grades, it is hard to determine what a score (e.g., 85 versus 79 in American
History) really means. The score is still only a summary of the student’s performance
based on what was collected as evidence for that student over a specified time period.
Also, percentage cutoffs at the various levels are arbitrary, which can blur the accuracy
and purpose of the system.
As was mentioned earlier in the “Assessment in Action” box, School District A uses
an eight-point scale, while School District B uses a 10-point scale. So an English paper
that received a score of 91 is considered a B in School District A, but in School District B,
it’s an A. Try to explain that to the student who goes to School District A. As you might
imagine, these cutoffs and ranges can be difficult to justify.
Pass-Fail Grades
Pass-fail is a simple binary system that only acknowledges two possible grading outcomes: (1)
the passing of a course or program or (2) the failure to pass a course or program. This type of sys-
tem is not used extensively in the P–12 arena, but it still can be found in postsecondary
instruction, where it originated as an alternative to grade-based classes. Within certain
academic majors, some classes can be taken on a pass-fail or credit-no credit basis, where
the student is required only to meet the essential requirements of the class in order to pass
the course. Although effective in some settings, the limited amount of information that is
provided on a student’s performance stands out as its major grading limitation. Unfortu-
nately, students are often motivated to only do the minimal level required to pass a class
when this system is in place.
Standards-Based Grading
Due to the impact of academic standards-based instruction on P–12 education, school dis-
tricts across the country have been working to design grading and reporting procedures
that track the progress of students’ achievement relative to required academic standards
and outcomes. As reported in Guskey and Bailey (2001), there are four primary steps
that are followed when developing a standards-based grading system (the evaluation of
student progress relative to academic content standards identified at each grade level). They include
the following:
1. Identify the major learning goals or standards that students will be expected to
achieve at each grade level or in each course of study.
2. Establish performance indicators for the learning goals or standards.
3. Determine graduated levels of quality (benchmarks) for assessing each goal or
standard.
4. Develop reporting tools that communicate teachers’ judgments of students’ learn-
ing progress and culminating achievement in relation to the learning goals or
standards.
As part of the academic standards-based movement within P–12 education, all states
but Iowa have recognized grade-level learning goals and standards within the primary
content areas of instruction (e.g., math, reading, language arts, social studies, science).
This allows teachers to focus on performance or “evidence” indicators for those standards
as benchmarks or “learning mile markers” are reached that document the learning prog-
ress of the students. Reporting methods and progress reports attempt to record student
progress relative to identified standards as well as the specific student learning skills and
outcomes that are embedded within those standards (Figure 9.1).
With standards-based reporting methods, performance-level descriptors are used
to describe student progress relative to grade-level standards. Student progress can be
described in several ways, which can involve numerical ratings (e.g., 1—at grade level;
3—exceeds grade level), achievement descriptors (e.g., “emergent,” “proficient”), or even
behavioral explanations (e.g., “rarely evidenced,” “consistently evidenced”). This allows
the teacher to provide more detailed feedback to the students regarding their progress,
as well as to their parents. This is important, as many listed skills within standards may
require the entire academic year before they are acquired and mastered.
Standards-based grading can generate considerable information regarding a student’s
learning progress. It can provide a meaningful picture of a student’s achievement, as
well as provide clarity to skill areas that require more attention. From an instructional
perspective, this reporting format is much more informative than the traditional letter or
numerical grade systems.
This system does require considerable time and effort on the teacher’s part to orches-
trate, document, and evaluate student performances. This grading method is more com-
plicated and may not be well understood or received by parents. It may provide too
much information, especially for parents who are trying to find out “What grade did
my kid get?” Also, standards-based grading “fits” better at the elementary level because
of the skill emphasis, whereas at the middle and high school levels, instruction is more
content-focused. Due to a greater range of courses and programs of study, a standards-
based reporting system is more difficult to use effectively at the high school level. Another
potential drawback is the use of grades after high school (e.g., transcripts to college,
Narrative Grading
The narrative grading format typically involves an individualized written review of a student’s
performance that focuses on his or her particular learning accomplishments or skill acquisitions, as
well as continuing needs or learning issues that must be addressed. Few school districts utilize
a complete narrative approach to grading. However, more schools have started to use
brief narrative response options to accompany a more traditional grading and reporting
system.
The distinct advantage to the narrative approach is that rich, detailed information
on student performance is obtained, particularly when feedback is provided relative to
specific skills or goals, and can be used to direct further instruction. Unfortunately, pro-
viding this kind of feedback requires time that teachers may not have in their already
overloaded schedules. Standardized comment menus are available with computerized
grading programs that allow teachers to select and print specific comments on students’
report cards. However, be aware that these responses are sometimes perceived by parents
and students alike as impersonal, as the same comments can be reported by different
teachers for the same child.
♦ class attendance
♦ projects
♦ work habits
♦ attitude
♦ extra credit
♦ portfolios
♦ improvement
♦ behavior
♦ other elements?
What elements did you select? Are there elements that were not included in this list?
More importantly, why did you select them? The reasons behind the selections will pro-
vide you with insight into what you, as a professional educator, value in the evaluation of
your students’ work. That is really the most critical issue, because the reasons behind the
selections must be evident to a teacher as well as any other interested audience, includ-
ing students, parents, colleagues, or other community members. Teachers must be clear
about their evaluation system before they can make it clear to anyone else. In addition,
consistency among evaluation elements must exist from teacher to teacher and grade to
grade. Only then will a standard measuring stick exist for students within a school build-
ing or district.
© Royalty-Free/CORBIS
Given that a single grade is usually inadequate to provide all the information that
needs to be generated about a student’s performance, this multimeasure system can pro-
vide information along three criteria levels, including the product level, which serves
as the achievement measure and includes only evaluation elements (e.g., tests, quizzes,
papers, projects, presentations) that ref lect a student’s actual learning accomplishments.
Also, since the product measure serves as the achievement index, a student’s grade or
mark would be based on his or her performance on this dimension. The process level
can examine and report on important nonachievement factors such as effort, participa-
tion, homework completion, attitude, or attendance. Information and description on the
student’s actual progress during the instructional period can also be provided. With this
approach, a much more accurate and detailed evaluation of a student’s performance is
possible.
GUIDING PRINCIPLES
As an educator it is critical to have a clear idea of what grading system you will imple-
ment and how it will work in your classroom, and that starts with identifying the guid-
ing principles in the construction of your grading system. You can begin this journey by
asking yourself these basic questions.
1. What do you want and need your grading system to do for you as a teacher?
The fundamental purpose of the grading system must be identified so that it can be
designed and implemented to fulfill that purpose. If it is to serve several purposes, and
many grading systems do, then it will need to be more complex and may involve mul-
tiple levels. Whatever purposes must be met, they must be acknowledged by you as the
teacher.
As mentioned earlier in the chapter, a primary purpose of grading should be to pro-
vide a clear picture and measure of a student’s academic achievement and progress. That
sounds good, and it appears reasonable, since most (if not all) teachers would desire that
kind of outcome with their grading system. But that clear picture of academic achieve-
ment and progress must be specifically identified and articulated, complete with a listing
of what constitutes that picture and the clarity it should provide relative to the quality of
the student work that is demonstrated in the classroom.
“For you as a teacher” is an important add-on in this question. Grading must serve a
direct purpose for you. It should exist because it answers questions or helps to confirm
your observations or contentions about the progress of your students. Grading is certainly
a professional requirement, but more importantly, it should hold a personal purpose for
you and the students you evaluate.
Let’s assume in your review of this question that your grading system is to fulfill at
least three functions, which include the following: (1) serve as an official record of class-
room performance, (2) provide performance feedback to students, and (3) inform parents
of their child’s classroom performance and general status. At first glance, these functions
seem to be quite reasonable, but in order to fulfill these functions, focused work and
effort would be necessary on the teacher’s part to develop a grading system that meets
these identified needs.
2. How is the collected information from this system to be used?
This question addresses the purpose issue directly. The grades or marks that are gener-
ated from your system should be used for their identified purposes. Three potential func-
tions were identified in your grading system, and they represent three distinct purposes.
Because of this, the selection of student performance information, how that information
is presented, and what is communicated to targeted audiences becomes very important.
For example, grades typically serve as the official evaluation marks of students’ class-
room performances. Therefore, the district-approved grading scale (most school districts
have approved scales), complete with identified grading elements and a scoring system,
would need to be used in the review of all student academic performances. Grades would
then be generated at specified times (e.g., every grading period) as evidence of the quality
of the student’s academic achievement in the classroom and officially recorded as part of
the student’s academic record or grade transcript.
A grade record, now commonly provided in electronic programs such as Gradebook
or Edline, lists the grades received within a specific class during a particular semester or
year. Additional indices, such as grade point average (current, yearly, and cumulative),
credits received, attendance record, and class rank may also be provided as part of the
record. This report serves as a record of performance and can be used in administrative
and other school decisions (e.g., calculation of grade point average, honor roll selection,
academic awards, credits earned).
However, a traditional grade report provides no specific information regarding a stu-
dent’s actual performance, current skill levels, learning aspects that require continued
© Royalty-Free/CORBIS
work and attention, or any other individualized detail connected to his or her perfor-
mance in the classroom. Providing feedback to students on their classroom performance
is a very different purpose. It’s not really possible to get an individualized performance
profile from a standard grade or mark. If that kind of detail is desired, a qualitative
response such as a written narrative or a face-to-face discussion as part of grading meet-
ings is likely required.
Written comments describing a student’s general performance, including successes and
areas that still need to be worked on, can prove to be informative to a student regarding
his or her continued academic progress. As you might expect, most grade reports are not
designed for narrative responses, or at least for detailed responses due to time limitations.
Consequently, if that is a genuine purpose of your grading system, changes would be
necessary so that kind of information could be provided.
Informing parents of their child’s performance is an important purpose and one that
should not be downplayed. In general, most parents are truly interested in their children’s
school experiences, and that is especially the case when it comes to grades. Information
can be provided to parents through the qualitative student feedback examples that have
just been described. Receiving information at the end of a grading period, however, is
not preferable to receiving it throughout the entire grading cycle. That is why many
school districts encourage or require teachers to use electronic grade books, which pro-
vide current progress reviews to parents and students on a 24/7 basis.
3. Who is to benefit from this information?
Who are the identified beneficiaries of the information that is generated from your
grading system? Who needs this information and for what purpose will it be used? As
you think about this question, I encourage you to place yourself at the top of that list of
beneficiaries. Why? To start with, your own professional security is as stake here. Issu-
ing grades and evaluating students puts you directly in the line of fire, and because of
that, you will be placed in potentially volatile situations. As a teacher, it is important to
FOCUS ON ASSESSMENT
operate from a grading system that is designed to give you the most accurate information
possible so that the best decisions can be made. Also, by determining the purposes of the
system, the audiences connected to those identified purposes become more evident. Just
make sure you, as the teacher, benefit from the information that is collected.
4. What measurement procedures will be used in determining the grades of my
students?
The composition of the eventual grade a student will receive and how it is calculated
is important for you as a teacher and for your students. The elements that make up the
grade serve as evidence of student learning and are selected by the instructor as the best
measures of that learning. For example, if it is determined that a grade will ref lect stu-
dent achievement, and nothing more than student achievement, identified measures that
ref lect only student achievement should constitute that grade.
In addition to the selection of the activities that make up a grade, the relative impor-
tance and impact of those activities toward the calculation of that grade must be deter-
mined. If all are equally important, then they share or maintain equal value and weight.
However, if some elements or activities are more important than others (as determined by
the instructor), stronger value or relative weight should likely be given to those identified
activities. For instance, if a final exam covers more content and requires more knowledge
integration and skill usage then a midterm test, there is justification for giving the final
exam more value and relative weight in the calculation of the semester grade.
Once the elements and their relative weights are determined, a decision must be made
on how a summative or cumulative mark will be reached. The common practice of
averaging all grades and scores acquired over a grading period into a final average score
proves to be an inaccurate measurement practice (Guskey, 1994; Wright, 1994). The
problem with this practice involves the assumption that test scores are interval data points
and that equal intervals exist between the scores that make up that grading scale. In
actuality, test scores ref lect ordinal data and only show relative position (e.g., high or low
relative to other scores) with no uniform intervals between those positions.
Unfortunately, absolute precision does not exist with grading scales. Imagine using a
ruler where the distances between various points (e.g., inch markings, half-inch mark-
ings) do not have any uniformity to them. This is closer to the reality of a grading scale
and because of that, other more appropriate selection options should be considered. For
example, since test scores are ordinal data, selecting a representative grade based on the
median (50 percent of the grades fall above this score and 50 percent fall below this score)
is a more accurate and defensible measurement practice than using a composite average.
In this way, test scores are ranked from lowest to highest, with the middle score used
as the representative performance score for the student. However, Guskey (1994, 2002)
contends that using a student’s most recent performance is the best and most defensible
option for selecting a grade, since it is the most accurate measure of what is currently
understood.
Also, the recording of zeros can have a tremendous impact on a student’s overall grade.
It is not an uncommon practice for teachers to record a zero for missing work or for work
that is not turned in on time. Moreover, if a zero is recorded for an activity that has com-
parable weight to other evaluation elements, this practice can literally destroy a student’s
grade, especially if an average is used in the determination of that grade. Other practices,
such as deleting one score from a student’s grade, the use of extra credit, the impact of
homework, make-up work, etc., must be carefully reviewed before they are integrated
into an instructor’s grading practices. Remember that there should be clear and logical
reasons and operating procedures used in the selection of the grading components for the
student evaluation process, and everything that is selected must be defensible.
PROFESSIONAL CONSENSUS
As difficult as it is to select and construct your own grading system, try to imagine what
it would be like to have all the teachers in your building (or for the teacher candidate col-
leagues in your college classrooms) go through this process and fundamentally agree on
a grading system that they would all use when evaluating the progress of their students.
Sound difficult? Well, of course it is, but that is exactly what should happen.
Separate grading systems across teachers, grades, and schools only create ineffective
and inconsistent evaluation of students and their work (Guskey & Bailey, 2001; Marzano,
2006). Grading is a district-wide process that cannot afford to operate in isolation. Col-
legial collaboration and grading unification work must happen on a large scale if it is to
translate into a truly effective system of judging student work and academic progress.
If grading measures and systems differ consistently from teacher to teacher, significant
grading variation will exist within and across grade levels, school buildings, and even an
entire school district. This creates a situation where students are graded and judged based
on different criteria for every course they take, and thereby eliminates any chance of a
uniform student evaluation system. Little if any validity exists in this situation. It is very
possible that a student’s tenth-grade English teacher will evaluate progress and achieve-
ment based on factors that are fundamentally different from the same student’s tenth-
grade history or science teacher.
Grading variation is to be expected to some degree and can be supported given differ-
ences in content and skill requirements and the individuality of the teachers involved in
the process. However, a core set of evaluation elements must exist as part of the grading
process, along with a shared purpose. Inconsistent element selection in grading practices
within a district creates an evaluation accuracy problem where grades are not comparable
under any circumstances.
School districts will usually identify a particular grading scale and reporting method
that is to be followed in the district and used for official school business such as determin-
ing grade point average, academic awards, or class rankings. However, in many school
districts, the subject of grades and how they are constructed does not appear on the
“we-need-to-look-at-this-now” priority list. In general, most school administrators only
require that a teacher have some documentation of how grades are generated, and grad-
ing practices among faculty are seldom critically examined. This makes your work in
determining an effective grading and reporting system all the more important. Grading
synchronicity is essential and needs to exist within all classrooms, across all grade levels,
and in all school buildings within a school district. Obtaining grading consistency should
be initiated through conversations with your teaching colleagues at all levels (grade,
building, and district), and those conversations should not end until grading uniformity
exists.
COMPARISON TYPES
Along with identifying a grading scale and the composition elements of what a grade will
represent in your classroom, an overall framework is needed to give a grade perspective
and meaning. The most common grade comparison models include norm-referenced,
criterion-referenced, and aptitude comparison or self-referenced models.
Norm-Referenced Grading
Norm-referenced grading involves comparing a student’s performance to a specific group,
such as a classroom or a grade level, in order to see where that performance falls relative to
the performance of the entire group. With this approach, specific percentages of grades
are typically preset so that only a certain number or quota of A’s, B’s, C’s, D’s, and F’s
are given by the teacher. This is known more generally as “grading on a curve,” and this
approach is used in classes when instructors want to limit the number of high marks,
distribute student performances, and avoid grade inf lation. It also frees the teacher from
any grading dilemmas and subjective calls, since the number of grades is already prede-
termined as part of the system. This approach is also erroneously viewed by some teach-
ers as ref lecting a higher grading standard and rigor. A common norm-referenced grade
breakdown is shown in Table 9.2.
The selected percentages are based on the spread of scores, assuming that they are dis-
tributed along the boundaries of a normal distribution or bell-shaped curve. Fewer scores
are expected at the ends of the distribution (i.e., A’s and F’s) so the smallest percentages
are listed there, with the highest percentages listed for the middle grades. As a student
in this class your sole motivation is to score higher than your classmates, because if you
score high on a test, your chances of receiving a high mark are good, but if you obtain a
low score and your classmates score higher, you will certainly receive a low grade.
When considering this model, limitations must be acknowledged. To start with, the
norming group that is to be used as the comparison is very important, and of course
reference groups can vary considerably. For example, if you are an average student and
you walk into a classroom with a group of academically strong students, you will face
a challenging grading competition experience. On the other hand, if you are fortunate
enough to enter a class made up of students at the bottom of the educational food chain,
this is your lucky day. Clearly each outcome is highly dependent upon the constituency
of the groups.
Most importantly, norm-referenced grades are based on a student’s positioning within
a group and are not linked to the acquisition of specific learning outcomes. As a student,
you could theoretically do well with a low comparison group and not actually know the
material very well, or conversely not do well relative to a high comparison group but
really know your stuff. Also consider this: If you have students who consistently score
in the middle to the bottom relative to their peers (and you will have students like this),
that is where they will stay, and their grades will ref lect that position. In essence, they
become “stuck” in that relative position within the class, and there is no real incentive
to move beyond it simply because they can’t. The teacher also suffers, as a portion of his
or her students are always going to do poorly or fail no matter how well the content is
taught.
Finally, the percentage of students who can receive a specific grade is arbitrary and
completely under the control of the teacher. For example, 10, 12, 15, or 20 percent of stu-
dents could be assigned A’s, and the same variability can be demonstrated with the other
grades. Those determinations are controlled exclusively by the instructor, so the percent-
age cutoffs of certain grades can vary, which makes grade comparisons very difficult.
Criterion-Referenced Grading
Criterion-referenced grading involves comparing a student’s performance to a specific, pre-
determined performance level or standard of work. If a student meets a certain performance
level according to set learning standards (e.g., acquires a certain number of points based
on work completed within the class), then the student receives the grade associated with
that level. No preset number of grades is established, as with the norm-referenced model.
Therefore, all students who reach a certain level receive that grade (e.g., an A) regardless
of whether the entire class reaches that level. Conversely, many or most students could
also receive poor grades depending upon the performance levels that are demonstrated.
Criterion-referenced grading is recognized as the most commonly utilized grading
model in schools. The criterion-referenced aspect to the grading usually falls within two
general categories: performance-based criteria or percentage-based criteria. Performance-
based criteria involve the listing and completion of specific learning competencies that
students must demonstrate in order to receive a particular grade. These competencies can
take the form of specific work products (e.g., classroom projects, portfolios, reports, pre-
sentations) in which rubrics or other assessment measures are used to score each item based
on the appropriate criteria. The specific criteria that the students are to demonstrate are
clearly stated and then reviewed with the students as part of the instructional process.
Percentage-based criteria are based on the percentage of items answered correctly on
assessment measures that are constructed and aligned with identified learning targets.
Cutoff percentages are set at particular performance levels and used to designate certain
levels of mastery of the material. For instance, 90 to 100 percent of items answered cor-
rectly could serve as the highest level, and a student would receive a traditional letter
grade of A. A 10-point range could be used for the remaining grades, as well. However,
these cutoffs are arbitrary and are based on teacher judgment. In addition to providing
straightforward guidelines for assigning grades, the potential exists for many students to
do well with this model, and that’s exciting for both the students and the instructor. This
provides a genuine incentive for the teacher to provide the best possible instruction.
Caution is still in order with criterion-referenced grading, however. The grading
system is connected to recognized standards of performance, and this reality must be
understood by the students who are expected to acquire these performance standards.
All too often clarity is not provided and students are left clueless to the purpose of their
instruction. Also, effective teaching relative to content mastery and accurate assessment
of that learning progress is essential with this approach, because if neither one is accom-
plished, the resulting inconsistency provides no grading comparison value at all. Because
the potential for all students to do well exists with this model, the general public and
even some educators may view this as a less rigorous and effective grading system. This is
an inaccurate perception, but it is one that still exists.
levels. It’s our job as professional educators to make sure we effectively teach, assess, and
evaluate that learning progress.
possible. In general, classrooms are more heterogeneous (i.e., made up of students at dif-
ferent levels of achievement, knowledge, and skill) than they are homogeneous (i.e., made
up of students who are comparable and uniform in their level of achievement, knowl-
edge, and skill), unless honors, remedial, or advanced placement classes are involved.
assignments, as stipulated in a student’s IEP, all student work would be evaluated and held
to the same standards; in essence, an absolute standard of performance must be met in
order to receive a specified mark or grade.
For identified students that possess a strong academic background and skill set, this
approach may be seen as appropriate, as these students truly experience the full-ranging
impact of inclusion. The reality of this approach is quite drastic for the majority of
students with special needs, however, as considerable evidence demonstrates that these
students as a collective group demonstrate disproportionately higher rates of low to
failing grades compared to their non-disabled peers (Donahoe & Zigmond, 1990; Val-
des, Williamson, & Wagner, 1990). When it comes to traditional and current grading
procedures, a level playing field does not exist for students with disabilities, and they
are more likely to experience academic failure and are at high risk for dropping out of
school (Zigmond & Thorton, 1985). In addition to meeting the same standard of per-
formance, students with disabilities are also placed in the situation of acquiring those
standards within the same amount of academic instructional time as their nonidentified
peers. According to Gettinger (1989), significantly more engaged time is required for
low-achieving students, and with some educational outcomes, six times as much time
may be needed.
Therefore, since instructional time is a critical factor for all students but especially for
academically weaker students, the expected standard of performance must be carefully
aligned with and supported by the amount of engaged time that is provided in the class-
room along with the directed instruction. According to Gettinger and Ball (2008):
Students gain the most from their learning when they experience a balance of high
and medium success, with more activities targeted at a high-success level (at about 80%
accuracy). Thus, accurately measuring and ensuring success is critical for increasing
students’ academic learning time. Evidence of success may include answering ques-
tions correctly, completing work accurately, or demonstrating understanding through
other performance indicators (p. 1046).
With some students, more instructional time and practice is necessary in order to acquire
and master a skill, and that may apply to a wide range of students, whether they qualify
for special services or not.
Grading Adaptations
The grading adaptation approach involves teachers making grading adjustments or refine-
ments to their existing grading system for their students with disabilities. This is a common
approach among teachers (Bursuck, Polloway, Plante, Epstein, Jayanthi, & McConeghy,
1996; Polloway, Epstein, Bursuck, Roderique, McConeghy, & Jayanthi, 1994; Silva,
Munk, & Bursuck, 2005) and according to Guskey and Bailey (2001), the typical adapta-
tions teachers find useful include “grading on improvement, basing a grade on meeting
the objectives specified in an IEP, and assigning separate grades for product (achieve-
ment) and for process (effort and class behavior)” (p. 117). Adaptations fall into three
general areas: (1) changing grading criteria, (2) providing supplemental information, and
(3) using other grading options. Teachers will often use one or more of these options when
making grading adjustments for students with special needs. The lack of consistency and
organization surrounding this issue is not surprising, as most school districts lack uniform
policies regarding grading, and most teachers rely on themselves in regards to grading
considerations (Polloway, Epstein, Bursuck, Roderique, McConeghy, & Jayanthi, 1994).
When these approaches are collectively reviewed, Guskey and Bailey (2001) make the
case that the best and most equitable grading option is the same one that was described
earlier in the chapter. Since a single grade or mark is really not adequate to provide all
the information needed for the evaluation of a student’s performance, a multimeasure
grading system is optimal, where information on product (the achievement measure),
process (effort and behavior variables), and progress (improvement over the course of
instruction) can be recorded.
Determining what elements will make up these three grading domains and how they
will be configured become important considerations for the teacher and special educa-
tion team. In addition, given the unique learning issues for each student with special
needs, these grading domains would likely require adjustment and modification based
on the particulars of the IEP, since that should always direct the instructional process for
any identified student. Students with special needs represent genuine teaching and grad-
ing challenges, and the implementation of effective grading and reporting systems for all
students is possible for any school district willing to accept and address these challenges.
As educators, we are at a point where comprehensive grading and student evaluation has
become a necessity in the classroom of the 21st century.
and activities that students are asked to complete and that are used in the grading
process.
♦ Strongly consider using a multilevel grading system, where the achievement grade
or mark is separate from other factors that inf luence the learning process or prog-
ress that students evidence during their instruction.
♦ Always make sure the grading practices involve activities that are aligned with the
identified learning outcomes of the provided lesson. Instruction and assessment
alignment must always exist.
♦ If possible, avoid the use of zeros in the grading process and instead provide the
administration of at least a minimum failing score.
♦ Always share grading policies with students and parents and include grading infor-
mation as part of the class syllabus at the beginning of the class. This can be pro-
vided as a hard copy or posted on the teacher’s Web site or as part of the school’s
electronic grade reporting system.
♦ As good as your grading system becomes, remember that there is no substitute for
face-to-face contact. Grades and records provide information, but the need for per-
sonal contact will never go away. Phone calls and e-mails are important and may
be sufficient in some situations, but sometimes students and parents just need to sit
down with the teacher.
♦ When you break down grading and evaluation into its most basic form, it’s really
all about communication: accurate, detailed, and personalized feedback about a
student that is clearly conveyed to all involved parties.
♦ Within most preparation programs, teachers receive very little specific instruction
on grading and the role evaluation plays in the classroom. All too often it becomes
a professional skill that is “picked up” on the job. Many teachers blindly accept the
methods they encounter when they complete their student teaching assignments or
first teaching position in a local school district. Please recognize that this acquired
“blindness” does not absolve a teacher of his or her professional responsibility to
provide accurate and effective assessment of students and their learning progress in
the classroom.
♦ Resources and school efforts (e.g., teacher in-service or faculty training) devoted
to faculty development on the effective measurement of student learning and the
documentation and evaluation of that learning are usually limited. But if teachers
indicate a desire to know more and to become more proficient in this skill area, it’s
likely school administrators will pay attention and do something about it.
♦ Despite professional controversy, disagreement, and the reality of grading incon-
sistency, the general public expects teachers to be knowledgeable of and compe-
tent in constructing their own grading and evaluation systems, complete with a
strong rationale for their construction and implementation. Make sure you meet
this expectation.
♦ Grading uniformity must exist in a school district, and there should be general
agreement on the following: what evaluation elements will constitute a grade;
whether multiple grading levels (e.g., product, process, and progress) are to be
provided; how a grade will be expressed (i.e., general scaling and interpretation);
and what purpose the grading system will serve. This continuity starts with every
teacher becoming part of a professional dialogue within a district as opposed to
remaining isolated, with everyone following their own unique grading process and
pathway.
Be comfortable and confident enough to discuss your grading practices with your col-
leagues, since as educators we can all benefit from meaningful discussions on this topic.
Whether these dialogues focus on new information or approaches or on something else,
it doesn’t matter; having the discussions is what really matters.
CHAPTER REFLECTION
At the beginning of the chapter you answered several grading questions in the Chapter
Activity based on your experiences as a student. Now address the following questions
based on your future position as a teacher:
1. What functions do grades perform for you as a teacher?
2. How do you want to evaluate the work of your students?
3. Is the evaluation process necessary? Why or why not?
CHAPTER REVIEW
Conclusion
The issues that existed in the early 20th century surrounding the use and reporting of
grades remain with us today. Clearly, grades as a form of student evaluation are a central
part of the educational experience and that will not likely change. Due to the univer-
sality of the evaluation and reporting of student work, teachers must be confident in
the manner in which grades are assigned in their classrooms. In addition, this system
must serve all students, even those who have identified disabilities. Grading confidence
can only manifest itself if teachers receive effective postsecondary instruction, continue
to participate in professional development opportunities, and work in an environment
where dialogue regarding this topic is supported.
to needs, experiences, and backgrounds, the integration of the three grading domains
(product, process, and progress) advocated by Guskey and Bailey (2001) into one’s grad-
ing system appears to be warranted.
As a system is selected and refined, remember to clearly identify what purposes it must
serve and who the beneficiaries will be. As mentioned at the beginning of the chapter,
grading is always personal. Therefore, develop a system that possesses high communica-
tion value for all of those who are connected to it.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. Identify and describe the guiding principles that will be associated with the grad-
ing system you would like to use with your future students. Also, identify and
describe what grading elements will make up a grade in your classroom. How will
this system operate on a daily basis?
2. How does your grading system account for the evaluation of students with special
needs?
RESOURCES
One of the most complete references on grading and student evaluation is Thomas Gus-
key and Jane Bailey’s book Developing Grading and Reporting Systems for Student Learning,
published in 2001 by Corwin Press. This is a book that every teacher should have as
part of their professional library.
At the National Education Association Web site (www.nea.org), type the word “grading”
in the search box and a listing of individual teacher responses and practices, relative to
grading issues in the classroom, is provided.
REFERENCES
Brookhart, S. M. (1991). Grading practices and validity. Educational Measurement: Issues and
Practice, 10, 35–36.
Brookhart, S. M. (1993). Teachers’ grading practices: Meaning and values. Journal of Edu-
cational Measurement, 30(2), 123–142.
Brookhart, S. M. (2004). Grading. Upper Saddle River, NJ: Pearson Education.
Bursuck, W. D., Polloway, E. A., Plante, L., Epstein, M. H., Jayanthi, M., & McConeghy,
J. (1996). Report card grading and adaptations: A national survey of classroom prac-
tices. Exceptional Children, 62(3), 301–318.
Butler, S. M., & McMunn, N. D. (2006). A teacher’s guide to classroom assessment: Under-
standing and using assessment to improve student learning. San Francisco: Jossey-Bass.
Donahoe, K., & Zigmond, N. (1990). Academic grades of ninth-grade urban learning-
disabled students and low-achieving peers. Exceptionality, 1(1), 17–27.
Frary, R. B., Cross, L. H., & Weber, L. J. (1993). Testing and grading practices and opin-
ions of secondary teachers of academic subjects: Implications for instruction in mea-
surement. Educational Measurement: Issues and Practice, 12(3), 23–30.
Friedman, S. J., & Manley, M. (1992). Improving high school grading practices: Experts
vs. practitioners. NASSP Bulletin, 76(544), 100–104.
Gettinger, M. (1989). Effects of maximizing time spent and minimizing time needed for
learning on pupil achievement. American Educational Research Journal, 26, 73–91.
Gettinger, M., & Ball, C. (2008). Best practices in increasing academic engaged time.
In A. Thomas & J. Grimes (Eds.), Best practices in school psychology V (pp. 1043–1057).
Bethesda, MD: The National Association of School Psychologists.
Guskey, T. R. (1994). Making the grade: What benefits students. Educational Leadership,
52(2), 14–20.
Guskey, T. R. (2002). Computerized gradebooks and the myth of objectivity. Phi Delta
Kappan, 83(10), 775–780.
Guskey, T. R. (2006). Making high school grades meaningful. Phi Delta Kappan, 87(9),
670–675.
Guskey, T. R., & Bailey, J. M. (2001). Developing grading and reporting systems for student
learning. Thousand Oaks, CA: Corwin Press, Inc.
Individuals with Disabilities Education Act Amendments of 1997 (IDEA-97), 20 U.S.C.
1400 et seq. (1997).
Individuals with Disabilities Education Improvement Act of 2004 (IDEA-04), 20 U.S.C.
1400 et seq. (2004).
Marzano, R. J. (2000). Transforming classroom grading. Alexandria, VA: ASCD.
Marzano, R. J. (2006). Classroom assessment & grading that work. Alexandria, VA: ASCD.
McMillan, J. H., Myran, S., & Workman, D. (2002). Elementary teachers’ classroom
assessment and grading practices. The Journal of Educational Research, 95(4), 203–213.
Munk, D. D., & Bursuck, W. D. (2003). Grading students with disabilities. Educational
Leadership, 61(2), 38–43.
Nitko, A. J., & Brookhart, S. M. (2007). Educational assessment of students (5th ed.). Upper
Saddle River, NJ: Pearson Merrill/Prentice Hall.
No Child Left Behind Act of 2001 (ESEA), Pub. L. No. 107-110 (2002).
O’Connor, K. (2002). How to grade for learning: Linking grades to standards (2nd ed.). Arling-
ton Heights, IL: Skylight.
Polloway, E. A., Epstein, M. H., Bursuck, W. D., Roderique, T. W., McConeghy, J. L.,
& Jayanthi, M. (1994). Classroom grading: A national survey of policies. Remedial and
Special Education, 15(2), 162–170.
Popham, W. J. (2008). Classroom assessment: What teachers need to know (5th ed.). Boston:
Allyn and Bacon.
Reynolds, C. R., Livingston, R. B., & Willson, V. (2006). Measurement and assessment in
education. Boston: Pearson Education, Inc.
Silva, M., Munk, D. D., & Bursuck, W. D. (2005). Grading adaptations for students with
disabilities. Intervention in School and Clinic, 41, 87–98.
Stiggins, R. J., Arter, J. A., Chappuis, J., & Chappuis, S. (2004). Classroom assessment for stu-
dent learning: Doing it right—using it well. Portland, OR: Assessment Training Institute.
Stiggins, R. J., Frisbie, D. A., & Griswold, P. A. (1989). Inside high school grading prac-
tices: Building a research agenda. Educational Measurement: Issues and Practice, 8(2), 5–14.
Valdes, K. A., Williamson, C. L., & Wagner, M. M. (1990). The national transition study of
special education students (Vol. 1). Menlo Park, CA: SRI International.
Worthen, B. R., Borg, W. R., & White, K. R. (1993). Measurement and evaluation in the
schools. New York: Longman Publishing Group.
Wright, R. G. (1994). Success for all: The median is the key. Phi Delta Kappan, 75(9),
723–725.
Zigmond, N., & Thornton, H. (1985). Follow-up of postsecondary age learning disabled
graduates and dropouts. Learning Disabilities Research, 1(1), 50–55.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How can I use applied statistics to better understand student performance in my
classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Describe terms including distribution, range, normal curve, skewed distribu-
tions, variance, standard deviation, stanine, percentiles, mean, mode, median,
and standard scores.
♦ Compute the three basic measures of central tendency.
♦ Define variance and calculate a standard deviation from a set of test scores.
♦ Understand the importance of knowing the vari-
ance and central tendency within a distribution.
♦ Differentiate common standard scores (z-scores,
T scores, Deviation IQ scores, stanines) that are
used in educational and psychological tests and
measures.
♦ Describe how standard scores function within a
normal distribution.
♦ Explain the strength of a correlation coefficient
when a linear relationship is presumed.
♦ Calculate item difficulty and item discrimination
index levels for test items within a classroom test.
254
Introduction
It is important to understand that the area of descriptive statistics involves the use of
numbers and procedures to more fully describe a larger data set that is collected for some reason. This
is done in order to gain a better understanding of the findings that are generated from
that large data source. For example, in your classroom, you likely collect test, project, and
homework scores for each of your students. If you have 25 or more students in your class,
you end up collecting a lot of information on each student over a short period of time.
In order to get a better handle on the academic performance of each student, an overall
grade point average is typically generated from those recorded scores.
This summarization of information into one score (i.e., grade point average) is a good
example of how descriptive statistics have become a direct part of a teacher’s professional
practice. Now we can debate whether the averaging of scores is the best way to indicate
a student’s learning of the material in your classroom, but the process of averaging those
scores and generating a composite number definitely involves the use of descriptive sta-
tistics. Whenever you look at averages, graphs, or distributions (e.g., collections of scores)
of information, you are using and hopefully benefiting from statistics. In addition, with
a more developed knowledge base you also become a better consumer and reviewer of
statistical information, and that can assist you both inside and outside the classroom.
CHAPTER ACTIVITY
Statistics are a part of our daily life. For example, knowing the expected temperature range
(low and high temperatures for the day) is important for any planned outside activity. If
you are shopping for a new car, knowing the average price of that car as well as the gas
mileage (e.g., average miles per gallon in the city and on the highway) is necessary. If you
have plans to go sailing, it is important to check the average wind speed as well as the esti-
mated peak wind speed for the day. In addition, if you are a baseball fan, you need to check
the sports page for your favorite players and their current batting averages. Obviously many
more examples could be provided, but it is clear from this short list that statistics are part of
our everyday existence.
Now it is time to take a look at the role statistics plays in your immediate world. For one
entire day, record in journal entries the number of occurrences where statistics or statistical
procedures are implemented that directly impact you and your personal or pre-professional
activities. Activities could range from reading a professional journal article that includes
statistical information, the score range and average test grade in your math class, or the
price of gas you just paid at the pump compared to the national or regional average. Once
the journal entries are collected, break into small groups with your classmates (if possible)
and discuss and compare your findings.
in direct and meaningful ways has also intensified. Through the use of descriptive sta-
tistics, collected evidence on your students can be effectively recorded, managed, and
summarized, as well as presented to a variety of audiences (e.g., students, parents, admin-
istrators, the general public) depending on the need at the time. In addition, due to the
reality that basic statistical packages are commonplace in word processing programs,
computation of simple statistical procedures can be easily completed, which is a great
time saver.
However, it is important to recognize that statistical procedures, whether or not they
are computer driven, are only designed to analyze collected data. The actual educational
decisions, which are based on statistical reviews, are always executed by the professionals
in charge. Sometimes you may hear the expression that “the data determines the deci-
sion,” but the truth is that we as professional educators make the decisions, and hope-
fully those decisions have strong data support. The responsibility of the decisions that are
made, regardless of the collected data, is ultimately held by us.
Descriptive statistics allow a teacher to summarize, examine, and present student
learning evidence, which is desirable since it is expected that information will be reported
in direct and meaningful ways. Also, whether it is true or not, teachers are often seen by
parents and the general public as being well informed when it comes to measurement
principles and statistical practices. Therefore, teachers need to have a functional knowl-
edge base and an understanding of basic measurement procedures and principles. In order
to make effective decisions, a teacher needs to collect meaningful student data. Often-
times that evidence and data needs to be analyzed, and that is where statistics can help.
Scores Frequency
98 1
95 1
91 1
90 1
89 1
88 1
87 1
85 2
84 2
83 1
82 1
81 4
80 1
77 1
76 3
72 1
70 2
69 1
65 1
Distributions
With a simple frequency distribution, each separate score is listed in addition to how
many times each score was obtained. Just by looking at the Spanish test score list, the
highest (98) and lowest (65) scores can be identified, and when the lowest score is subtracted
from the highest score, this produces a variance measure (i.e., how much scores differ from each other),
which is referred to as the range. Most of the listed test scores were obtained only once,
except for five scores that were obtained at least twice or more. Also, it appears that most
of the scores fell in the middle to the lower part of the range.
As you can see, a simple frequency distribution can be useful in presenting data. How-
ever, as more data is collected and an expanded score range is encountered, the ability of
the simple frequency distribution to accurately display information is diminished. When
this occurs, the construction of a grouped frequency distribution is in order.
Grouped frequency distributions are based on the construction of predetermined bands
or intervals where individual scores reside (Figure 10.1). Since the range is obtained from sub-
tracting the lowest score from the highest score (98 – 65 = 33), this number can then be
divided into specific point intervals. For example, it has been arbitrarily decided that
seven intervals will be used to display Mrs. Smith’s test scores, which results in five-point
intervals. If the range is divided by the number of desired intervals, the actual size of each
interval is then generated. With seven intervals, the actual interval length of 4.7 needs
to be rounded up to 5. The number of intervals is up to the teacher and is based on how
the information is to be displayed. The selection of the number of intervals depends to a
Range
5 Interval length
Number of intervals
33
5 4.7 or 5
7
The Spanish test scores have now been placed in a grouped frequency distribu-
tion based on five-point intervals (see Table 10.2).
great extent on the range of your scores. With too few intervals, you don’t get a real sense
of the data, and with too many you may lose the distribution of the data altogether.
As is evident in Table 10.2, this approach compresses the data closer together and
displays where most of the scores fall within the distribution. A majority of the students
had scores that fell within the 79–83 range (26 percent of the class), and if the range is
expanded from 74–88, 63 percent of the students are included. These results could be
encouraging to a teacher, as it appears that a majority of students did reasonably well on
the test. However, if the teacher follows a mastery approach and is expecting a 100 per-
cent success rate, these results would be informative but probably not desirable.
Regardless of the interpretation, presenting scores within a grouped frequency dis-
tribution allows for a condensed view of the spread of the data. However, when the data
is grouped like this, it is not possible to look at individual scores or data points. If that
information is needed, a teacher would have to go back to the original data set to find the
specific scores within a certain interval.
FOCUS ON ASSESSMENT
Interval Frequency
98–94 2
93–89 3
88–84 6
83–79 7
78–74 4
73–69 4
68–64 1
The grouped frequency distribution provides for an effective organization and review
of a data set. This procedure may be all that is necessary for Mrs. Smith to understand the
performance of her students on the test. If not, there are other ways to display this infor-
mation. In particular, along with ordering data within distributions, providing visual
presentations of that information can be useful for anyone who examines it. That is why
graphing is so popular.
Two of the most popular graphs are the bar graph (or histogram) and the frequency
polygon. Both of these types of graphs can help provide clear visual displays of collected
data. The bar graph or histogram provides rectangular columns, and the height of those
columns represents the frequency of the scores that falls within a score interval (Figure 10.2). The
frequency polygon is actually a line graph that connects the frequencies of the intervals (Figure
10.3). However, with the frequency polygon, a midpoint is selected for each interval and
those points are connected together. Keep in mind that whenever data is involved, pro-
viding pictures and graphs to display and highlight information, along with a clear expla-
nation, helps to increase the understanding of the information for its intended audience.
Frequency
5
4
3
2
1
0
96 91 86 81 76 71 66
Score
FIGURE 10.3 FREQUENCY POLYGON OF MRS.
SMITH’S TEST SCORES
Mrs. Smith definitely has more information on the performance of her class than
when she started, but she also wants to know the average test performance of her stu-
dents. Before that can be answered, however, the properties of distributions need to be
investigated further. Both graphs of the test scores provide a visual profile for the distri-
bution of the scores. If the collection of similar Spanish test scores was repeated a number
of times, a smooth, balanced, and uniform shape would likely emerge. This shape reflects a
symmetrical distribution and is commonly referred to as a normal curve (Figure 10.4).
Although technically a hypothetical distribution, normal curves are useful, since
many occurrences in the world, such as height, weight, age, or test performances, are
distributed in a similar way. With this distribution pattern, most occurrences are located
in the middle or center, with fewer events occurring at the ends (referred to as the tails
of a distribution). As common as this type of distribution is, it is important to understand
that not all distributions follow this pattern. In fact, some distributions are asymmetrical and
are slanted in one direction; these are referred to as skewed distributions (Figure 10.5).
For example, if a distribution is positively skewed, that would indicate that most of the
data or scores are located on the left side of the distribution, with few on the right side. If test scores
evidenced this kind of distribution, the majority of the test scores would fall at the lower
end, with only a few scores at the higher end. Conversely, if a distribution is negatively
skewed, most of the scores are located on the right side of the distribution, with few scores on the left
side. With this distribution, most of the test scores would be at the higher end, with fewer
low scores. Other distributions can be f lat or look like a plateau. Distribution patterns
are dependent to a great extent on what is being measured and how often it is assessed, as
well as the unique qualities of the event being appraised.
directly inf luenced by those scores, resulting in a lower and less accurate overall measure
of central tendency.
The last measure of central tendency is the mode, which is the score that occurs most fre-
quently in a distribution. In Mrs. Smith’s class, the mode also happens to be 81, and more
than one mode can exist within a distribution. Furthermore, in a small data set like the
10 scores just listed, there may not be a mode at all. The mode is most effective when cat-
egorical or nominal data is involved. For example, a teacher has 10 different take-home
storybooks in her class and could use the mode to signify the most selected and popular
of those books.
Measures of central tendency allow teachers to better understand the nature of the dis-
tributions they may experience within the classroom. More importantly, with different
distributions, measures of central tendency are affected in different ways. For instance,
with a symmetrical or normal distribution, all three measures of central tendency possess
the same value. However, in skewed distributions, that is not the case. For example, in
a positively skewed distribution, the mean is higher than the median and the mode; in a
negatively skewed distribution, the mean is lower than the median, which is lower than
the mode. Being aware of these differences can help you to identify and understand the
different kinds of distributions that may be present in your classroom or in other educa-
tional settings.
to as the deviation score method (Table 10.3). Based on the deviation score method, the
standard deviation for the following set of test scores (97, 96, 94, 92, 85, 83, 82, 75) is:
The standard deviation can also be generated from the raw scores or test scores that
are collected without the generation of the mean. In this process, the square of each test
score is generated and both the test scores and squared scores are then added up; this is the
sums of squared scores method. Both approaches generate the same standard deviation.
More importantly, the size of the standard deviation provides a gauge as to how much
spread exists among the scores relative to the mean.
For example, if the standard deviation is small, the scores are packed tightly around
the mean, and if the standard deviation is large, the scores are more spread out and more
variance exists within the scores of the distribution. This becomes very important,
because when you possess information on the central tendency and variability of a distri-
bution, you have much a greater understanding of your data set and ultimately the class-
room decisions that will be made based on that data set.
Normal Distribution
Let’s turn our attention to the most well-known and utilized distribution in the field
of education, which is the normal distribution. The normal distribution is a sym-
metrical distribution that possesses characteristics useful in describing the ways scores vary from
the mean. With this hypothetical distribution, all measures of central tendency are the
same and are located at the center of the distribution (Glass & Hopkins, 1984). In addi-
tion, with a normal distribution, 34.13 percent of all scores fall between the mean and
one standard deviation from the mean. When one standard deviation above and below
the mean are considered, that expands to 68.26 percent of all the scores of a normal
distribution. Finally, 95.44 percent of all cases or performances, within the normal
distribution, are found two standard deviations above and below the mean (Kubiszyn
& Borich, 2003). These properties allow for the calculation of score percentages rela-
tive to the areas of the distribution. For example, with Mrs. Smith’s Spanish class, the
class mean is 81 and the standard deviation is 7.38 (rolled back to 7 for computation
purposes). Using this information and assuming these scores are normally distributed,
68 percent of the scores fall between 74 and 88, and 95 percent of the scores fall
between 67 and 95.
The properties of the normal curve allow for the calculation of the percentage of cases
that fall within specific sections of the distribution. Along with the determination of per-
centages, percentiles, or the percentage of individuals scoring at or below a given point, can also
be identified and set along a normal curve. For example, at the center of the distribution
is the 50th percentile, one standard deviation above the mean is the 84th percentile, and
two standard deviations above the mean is the 98th percentile. Going in the other direc-
tion, one standard deviation below the mean is the 16th percentile, and two standard
deviations below the mean is the 2nd percentile. This allows a teacher to place a student’s
performance within the distribution as well as compare performances within the group.
Comparisons to other test scores and performances are also possible with the normal
curve distribution (Figure 10.6).
STANDARD SCORES
Classroom test scores like those from Mrs. Smith’s class are referred to as obtained or
raw scores, and they usually represent the number of items that are correctly answered. These
terms are used because no statistical or “normalization” procedures have been utilized to
convert these scores into a more standard score set. When scores are converted they are
generally referred to as derived scores, which then allows them to be compared to scores
from other tests. We will now investigate the most typical standard scores and how they
are generated.
If we want to compare test performances within a normal distribution, a test’s raw
scores need to be converted to a common scale and into a set of standard scores. Stan-
dard scores are derived from an original set of raw or observed scores and are based on the standard
deviation of the distribution (Mason & Bramble, 1978). Two basic standard scores teachers
should be familiar with are z and T scores.
The z-score is one of the simplest and easiest standard scores to construct. A z-score is
generated by subtracting the mean from the raw score and then dividing that by the standard deviation
of the raw score distribution.
X 2 Mean
z 2 score 5
SD
The z-score has a center point of zero, and a score of 1 corresponds to one standard devi-
ation above the mean. A score of 2 corresponds to two standard deviations, and a score of
3 corresponds to three standard deviations. Conversely, a score of –1 corresponds to one
standard deviation below the mean, –2 corresponds to two standard deviations below the
mean, and –3 corresponds to three standard deviations below the mean.
Percentage of cases
under portions of
.13% 2.14% 2.14% .13%
the normal curve 13.59% 34.13% 34.13% 13.59%
Standard Deviations −4σ −3σ −2σ −1σ 0 +1σ +2σ +3σ +4σ
(SD or σ)
Cumulative Percentages 0.1% 2.3% 15.9% 50.0% 84.1% 97.7% 99.9%
Rounded 2% 16% 50% 84% 98%
Percentile Equivalents
1 5 10 20 30 50 70 80 90 95 99
Q1 Md Q3
Going back to the earlier score set, these test scores would generate the following
z-scores:
Raw scores z-scores
97 1.21 1 97 2 88 4 7.38 5 1.21 2
96 1.08 1 96 2 88 4 7.38 5 1.08 2
94 .81 1 94 2 88 4 7.38 5 .81 2
92 .54 1 92 2 88 4 7.38 5 .54 2
85 –.41 1 85 2 88 4 7.38 5 2.41 2
83 –.68 1 83 2 88 4 7.38 5 2.68 2
82 –.81 1 82 2 88 4 7.38 5 2.81 2
75 –1.76 1 75 2 88 4 7.38 5 21.76 2
Mean = 88
SD = 7.38
The benefit of this approach is that a teacher can place these scores on the normal
curve. For example, the 96 test score falls slightly above the first standard deviation and
exceeds the 84th percentile. In addition, the 75 test score falls below the first standard
deviation below the mean and falls below the 10th percentile.
T scores, a standard score scale with a mean of 50 and a standard deviation of 10, are trans-
formed scores, just like z-scores, except a different mean and standard deviation exist. T
scores can be converted from z-scores by multiplying the z-score by 10 and then add-
ing 50 (T score = 50 + 10(z)), as shown in Table 10.4. The T scores retain their exact
positioning relative to the z-scores within the distribution. They just ref lect a different
scaling, with numbers exceeding 50 above the mean and numbers less than 50 below the
mean.
Other standard score scales are associated with the normal curve and are used in the
fields of education and psychology (Glass & Hopkins, 1984). Normal curve equivalent
(NCE) is a standard score that has a mean of 50 and a standard deviation of 21.06, which allows
for a 1–99 range configuration just like the percentile equivalents in a normal distribution. In addi-
tion, NCEs and percentiles do match up at central points within the distribution (i.e.,
1, 50, 99); however, it is essential to recognize that NCEs are equal interval scores and
percentiles are not.
Wechsler or Deviation IQs reflect a standard score format with a mean of 100 and a stan-
dard deviation of 15. This scaling format is common with aptitude and achievement test
batteries. The College Entrance Examination Board (CEEB) standard scores that
are reported with the SAT Reasoning Test or the Graduate Record Exam (GRE)
include a mean of 500 and a standard deviation of 100.
Although all of these standard scores possess different means and standard deviations,
they share a common scaling, which allows scores from different tests to be matched
to the same locations within the normal curve distribution. For instance, if a student
receives a standard score of 115 (which is one standard deviation above the mean) on one
test and another student gets a CEEB score of 600 (which is also one standard deviation
above the mean), both of these scores will be located one standard deviation above the
mean or at the 84th percentile on the normal curve. The normal distribution provides a
common distribution reference map for standard scores no matter what standard scale is
used in the construction of various tests or by different test publishers.
A special type of standard score is the stanine, which is a term derived from the
compilation of the words “standard” and “nine” (Worthen, White, Fan, & Sudweeks,
1999). With this standard score, the normal curve is divided into nine separate segments,
or stanines, and within each of these stanines—which are half a standard deviation wide
except for 1 and 9—a set percentage of cases are found (Gronlund, 1985).
Stanine Percentage of cases
1 4
2 7
3 12
4 17
5 20
6 17
7 12
8 7
9 4
The stanine serves as a quick and easy reference point for a test performance along the
normal curve, with stanines 1, 2, and 3 representing the below-average range; stanines
4, 5, and 6 constituting the average range; and stanines 7, 8, and 9 ref lecting the above-
average range. Overall, the stanine approach provides a general indicator of performance
relative to other obtained performances, based on an assumed normal distribution.
a comprehensive final that all second-year Spanish students must take. They also took a
diagnostic skill test at the beginning of the year. Mrs. Smith wants to examine the rela-
tionship of student performance between the diagnostic test and the comprehensive final.
Did students who performed well on the early skill test also perform well on the final?
Conversely, did students who performed poorly on the skill test also perform poorly on
the final? As you can imagine, these are important questions, and a correlation coefficient
can examine and describe the relationship between those two different performance sets.
As was mentioned brief ly in Chapter 4, the examination of the relationship that exists
between two scores or measures is called a correlation coefficient. A correlation coeffi-
cient, recognized as r, is used to represent both the direction and the strength of an assumed
linear relationship between two events or variables. In the field of education, the Pearson
Product-Moment Correlation is typically used to generate a correlation coefficient. The
following formula (Figure 10.7) is used in generating that correlation coefficient:
N g XY 2 1 g X 2 1 g Y 2
r5
"Ng X2 2 1 g X 2 2 "Ng Y2 2 1 g Y 2 2
The good news is that teachers are rarely placed in a position where they must gen-
erate a correlation coefficient, and even if that was the case, graphing calculators have
made that task relatively easy. The more likely need for teachers is to be able to accurately
understand and interpret correlation coefficients that exist in current research, profes-
sional journal articles, or test manual reports.
The correlation coefficient value can range from +1.0 to –1.0. A coefficient of +1.0
represents a perfect positive relationship between the measures (both measures either
increase or decrease in the same direction). A coefficient of 0 represents no linear rela-
tionship at all between the measures, and a coefficient of –1.0 represents a perfect inverse
relationship between the measures (one measure increases as the other decreases).
The interpretation of the strength of a correlation coefficient depends on the type of
information that is collected, and no specific interpretation guidelines or rules exist.
However, the following guidelines serve as a good overall indication of strength for
correlation coefficients: A correlation coefficient at the level of +.80 or higher indicates
a strong positive relationship and a coefficient of +.51 to +.79 represents a moderately
strong relationship, while a coefficient at the level of +.50 or lower indicates a weak
positive relationship. The same strength indicators apply to negative coefficients. Conse-
quently, a correlation coefficient of –.65 has the same strength as a correlation coefficient
of +.65. The only difference is in the direction of the relationship among the examined
variables or factors.
In order to get a visual image of an examined relationship, a scatterplot is often com-
pleted (Figure 10.8). According to Glass and Hopkins (1984), “The chief purpose of the
scatter diagram is for the study of the nature of the relationship between two variables”
(p. 81). A scatterplot involves the graphic presentation of data that can be used to determine
whether a potential relationship (including strength and direction) exists, and is constructed by plot-
ting the scores from the two different tests along an x- and y-axis. The score for each student is
plotted at the intersection of the two scores, and then the overall pattern is examined.
100
90
80
The direction of the data pattern indicates what kind of relationship exists (e.g., positive,
negative, curvilinear, or no relationship).
For example, if a straight ascending line (scores or ratings increase or decrease in
the same direction on both axes) is generated, then a perfect positive correlation would
be demonstrated. If, however, a straight descending line is generated (scores or ratings
decrease on one axis while increasing on the other axis), then a perfect negative correla-
tion would be demonstrated. When reviewing the plotted points of a scatterplot, a line
is typically drawn through the points (referred to as a regression line, since the plotted
scores are examined as they regress or fall back toward this line) in order to help deter-
mine if the data points have a positive or negative direction.
When reading scientific studies or reviewing test manuals, correlation coefficients
are sometimes used to help support some finding or specific quality of a measure. It is
important to recognize, however, that correlation should never be interpreted as causa-
tion (that which produces a particular effect). Two factors or events may be related, perhaps
even strongly related, but that doesn’t mean one factor necessarily causes the other.
Also, correlation coefficients are based on the assumption that linear relationships
exist between the examined factors. Therefore, if a nonlinear relationship exists (e.g., a
curvilinear relationship), a correlation coefficient will not be an accurate measure of that
relationship. For that reason alone, it is important to recognize assumptions of expecting
a relationship to be linear when in reality it may not be. As a teacher you will rarely need
to calculate a correlation coefficient. However, you will certainly be expected to know
and interpret the strength and direction of correlation coefficients, especially as they are
used to support or explain various educational or psychological findings.
performance as if the same test or performance was generated time and time again. In order to
have confidence in a test score that is generated, it is desirable to have a small standard
error of measurement that ref lects a minimum of measurement error within the test and
in the generation of the test scores. A standard error of measurement is generated based
on the standard deviation of the test scores and the reliability of the test itself (Mason
& Bramble, 1978). This estimate is often reported in test manuals in the form of a +
or – band that surrounds the obtained score because it is not an absolute finding. For
example, a score of 100 with a SEM of 5 really means that this student’s true score, or
the score that would be obtained if the test and measurement were perfectly reliable, falls
somewhere between 95 and 105. The mathematical formula used to determine the SEM
is the following:
SEM 5 SD "1 2 r
SEM = standard error of measurement
SD = standard deviation
r = reliability of the test
Since the distribution of error scores ref lects a normal distribution (Kubiszyn &
Borich, 2003), a score range that contains the hypothetical true score and the confidence
associated with the range of scores that surround the actual obtained score can be deter-
mined. With the example just provided, if we use +1 or –1 SEM from the obtained score
(obtained score = 100, 95–105 SEM score band), you can be assured that 68 percent of
the time the true score exists within that score band or confidence interval.
All too often we as educators interpret test scores as accurate scores, without any mea-
surement error or any imprecision of a student’s performance on some academic measure.
The reality is that at best we can be only reasonably confident of the performance range
of a student based on the SEM. This is important to recognize for teaching purposes as
well as for how the test scores are used for other educational decisions. In particular, with
high-stakes testing, students are often required to obtain certain scores in order to pass or
be seen as competent within certain academic content domains. Therefore, the determi-
nation of cutoff scores for those tests must be carefully selected, along with the recogni-
tion that a score band, which accounts for measurement error, is the most accurate way to
identify a student’s test performance.
In regards to Mrs. Smith’s Spanish test scores, 27 students completed the test, and
all measures of central tendency were the same (81), with a standard deviation of 7.38
reported. Roughly 64 percent of all scores fell within the score range of 74–88 (one
standard deviation above and below the mean). A few very strong performances (90 and
above) were demonstrated, along with a couple of very low performances. At this point,
this is all Mrs. Smith knows about the Spanish test results.
However, Mrs. Smith wants to know how well her students did relative to the test
items and the demands of the test itself. Were some items easier than others? Were there
some items that most of the students missed, and if so, why? At this time, Mrs. Smith
does not have any information on how well the students did on each item of the test
itself. And that is important data to have, because the internal accuracy and consistency
of the items of the test determine the ultimate accuracy and utility of the test scores.
Before an internal review is conducted, it must be assumed that effective test construc-
tion practices have been followed with this test and with all tests that Mrs. Smith generates.
That would include ensuring that proper content coverage has occurred, that good test
item construction practices have been followed, and that instruction-learning-assessment
alignment exists, as well as additional factors like administration issues (e.g., adequate time
to take the test, conducive testing conditions) and avoidance of bias within the test itself.
All of this is important and must exist before a test is given to any group of students.
The internal review of a test is performed because it provides important information to
a teacher. As the teacher, you have a primary interest in knowing how well your students
did on the items of the test. This is necessary for the sake of understanding their progress as
well as whether they have acquired the knowledge and skills identified as essential for their
continued progress within an area of instruction. Consequently, in order to complete this
instructional audit, certain item-analysis procedures need to be used in the review process.
answer by chance alone. For a multiple-choice item, that drops to a 25 percent chance
when a four-response answer set is used. With Item 20 (a multiple-choice item) on the
Spanish test, 24 out of the 27 students got that item correct. The difficulty level of that
item would be .88, and only 12 percent of the students got this item wrong.
So if an item has a high p value (.88), what does that mean? One could determine
that since 88 percent of the students got the item right, it was an easy item, so perhaps
it should not be included in the next test. Assuming that this is not a “good” item,
however, might be incorrect. In fact, it could be a very effective item that differentiates
students who know the information from those who do not. High student performance
on this item may ref lect the fact that the information it measures was taught well; conse-
quently, almost all of the students understood and learned the presented material.
Conversely, with Item 19 (a multiple-choice item) on the Spanish test, only 14 out of
the 27 students, or a little over half, got the correct answer. The difficulty level of that
item would be .52. Is this test question too hard, and should it be eliminated for that
reason? This is a reasonable question to ask, but before any decision is reached, this item
should be reviewed in order to determine whether it possesses any technical f laws or
ambiguity (e.g., confusing wording, missing words) in its structure.
When evaluating the quality and testing intent of test items, it is important to ask
whether you as the teacher would predict that over half your students would not know
the answer to this or any question you are reviewing. Is the item unclear or poorly
worded? Would it mislead or confuse the reader? Remember that if an item is intention-
ally designed to confuse students, the item is not measuring the knowledge or skills of
your students as much as evaluating how well they can recognize and deal with misdirec-
tion. Moreover, navigating confusing test items is not typically listed on a lesson or unit
plan as an educational objective for students to master. Along with writing a clear and
meaningful test item, it is important to make sure other effective test construction and
administration practices have not been violated.
If technical issues are not at fault, the problem may not reside with the test item and
may have more to do with instructional delivery issues. For example, it may be pos-
sible that the information this item covers was not taught in adequate detail or practiced
enough in order to answer the item correctly. If other test items that cover the same or
similar material also have low p values, this possibility becomes a more likely explanation
for the observed results. Instruction-learning-assessment alignment must manifest itself
through the identical match-up of what is taught to what is tested. If test items are con-
structed to measure something different from what was taught, a mismatch will exist that
can be evidenced through test performances of the students.
If the test items look good and the instruction was solid and effective, student factors
need to be considered. Student factors (e.g., motivation, health, test review and prepara-
tion) play an important role in any assessment and must be considered when any test is
reviewed. For instance, as much as we would like to believe that students prepare for
tests, sometimes they don’t, and that can result in less-than-desirable test scores. Some-
times test items are missed because the students simply didn’t know the correct answer,
which is an educational reality. As a teacher you just want to make sure, as much as pos-
sible, that student preparation is the primary deciding factor when evaluating student
achievement and performance in your classroom.
ITEM DISCRIMINATION
Another useful item analysis approach teachers can employ is the use of an item discrimi-
nation index. The item discrimination index involves a comparison of a student’s complete
performance on a test to his or her accuracy on particular items on that test. When an item is
reviewed, it can be identified as one of three types:
♦ A positive discrimination item is more likely answered correctly by students who per-
form better on the entire test (high group) as opposed to those who perform poorly on the test
(low group).
♦ A negative discrimination item is more likely answered correctly by students who
perform poorly on the entire test (low group) as opposed to those who perform better on the test
(high group).
♦ A nondiscriminating item is an item where there is no difference in accuracy between
the test performances of high- and low-ranking groups.
The procedure for determining an item’s discrimination value is straightforward, and
only a small amount of time is required to generate a value. Consistent with the practice
of placing data or scores in some logical order, the student performances based on the
total test score should be placed in order from the highest to lowest score. Two groups are
selected (high and low) based on their scores (highest scores go into the high group and
lowest scores into the low group). A p value is generated for each of the groups based on
the test item being examined. Then the p value from the low group is subtracted from
the p value of the high group. The resulting value is the item’s discrimination value or
index. The discrimination index range is just like the p value range of 0 to 1.
Returning to Mrs. Smith’s Spanish exam, after ordering the 27 scores, the highest 13
are placed in the high group and the lowest 13 are placed in the low group. Since an odd
number of scores exist, the middle score was removed and will not be used to determine
the index. On Item 18 of the test, 12 of the 13 students in the high group answered cor-
rectly. For the low group, only 5 out of the 13 students got this item correct. Subtracting
the p value of the low group from the high group (.92 – .38 = .54) generates an item
discrimination value or index of .54. Most of the students who did well overall on the
test also got this item correct, as compared to the students in the low group, so this item
would be considered a positive discrimination item.
According to Hopkins, Stanley, and Hopkins (1990) (as cited in Worthen, Borg, &
White (1993)), the item discrimination index bands provided in Table 10.5 serve as use-
ful guides in reviewing test items and their relative effectiveness in accurately measur-
ing an intended concept. As the discrimination index increases, so does the value of the
items and their collective ability to differentiate performance among high- and low-
performing students. In general, these values serve as useful markers when reviewing the
discrimination values associated with test items.
As a teacher, you expect to find positive discrimination items, like the example just
provided, within your test, which makes sense since students who study and prepare well
will perform well on the majority of test items. However, when negative discrimina-
tion or nondiscriminating items emerge, further analysis is definitely required. Since
it is unlikely that the least-prepared students would outperform the most prepared stu-
dents, something about the item is likely causing some kind of confusion. Because of this
unusual response pattern, these items are likely candidates to be dropped from the test.
With nondiscriminating items there is essentially a no-difference factor between the
students in the classroom. This is also an unusual occurrence, since some variation would
naturally be expected given the different preparation levels and efforts of students. Non-
discriminating items are sometimes seen as items that are too easy (most of the students
get them right) or too difficult (most of the students get them wrong). However, if a
mastery approach is taken (where all students are trained and expected to demonstrate a
high level of proficiency), nondiscriminating items ref lecting high accuracy (all students,
regardless of grouping, do well on the items) would not necessarily be seen as a limitation
of the test. Overall it becomes important to develop a professional habit of reviewing test
items, especially the negative discrimination and nondiscriminating items, to make sure
they are providing accurate and useful information about your students.
ASSESSMENT IN ACTION
CHAPTER REFLECTION
In the Chapter Activity exercise, you examined how pervasive descriptive statistics are in
everyday life. Beyond the daily impact, as a future educator the role of statistics plays a sig-
nificant part in the professional literature and research that inf luence instructional delivery
and student learning. In addition, statistical data and review procedures are critical for class-
room teachers when examining, evaluating, and grading student work and performances.
As a teacher, you will be using statistical procedures and rendering decisions based on
data generated from those procedures. Since a skill is only developed through practice,
it becomes essential for every teacher to gain a comfort level with statistics, and that can
only happen if this tool is implemented on a routine basis in the classroom.
CHAPTER REVIEW
Conclusion
The collection, review, and analysis of student performance evidence all require the
understanding and use of certain basic statistical and measurement procedures. Under-
standing the qualities of distributions, especially the normal distribution, central ten-
dency, correlation, and variability, allows teachers to be better informed when it comes
to looking at student test scores in the classroom and understanding test results from stan-
dardized achievement batteries, state-mandated tests, individual student evaluations, or
other test information sources. In addition, the use of item analysis procedures, including
item difficulty level (p value) and discrimination value, provides for the internal review
and examination of items that constitute tests and the scores that are generated from them.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. Why is organizing data the first essential step in understanding it?
2. If a colleague asked you to explain and compute the measures of central tendency
and variability (e.g., range, standard deviation) of a classroom test, how would you
do it? Why is it important to know this information?
3. Identify and describe the characteristics of the normal distribution. How is this
distribution helpful to you as a teacher? For example, if a student’s score falls one
standard deviation above the mean in a normal distribution, what does that mean?
4. Identify common standard scores used in educational and psychological tests and
how they are constructed.
5. In a professional article you are reading, a scatterplot is given and a correlation
coefficient of .65 is reported. Describe that coefficient in regards to both direction
and strength.
6. Explain how difficulty and discrimination values are generated for items in your
tests. Why are these procedures useful in examining a classroom test?
RESOURCES
The American Statistical Association (ASA) at www.amstat.org provides useful informa-
tion for K–12 educators. Click on the “Education” button for information on work-
shops, links, and resources designed for teachers to increase their statistics knowledge
base. Ideas for potential lessons and activities that can be used to integrate statistics into
the classroom curriculum are also offered.
The National Center for Education Statistics at nces.ed.gov provides access to elementary
and secondary education data from across the country, as well as educator resources
designed to enhance the use of data in the educational experience for students at the
local, state, and national levels.
X ± M = x x2 z-score
70 79 –9 81 –.97 1 70 2 79 4 9.3 2 5 2.97
66 79 –13 169 –1.4 1 66 2 79 4 9.3 2 5 21.4
65 79 –14 196 –1.51 1 65 2 79 4 9.3 2 5 21.51
63 79 –16 256 –1.72 1 63 2 79 4 9.3 2 5 21.72
1,804
REFERENCES
Glass, G. V., & Hopkins, K. D. (1984). Statistical methods in education and psychology (2nd
ed.). Englewood Cliffs, NJ: Prentice-Hall.
Gronlund, N. E. (1985). Measurement and evaluation in teaching (5th ed.). New York: Mac-
millan Publishing Company.
Hopkins, K. D., Stanley, J. C., & Hopkins, B. R. (1990). Educational and psychological mea-
surement and evaluation (7th ed.). Englewood Cliffs, NJ: Prentice Hall.
Kubiszyn, T., & Borich, G. (2003). Educational testing and measurement: Classroom application
and practice (7th ed.). Hoboken, NJ: John Wiley & Sons, Inc.
Mason, E. J., & Bramble, W. J. (1978). Understanding and conducting research: Applications in
education and the behavioral sciences. New York: McGraw-Hill, Inc.
Worthen, B. R., Borg, W. R., & White, K. R. (1993). Measurement and evaluation in the
schools. New York: Longman.
Worthen, B. R., White, K. R., Fan, X., & Sudweeks, R. R. (1999). Measurement and
assessment in schools (2nd ed.). New York: Longman.
STANDARDIZED TESTS
AND HIGH-STAKES TESTING
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How can results from standardized tests be used to understand and facilitate stu-
dent learning and progress?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Describe the assessment and accountability compo-
nents of No Child Left Behind (NCLB).
♦ Comprehend testing terms such as norm, standard-
ized, norm- and criterion-referenced, age and grade
equivalents, and value-added model.
♦ Interpret test score profiles from commercial
achievement reports.
♦ Distinguish the use of value-added growth models
in the evaluation of student progress.
♦ Understand the importance of test preparation and
test-taking strategies for students.
♦ Facilitate the impact of parent-teacher conferences
by conveying meaningful student performance
information to parents.
Comstock/PictureQuest
280
Introduction
The use of standardized achievement tests to systematically evaluate student learning can
be traced as far back as the 1920s and 1930s. Interest in collecting student test scores across
grade and subject areas has only increased over the decades to the present day, where group
achievement tests with “high stakes” implications exist throughout the country. Over the
past 20 years, the standards reform movement has forced states to establish academic con-
tent or proficiency standards for students, along with connected assessments and tests based
on those standards. At no other time in the history of education has greater scrutiny of test
scores and their use existed in the schools. Consequently, your students are being educated
during the era of accountability, and high-stakes tests have taken center stage.
These tests are identified as high-stakes tests because serious consequences are attached
to them for students, schools, and even states. According to Braden and Tayrose (2008):
Federal laws and policies mandate statewide tests, and mandate consequences for
schools and local education agencies based on test results. Individual consequences
such as grade promotion, the attainment of a high school diploma, and tracking or
allowing students’ achievement grades to dictate the school, program, or class to
which they are assigned are not specified by federal laws or policies, but about half
of the states in the United States elect to attach such consequences to their testing
programs (p. 576).
Pre-service teachers are entering a profession that must deal directly with the focus
and impact of high-stakes testing. This has huge implications for how teachers approach
their instruction, teach, and prepare for required measures and standards of student
achievement. Obviously, it has even greater implications for the students who must pass
these tests and evidence competency across an ever-expanding array of subject fields and
skills.
ASSESSMENT IN ACTION
disadvantaged children who have not previously shown significant academic progress
relative to other groups, particularly on group achievement test measures or other exter-
nal indices of academic accomplishment. This legislation is broad in scope and involves
several interconnected areas (e.g., highly qualified teachers, evidence-based practice,
expanded parental options). However, the most important area, relative to the area of
assessment, is the accountability of student progress and achievement. This accountability
includes several key components including annual testing of students, the documentation
of adequate yearly progress (AYP) of students, and state academic content standards.
Specifically, NCLB requires:
♦ that yearly testing be completed by all students, including those with identified dis-
abilities or limited English proficiency
♦ the testing must be connected to the academic standards that are required in the
state in which the child is receiving his or her education
♦ at a minimum, all states must provide standardized testing in reading and math in
grades 3 through 8 and at least once in high school
♦ testing in the area of science must be conducted at least once in grades 3–5, 6–9,
and 10–11
♦ each state must provide a sample of fourth and eighth graders who have completed
the National Assessment of Educational Progress program in reading and math
every other year
♦ ultimate accountability is attached to mandated testing, in that every state is
required to set its own yearly proficiency goals with the ultimate milestone of all
students who take the test meeting the proficiency standard set by their respective
state by the 2013–2014 school year
Student progress is identified as adequate yearly progress (AYP), as defined by scor-
ing at the proficient or higher levels on state-required tests. This status level must be
demonstrated by all students. Schools must report this progress as a unit as well as for key
groups of students, such as ethnic minority students, students with identified disabilities,
English-limited students, and others. If any group of students does not make the expected
progress over a two-year span (or more), sanctions can be administered to the school and
the school district (Table 11.1). Intense pressure is placed on schools to demonstrate real
and sustained test score improvement for all students but especially for those who histori-
cally have had limited academic success as measured by group achievement tests.
Along with the NCLB expectations of all students learning and making appropriate
academic progress, the use of high-stakes testing as the major indicator of student accom-
plishment in educational decisions has become the norm. Federal legislation requires
annual testing, and the public expects students to do well on these group achievement
tests, since they are based on academic standards that students are expected to master as
part of their classroom instruction. As reported by Braden and Tayrose (2008), “Most
(44) states use tests custom designed to state standards” (p. 578). So teachers as well as
students are involved in a high-stakes learning environment with high-stakes implica-
tions. Make no mistake: the standards-based tests that students are mandated to take and
pass across each respective state directly inf luence the curriculum and instruction that
Consecutive
years failing
AYP Schools Local education agencies
2 School choice (students may choose to Develop a plan for improvement
attend another school in the district;
district must provide transportation)
3 Supplemental educational services Implement plan and state may
(schools must provide free tutoring direct corrective action
services to students who elect to
remain)
4 Corrective action (adopt scientifically- Continue plan and state may
based methods of instruction, train staff direct corrective action
to use effective methods)
5 Develop plan for alternative governance Mandated corrective action (e.g.,
(plan for radical restructuring, external state takeover, dissolution of
contracting, reconstitution) local education agency)
6 Restructuring (essentially closes the
school; may reopen as Charter, magnet,
or other alternative)
Copyright 2008 by the National Association of School Psychologists, Bethesda, MD. Reprinted by permis-
sion of the publisher. www.nasponline.org
takes place in every classroom. That is one of the major reasons why teachers need to be
informed and possess a high level of assessment competency so that the most effective
instructional decisions are made for their students.
NORM-REFERENCED
STANDARDIZED TESTS
Teachers recognize the need to review and interpret student evidence that is collected
in their classrooms. However, that assessment competency must also extend to measures
and tests that are administered outside the classroom, and the primary tests that teachers
encounter and will have to interpret are standardized achievement tests. A standardized
test is any test that is administered and scored in a uniform and consistent way. This is done in
order to maintain reliability throughout the process as well as maintain high confidence
with the test results. The test results are also reviewed and presented within a standard
format so that findings are explained in a comparable fashion despite the variation of
CHAPTER ACTIVITY
Given the high personal impact associated with standardized tests, as a future teacher it
is important to develop a framework and format that can be used when reviewing and
discussing student test results with various audiences (e.g., students, parents, administra-
tors). As you begin to think about what your format will look like, consider the following
components.
• Testing description—a brief explanation of the tests that are administered.
• Covered performance areas—what students are required and expected to do given the
provided test items and tasks.
• Accuracy and confidence—brief statements regarding the validity of the findings and
your level of confidence that these results are accurate estimates of the child’s true skill
and knowledge levels.
• Test profile review—a clear and meaningful explanation of the student’s performance,
with particular emphasis on relative strengths and areas or skills that require continued
effort and practice; in addition, the student’s success in meeting the expected academic
standards and learning goals that the test is designed to measure must be reported and
understood by all parties.
• Learning impact—test results are only helpful if they are used to make decisions that
improve student learning; therefore, how will test results be used to directly impact
student learning in the instruction that you provide?
• Other—what additional components (if any) do you feel are necessary for your test
results script?
As you continue this ref lective process, the framework components will likely evolve
over time. However, at this time, begin to identify and refine the elements that you believe
are necessary in order for your students, their parents, and other interested parties to fully
understand the testing information that will be provided.
individual student performances. The most common examples of standardized tests are
the commercially available achievement and aptitude tests, which are produced and sold
by private test publishers (see Figures 11.1 and 11.2 later in this chapter). In addition,
some state departments of education (e.g., Arizona, Hawaii, Iowa, Tennessee) implement
their own standardized achievement tests as part of their statewide assessment programs.
Most standardized tests are constructed to provide norm-referenced information,
which allows a comparison to a specific age or grade that are usually based on national norms. A
norm group is an identified reference group that serves as a comparison for students who complete
the same test. For example, the standard format for national norms is typically based on
several key variables including age, grade level, ethnic status, and gender. More special-
ized norms can be constructed to be used in student comparisons, and they are usually
identified as specialized norms in the results section of the test. In particular, additional
factors such as school size, suburban versus rural location, and socioeconomic status can
be used in the construction of norms. In addition, local norms that are specific to a
particular school district can also be generated. This allows for important comparisons
among students at individual schools within a district.
With a nationally norm-referenced standardized test, students who take the appro-
priate test for their age or grade level are compared to a corresponding national sample,
which allows them to examine their relative performance as well as identify particular
skill strengths and weaknesses. Therefore, it is important to make sure that the appropri-
ate norms are used with the selected achievement test. It is also essential to make sure the
most recent test norms are used, as this provides for the most accurate comparison of your
current students. The more years that separate the norm group from the students who are
taking the test, the weaker the comparability between the two groups.
When conducting student and score comparisons, it is important to recognize that the
norms that exist within a norm-referenced standardized achievement test serve only as an
academic reference point and do not ref lect the focused acquisition of particular learning
targets within designated academic standards. A performance profile may be provided,
including the number of items passed within a particular skill domain (e.g., word study
skills). However, no extensive or meaningful information within specific skill areas or
on the required performance levels is typically provided. Only a criterion-referenced
standardized test that is comprised of expected skills and performance levels directly connected to
specific content standards is designed to address that issue. Consequently, the emphasis of the
criterion-referenced test is on how well each student does relative to the content of the
test and not on student-to-student comparisons.
on, and not on student-to-student reviews. The natural focus is on learning the skills and
not on “looking at everyone else” during the learning process. A criterion approach also
forces careful instructional preparation on the teacher’s part, as the necessary elements
and skills must be selected and taught in the classroom before they are executed as part of
a formal assessment.
this student got 85 percent of the material correct; it only indicates the positioning of this
score relative to the other scores in the distribution.
In addition, based on the normal curve, the percentage of cases that fall within spe-
cific sections of the normal distribution can be determined (e.g., one standard deviation
above and below the mean = approximately 68 percent of all the cases), along with scor-
ings at particular points along the distribution. The center of the distribution is at the
50th percentile; one standard deviation above the mean is the 84th percentile; and the
98th percentile is two standard deviations above the mean. One standard deviation below
the mean is the 16th percentile, and the 2nd percentile is two standard deviations below
the mean. This approach allows for a student’s performance comparison within the dis-
tribution, in addition to a performance comparison relative to his or her peers.
Standard scores, derived from the student’s set of raw scores, serve as the key report-
ing scores for the test battery. Different standard score formats exist (e.g., T scores [mean
= 50, standard deviation = 10] or Deviation IQ [mean = 100, standard deviation = 15]),
but they share a common scaling that allows for test comparisons between different tests
within the normal curve distribution.
In order to help you become more familiar with various test profile formats, scores,
and interpretations, test profile copies belonging to the author’s daughter are provided
here. Examine the score profile in Figure 11.1 on page 291. At this time, Ashleigh was
an eight-year-old second grader who was given the Stanford Achievement Test Series.
At the top of the score report, information on the student and her scores, along with
her percentile performance bands, are provided. She completed several tests, including
three subtests (Word Study Skills, Reading Vocabulary, and Reading Comprehension)
in the area of Reading, two subtests (Problem Solving and Procedures) in the area of
Math, and additional testing in Language, Spelling, Environment, and Listening. The
selection of these administered subtests was determined by the school that she attended.
As more subtests are included, the overall cost of the test, both in time and money, also
increases.
The number of items for each subtest are provided, along with the raw score (how
many she answered correctly out of the total number of items) and the percentile ranks
and stanines that correspond to each of the subtest performances. Below is a short list of
Ashleigh’s subtest performances.
Subtest Items Raw score Percentile Stanine
Spelling 30 26 88 7
Total math 74 61 86 7
Total reading 118 91 78 7
Language 44 28 58 5
Listening 40 15 8 2
Based on this information, what are Ashleigh’s relative academic skill strengths and
weaknesses? Strong performances were noted in the spelling, total math, and total read-
ing subtests. In particular, she exceeded most of her contemporaries (88 percent in spell-
ing, 86 percent in math, and 78 percent in reading), and each of these score placements is
located in the above-average stanine of 7. Her language performance fell in the average
range yet her listening performance was extremely low, as she only exceeded 8 percent of
her peers on this particular subtest.
Specific subskill performances listed within the content clusters indicate average to
above-average attainments for all subskill areas in spelling, math, and reading. Overall,
she was able to successfully complete the grade level subtests; however, she clearly dem-
onstrated difficulty with the listening subtest. Of the seven subskills that were evaluated,
six of the performances fell in the below-average range.
An Achievement/Ability Comparison is also provided, which compares her achieve-
ment performances against her estimated abilities to see if her academic achievement is
below, comparable to, or exceeds her estimated ability level. Along with the measure-
ment of achievement, some standardized tests include a general measure of ability or
learning capacity. This is provided so the student’s academic performance can be com-
pared to his or her estimated ability to learn. The ability scores are often reported as stan-
dard scores (mean = 100, standard deviation = 15), and age and grade level comparisons
may also be provided.
The provided results present the picture of a high-achieving second grader. In fact,
her performances across all measured areas fall in the high-average to above-average
range except for the listening area, which was extremely low relative to her other perfor-
mances. Based on her estimated abilities as measured by the Otis-Lennon School Ability
Test, both her verbal and nonverbal learning skills fall within the average range relative
to age and grade norms. Given this information, Ashleigh presents as an “overachieving
student,” where her actual academic achievement exceeds her estimated abilities based on
the items from this test. She evidences a consistent achievement profile, except for her
listening performance, where her accomplishments were consistently low in all listening
subtasks. The information provided in the bottom sections gives more detail on what
specific skills were examined and how she performed on each one of those skills. In addi-
tion to national percentile rankings, a qualitative scoring rating of below average, aver-
age, and above average is also provided for each of the separate skill areas.
If you were Ashleigh’s teacher, how would you react to the following score summary?
Would you have any concerns? What about the poor listening performance? Is that a
real issue to examine or just a strange artifact of the testing? As you can tell, this was
an exhaustive testing (i.e., the complete battery was given), and a poor performance on
one or more tests is always possible, especially for young students who are involved in a
formal testing routine over a number of school days. As mentioned earlier, however, it
is essential to consider all pieces of evidence when examining the learning progress of
students.
In this particular case, the low listening performance makes sense and is supported
when other information about Ashleigh is revealed. At the time of the testing, Ashleigh
was receiving speech and language therapy to improve her auditory processing skills.
As a young child, she had a history of significant ear infections that affected both her
hearing and language skills. Although she did not require hearing aids, she did have
interference problems with background noise and not being able to clearly differenti-
ate sounds or spoken words. Consequently, she started receiving language services in
first grade to improve her listening skills. This performance profile makes sense when
the testing results and other personal information are examined, and that’s where the
real benefits are obtained: when all available information is used, it provides for greater
understanding, which allows for more accurate and meaningful decisions to be made in
the classroom.
estimated ability to “average” students and even those who can be classified as gifted. For
that reason alone, great care, restraint, and professional ref lection are needed before this
kind of information is discussed for educational purposes.
Now examine and review Ashleigh’s test scores that were obtained when she was in
the seventh grade (Figure 11.2). Are there any similarities between her second grade and
seventh grade test performances? How would you describe her overall performance? As
her teacher, do you have any instructional concerns, and if so, what are they? How would
you describe these results to other parties, such as parents or other professionals?
Ashleigh’s composite score performance fell in the average range, with most of her
academic skill performances falling in the average to low-average range. However, a
relative strength was noted in her science performance area. In regards to her estimated
abilities, her academic performances are in line with her overall ability estimate. Relative
processing strengths were noted in the quantitative and nonverbal reasoning areas. Her
weaker skills were demonstrated in the verbal ability area.
VALUE-ADDED MODELS
Standardized achievement test scores provide a considerable amount of information on
a student and his or her performance in regards to the particular skills that are measured
as part of the testing. However, the information that is collected provides an educational
“snapshot” of a student’s accomplishments at one specific point in time. Even when test-
ing is completed every year and multiple “pictures” are collected over time, those images
are still separate events and measurements that are not necessarily connected to each
other or to the student in a continuous and uniform fashion. On the other hand, growth
or value-added models (VAMs) are designed to measure the growth of student learning over
multiple points in time in order to examine the learning progress of a student over that period. In
particular, the yearly value that contributes to a child’s learning is measured and can be
compared to other students or from classroom to classroom or from school to school.
The VAM measures a student’s improvement from one year to the next through the
examination of what are referred to as “gain scores” that are generated for each student.
Student gain scores (i.e., individual student progress or growth in achievement over a
specified period of time) can be collected and combined for a classroom or for a particu-
lar teacher, as well as for classrooms within a certain grade level in a school. By following
a value-added approach, it is possible to compare the combined or aggregate gains of stu-
dents across time so that comparisons between classrooms or schools can be made. The
value of a student having a certain teacher can be evaluated based on the growth gains of
the students in that teacher’s class, which is referred to as “teacher effect” (Ballou, Sand-
ers, & Wright, 2004).
VAMs have become popular because they can effectively address the NCLB require-
ments of collecting and confirming the learning progress of all students, especially low-
achieving students. School administrators and officials need to know which students are
making progress and which ones are not across each individual school in a school district.
In addition, at the district level, teachers who are highly effective, as illustrated by the
strong learning gains of their students, can be identified, as compared to those who are
less effective. This information can be used to help less-effective teachers become more
effective through professional development and other district supportive services.
difficult to sort out differences due to student and class selection issues and differences
that are actually due to the impact of a teacher’s instruction. Braun’s message regarding
the potential of VAMs is consistent with the reported reviews. However, his warning to
not rely on a VAM as the only evaluation method for teachers (and their instructional
effectiveness) is warranted and needs to be recognized by all school officials.
TEST-TAKING STRATEGIES
In regards to test-taking strategies, when answering selected response questions (e.g.,
multiple choice), have your students consider these basic selection suggestions.
♦ As they read the test questions, try to have them identify the answers in their heads
before they look at the responses that are provided. If this is done, then their answer
FOCUS ON ASSESSMENT
3. Complete easy questions first and then answer the difficult questions. As
students go through test items, they need to complete those that are easy and those
they are sure they can answer correctly first. This helps build their confidence and
focus their thoughts on related concepts that will likely be examined in other items.
If an item is difficult and the answer is unclear at that moment, skipping that item and
returning to it at a later time is the best option. Upon returning to the item, students
should eliminate any answers that are clearly wrong or don’t fit with the question itself
(for selected response items). They should try to reduce the number of options to two
responses, then go back to the question stem and determine the best answer from the
two options. When dealing with essay or extended response items, students need to
create an outline for their answer and follow a logical order with the points that are
made in the response.
4. Students need to monitor their progress and the testing time. In general,
teachers are good about indicating how much time is left when taking a test. However,
the student is the only one who knows what he or she has done and not done during
a test, so the available time to complete work is a huge factor. Students need to check
their progress relative to time so that they can be assured of completing the test within
the specified time limits (if there are time limits). Fractional time review (e.g., half of
the questions are completed with half of the hour remaining) can be done rather eas-
ily; it just requires a habit of checking progress at specified time intervals (e.g., first 15
minutes, first 30 minutes) on the student’s watch or the clock in the room. In general,
students need to be mindful of the time during the test and not rely on the teacher as
the only source of that information.
5. Students need to review their work. Although the natural inclination is to hand
in a test as soon as it is finished, your students need to learn to resist that urge and
to review their test answers. They need to check to make sure they have answered
all questions. Also, inaccurate markings can occur (e.g., selecting two answers for a
multiple-choice item), and other careless mistakes (e.g., forgetting a decimal point, not
writing out the entire number correctly, missing a punctuation mark) can be identified
and changed during the final test review. Confirming that the student’s name is on the
test is also essential and should be done before the test is handed in.
will hopefully match up with one of the provided answers, which will likely be the
correct answer.
♦ Have them review all the provided responses and eliminate the ones they know
aren’t correct. Once they have selected an answer, encourage them to stick with it,
as their first choice is generally the correct choice.
♦ Make your students sensitive to question options that do not grammatically fit the
question stem. Also, if an option is unknown to them and seems to be out of con-
text with the question, they should disregard that item. If two options are given
that are direct opposites of each other, there is a strong possibility that one of those
responses is correct.
© Royalty-Free/CORBIS
♦ The use of negatives in test items can be confusing, so encourage your students
to get in the habit of dropping the negative and then reading the statement and
deciding whether it is true or not, and then reexamine the statement in the reverse
position.
When providing constructed responses (e.g., written responses for short answer, essay,
or extended response questions), your students need to be aware of the following sugges-
tions and strategies.
♦ Make sure they pay close attention to the directions so that they are providing what
is asked for. For example, they need to know whether they are supposed to answer
all of the essay questions or just a certain number of them (e.g., answering three of
the four questions).
♦ They should always make sure they know what a question is asking them to do. If
they don’t understand any part of a question, they should be encouraged to ask the
teacher for further clarification.
♦ Make sure they directly answer the question. Don’t have them provide information
(e.g., personal ref lection or opinion, unrelated facts) that is not asked for or is not
needed to answer the stated question. They need to provide the requested informa-
tion in a simple and straightforward manner that is easy to understand and follow.
Also, if answers are handwritten, they should write or print carefully because messy
responses are harder to read and are consequently likely to receive lower grades.
♦ Have them state their main point in the first sentence of their response, followed
by supporting information and evidence of that statement. Make sure the specific
information that is provided (e.g., basic information, appropriate research) is useful
and connected to the question. Responses that are concise, clear, and well orga-
nized are well received by teachers.
♦ If time permits, have them review and proofread their work. Identifying gram-
matical errors and sentence structure issues always provides for a better product and
a more meaningful written response.
In recognition of this impact, teachers need to encourage and cultivate interest and
support for parents within their classrooms. Being connected is a critical component,
and the strength of that connection will be tested each time a parent-teacher confer-
ence takes place. Moreover, under both NCLB and Individuals with Disabilities Educa-
tion Improvement Act (2004) legislation, parent involvement is mandated; they must be
directly involved in numerous and meaningful activities that directly relate to the educa-
tion of their children (Miller & Kraft, 2008).
Even though electronic communication has become prevalent, when parents desire
genuine understanding and direct answers to their questions, they will often seek out
face-to-face contact with their child’s teacher. In fact, the most likely situation in which
teachers will be asked in-depth questions about test scores and other school performance
issues is a parent-teacher conference. Therefore, it becomes very important to use these
conferences as opportunities to strengthen the family–school connection and the parent–
teacher bond in particular.
A parent-teacher conference is recognized as a formal school meeting where a teacher com-
municates with parents about their child. Although the parents and teacher can meet for any
reason, the usual reason for meeting would be to discuss school progress and classroom
issues. Through these conferences, a professional foundation is established, and hopefully
a solid relationship starts to develop that helps to serve and support the academic success
of the child. When conducting a parent-teacher conference, there are some essential steps
you’ll want to take. In preparing for a conference where test scores and other sources of
learning evidence are discussed, consider the following suggestions and then determine
how they might be utilized in your future parent-teacher conferences.
♦ Always greet parents with genuine warmth and acceptance. The success of a con-
ference, or any meeting for that matter, will be determined to a great extent by
how well everyone works together. There will always be hidden agenda items,
but everyone at the meeting should have one primary goal in mind, and that is to
provide the best opportunity for “Johnny” or “Mary” to be successful in school.
In fact, it is strongly suggested that you open your meeting with that statement or
something similar to it, and that you keep this message as the central focus of the
meeting, particularly if divergent opinions and positions emerge.
♦ You will be talking about someone’s child, so be careful with your words and
descriptions. As a parent, it is sometimes very hard to hear things about your child,
especially if the information that is shared is not positive. The key is to be caring,
yet focused on the purpose of the meeting. Provide objective and useful informa-
tion so that all parties are fully informed and can work constructively together. Be
clear in your own mind about the purpose of the meeting and stay focused on that
purpose. Allow for discussion and the exchange of ideas, but don’t lose control and
discuss unrelated events. Time is a valuable commodity for everyone, so you will
need to keep the discussion on task and moving in a purposeful direction.
♦ Be sensitive to the real possibility that for some parents, parent-teacher conferences
are going to be uncomfortable, and that this may having nothing to do with you
personally. Oftentimes parents relive their own school experiences through their
children, and these kinds of meetings can spark old memories and hard feelings that
still may exist. Also, you will likely meet parents who have difficulty with the English
language and who will have difficulty following what you are saying. Make sure you
use understandable words when talking and especially when explaining test results
and what they mean. Having information available in a clear written form (e.g., out-
line, student performance summary, visual diagrams) can also be very helpful.
♦ As the teacher you are in the best position to provide accurate information about a
student, and that includes standardized test results. You can provide the best overall
picture of a student’s progress. You can indicate how current test results are consis-
tent, or in some cases inconsistent, with what is observed and demonstrated in the
classroom on a daily basis. What you say will matter, so make sure you are clear in
what you want to say and how you say it.
♦ Preparation for parent-teacher conferences is essential. Before you meet with par-
ents, make sure you understand all the information that will be shared during the
meeting. The level of preparation should be such that you can easily describe the
test results without having to refer to any particular section of the test or any refer-
ence material to explain the findings. Also, be sure of where you want to go dur-
ing this meeting. Is this meeting just to review the obtained test results, or is there
some other purpose? Be clear in your mind what you want to be able to accomplish
during this time and follow a plan that will allow that to happen. As the teacher, it’s
important to listen to and answer any questions that parents have, but you will also
need to direct this meeting and make sure the goals are met.
♦ If test results are to be reviewed, it is important to have several copies of those
results available, along with any additional information (e.g., handouts, pamphlets,
textbooks) that could be used to provide further explanation for various scores or
testing terms. For instance, sometimes drawing a normal distribution and placing
scores accurately within that distribution provides a helpful visual guide for parents
in understanding where their child’s performance falls relative to other children
who have taken the test.
♦ Operate on the assumption that most parents will have little (if any) knowledge
about test scores and how to interpret them. This is usually a correct assumption.
One of the main tasks you will have in a meeting like this will be to provide a
mini-lesson on how to read and understand test scores, because the parents will
be relying on you to do just that. They may even say they know what the scores
mean and how they are obtained, but many actually don’t. So if the parents you are
meeting with are limited in their understanding of test scores, they will need your
help and expertise. And in order for you to do this well, you must first be comfort-
able with this material and what it means.
♦ When test results are discussed, it is important to provide information on both the
student’s strengths as well as areas that need improvement. It is natural to point out
the weaknesses, but remember to provide a balanced presentation, because every
student has strengths that need to be acknowledged and recognized by others.
Also, it is also important to emphasize that test scores, although they may be well-
constructed measures, are only (at best) good estimates of what a child understands
and is currently able to do within the measured skill areas. Scores should not be
viewed as permanent, “carved-in-stone” findings, or as perfect ref lections of stu-
dent performances.
♦ When describing test scores, percentiles and standard scores are most useful. A
normal distribution can be used to help explain the location of various test per-
formances. In regards to percentiles, a quick note of caution: Many individuals
confuse the term “percentile” (the percentage of cases that fall at a particular point)
with “percent correct,” which are not the same thing, so be sure to provide a clear
explanation of a percentile.
♦ Since most standardized tests generate information on a student’s achievement and
learning ability, these areas will need to be clearly differentiated and explained.
Information on how the ability scale is determined should be available. Also, a brief
explanation of the expected relationship between a student’s estimated abilities and
academic performance (i.e., a student should be working up to his or her estimated
abilities) may prove helpful for parents.
♦ If the conversation moves from reviewing the test scores and results to a discussion
of what things need to be done to improve the child’s performance in the class-
room, it is important to make sure that any plans that are made go beyond words
and into direct actions. Identify the goal that is to be obtained and the required
actions that will be needed in order to reach this goal, as well as the people respon-
sible for helping to reach this goal. Always end on a positive note and thank the
parents for their participation.
♦ Should the child be included in a parent-teacher conference? That’s a good ques-
tion, and it depends on the age of the child and the nature of the meeting. In
general, it’s not a bad idea as long as everyone knows the child will be expected
to participate in the meeting and not just sit there. Also, the child doesn’t have to
be involved in all meetings, so a selective attendance policy could be followed.
Whether you choose to include or not include the child in the meeting, remember
that strong, positive parental involvement sends an important and powerful mes-
sage to a child.
♦ Always remember that teachers need to conduct themselves in a professional
manner throughout a meeting. That may sound easy, but there will be times
when discussions get “hot,” and blame might be thrown around. The best thing
that can be done in those circumstances is to remain calm and stay focused on the
primary goal of the meeting: What needs to be done to make Johnny or Mary
as successful as possible in the classroom? Part of a teacher’s professionalism starts
with demonstrating a genuine interest in the child and the family. Being orga-
nized and objective, yet caring, during the conference helps to ref lect the impor-
tance of the meeting and in most cases gains the appreciation and respect of the
parents.
CHAPTER REFLECTION
As part of the Chapter Activity exercise, you were encouraged to develop a testing results
script and framework that can be used when reviewing standardized test results with
students and parents. What test review elements do you plan on covering when you meet
with students or parents? Within any chosen framework, providing a clear and accurate
picture of a child’s performance is essential. However, communicating those results is a
very human enterprise, and how information is presented to students and parents is just
as important as the information itself. A genuine, caring approach is needed in order to
develop and maintain parental cooperation, especially when the results indicate progress
problems or difficulties. As teachers we must recognize that communicating is one of our
most important professional skills, and that is especially the case when providing mean-
ingful yet sensitive test performance findings.
CHAPTER REVIEW
Conclusion
Due to the high impact and political energy associated with test scores, teachers need
to be knowledgeable about standardized achievement tests and the testing profiles and
performance information that are generated from them. Since all students are involved in
mandated district testing, teachers must be competent to interpret the test results that are
generated. In addition, they must be able to answer testing questions from parents, and,
more importantly, to determine the impact of those test scores on their students as they
relate to intended academic learning goals and outcomes. With continued practice, test
score interpretation skills directly benefit every child and parent.
As VAMs are further integrated into the formal evaluation of student learning prog-
ress, teacher awareness, knowledge, and understanding of these models will need to
increase. In order to obtain the most accurate and optimal test results, the testing stage
needs to be carefully constructed, and test preparation should always be an important
part of any formal assessment process. Students need to be as fully informed as pos-
sible regarding a testing event, and test preparation practices allow for greater student
understanding and awareness. Finally, parent-teacher conferences ref lect an important
part of the service that teachers provide to the general community. This dynamic inter-
play serves as an important relationship conduit. For a teacher, developing and enhancing
a positive relationship with parents and the general community is time very well spent.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How has NCLB inf luenced the current instructional practices of the classroom
teacher? As a teacher candidate, do you feel that you are being fully prepared for
the testing and accountability challenges that accompany this legislation? Why or
why not?
2. If asked by a colleague or by a parent, how would you describe the basic test scores
that are typically reported with standardized achievement tests? How should
information from formal tests be integrated with other sources of student perfor-
mance evidence?
3. Why have value-added models become so popular in education? What do you
need to do in order to be fully informed about these models?
4. How do you make test preparation content meaningful to students in the
classroom?
5. Do you feel that you could conduct a parent-teacher conference if you had to?
Construct an outline of the plan that you would follow in that conference. What
steps would you follow and why?
RESOURCES
The National Center for Research on Evaluation, Standards, and Student Testing
(CRESST) at www.cse.ucla.edu is a valuable educational resource that includes infor-
mation on standardized testing and current high-stakes testing issues. In addition,
current educational issues and research are provided that would benefit every educator.
At the Council of Chief State School Officers (CCSSO) site at www.ccsso.org, consider-
able information is provided on value-added models that are being implemented in
various states, along with the current progress of those state models, general findings,
and future directions. To access this information, type “value-added growth models”
into the search box.
The TeacherVision site at www.teachervision.fen.com provides considerable information
for teachers, including useful parent-teacher conference materials. Type “parent con-
ference resources” in the search box for an extensive list of materials and articles to help
better prepare for parent-teacher conferences.
REFERENCES
Ballou, D., Sanders, W., & Wright, P. (2004). Controlling for student background in
value-added assessment of teachers. Journal of Educational and Behavioral Statistics, 29(1),
37–65.
Braden, J. P., & Tayrose, M. P. (2008). Best practices in educational accountability: High-
stakes testing and educational reform. In A. Thomas & J. Grimes (Eds.), Best practices
in school psychology V (pp. 575–588). Bethesda, MD: National Association of School
Psychologists.
Braun, H. I. (2005). Using student progress to evaluate teachers: A primer on value-added models.
Princeton, NJ: Educational Testing Service.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How can I effectively assess the learning progress of students with identified learn-
ing needs in my classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Identify the federal legislation that
has mandated the provision of special
education services to qualified students.
♦ Understand the role and purpose of an
individualized education plan (IEP) for
each identified student.
♦ Distinguish the major special education
disability categories.
♦ Identify the potential assessment chal-
lenges connected with students identi-
fied with special learning needs.
♦ Apply assessment modifications in the
areas of presentation, time, setting,
and response for students with special
learning needs.
309
Somos RF/Getty Images
Introduction
For a regular education teacher, this chapter and its content may seem out of place. Why
should regular education teachers who teach in the major content areas (e.g., science,
math, English, social studies) or specialty areas (e.g., art, music, health, physical educa-
tion) need to be concerned about assessment issues that involve special education stu-
dents? Don’t special education teachers work with students with identified learning needs
in their classrooms? The answer is yes; however, these students are also served by regular
education teachers in the general education classroom.
Special education teachers help serve the learning needs of students who evidence a
wide variety of disabilities and conditions. However, while some of these students are
served in special classrooms, for the majority of their instruction during the school day
many are served in the regular classroom by way of instructional collaboration between
special and regular classroom teachers. If you think about it, all students have unique
learning needs in the classroom; it’s just that these particular students meet legal qualifi-
cations to receive special instructional services and resources.
The seclusion model that characterized special education delivery in the past has been
replaced by an inclusion-focused model (mandated through the Individuals with Dis-
abilities Education Act [IDEA], which was first passed 1990, reauthorized in 1997, and
refined again as the Individuals with Disabilities Education Improvement Act [IDEIA] in
2004), where instructional service delivery occurs in the regular classroom to the fullest
extent possible. Because of this instructional focus, if it is in the best interest of an iden-
tified child to be served in the regular classroom, that is where he or she must receive
instruction, with the support of both special and regular education teachers. Therefore,
all regular education teachers need to be as informed and skilled as possible in order to
deal with the wide range of student needs that will likely exist in their classrooms.
Moreover, No Child Left Behind (NCLB) works in conjunction with IDEIA, as stu-
dents with identified disabilities are required to meet general education standards and
requirements. Remember that special education is really about services that are provided
to students, and not about a specific place or class. More importantly, special needs stu-
dents have the same constitutional right to be educated in the regular education class-
room as their nondisabled peers do.
Regular education teachers contend that they do not receive enough information, train-
ing, or support in regards to special education students or at-risk children (Bennett, Deluca,
& Bruns, 1997; Brice & Perkins, 1997). This is a national phenomenon, and given the rise
of inclusionary practices, the instructional and assessment pressures on regular education
teachers have only increased. Consequently, it is essential that regular education teachers
continue to increase their knowledge and skill base in regards to the complexities of serv-
ing special education students within the regular classroom setting. In order to help address
this important need, this chapter provides basic information about IDEIA and other legisla-
tive mandates (e.g., the Rehabilitation Act of 1973 Section 504), individualized education
plan implementation issues, general characteristics and needs of special education students,
and modifications of the assessment process for these students.
In addition, with the nationwide increase in the number of students who are identi-
fied as bilingual and for whom English is not the dominant language (English as a second
language [ESL] students or English language learners [ELL]), teachers need to be prepared
to address the unique learning requirements for these students. Within the field of educa-
tion today, almost every general education classroom contains students who represent a
broad range of abilities, experiences, and acquired skills, so significant instructional chal-
lenges exist for every teacher. When it comes down to it, regular classroom teachers need
to be able to serve the instructional and assessment needs of all students, whether they
qualify for special services, school assistance programs, or any other educational resource.
An individualized education plan (IEP), the official school document and action plan
designed to facilitate the educational progress of a student identified with special learning needs, is
generated by a committee that typically includes the student’s parents, regular and spe-
cial education teachers, educational support staff (e.g., school psychologist, counselor,
speech/language therapist), administrators, and any other personnel that are important to
the child’s case. The IEP itself contains several components:
♦ the current functioning status of the student, complete with estimated knowledge
and skill levels and short-term learning objectives
♦ annual and long-term goals (e.g., post-secondary instruction, vocational or trade
school training)
♦ types of special services provided (when, for how long, and by whom)
♦ evaluation findings and progress monitoring results
♦ the student’s qualification and program delivery
Teachers contribute to the development of the IEP by helping to create a clear picture of
current skills and progress (complete with identified strengths and weaknesses), so that
immediate and future learning targets specific to the individual needs and learning pro-
file of the student are generated.
The IEP serves as a learning contract for the student and his or her teachers. All teach-
ers, including special and regular education, are held responsible for making sure that it
is followed and that progress is monitored as ref lected in the short-term learning goals.
Assessment of student progress is a major requirement, as progress toward the identified
goals must be made on a regular basis (e.g., formal student review must occur at least once
a year). The assessment of the learning progress of the identified student may involve the
use of standard measures that are provided to all the students in the classroom, as well
CHAPTER ACTIVITY
As a regular education teacher, you may have wondered what is special or unique about
the training that a special education teacher receives. In a special education program, both
general and specialized courses in special education are provided, and many programs are
structured along a fifth-year post-secondary or graduate instructional level. Within the
program, specific courses in educational psychology and child growth and development,
along with training in specific strategies for teaching students with disabilities, are typi-
cally offered. A half year or in some cases a full year is spent student teaching in a classroom
supervised by a certified teacher.
In order to better understand the background and training of your special education
colleagues, make an effort to talk with current special educators in the field. When an
opportunity to interview a special education teacher presents itself, find out what instruc-
tion or experiences they’ve had that help them address the needs of special education stu-
dents. Also, ask them what advice, if any, they would give as you begin to prepare for the
special needs challenges that will face you in the classroom.
as modified instruments or procedures that are designed for the particular needs of that
student.
If warranted, testing accommodations may also be made as part of the IEP for stan-
dardized achievement tests that exist as part of state- or district-wide testing programs.
Because of IDEIA, a student’s IEP must document how he or she will participate in
the mandated testing. There are three options: (1) participate with no modifications,
(2) participate with modifications, or (3) participate in an alternative assessment (Braden
& Joyce, 2008).
As a pre-service regular education teacher, it is important to recognize that you will
be teaching in one or more classrooms that will include students who have been identi-
fied with special needs. No matter your content expertise or grade range (e.g., elemen-
tary, middle, or secondary), you will be teaching special needs students who may present
a variety of disabling conditions and learning issues. And even when a group of students
share a similar identification such as a learning disability, considerable variation will
exist among those students in regards to their particular learning problems and instruc-
tional needs. It is essential that you keep this in mind as the basic disability categories are
described in the next section.
3. Developmental Delay
4. Emotional Disturbance
5. Hearing Impairment
6. Mental Retardation (Intellectual or Cognitive Disability)
7. Multiple Disabilities
8. Orthopedic Impairment
9. Other Health Impairment
10. Specific Learning Disability
11. Speech or Language Impairment
12. Traumatic Brain Injury
13. Visual Impairment
To qualify for special education services under these disability categories, it must be
demonstrated through a formal review process that the severity of the condition is to
such an extent that it adversely affects a student’s educational performance and cannot be
adequately addressed with the boundaries of the regular classroom setting.
It is important for teachers to be informed about these conditions, since they will likely
be instrumental in helping to identify children for potential services and be involved in
the review and evaluation process itself, as well as potentially be teaching and assessing
students with these special needs. The more knowledgeable and confident you are about
special education issues, the more effective you will be in your own classroom. A brief
look at each of these disability categories, as defined and described within the federal
qualification guidelines, follows.
Autism
Autism is a pervasive developmental disability that is typically evident by the age of three
and is typically characterized by significant difficulty in communicating with others and
understanding and responding to spoken language, as well as relating socially to others
or to common objects or events (Sattler, 2002). Additional behaviors associated with
autism include repetitive body movements or behavior patterns, unusual play with toys
and other objects, resistance to environmental changes or daily routines, extreme behav-
ior f luctuation, and atypical responses to sensory experiences or stimuli.
Early diagnosis and supportive educational instruction within the regular classroom
setting are very important for students with autism. Classroom learning, through mul-
timodal instruction (e.g., visual, auditory, kinesthetic), most often centers on the main
goals of effective communication, social interaction, behavioral control, and daily func-
tioning. Providing a consistent and organized classroom and home routine are essential
for children with autism. Related service personnel (e.g., speech and language therapists)
often play a major role in the educational process of autistic students, and their services
are often provided within the daily routine and confines of the classroom. With effective
classroom instruction and community support services, adults with autism can transition
effectively to living and working independently within the community setting. That is
especially the case for individuals who are diagnosed with Asperger' s syndrome, a
condition associated with the high-functioning end of the autism spectrum, as they do not evidence
the speech and cognitive issues that are typically associated with autism and generally
demonstrate greater success with the academic and social demands of school (Volkmar &
Klin, 2000).
Deaf/Blindness
Deaf/blindness includes students whose combination of hearing and visual impairments
are so severe and generate such problematic communication, developmental, or educa-
tional needs that they must be served in programs that address both sensory impairments.
Students who qualify under this category usually require intensive instruction, extra
resources, and highly-trained teachers and support personnel.
Integration into regular education classrooms is more challenging with students who
demonstrate these needs; however, it can be successfully done. Services that can be pro-
vided can range from speech and language therapy and auditory training to hearing aids
and the use of sign language and the reading of lips. Closed-caption films and videos,
as well as the use of aides to take notes for the student during instruction, can also be
part of the services that are provided for students with a hearing disability. In the area of
teaching, a combined or total communication approach that involves oral communica-
tion (e.g., speech and lip reading) and manual communication (e.g., signing and finger-
spelling) is followed in most classrooms.
Developmental Delay
The qualification of developmental delay is restricted to preschool children ages three
through five who have documented developmental delays, as measured by diagnostic
instruments and age-based measures, in one or more of the following areas: physical
development (fine and gross motor skills), cognitive development, communication devel-
opment (speech and language skills), social development, emotional development, or
adaptive development (self-care skills).
Since young children develop at varying rates, the occurrence of developmental delays
is not unusual. It is hoped that with early intervention services, a developmentally-delayed
child can catch up with and maintain the progress of his or her peers. By the time the
child reaches school age, a reevaluation must be completed, and if services are still war-
ranted the child must meet the qualifications of one of the other existing disability cate-
gories. Early assistance and intervention services are provided so that the majority of these
children will not need special education services when they enter elementary school.
Emotional Disturbance
Under the qualification of emotional disturbance, a student must demonstrate one or
more significant problems over a long period of time (generally over a period of several
months) to a severe degree, and it must directly affect his or her educational performance.
In addition, a student must demonstrate one or more of the following characteristics:
♦ learning problems that are not due to intellectual, sensory, or health-related conditions
♦ difficulty making or maintaining appropriate and necessary interpersonal relation-
ships with those around them, such as peers and teachers
♦ experiencing feelings or demonstrating behaviors that are considered inappropriate
for a particular setting or situation
♦ a consistent state of unhappiness or depression or a tendency to demonstrate physi-
cal symptoms and fears connected to a student’s personal or school-related problems
This can include a student who is identified as schizophrenic or someone who can-
not differentiate reality from nonreality on a daily basis. However, it does not apply to a
socially maladjusted student (i.e., someone who understands and is in control of his or her
decisions and voluntarily chooses to act against others), unless it is decided that the nature
of the student’s difficulty emanates from a true emotional disturbance and the social mal-
adjustment is a result of that condition.
Given the potential psychological severity of this condition, students with an emotional
disturbance usually require direct and intensive services to address their emotional and
behavioral needs while maintaining their academic progress within the classroom. Coun-
seling and training in developing effective social skills, including self-control strategies,
improved self-esteem, and behavior management, may be necessary. Since this child’s
behavior may impede his or her learning, and may impact the learning of others, behavior
control and management is a prime consideration both in the school setting and at home.
Hearing Impairment
A hearing impairment, whether caused by a birth defect or acquired as part of one’s life
experience, results in a decreased ability to hear and discriminate sounds. Some hear-
ing impairments can be traced to genetic defects, gene mutations, or infectious diseases
(e.g., rubella), or other teratogenic factors during pregnancy. In addition, head trauma
or childhood infections (e.g., chickenpox) can cause hearing impairments, along with
significant and recurring ear infections that result in damage to the eardrum and the
auditory system.
Hearing impairments occur in various ways and affect different components of the
auditory system. However, regarding the eligibility guidelines for special education ser-
vices, there are three major types of impairments that are recognized, including (1) con-
ductive, (2) sensorineural, and (3) mixed. Conductive hearing impairment or loss
involves the interference of sound transmission through the outer or middle ear to the inner ear.
This can be caused by blockages in the ear canal, infections in the middle ear (e.g., otitis
media), f luid in the Eustachian tube, bone vibration problems within the middle ear, or
other related issues. Some conductive hearing problems (such as ear infections) are tem-
porary, while others may be chronic and contribute to actual hearing loss. Since conduc-
tive hearing issues center around sound collection and amplification, the use of a hearing
aid is often recommended.
Multiple Disabilities
A child who qualifies under the category of multiple disabilities demonstrates a severe
learning condition due to simultaneous handicapping conditions (e.g., mental retarda-
tion and blindness, mental retardation and an orthopedic impairment), which require ser-
vices that exceed the resources of a special education program designed for only one of
the identified impairments. This category does not include the category of deaf/blindness.
Although a small percentage of students—approximately 2 percent of those receiving spe-
cial education services—are identified with this category, given their unique learning and
functional needs, considerable resources and classroom arrangements are often necessary.
Orthopedic Impairment
An orthopedic impairment involves a severe injury to the skeleton or limbs of a student
that adversely affects his or her educational performance. This can include genetic-based
diseases (e.g., muscular dystrophy), congenital abnormalities (e.g., spina bifida, clubfoot),
disease-related impairments (e.g., bone tuberculosis), or other problems such as cere-
bral palsy, spinal cord injuries, or other structural issues. Students with these conditions
often demonstrate physical motion problems, such as isolated or general paralysis, poor or
unsteady gait, weak muscle control, loss of limb movement, and other related problems.
Speech and language production can also be affected.
One of the more common conditions recognized under this category is cerebral palsy.
Cerebral palsy is used to refer to neurological-based disorders that emerge early in a person’s
life that affect body movement and coordination. Common symptoms include poor mus-
cle coordination (referred to as ataxia), extremely tight or stiff muscles and uncontrollable
motor movements (spasticity), and an atypical walking gait (e.g., toe or foot dragging, toe-
walking). There is no cure for cerebral palsy; however, with the use of various treatment
options, including physical and occupational therapy, medication, corrective surgery, and
assistive devices (such as braces or wheelchairs), the condition can be effectively managed.
Richard Hutching
This formal determination does not usually occur until students reach school age,
since these skills only begin to emerge at that time. Despite possessing average or above-
average intelligence, skill delays or unusual processing tendencies are typically noticed
first by teachers, as well as parents.
In the past, in order to qualify for services, a significant discrepancy between one’s
ability and actual achievement (i.e., aptitude-achievement discrepancy model) had
to be demonstrated. However, that approach is now being replaced by an alternative,
intervention-based model (referred to as response to intervention [RTI]) that evaluates
the effectiveness of research-based intervention on student achievement and progress.
RTI is now considered the selection model of choice in the schools. With this approach,
if a student does not respond to specific and leveled academic or behavioral intervention
within the regular classroom setting, special education services would be considered.
The disability determination is made at the school level, and regardless of the identi-
fied academic area (basic reading skills, reading comprehension, written expression, oral
expression, listening comprehension, mathematics calculation, or mathematics problem
solving), if the evaluation team decides that a student qualifies as having a learning dis-
ability, he or she is eligible for special education services. Specific learning disability
represents the largest special education category, as approximately half of all students in
special education are identified as having a learning disability (Lichtenstein, 2008).
in providing intervention services and therapy. These services are often provided in the
classroom setting as part of the normal instructional day.
Visual Impairment
A student can qualify for a visual impairment, ranging from partial sight to blindness,
if even after corrective measures (e.g., glasses) the condition adversely affects his or her
learning. Visual impairments can range from partially sighted, with some field of vision
limitation, to low vision, which is considered a severe visual impairment (e.g., unable to
read newsprint with the aid of glasses), to legally (i.e., less than 20/200 vision) or totally
blind. A visual impairment or vision loss must be diagnosed by an ophthalmologist or
optometrist.
Children with visual impairments are usually identified early, and classroom supports
such as special computers and programs, large-print materials, books on tape, and Braille
books are available that allow for a considerable amount of their instruction to take place
in the regular classroom. In addition, children with visual challenges can and should be
commensurate with their peers if resources and effective early intervention services are
provided (Sattler, 2002). Therefore, the collaboration between regular and special educa-
tion teachers is vital for the success of these students.
© Compstock/PunchStock
Section 504
Another piece of federal legislation that can potentially serve any student is Section 504
of the Rehabilitation Act of 1973. Section 504 protects individuals with handicapping condi-
tions from being discriminated against, and that includes students and staff in school districts. Under
ASSESSMENT IN ACTION
• Daily notes and feedback were sent back and forth between Mrs. Wood and Bob’s
mother.
The assessment adjustments, which were relatively minor and easy to implement, proved
to be essential for Bob as he began his important transition back into the classroom. As you
might expect, Bob’s problems didn’t disappear at the end of second grade. In fact, chal-
lenges continued to follow him through elementary and middle school. But the assessment
accommodations and modifications that were implemented by Mrs. Wood established an
important foundation for Bob and his future academic progress.
this federal law and within the educational arena, reasonable and appropriate accom-
modations must be provided so that students with identified disabilities can access and
benefit from the same educational opportunities as any other student.
It is also important to recognize that Section 504 defines “handicapped” as “a physical
or mental impairment which substantially limits one or more of such person’s major life
activities,” and major life activities are defined as “functions such as caring for one’s self,
performing manual tasks, walking, seeing, hearing, speaking, breathing, learning and
working” (Rehabilitation Act of 1973, § 706).
In regards to legal coverage, Section 504 covers all students with handicaps, and that
includes those who are identified through IDEIA as well as through the Section 504
review process. However, unlike special education qualification, a student’s educational
performance does not have to be adversely affected by the disability in order to be eli-
gible for accommodations within the classroom (Miller & Newbill, 1998). The student
only has to be recognized as meeting the definition of handicapped in order to receive
potential modifications. As part of the Section 504 review process, a school evaluation
must be conducted, and if it is determined that a student meets the requirements of hav-
ing a handicapping condition, an action plan to address his or her needs must be gener-
ated and implemented.
ASSESSMENT MODIFICATIONS
In today’s classroom, a teacher faces the challenge of teaching a wide variety of students
who present with unique learning needs. Therefore, a teacher’s competency in adjusting
instructional delivery while simultaneously assessing learner progress has never been more
important. For instance, the increase of ELL students has required schools and teachers to
incorporate a broader array of instructional options. Some school districts pursue immer-
sion programs at all program levels (e.g., elementary, middle, and secondary) where ELL
students may be involved in small guided reading groups during the school day or
have extended instructional time provided to them (e.g., extended kindergarten pro-
grams) during or after school hours. At higher grade levels, ELL students may par-
ticipate in one or more classes of intensive English learning, combined with online
FOCUS ON ASSESSMENT
have been created as part of the assessment activity. Following uniform or standardized
procedures during an assessment allows for the most accurate interpretation of a student’s
educational performance.
On the other hand, some standardized procedures may not be appropriate for all stu-
dents and may actually disadvantage certain students (e.g., a student with a disability,
ELL) in regards to the accurate measurement of their learning progress or skill levels. For
this reason, it is important to examine whether assessment accommodations are appropri-
ate and necessary. Accommodations should be viewed as deliberate attempts to provide
a level playing field for assessment so that a student with a disability is able to perform
the same task as someone without a disability. Also, since many students with identified
disabilities may have predetermined testing modifications stipulated in their IEP, as may
students with Section 504 action plans, teachers must be able to make those required
adjustments within the assessment measures and procedures in the classroom. Therefore,
whether the modifications are mandated through an IEP or become evident upon review
of the assessment process itself, when considering assessment accommodations, make sure
that these actions are completed:
♦ develop a justified reason to modify the assessment process
♦ measure expected knowledge and skills
♦ ensure the accommodation is appropriate and fair to the student
Outcome Measurement
When dealing with accommodations, the standard assessment process is changed to
address the particular needs of the student. These adjustments can take a variety of forms
(e.g., use of a computer instead of requiring handwritten responses, an oral quiz instead
of a written one). However, the modified procedures still must measure how effective
the student is in obtaining the same or comparable learning outcomes, just as the original
assessment process is designed to do. If they do not, two different sets of learning objec-
tives and outcomes will exist and will need to be documented and monitored within
your classroom.
Different learning or performance standards can exist if a student’s IEP stipulates
learning objectives that are different than the ones typically required within a specific
content area and grade level. That is why careful examination of every identified child’s
IEP is a necessity for every teacher. By law a teacher must ensure that a student’s learning
progress is measured against the listed educational objectives and outcomes in the IEP.
Nevertheless, because of the strong emphasis of NCLB and other school accountability
measures, the expectation is that all students, including those with special needs (ELL,
Section 504, etc.), will meet the same or comparable learning and academic content stan-
dards as other students. Only in extreme situations are exemptions from standardized
testing or other formal measures of student achievement approved, and even in those rare
cases, other forms of evidence must be provided to show that the exempted student is
making appropriate and measurable learning progress.
For the classroom teacher, the reality is that every student, whatever his or her unique
learning challenges may be, must demonstrate consistent and measurable growth in
regards to academic and performance standards stipulated within individualized learn-
ing plans (e.g., IEP or Section 504) or learning outcomes that have been approved by
each respective school district. Therefore, even when accommodations are permitted for
special needs students, it is important for teachers to require the same high level of excel-
lence and skill proficiency that is expected for the entire classroom.
Assessment modifications should eliminate or greatly minimize the effect of the
child’s disability but still require the attainment of the desired learning outcomes.
When attempting to measure those learning outcomes, teachers can change different
components of the assessment process. For example, how the assessment is presented
or set up can be changed. The administration time and testing setting, as well as the
actual student response, can be modified based on the needs of your students. These
modifications are consistent with the assessment practices and standards that profes-
sionals (e.g., psychologists) must follow, as ref lected in Standards for Educational and Psy-
chological Testing (American Educational Research Association, American Psychological
Association, & National Council on Measurement in Education, 1999).
basically two types of tests: speed and power. A speed test is designed to see how many
items can be completed within a specified amount of time. The focus is on getting as
many items answered as possible with an emphasis on completion speed. A good example
of a speed test is the completion of mental math facts. Here students answer problems
involving all multiples of a number (e.g., 4 × 1, 4 × 2, 4 × 3) within a specific time
limit (e.g., 60 seconds).
On the other hand, a power test is not designed around a time factor as much as it is
centered on obtaining the optimum performance from a student. Consequently, time is
not the deciding factor when examining student performance. A power test utilizes an
extended time range where most if not all students would complete the test. Providing
additional time to complete a test (e.g., in class or after class) also ref lects a power per-
spective. With this approach, an emphasis is on the quality of the responses, especially
the student-generated responses, across all the items in the assessment. Clearly, when
students don’t feel the overwhelming pressure of time during an exam, they can concen-
trate on providing high-quality work, and that is true for all students, not just those who
receive special accommodations.
Identified students may also need to complete their assessments and work in a special
or modified setting. This could be as simple as placing a student in a corner or next to the
teacher to minimize distractions during a formal testing event. Another common option
is changing a group assessment experience to an individual assessment experience. Some-
times, students are best served outside the classroom. For instance, a highly distractible
student may require the supervision of an aide in an isolated or empty classroom in order
to complete a test. This approach can mutually benefit the student as well as the rest of
his or her classmates. However, proctoring a test when the setting is switched can be a
little tricky. For example, in most cases an aide is allowed to proctor a group test, but
a licensed teacher must proctor formal or state achievement tests for special education
students. As a teacher, creating a classroom environment that allows for and supports the
effective testing and evaluation of students is certainly an important goal. A classroom
free from environmental distractions and that has a positive testing atmosphere serves all
students well.
Response modifications should ref lect good common sense in the selection of alter-
native ways of having students respond to test items or questions. For example, teachers
may give oral instead of written exams, use a tape recorder to collect answers, or break
a test into two or three parts for students with attention-based issues. In essence these
students are ending up at the same educational destination as everyone else; they are just
taking a different route in getting there. With this approach, everyone in your classroom
will possess the opportunity to perform at their highest levels regardless of their indi-
vidual needs and experiences.
Sometimes teachers require identified students to only complete a portion of a test or
assignment, because fatigue or motor production issues make completing the entire test
difficult. However, this option should be carefully considered, because when sections of
a test are not completed, it alters what is truly being measured by the test. If the decision
is made to complete a smaller version of the test, it is vital to make sure all areas of the
test are still represented, only with fewer items for each section. Even with this approach,
it is still likely that the smaller test version will not adequately measure the knowledge
and skills that are utilized in the original test. Instead of considering a reduced test for-
mat, administering the original test in sections over two or more testing periods may be
the better option. This way the integrity of the test and what it is assessing is maintained,
while the student’s test completion difficulties are minimized or even eliminated.
are not receiving the accommodation are not disadvantaged by the recording modifica-
tion. In essence, with the administration of this test, all of the students have an equal
opportunity to perform at their highest level. It’s also important to recognize that just
because a student has an identified disability, that does not automatically mean that he
or she needs to receive assessment accommodations. It will depend on each individual
case, and if it is determined that assessment modifications are needed, then they need to
be modified for each respective case. Obviously one size doesn’t fit all, and that is par-
ticularly the case when considering accommodations for students identified with special
learning needs.
The Universal Design for Instruction approach is just another example of how assess-
ment process and practices have continued to integrate with instructional delivery. When
you talk about teaching, you can’t help but talk about assessment. And that is particularly
the case when teaching students with special learning needs. Assessing and monitoring
student progress provides the essential evidence that is necessary in order to document
learning status and the effectiveness of the instruction that they receive. In this way,
students with special learning needs are provided a genuine opportunity to demonstrate
what they have learned and to be assessed in a fair and equitable manner.
CHAPTER REFLECTION
At the beginning of the chapter, a description of the background and training that special
educators complete was provided. In addition, you were encouraged to consult with your
special education colleagues, particularly in matters that deal with students who have
special needs.
Sometimes, consultation opportunities are rare in schools. Therefore, a chance to
receive professional advice is now provided. Examine the following suggestions from a
special education teacher, Ms. Coles, as you prepare to teach and serve students with spe-
cial learning needs in your classroom.
♦ Be willing and approachable in collaborating with special education teachers on
classroom assessments.
♦ Tests and other assessment activities may need to be adjusted in order to fit the
needs of the student, so accommodations will likely be necessary.
♦ Use discretion when providing accommodations to students with special needs;
they may feel awkward about the process, so any attempts to minimize public
awareness of it would be beneficial.
♦ Use the data from the classroom assessments that are conducted; make the assess-
ments relevant so that the information can be used to guide your instruction.
♦ Regarding student performance, continue to enrich those students who exceed the
target performance as well as support and remediate those that fall short of the tar-
get performance.
How would you incorporate the information and advice from Ms. Coles into your
instructional delivery model? How will this inf luence your assessment system and collec-
tion of student learning evidence?
CHAPTER REVIEW
Conclusion
Regular education teachers must be prepared to teach children who present with special
learning needs. It is imperative that all teachers become as informed and knowledgeable as
possible so that the best possible services can be provided. Awareness of disability catego-
ries, student issues and needs, and the importance of the IEP, along with potential assess-
ment modifications, will be necessary in order for teachers to provide those services well.
progress. Students with disabilities typically bring unique learning and assessment issues
to the classroom setting. Being as fully informed about each child, his or her unique
strengths and weaknesses, and the specific details of his or her IEP is vital in recording
the learning progress of each identified student.
Making appropriate changes in the presentation of information, increasing time allo-
cations for assessments, providing alternative assessment settings, and making student
response modifications are all viable considerations if a justified reason exists for modi-
fying the existing assessment procedures in the classroom. The key is ensuring that the
recommended change is fair to the student receiving the modification as well as to his or
her classmates who are still following the standard assessment procedures.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How has special education legislation inf luenced the instructional practices of the
regular classroom teacher?
2. As a teacher candidate, do you feel that you are fully prepared to meet the instruc-
tional needs of students with special learning needs? Why or why not?
3. If asked by a colleague or by a parent, how would you describe the purpose and
function of an individualized education plan or IEP?
4. Brief ly describe the major disability categories and potential instruction-learning-
assessment challenges that may exist in serving students that fall under these
categories.
5. Describe how you could provide assessment modifications relative to presentation,
time, setting, and response for students with disabilities.
RESOURCES
General Disability Resources
The Council for Exceptional Children at www.cec.sped.org provides links to a variety of
resources for professionals as well as parents and children. Also, divisional sites associ-
ated with various disabilities are available through this Web site.
Autism Resources
Wrong Planet at www.wrongplanet.net is a Web community centered on issues sur-
rounding individuals with autism, Asperger’s syndrome, ADHD, and other pervasive
developmental disorders.
Deaf/Blindness Resources
The National Federation for the Blind Web site at www.nf b.org is the largest membership
organization of blind people in the United States. This site provides information on
a wide range of topics including education, research, and technology applications for
people who are blind.
disabilities. It provides services and support for families and individuals, including
over 140,000 members connected to more than 850 state and local chapters across the
nation. A family resource guide, policy and legislative issues, employment opportuni-
ties, insurance coverage, and more are available on this site.
The Parentpals.com Special Education Guide at www.parentpals.com is an extensive
interactive resource where parents and professionals share information and support on
various subjects surrounding mental retardation.
ADHD Resources
Children and Adults with Attention-Deficit/Hyperactivity Disorder (CHADD) at www
.chadd.org is a comprehensive Web site that provides information and updates on areas
such as research and treatment on individuals with ADHD. Professional publications
for teachers and parents are also available from this site.
REFERENCES
American Educational Research Association, American Psychological Association, &
National Council on Measurement in Education. (1999). Standards for educational and
psychological testing (3rd ed.). Washington, DC: Author.
American Psychiatric Association. (2000). Diagnostic and statistical manual of mental disor-
ders—Text Revision. Washington, DC: Author.
Barkley, R. A. (2006). Attention-defi cit hyperactivity disorder: A handbook for diagnosis and treat-
ment (3rd ed.). New York: Guilford Press.
Bennett, T., Deluca, D., & Bruns, D. (1997). Putting inclusion into practice: Perspectives
of teachers and parents. Exceptional Children, 64, 115–131.
Braden, J. P., & Joyce, L. B. (2008). Best practices in making assessment accommodations.
In A. Thomas & J. Grimes (Eds.), Best practices in school psychology V (pp. 589–603).
Bethesda, MD: National Association of School Psychologists.
Brice, A., & Perkins, C. (1997). What is required for transition from the ESL classroom to
the general education classroom? A case study of two classrooms. Journal of Children’s
Communication Development, 19, 13–22.
DuPaul, G. J., & Stoner, G. (2003). ADHD in the schools: Assessment and intervention strate-
gies (2nd ed.). New York: Guilford Press.
Education for All Handicapped Children Act of 1975 (Public Law 94–142), 20 U.S.C.
1400 et seq. (1975).
Faul, M., Xu, L., Wald, M. M., & Coronado, V. G. (2010). Traumatic brain injury in the
United States: Emergency department visits, hospitalizations and deaths 2002–2006. Atlanta,
GA: Centers for Disease Control and Prevention, National Center for Injury Preven-
tion and Control.
Individuals with Disabilities Education Act of 1990, Pub. L. No. 101-476, 104, Stat. 1142
(1990).
Individuals with Disabilities Education Act Amendments of 1997, 20 U.S.C. 1400 et seq.
(1997).
Individuals with Disabilities Education Improvement Act of 2004, 20 U.S.C. 1400 et seq.
(2004).
Katz, J., & Mirenda, P. (2002). Including students with developmental disabilities in gen-
eral education classrooms: Educational benefits. International Journal of Special Education,
17(2), 14–24.
Lichtenstein, R. (2008). Best practices in identification of learning disabilities. In A.
Thomas & J. Grimes (Eds.), Best practices in school psychology V (pp. 295–317). Bethesda,
MD: National Association of School Psychologists.
Miller, L., & Newbill, C. (1998). Section 504 in the classroom: How to design and implement
accommodation plans. Austin, TX: PRO-ED, Inc.
No Child Left Behind Act of 2001 (ESEA), Pub. L. No. 107–110 (2002).
Rehabilitation Act of 1973, 29 U.S.C. 701 et seq. (1973).
Sattler, J. M. (2002). Assessment of children: Behavioral and clinical applications (4th ed.). San
Diego, CA: Jerome M. Sattler, Publisher, Inc.
Scott, S., Shaw, S., & McGuire, J. (2003). Universal Design for Instruction: A new para-
digm for adult instruction in postsecondary education. Remedial and Special Education,
24, 369–379.
Volkmar, F. R., & Klin, A. (2000). Diagnostic issues in Asperger syndrome. In A. Klin,
F. Volkmar, & S. Sparrow (Eds.), Asperger syndrome (pp. 25–71). New York: Guilford
Press.
Witte, R. (1998). Meet Bob: A student with traumatic brain injury. Teaching Exceptional
Children, 30(3), 56–60.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How can response to intervention (RTI) be used by teachers to help increase stu-
dent success in the classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should
be able to:
♦ Identify the intervention process of
RTI in addressing student learning and
behavioral issues in the classroom.
♦ Discuss the legislative requirements of
evidence-based practice in regards to
classroom instruction.
♦ Follow the problem-solving steps and
tiered levels of RTI implementation.
♦ Apply the RTI process when confronted
with academic or behavioral concerns in
the classroom.
♦ Describe the instructional necessity of
RTI and action research by educators in
documenting intervention effectiveness.
© Creatas/PunchStock 339
Introduction
Federal legislation has now mandated the use of evidence-based practices in meeting the
instructional needs of children in the classroom. For example, No Child Left Behind
requires scientifically-based strategies for instruction and intervention when students
experience learning difficulty in the classroom. As Reschly (2008) has identified, there
are several critical impact features associated with NCLB, including:
(a) emphasis on improving low achievement in general, remedial, and special edu-
cation; (b) endorsement of scientifically based instruction (SBI) and scientifically
based reading research principles (SBRR); (c) focus on prevention of achievement
problems and antisocial behavior patterns; and (d) improving general outcomes for
all children and youth including SWD (students with disabilities) in areas such as high
school completion, passing high-stakes state tests, and entering post-high school
educational and career preparation (p. 9).
In addition, the Individuals with Disabilities Education Improvement Act (2004) (IDEIA)
requires early and documented intervention services before any referral can take place.
Also, the evaluation of a student’s response to intervention services must occur as part of
that process.
In response to these issues, a systematic and verifiable approach to address learning
problems, complete with verifiable evidence, was considered necessary. Response to
intervention or RTI, a method of school intervention based on the review of student data in
making educational decisions (sometimes written as RtI), has emerged to assist with the
evidence-based mandates that have accompanied the recent NCLB and IDEIA legisla-
tion. As reported in Tilly (2008):
[A]n emerging evidence base suggests that implementation of a three-tiered system
is effective at remediating academic and behavioral problems for a significant num-
ber of students (e.g., Burns, Appleton, & Stehouwer, 2005; Colvin, Kame’enui, &
Sugai, 1993; Lau et al., 2006; Torgesen et al., 2001; Vellutino et al., 1996; Walker
et al., 1996). In addition, ongoing research is demonstrating that tiered models of
service delivery can produce important improvements for special populations such
as English language learners (e.g., Healy, Vanderwood, & Edelston, 2005; Linan-
Thompson, Vaughn, Prater, & Cirino, 2006) and minority populations (e.g., Hosp
& Reschly, 2004; Marston, Muyskens, Lau, & Canter, 2003; Shinn, Collins, &
Gallagher, 1998) (p. 33).
Beyond the legal mandates, the expectation of providing evidence of effectiveness is now
held within the education field as it relates to the provision of instruction and services to
children. Because of its preventive and proactive design, RTI provides direct evidence
on student performance and improvement toward identified learning outcomes, which is
helpful to all involved, including students, teachers, and parents.
Based on the major policy work Response to Intervention: Policy Considerations and Imple-
mentation (Batsche, Elliott, Graden, Grimes, Kovaleski, Prasse, et al., 2006) that was pub-
lished by the National Association of State Directors of Special Education, several core
principles recognized as essential to the RTI process are examined throughout this chap-
ter. These ideals ref lect the following beliefs and actions:
♦ all children are capable of learning and we as educators can provide effective
instruction during a learning event
♦ intervention must occur early in the educational careers of students
♦ a leveled or tiered intervention model responsive to the learning needs of the stu-
dents can address problems early on in the learning process
♦ RTI is centered around a problem-solving process whereby concerns are clearly
identified and substantiated, interventions implemented, and progress monitored in
order to determine the value of the intervention services provided to students
♦ in regards to interventions, the use of supported research-based, scientifically vali-
dated treatments need to be followed to ensure the highest probability of success
♦ student progress must be monitored in order to confirm learning progress and
direct instruction in the classroom
♦ educational decisions must be based on the collection of relevant student progress
evidence or data
♦ assessment should focus on three purposes: (1) screening to identify students who
require support and assistance; (2) diagnostic review to determine specific student
skill performances; and (3) progress monitoring to evaluate the effect(s) of provided
interventions
CHAPTER ACTIVITY
RTI is a relatively new area of practice for teachers. It is an evidence- or data-based pro-
cess that is used to determine the effectiveness of instructional changes for students in the
classroom, and is connected to the NCLB legislation and requirements of providing early
support for school-age children who experience learning difficulties. During your next
school visit, check out how RTI is being implemented in that school or district. If it’s not
presently being implemented, find out where school officials are in this process. As a future
educator, become familiar with the activities and approaches that are being used to help
students at all levels of need. As you review the school’s procedures, try to answer the fol-
lowing questions.
• Are intervention phases or levels clearly identified and fully explained?
• What are the primary interventions being used within the system?
• Does a school-based team review and coordinate interventions for the students?
• Do written or Web-based materials exist that explain the RTI or other intervention
process at this school?
• How and when are parents informed and how do they become involved in the RTI or
other intervention process?
Also, try to determine how professional development is being implemented (e.g., fac-
ulty in-service) to help build awareness and support for RTI among teachers in that school.
© Creatas/PunchStock
Was the
What
intervention
intervention
successful in fixing
needs to be
the problem? If
implemented in
not, consider a
order to address and
new or modified
fix this problem?
plan of action.
FIGURE 13.1 PROBLEM-SOLVING/CLASSROOM
INTERVENTION CYCLE
progress review meeting is convened; the establishment of a timetable for collecting stu-
dent data and the review of that data; and the identification of the persons responsible
for intervention implementation and progress monitoring. The identification of school
resources and materials is necessary, as they will be directly connected to the action plan.
Some of the most important elements of the intervention program are the performance
indicators that will be collected and used to measure task completion and effectiveness
of the intervention process. Employing valid and reliable measures that provide accurate
estimates of student accomplishment and progress are essential in determining the overall
effectiveness and utility of the current intervention approach.
With the data generated from these measures, the critical element of evaluating the
intervention can then take place, which is really what this process is all about. Was the
plan implemented as it was designed, and is it working? For school building A, are 70
percent or more of the students meeting their expected academic standards?
In this step, the implementation of the action steps is examined, along with the col-
lected data on the student performance outcomes. If the goal was met, the interven-
tion will likely continue, with student progress being monitored and reviewed. If the
goal was not met, decisions to continue or modify the plan (indicating new actions to
be taken) will need to be made. This review can involve any or all of the aforemen-
tioned steps. The key is constructing and implementing an intervention that generates
the desired effect where the number of students meeting the academic performance stan-
dards increases and is maintained.
As an intervention-based approach, Tilly (2008) describes RTI as a leveled support
system consisting of three tiers:
1. Tier 1: Core instruction for all students, which is preventive in nature and involves
effective grade-level instruction relative to the standards and benchmarks assigned
for those students (serves approximately 80 to 90 percent of all students)
2. Tier 2: Supplemental instruction/intervention involving identified at-risk students who
receive specific assistance with identified needs, typically through small group
instructional activities and support (serves approximately 10 to 15 percent of all
students)
3. Tier 3: Intensive instruction that is designed to provide high-intensity, individually-
designed learning supports for each student who reaches this level (serves approxi-
mately 1 to 5 percent of all students).
This system applies to both academic and behavior issues. With this approach, all stu-
dents are served, with only those students whose needs require more intensive services
moving to the higher tiers in this model (Figure 13.2).
Tier 1
As a future teacher, you are dedicated to providing the best possible instruction to all your
students. Even with the best of intentions and resources, however, one or more of your
students will likely experience learning problems or may demonstrate behavior problems
that interfere with their learning progress and that of their classmates. Unfortunately,
students may present with low achievement or inconsistent learning progress due to
factors such as illness, frequent school changes, family issues, or other related factors.
Because of this reality, schools must be able to identify students who are not making the
expected academic or behavioral progress compared to their classmates. Consequently,
the data from screenings and other recognized measures (e.g., state assessment data,
curriculum-based information, or early reading intervention programs like the Dynamic
Indicators of Basic Early Literacy Skills [DIBELS]) are used to help make instructional
decisions about those students (estimated to be 15 percent or fewer of all students within
a district) who are not meeting expected grade level standards, despite the benefits of
effective grade-level instruction, and are in need of more focused intervention. For these
at-risk students, supplemental instruction and support is usually provided in order to
enhance their learning progress so that they catch up and maintain the progress level of
their classmates.
At the Tier 1 level, the emphasis is on providing school-wide instruction and sup-
port for all students. For instance, many states have implemented school-wide positive
behavior support systems in order to teach positive behavior to their students and to
prevent later serious behavior-related problems. This ref lects an instructional component
of desired behaviors that are expected from children and provides corrective feedback
during “teachable moments,” when incidents arise or problems are documented. On the
academic side, school-wide support and intervention can take various forms, but they are
directed at increasing the academic skill base of the students in order to ensure that the
expected academic standards are met, as evidenced by the data that is collected through
achievement performances.
These kinds of activities and an educative and preventive focus on the entire stu-
dent body are ref lected in Tier 1. Basic intentions at this level are to maximize the
learning of all students and to minimize the number of students who will require
intervention assistance. Individuals who are identified as needing supplemental sup-
port (i.e., Tier 2) are typically identified through the periodic academic screenings and
progress-monitoring collection points that are followed as part of the instructional pro-
cess, particularly as it relates to reading instruction. The focus is on the core curriculum
and daily instruction so that all students are learning and acquiring their intended learn-
ing outcomes, standards, and benchmarks. This can be achieved through quality general
education instruction that maximizes the learning opportunities for all students. Infor-
mation on student progress gathered in the classroom, as well as by school-wide sources,
can be used to confirm these learning achievement and behavioral accomplishments.
Tier 2
At this level, the core curriculum and instruction model is still maintained, but supple-
mental help is provided to a relatively small percentage of students (10 to 15 percent or
fewer). As Tilly (2008) has stated, “It is important to note at this level that both core
instruction and something supplemental are received by students. That supplemental
support can range from additional time in the core curriculum to additional opportu-
nities and/or more time to learn, all the way through additional, strategically planned
supplemental instruction” (p. 32). This is a combination general education/instruction
approach, where the children who receive it are viewed as requiring more instruction
before they become proficient with the material. Moreover, Tilly (2008) emphasizes that:
Supplemental instruction in all cases is put in place in addition to core instruction. It
does not replace it. It is usually delivered in groups of three to six students. Often,
30–45 minutes of supplemental instruction are provided to students. Usually 10 weeks
of supplemental instruction might be provided in a cycle with the option of an addi-
tional 10-week cycle being available if student performance warrants it (p. 32).
Students who are identified (usually through district screenings or some kind of
school-based indicator) for Tier 2 intervention are those who are viewed as at risk for not
reaching the academic or behavior standards that exist for their school-age peers. At this
intervention level, supplemental instruction and programs are followed that are identi-
fied as being culturally responsive and scientifically supported. Usually the programs and
lessons are very specific and are customized in the content that is provided, and they are
also directly linked to the core curriculum. As mentioned above, instruction is provided
within the format of f lexible small groups consisting of students who share the same or
similar academic or behavior needs. The instruction-intervention match is critical for
maximizing the potential progress of Tier 2 students. The instruction that is provided
is focused and explicit to the needs of the students, and their resulting progress is moni-
tored frequently and consistently (once or maybe twice a week in some classrooms). This
supplemental instruction is considered a general education intervention and is most often
Tier 3
Children who are identified for Tier 3 intervention services still receive the core instruc-
tion, but they also receive intensive, individual instruction to address their particular
needs and skills. Five percent or fewer of students should theoretically receive this level
of intervention. However, this level does not provide any special status or classification.
Tilly (2008) is quick to point out that:
[I]ntensive instruction does not connote special education. Special education is one
service that might be brought to bear to meet some students’ intensive instructional
needs. However, there will be students who have intensive needs that will not qual-
ify for nor would it be appropriate to provide them special education services. So,
for example, there may be talented and gifted students who need intensive instruc-
tional services who do not qualify for special education. In another example, a student
whose academic difficulties stem from the fact that he is learning English as a second
language may need intensive instructional support, though he may not qualify for
special education services. Tier 3 refers to the need for intensive instruction, not for a
particular program (p. 33).
At this intervention level, a student should possess a written plan that outlines his
or her individualized, research-based supports and instruction. These supports would
be specifically designed to provide explicit and systematic instruction by highly-trained
educational specialists. Each student’s progress is carefully monitored (e.g., weekly) and
reviewed in order to determine the effectiveness of the implemented interventions. In
regards to skill development, especially in the area of literacy, daily sessions involving
one-on-one or small group opportunities to practice emerging skills are usually expected
as part of an intensive intervention program.
ASSESSMENT IN ACTION
(continued)
130
120
110
100
90
80
70
60
50
40
30
20
10
0
The grade-level passages are theme-based (e.g., a day at the zoo) and require the stu-
dents to read a series of paragraphs and record the number of words read correctly to mea-
sure oral f luency. During the reading activity, she often has her students work in pairs, so as
one student is reading, the other student maintains an account of the words read correctly
as well as those that were missed. Mrs. Flower monitors the reading activity, provides cor-
rective feedback when needed (as do the reading partners), and also completes random
reading checks, especially for those students who may not be consistently accurate in their
recordings. The readings are repeated until mastery is demonstrated. Students then move
on to new stories. Each student is expected to graph his or her own daily progress on indi-
vidual reading graphs. According to the students and Mrs. Flower, this activity is effective
and enjoyed by the students. In particular, the students like to see and be able to demon-
strate their reading progress through their daily data graphs. The reading performances of
the students are conducted three times during the school year. This is a simple but effective
activity, and its instructional impact is demonstrated in Figure 13.3.
In Mrs. Flower’s school district, DIBELS is used, and her second-grade class results for
oral reading f luency indicated that 74 percent of the students met or exceeded the bench-
mark goal of 90 words correctly read within a minute. As shown in Figure 13.3, complete
data sets (all three screenings completed) were obtained for 15 of the 21 students in Mrs.
Flower’s class. Data on the third screening was only provided for six students, four of who
exceeded the benchmark and two who did not. These students were likely absent for the
other screenings or moved into the district late in the academic year. What is most impres-
sive is that almost half of the students demonstrated low oral f luency rates (e.g., low 30s) at
the start of the school year. However, considerable progress from the beginning to the mid-
dle measurements was evidenced, while less progress was noted from the middle to the end
of the school year. Perhaps less focus and energy were directed toward this activity at the
end of the school year, or student motivation was not strong. In addition, a few students did
not reach the benchmark goal, so their progress would need to be carefully monitored at
the beginning of third grade. This becomes vital information that Mrs. Flower can review
for her own instructional purposes as well as for the third grade teachers.
Collecting actual student performance data is a very useful tool for a teacher. A class-
room progress “picture” helps to pinpoint and substantiate direct help for those in the most
need. It also becomes a powerful visual presentation of student progress over the course of
the academic year. Data-based activities like the one just described can truly make a tre-
mendous difference in the academic achievement of students and in the documentation of
that progress.
RESEARCH-SUPPORTED PRACTICES
The Dynamic Indicators of Basic Early Literacy Skills (DIBELS), a research-
supported early reading literacy program, monitors student progress (i.e., benchmark assess-
ments conducted at the beginning, middle, and end of each school year) within the basic
foundation literacy skills for reading. This program is designed to measure the growth of
foundational reading skills (initial sound f luency, phoneme segmentation f luency, non-
sense word f luency, oral reading f luency, word use f luency, and retell f luency) in order
to promote reading success and avoid student reading failure by identifying reading dif-
ficulties early on and then providing skill-specific interventions. Reading research has
identified the critical importance for young emerging readers in acquiring the founda-
tional reading skills for later reading proficiency. According to Kaminski, Cummings,
Powell-Smith, and Good (2008):
In the area of beginning reading, Basic Early Literacy Skills include (a) phonemic
awareness or the ability to hear and manipulate the individual sounds in words,
(b) alphabetic principle or the mapping of prints (letters) to speech (individual sounds)
and the blending of these letter sounds into words, (c) accuracy and f luency with
connected text, (d) vocabulary and oral language including the ability to understand
and use words orally and in writing, and (e) comprehension (Adams, 1990; National
Reading Panel, 2000; National Research Council, 1998; Simmons and Kame’enui,
1998) (p. 1182).
As part of early reading instruction and intervention programs, DIBELS has been
incorporated within the reading programs of many school districts across the country,
especially in kindergarten through the second or third grade. The DIBELS measures
Bananastock/AGE Fotostock
serve as powerful indicators of reading progress and help teachers stay on track to pro-
duce effective and successful readers in their classrooms. As stated in Kaminski et al.
(2008):
[This] model is prevention oriented and is designed to preempt reading difficulties
and support all children to achieve adequate reading outcomes by the end of third
grade. The model incorporates conceptual foundations regarding crucial early literacy
skills for assessment and instruction and is focused on empirically validated outcomes
for each early literacy skill. The model builds on reliable and valid measures of essen-
tial early literacy skills (DIBELS) to be used to document growth toward outcomes,
as well as a set of steps for using the data generated by the measures at the both the
individual and systems levels. At the individual level, DIBELS may be used to evaluate
response to intervention within an RTI approach. At the systems level, DIBELS data
may be used to evaluate the overall effectiveness of the system of support that is being
implemented (p. 1199).
Curriculum-Based Measurement
Curriculum-based measurement (CBM) involves the use of quick, standardized curricu-
lum skill assessments to measure student learning progress in the basic skill areas of reading, math,
writing, and spelling. Although originally developed for special education teachers to help
construct IEP goals and document progress, Shinn (2008) reports:
[I]ts use beyond special education quickly expanded to give all educators simple, accurate
tools to scale achievement for both universal screening and progress monitoring within
a problem-solving model (Deno, 1995, 2002, 2003; Germann & Tindal, 1985; Tindal,
Wesson, Deno, Germann, & Mirkin, 1985). More than a quarter century of scientific
evidence has accumulated demonstrating that CBM provides reliable and valid mea-
sures of general achievement (e.g., general reading achievement) that is sensitive to student
improvement and that improves student achievement when used to monitor progress
(Deno, Fuchs, Marston, & Shin, 2001; Fuchs & Fuchs, 1986, 1988, 1999, 2004) (p. 243).
CBM requires the use of probes, which are short skill tests that students complete in
order to examine their skill performances in particular content areas (reading, math, writ-
ing, or spelling). These learning probes, constructed from the instructional curriculum that
is to be learned, serve as learning evidence that students are acquiring the desired skills and
performance levels so that more advanced material and skills can be introduced. They are
generally administered weekly, and the results are visually graphed so that the performance
data can be compared to a set goal or performance level. The use of CBM is particularly
reliable at the elementary level, which is where it is most often used. These authentic mea-
surements of student progress fit within the curriculum of the classroom as they provide
accurate estimates of skill mastery involving both short- and long-term instructional goals.
As reported in Shinn (2008), the core CBM measures consist of six foundation mea-
sures. The Reading CBM serves as a word f luency measure, while the Maze CBM is a
cloze reading task (a reading task where a student must correctly identify the word that
is needed to complete blanks or gaps that are deliberately set within a story or passage).
The Spelling CBM is a straightforward spelling construction task of grade-level words.
The Written Expression CBM requires students to write a story, with credit given for
correctly written and spelled words. The final two measures are math probes. The Math-
ematics Computation CBM involves the completion of grade-level computational prob-
lems, while the Math Applications CBM requires students to provide the correct answers
to presented math problems.
CBM represents the measurement of a student’s performance relative to the instruc-
tional curriculum he or she is taught. It represents the direct assessment of a student’s
academic skills that must be learned and mastered. Educators have come to recognize the
importance of this approach and its utility in helping to monitor academic skill progress
in the classroom.
FOCUS ON ASSESSMENT
Consistent with the RTI approach, PBS is a data-based, problem-solving approach that
operates as a functional extension of applied behavior analysis. McKevitt and Braaksma
(2008) describe PBS as
a structured way to promote positive relationships in schools and to provide stu-
dents with social and behavioral skills to be successful learners and school citizens.
Researchers have demonstrated positive outcomes for staff in terms of regained time
for administrative duties and instruction as a result of not having to address behavior
problems (e.g., Scott & Barrett, 2004) and for students in terms of reductions in chal-
lenging behaviors at all grade levels (e.g., Bohanon et al., 2006; Duda, Dunlap, Fox,
Lentini, & Clarke, 2004; Luiselli, Putnam, Handler, & Feinberg, 2005; McCurdy,
Mannella, & Eldridge, 2003) and in various school locations (e.g., Kartub, Taylor-
Greene, March, & Horner, 2000; Todd, Haugen, Anderson, & Spriggs, 2002) (p. 736).
Through the guiding direction of an identified school team, PBS attempts to change
the existing behavioral climate of the school building while at the same time educating
students according to a new and expected behavior curriculum. The school PBS team
takes a leadership role in implementing this change. For example, the PBS team may
be asked to complete tasks such as assessing the behavioral needs and expectations that
are to be followed in the school, conducting staff training, and monitoring and analyz-
ing behavioral data and information that needs to be reviewed on a consistent basis.
Although school programs vary in their practices and designs, as reported in McKevitt
and Braaksma (2008):
Researchers at the OSEP [Office of Special Education Programs] Center on Positive
Behavior Interventions and Supports (OSEP, 2004) have identified the core features of
a school-wide system as (a) establishing and defining clear and consistent school-wide
expectations, (b) teaching the school-wide expectations to students, (c) acknowledg-
ing students for demonstrating the expected behaviors, (d) developing a clear and con-
sistent consequence system to respond to behavioral violations, and (e) using data to
evaluate the impact of school-wide efforts (p. 740).
Consequently, schools need to identify a set of positively stated behavioral expecta-
tions (three to five behavioral statements) that are taught directly to students. These must
be specific and observable behaviors that are expected to be demonstrated at various
locations throughout the school grounds. Direct instruction of these school rules needs to
take place in the classroom and should be approached like a content lesson, with coverage
of the material, skill practice and demonstration (e.g., examples of desired and undesired
behaviors), corrective feedback, and so on in place to ensure that the students understand
what is expected of them in regards to their behavior. A positive reinforcement system is
recommended in order to encourage students when they demonstrate the desired behav-
iors in school. In addition, a recognized consequence system for violations must exist and
be well understood by the students. The collection of behavior-related data (e.g., disci-
pline referrals, attendance reports, incidence of tardies) can then be used to evaluate the
effectiveness of the PBS program.
the teacher serves as an active researcher in identifying and addressing the instructional
needs of his or her students through this process.
Babkie and Provost (2004) contend that:
When a teacher considers evidence-based practices in making decisions or provides
content based on his or her knowledge of a particular strategy or methodology’s effec-
tiveness, that teacher is conducting research. When he or she groups students based on
performance, the teacher is using research. When he or she evaluates the effectiveness
of different interventions for changing a student’s behavior, it is a form of research
(p. 261).
They go on to say that teachers need to embrace this role as a classroom researcher in
order to address the legal mandates of NCLB for data-driven interventions and evidence-
based practice. However, beyond the required mandates, having a reliable accountability
system that can examine and document student learning progress provides teachers with
a career-long foundation skill. This skill needs to be used, refined, and embedded as part
of the daily classroom experience.
Babkie and Provost (2004) have identified a series of seven practical steps for teachers
to follow when conducting action research and addressing problems in their classroom
(see the box “Action Research Steps For a Teacher to Follow in the Classroom”). Consis-
tent with most problem-solving models, the first step involves identifying the concern, in
objective and measurable terms, that needs to be addressed. When searching to identify
the problem, Babkie and Provost (2004) report:
Probably the best way to identify the problem is to ask, ‘What isn’t working?’ or
‘What can I change?’ This may be based on thoughts about a particular lesson, con-
cern regarding students’ lack of responsiveness to instruction, worry about poor time
on task of one or more students, realization that a student is having difficulty in
SOURCE: Babkie, A. M., & Provost, M. C. (2004). Teachers as researchers. Intervention in School and
Clinic, 39, 260–268. Reprinted with permission.
learning a specific skill, or any other of a number of classroom concerns. The point is
that the question comes out of what is happening in the classroom and is an extension
of what you already do (p. 262).
When collecting and reviewing information, it is important to examine several
sources, including student work samples, error patterns, personal interviews, and direct
observations. Depending on the concern that is being examined, however, some data
collection measures (e.g., observation techniques) may be more useful than others in
determining the nature, extent, and intensity of the identified problem. When exam-
ined, the collective findings should indicate a general trend or pattern in how the student
is responding (or not responding) to the instructional and behavioral demands in the
classroom. Once information is collected, it needs to be analyzed and reviewed. At this
point it is important to have a clear and specific problem identified, with collected evi-
dence that can verify the problem exists.
After the collected information and evidence have been reviewed, an intervention and
plan of action need to be implemented in order to change the current results. Once the
intervention is designed, data needs to be collected during the implementation phase to
determine if the desired effects are being obtained. In other words, is the new approach
working as desired, and does the collected data support its continued usage? If it is not
working as expected, the intervention needs to be adjusted or discarded in favor of other
approaches. The collection and use of authentic data to help make sound educational
decisions is the cornerstone of all evidence-based practices being utilized within school
settings.
Action research methods can be used to critically examine and review daily instruc-
tional events within the classroom. For example, imagine that your fifth-grade class has
just taken a quiz on Newton’s fundamental laws of gravity. In the quiz, the students had
to identify the three basic laws, identify the principles associated with each, and apply
these laws to common events. Overall, the class performance on the quiz generated a
class average and median of 62, which is a much lower than expected academic perfor-
mance for the class.
Based on Babkie and Provost’s (2004) teacher as researcher model, the class perfor-
mance on the gravity quiz constitutes the problem and also serves as the functional base-
line regarding the current learning progress for this specific content area. Through an
item analysis of the quiz, it is discovered that almost all the students missed the items that
dealt with the application of the laws of gravity. In contrast, the test questions that dealt
with identifying and describing the laws were consistently passed. An exit slip completed
on the day of the quiz by your students indicated confusion and difficulty with the grav-
ity application questions. The next day, a class discussion of the quiz generated additional
student comments to support this finding.
After the review of the collected information is completed, you conclude that a lim-
ited amount of time was devoted to examining how Newton’s laws of gravity apply to
our general lives and events. Consequently, the students did not obtain a solid under-
standing of this information, which was ref lected in their weak quiz performance. As
part of your intervention plan, you decide to go back and review that information again,
with an emphasis on the application of those laws. After that is done, a comparable quiz
(covering the same content and skill domains as the first quiz) is administered. The class
performance on this quiz is higher (mean = 85), and item analysis indicates a much
higher pass rate on the previously missed application-type items. Based on the perfor-
mances of these two quizzes, which were designed to serve a formative function, you are
now convinced that the students understand the material and are ready to complete the
unit summative exam that covers this information. By following these steps, you have
made an important data-based intervention that was designed to promote student learn-
ing in your classroom.
In addition, this model can be followed when individual student issues need to be
addressed. Imagine that you have a fifth-grade student who has difficulty staying seated
in the classroom. First, a clear definition of his out-of-seat behavior must be generated
(e.g., no placement in and physical contact with his desk) along with verification by way
of direct evidence that, indeed, he is out of his seat and to a much greater degree than his
peers. Assume that a series of observations, both in the morning and the afternoon, have
been completed to document this occurrence, and he is out of his seat almost 60 percent
of the time during any instructional activity. This rate is much higher than his observed
peers (who are unseated approximately 10 to 20 percent of the time during any instruc-
tional activity).
Unfortunately, his out-of-seat behavior is interfering with his academic performance,
so as the teacher you need to design an intervention that will reduce his out-of-seat
behavior and improve his classroom academic performance. Since we know he is out of
his seat 60 percent of the time during any instructional activity, this level serves as his
baseline, or the level of the behavior before the intervention is implemented.
Now that the problem has been identified and verified within the classroom setting,
an effective intervention plan must now be devised to address this behavior. In particu-
lar, it has been determined that this student responds well to reinforcement, so positive
reinforcing contingencies to be given when he remains in his seat are identified (such as
student praise, behavior points that can be redeemed for tangible items such as candy,
classroom privileges, etc., or other desirable items like being named “student of the day”),
and are provided as part of his intervention plan. That plan and its effects are closely
monitored, and periodic behavioral observations are completed in order to measure the
percentage of time he remains in his seat during instructional activities throughout the
school day. Based on the success, or lack of it, this plan will be maintained or adjusted in
order to obtain the desired outcome, which for this student is to be to out of his seat for
no more than 20 percent of class time (the unofficial out-of-seat class average).
© Royalty-Free/CORBIS
With the passage of IDEIA in 2004, local school districts are able to identify students
with learning disabilities through a response to a scientific, research-based intervention
model like the DIBELS model that was brief ly described. This approach is favored by
special education professionals because it involves the systematic assessment of a student’s
skill progress when provided with quality, research-based general education instruction.
In regards to special education, the identification of students with learning disabilities is
critical, since as reported in Lichtenstein (2008), “As of 2005, the most recent year for
which federal data were available, this amounted to approximately 2.8 million children
in the United States, 5.3% of the school-age population, and 45% of students with dis-
abilities (57% for ages 12–17)” (p. 295).
Regular classroom teachers typically play a major role in helping to identify, through
collected classroom evidence, students who may be eligible for special education ser-
vices. In addition to the data that is collected as part of regular classroom intervention,
additional information through observations, rating scales, and other measures may also
be collected as part of the identification and qualification process. Regular education
teachers also assist in the development and implementation of the IEPs of those identi-
fied students. In fact, they are specified by IDEIA (2004) as a legal part of an IEP team.
Therefore, input from regular education teachers is vital to the process. Since regular
education intervention services are tied directly to students who may eventually qualify
for special education services, it is essential for regular education teachers to acknowledge
their important role in serving all students in their classroom, regardless of any special
services they may eventually require.
CHAPTER REFLECTION
In the Chapter Activity, you were encouraged to discover how RTI was being imple-
mented in surrounding school districts. As school districts continue to move toward and
refine their intervention-oriented service delivery model, the question of who directly
benefits from this approach will inevitably surface. Some will contend that the needs of
certain students (in this case at-risk students) are taking precedence over other students
within the system. Although RTI can be connected to at-risk students or students with
special needs, its first and primary focus is to help all students reach their educational
goals and objectives through providing the best instructional services possible.
Special education eligibility decisions can be a product of the RTI process, but are not
the primary goal. As stated by Bill East, executive director of the National Association
of State Directors of Special Education, regarding myths surrounding RTI implementa-
tion, “The major issues in RTI involve the need to enhance the range and diversity of
academic and behavioral interventions in general education (Tier 1) and to increase the
impact of supplemental interventions in Tiers 2 and 3, not how to make eligibility deci-
sions that divorce special education from general education” (East, 2006). RTI serves all
students, and therefore all teachers need to know and understand the RTI process. RTI is
not just a “special education thing.” In reality it is truly a “general education thing” that
also happens to include addressing the needs of certain identified students.
CHAPTER REVIEW
Conclusion
Due to legislative mandates and professional expectations, teachers need to be able to
provide academic and behavioral intervention for their students, and the use of the RTI
model provides for the needs of all students. A multitiered intervention approach allows
for the provision of support at the level that it is needed. All are served and consequently
all benefit, with no particular individuals singled out in the process. The hallmark of this
approach is that direct student evidence is used to decide which interventions are effec-
tive within the classroom setting.
RTI serves as an intervention safety net that has really never existed for regular class-
room teachers. It provides a mechanism by which all students can be served, and teachers
supported and assisted, within the regular classroom setting. And when every student
gets what they need in the classroom, student success is inevitable.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How has recent accountability legislation (e.g., No Child Left Behind) expanded
the intervention role and practices of the regular classroom teacher?
2. A new colleague arrives in the district and is not familiar with RTI and how it is
implemented. How would you describe the RTI process and the three-tier model?
3. Do you believe RTI will play a significant role in your classroom in regards to
academic or behavioral issues? Why or why not?
4. Of all the core RTI principles, which ones are most relevant and important to you
as a future classroom teacher? Why?
RESOURCES
Response to Intervention (RTI) Resources
The RTI Action Network at www.rtinetwork.org offers evidence-based information and
resources that allow professionals from general education, special education, and school
psychology to provide practical information and support for implementing RTI ser-
vices from preschool to high school. Specific directions are provided on how to create
your own RTI plans, while integrating specific elements (e.g., ELL, learning disabili-
ties identification, student assessment) into that plan.
The National Association of State Directors of Special Education at www.nasdse.org pro-
vides valuable resources on a number of topics, including an RTI blueprint series. Free
documents on how to set up RTI within a building or district are available, along with
additional RTI papers (for a fee). A superb listing of resource links is provided on this site.
The Technical Assistance Center on Positive Behavioral Interventions & Supports (PBIS)
at www.pbis.org is the Office of Special Education Programs site that provides profes-
sional information, research, and support about positive behavioral systems to educa-
tors and schools throughout the country.
The School-Wide Information System (SWIS) at www.swis.org is a Web-based informa-
tion system that can be used to collect and analyze problem behaviors, through office
referral data, within a school as part of a systemwide intervention program. SWIS can
provide useful information on the times and locations of problem behaviors, which can
then be used in generating either individual or group-based behavior interventions.
Intervention Central at www.interventioncentral.org is a private Web site that provides
information on academic and behavioral intervention strategies for the classroom. In
addition, this site contains an RTI wire link that provides information on how to
obtain free and high-quality RTI resources. For instance, click on the “CBM Ware-
house” button for a listing of CBM resources across all content areas (e.g., reading,
math, and written language), as well as useful forms and manuals.
The National Center on Response to Intervention at www.rti4success.org is a national
Web site supported by the U.S. Department of Education, and is devoted to the dis-
semination of RTI information and resources. Included on this site is an RTI library
and an interventions section that rates various programs and approaches according to
key assessment indicators (e.g., reliability, validity, efficiency) for both screening and
progress monitoring purposes. Continuing professional development activities, includ-
ing webinars and summer institutes, are highlighted on a master schedule.
The National Center on Student Progress Monitoring at www.studentprogress.org is a
national Web site connected to the Office of Special Education Programs. A vari-
ety of resources are available, including online training, family support, topic articles
(e.g., CBM, data-based decision making, family resources, math, reading, written lan-
guage), and professional development opportunities. Additional classroom curriculum
resources can be obtained by using the search box and typing in “CBM.”
Academic/Reading Resources
The DIBELS Web site at dibels.uoregon.edu provides information on how to use the data
system and obtain grade-level reading measures and benchmarks, as well as technical
resources. The DIBELS measures are designed to provide fluency checkpoints regard-
ing the development of pre-reading and early reading skills.
AIMSweb at www.aimsweb.com is a Web-based data management and reporting service
that collects and reviews student performance data as part of an evidence-based and
data-driven formative assessment system. Data collected through CBM or DIBELS can
be reviewed and evaluated through this program.
CBMnow at www.cbmnow.com provides K–12 teachers and other professional educators
with user-friendly information and support in using curriculum-based measurement,
including subject areas, standards, data usage and management, and other connected
topics.
REFERENCES
Adams, M. J. (1990). Beginning to read: Thinking and learning about print. Cambridge, MA:
MIT Press.
Babkie, A. M., & Provost, M. C. (2004). Teachers as researchers. Intervention in School and
Clinic, 39, 260–268.
Batsche, G., Elliott, J., Graden, J. L., Grimes, J., Kovaleski, J. F., Prasse, D., . . .Tilly,
W. D. III (2006). Response to intervention: Policy considerations and implementation. Alexan-
dria, VA: National Association of State Directors of Special Education.
Bohanon, H., Fenning, P., Carney, K. L., Minnis-Kim, M. J., Anderson-Harriss, S.,
& Moroz, K. B. (2006). School-wide application of positive behavior support in an
urban high school: A case study. Journal of Positive Behavior Interventions, 8, 131–145.
Burns, M., Appleton, J. J., & Stehouwer, J. D. (2005). Meta-analytic review of respon-
siveness-to-intervention research: Examining field-based and research-implemented
models. Journal of Psychoeducational Assessment, 23, 381–394.
Colvin, G., Kame’enui, E., & Sugai, G. (1993). Reconceptualizing behavior management
and school-wide discipline in general education. Education and Treatment of Children,
16, 361–381.
Deno, S. L. (1995). School psychologist as problem solver. In A. Thomas & J. Grimes
(Eds.), Best practices in school psychology III (pp. 471–484). Washington, DC: National
Association of School Psychologists.
Deno, S. L. (2002). Problem-solving as best practice. In A. Thomas & J. Grimes (Eds.),
Best practices in school psychology IV (pp. 37–55). Bethesda, MD: National Association of
School Psychologists.
Deno, S. L. (2003). Developments in curriculum-based measurement. The Journal of Spe-
cial Education, 37, 184–192.
Deno, S. L., Fuchs, L. S., Marston, D., & Shin, J. (2001). Using curriculum-based mea-
surement to establish growth standards for students with learning disabilities. School
Psychology Review, 30, 507–524.
Duda, M. A., Dunlap, G., Fox, L., Lentini, R., & Clarke, S. (2004). An experimental
evaluation of positive behavior support in a community preschool program. Topics in
Early Childhood Special Education, 24, 143–155.
East, B. (2006). Myths about response to intervention (RTI) implementation. Retrieved
from http://www.rtinetwork.org
Fuchs, L. S., & Fuchs, D. (1986). Effects of systematic formative evaluation: A meta-
analysis. Exceptional Children, 53, 199–208.
Fuchs, L. S., & Fuchs, D. (1988). Curriculum-based measurement: A methodology for
evaluating and improving student programs. Diagnostique, 14, 3–13.
Fuchs, L. S., & Fuchs, D. (1999). Monitoring student progress toward the development
of reading competence: A review of three forms of classroom-based assessment. School
Psychology Review, 28, 659–671.
Fuchs, L. S., & Fuchs, D. (2004). What is scientifi cally based research on progress monitoring?
Washington, DC: National Center on Progress Monitoring, American Institute for
Research, Office of Special Education Programs.
Germann, G., & Tindal, G. (1985). An application of curriculum-based assessment: The
use of direct and repeated measurement. Exceptional Children, 52, 244–265.
Healy, K., Vanderwood, M., & Edelston, D. (2005). Early literacy interventions for Eng-
lish language learners: Support for an RTI model. The California School Psychologist, 10,
55–63.
Hosp, J. L., & Reschly, D. J. (2004). Disproportionate representation of minority students
in special education: Academic, demographic, and economic predictors. Exceptional
Children, 70, 185–199.
Individuals with Disabilities Education Improvement Act of 2004, 20 U.S.C. 1400 et seq.
(2004).
Kaminski, R., Cummings, K. D., Powell-Smith, K. A., & Good, R. H. III (2008). Best
practices in using dynamic indicators of basic early literacy skills for formative assess-
ment and evaluation. In A. Thomas & J. Grimes (Eds.), Best practices in school psychology
V (pp. 1181–1203). Bethesda, MD: National Association of School Psychologists.
Kartub, D. T., Taylor-Greene, S., March, R. E., & Horner, R. H. (2000). Reducing
hallway noise: A systems approach. Journal of Positive Behavior Interventions, 2, 179–182.
Lau, M. Y., Sieler, J. D., Mayskens, P., Canter, A., Vankeuren, B., & Marston, D. (2006).
Perspectives on the use of the problem-solving model from the viewpoint of a school
psychologist, administrator, and teacher from a large Midwest urban school district.
Psychology in the Schools, 43, 117–127.
Lichtenstein, R. (2008). Best practices in identification of learning disabilities. In A.
Thomas & J. Grimes (Eds.), Best practices in school psychology V (pp. 295–317). Bethesda,
MD: National Association of School Psychologists.
Linan-Thompson, S., Vaughn, S., Prater, K., & Cirino, P. T. (2006). The response to
intervention of English language learners at risk for reading problems. Journal of Learn-
ing Disabilities, 39, 390–398.
Luiselli, J. K., Putnam, R. F., Handler, M. W., & Feinberg, A. B. (2005). Whole-school
positive behavior support: Effects on student discipline problems and academic perfor-
mance. Educational Psychology, 25, 183–198.
Marston, D., Muyskens, P., Lau, M., & Canter, A. (2003). Problem-solving model for
decision making with high-incidence disabilities: The Minneapolis experience. Learn-
ing Disabilities Research & Practice, 18, 187–200.
McCurdy, B. L., Mannella, M. C., & Eldridge, N. (2003). Positive behavior support in
urban schools: Can we prevent the escalation of antisocial behavior? Journal of Positive
Behavior Interventions, 5, 158–170.
McKevitt, B. C., & Braaksma, A. D. (2008). Best practices in developing a positive behav-
ior support system at the school level. In A. Thomas & J. Grimes (Eds.), Best practices
in school psychology V (pp. 735–747). Bethesda, MD: National Association of School
Psychologists.
National Reading Panel. (2000). Teaching children to read: An evidence-based assessment of the
scientifi c research literature on reading and its implications for reading instruction: Reports of the
subgroups. Bethesda, MD: National Institute of Child Health and Human Development.
National Research Council. (1998). Preventing reading difficulties in young children. Washing-
ton, DC: National Academies Press.
AFFECTIVE ASSESSMENT
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
How do I assess affective and personal aspects of my students in the classroom?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Acknowledge the boundaries of data collected through affective assessments.
♦ Identify major affective assessment areas of investigation, including student
attitudes, interests, beliefs and values, and personal viewpoints.
♦ Follow and complete all required steps in constructing a self-report affective
assessment measure.
♦ Select effective items for
a self-report measure
designed to collect per-
sonal information from
students.
♦ Discuss the importance
of integrating affective
assessment and feedback
into the instructional pro-
cess in order to enhance
future teaching practices.
364
Introduction
This chapter is devoted to the area of affective assessment, which involves the measure-
ment of personal qualities or aspects such as motivation, attitudes, perceptions, and belief
systems that contribute to and inf luence student work and performance in the classroom.
Affect is defined as emotion, so “affective” implies a connection with feelings that are
generally attached to an idea, issue, or action. Emotion is connected to the physiological
changes, subjective experiences, and expressive behaviors that are involved in feelings
such as love, grief, anger, and joy. These personal reactions impact how we conduct our
interaction with the world. Attitudes or perceptions do ref lect students’ emotions, but
they can also involve other dimensions such as cognition. It’s important to recognize and
understand that a student’s global perceptions, viewpoints, and attitudes (e.g., doing well
in school is important, I like my classes, I am a good problem solver) are inf luenced by
thinking factors as well as the emotional analysis and connections that go with them. In
fact most, if not all, educational experiences involve both cognitive and emotional pro-
cessing and connections.
It has been well documented that sensory information (except for smells) goes through
particular structures of the brain that are directly connected with emotion (in a matter
of milliseconds, by the way) before that information is reviewed and evaluated by higher
cortical regions in the cerebral cortex that are commonly recognized as the thinking part
of the human brain (Bloom, Lazerson, & Hofstadter, 1985; Sousa, 2001). Technically
you perceive and “feel” stimuli such as a loud bang or disruption at the back of the room
as you determine what they truly are and if they are important or not. Input is routed
through the sensory integration systems of the brain that are connected to other brain
structures responsible for emotional tone analysis and response. The interpretation of the
world isn’t only emotional or cognitive but an integration of both. This explanation is
provided since the term “affective” implies that only emotional involvement exists. It is
important to recognize that “affective” is a good but not totally inclusive descriptor for
the personal factors that will be examined. With that shortcoming acknowledged, the
limits and restrictions of this area of assessment also need to be reviewed.
assessment area. But there are limits and boundaries to affective assessment that need to
be acknowledged (see the box “Cautions with Affective Assessment”).
Affective assessment should never be associated with or considered as personality
assessment or any other in-depth analysis of a student’s personal attributes, qualities,
characteristics, or temperament. That type of analysis is reserved for psychologists and
other professionals who have a license to conduct psychological assessments and advanced
training to conduct such an evaluation. As a teacher you are not in the business of per-
sonality testing, and you don’t want to hold the expectation that affective assessment can
actually give you that kind and level of information about your students.
As with any assessment method, the reliability and validity of the measures that are
used must always be critically examined and substantiated, but more so here, given the
complexity and potential ambiguity of some of the constructs (e.g., general motivation,
attitudes toward school) and the ranging interpretations that will likely occur when
examining these kinds of domains, particularly in regards to their importance and impact
in the classroom. Also, since these measured concepts will likely change over time, the
need for multiple data sources and episodic measurements (e.g., once or twice a month or
semester) is critical in order to obtain and retain confidence with the results.
When dealing with personal and emotionally-focused information, it is important not
to overanalyze and overgeneralize the findings and the implications that are drawn from
those findings. As professionals, it is important to refrain from jumping to any conclu-
sions, and findings must always be kept in perspective. Conclusions should be based and
appropriately balanced on the amount and strength of the collected student evidence.
For instance, if a student rates himself or herself as being very nervous on some rat-
ing measure, the student may actually be nervous about his or her school performance
or may actually be just fine. It may be a genuine rating or it may not be, and it can be
inf luenced by a number of factors and issues (e.g., poor understanding of the questions,
selection of inaccurate ratings, inf luences or events outside of school, fatigue due to sleep
problems, caffeine consumption) so caution is warranted here. What is interesting is that
in the eyes of this student, he or she is nervous or anxious (regardless of whether or not
evidence supports this contention), and the recognition of that perception is important.
Also, the areas that are to be assessed should be educationally relevant and clearly
defined. In fact, given the limited time and resources available to most teachers, a par-
ticular area should be carefully pre-selected based on the information that you want and
need. For example, as a teacher you embed a lot of collaborative projects and activities
in your classroom, and mutual cooperation is a necessary skill. You want to gauge how
receptive your students are to working together and sharing responsibilities, so you give
the students an inventory consisting of collaboration questions. This would be useful
information to have as their teacher. There are a wide variety of potential affective areas
to examine and many, if not most, are interesting. However, it is most important to make
sure the collected information is actually needed and worth the required collection and
review time.
SELECTION OF AFFECTIVE
ASSESSMENT AREAS
A wide range of potential areas can be investigated with affective assessment. However,
the primary focus of an affective assessment usually centers on how your students feel,
react to, or view a particular topic, event, or situation. Oftentimes, teachers will try
to gain a better understanding through a measure that is designed to determine how
learning experiences are being perceived, defined, and processed by the students in the
classroom.
Affective aspects are also connected to the continuation of current and possible future
behavior. For example, if you hold a favorable attitude toward something or someone,
you are more likely to engage in that particular behavior or enjoy the company of that
person. Consequently, enthusiasm, motivation, and genuine interest usually follow and
are associated with those events and conditions. Conversely, negative attitudes carry with
them qualities that limit the occurrence or frequency of a behavior to a point that it may
be totally nonexistent. Obviously, affective feelings inf luence our behavior, and that cer-
tainly applies to learning. Most educational topics that are examined within the affective
domain can be grouped into one of four basic categories:
♦ student attitudes
♦ student interests
♦ personal beliefs or values
♦ personal perceptions or viewpoints
Student Attitudes
An attitude involves one’s personal opinion or feelings on some subject. Tapping into
students’ attitudes can yield interesting and useful information, so surveys or question-
naires are sometimes completed in the quest of this kind of information (Figure 14.1).
Potential areas that might be investigated include the following:
50. Have other kids at school ever teased you about what your body looks like?
A) No
B) Yes
51. How many days each week do you exercise, dance, or play sports?
A) 0 days
B) 1 day
C) 2 days
D) 3 days
E) 4 days
F) 5 days
G) 6 or 7 days
52. When not exercising, do you ever have trouble breathing (for example, shortness-of-breath,
wheezing, or a sense of tightness in your chest)?
A) No
B) Yes
53. Has a parent or some other adult ever told you that you have asthma?
A) No
B) Yes
54. Yesterday, how much time did you spend watching TV or playing video games?
A) None, I didn’t watch TV yesterday
B) Less than 1 hour
C) About 1 hour
D) About 2 hours
E) 3 or more hours
© Brand X Pictures/PunchStock
Obviously there are many more possibilities, but the assessment focus remains con-
stant in wanting to get the students’ viewpoints on a specific area, measured by several
items or questions, in order to better understand and view the identified issue through
the eyes of the students. If positive ratings and responses are generated, confidence would
grow that existing procedures or conditions are successful in supporting the development
of the students and their views.
Conversely, if negative ratings and responses are generated, and if these results are
supported by other findings, a review of classroom procedures would likely be in order.
Student attitudes ref lect a collective mindset and the students’ perceptual reality. There-
fore, this kind of information can be extremely valuable in confirming what is working
or seen as valuable—and what is not—in the eyes of the students in the classroom.
Student Interests
An interest deals with the unique quality of a topic or issue that draws one’s attention
toward it. It’s what students are “into” and usually like and enjoy. Within the classroom
setting, this can range from subject or content interest, enjoyment of different learning
activities, using computers, or even riding the bus. In regards to instruction, it is impor-
tant for a teacher to know how receptive his or her students are regarding the content and
information that will be covered as part of daily lessons or more extended unit lessons.
For example, if you are a history teacher (or any other content-focused teacher for that
matter), you know that some of your students may not be very excited about studying
history, since relevance and personal connection may be limited, especially with young
students. Consequently, it may be important to survey your students to find out how
interested (or more likely how disinterested) they are in examining the past and how
these historical events have inf luenced their current lives.
If minimal interest exists (and even if it doesn’t) as ref lected in the assessment results,
you could address that problem by creating unique learning events (e.g., living history
demonstrations), meaningful interaction and collaboration opportunities (e.g., contact-
ing and interviewing historical figures or participants), and other stimulating activities
that require those students to become part of history and its exploration. Otherwise, the
potential for a numbing and relatively ineffective learning experience for the students and
the instructor is very likely. Clearly, when we want to know what students like or enjoy,
obtaining student interest information is the assessment ticket.
must be exercised in the selection process, and it is always important to make sure the
selections are in the best interests of the students.
CHAPTER ACTIVITY
In regards to classroom instruction, one of the most important areas to have information on
is classroom climate. Classroom climate is recognized as the quality and responsiveness of a
learning environment that exists for students (Fraser, 1998; Freiberg, 1999). And a positive
and supportive classroom climate is associated with academic achievement (Griffith, 1995;
Freiberg, 1999; Madaus, Airasian, & Kellaghan, 1980). Teachers by design try to create
a positive learning setting where all students are supported and challenged to perform at
their highest levels. When a positive classroom climate exists, students feel safe, supported,
valued, and encouraged to learn. Risks can be taken without the fear of teacher retribution
or social embarrassment. On the other hand, when this kind of climate does not exist, stu-
dents can become disinterested, uncomfortable, and may even “downshift,” thereby gen-
erating minimal connection with the teacher, the instructional material, and the learning
expectations of the class.
At the present time, ref lect on the classroom climate of the assessment course that you
are currently completing, and answer the following questions.
1. As a student, how would you describe the learning climate of this class? Is it positive?
Negative? Or is it a combination of both? What makes it that way?
(continued)
2. Do you feel safe, supported, valued, and challenged in this classroom? What actions or
procedures generate your reactions?
3. What would your collective responses tell your instructor about the learning environ-
ment in this class? Is he or she doing the “right” things, or are changes needed, and if so,
what changes should be made?
Once you have completed your responses, if possible, share them with your classmates in
order to identify similar observations and trends.
Now the next and most interesting question centers on sharing this information with
your current instructor. Should this be done? Why or why not?
ASSESSMENT IN ACTION
MATHEMATICS SURVEY
Dear Student,
Please help us find out about how students in your district feel about mathematics. To do that,
please think carefully about each sentence below and decide whether you agree or disagree with
the sentence. Then tell us how much you agree or disagree by circling the number that best
expresses how you feel about what the sentence says. We will take your answers and combine
them with all the rest of the students. Remember, DO NOT put your name on this form.
Thank you for your help.
4 means you strongly agree with what the sentence says.
3 means you agree but not really strongly.
2 means you disagree but not really strongly.
1 means you strongly disagree.
Strongly Agree Disagree Strongly
Agree Disagree
one of the most important skills that a teacher can possess.” From the following options
(Strongly Agree, Agree, Not Sure or Neutral, Disagree, or Strongly Disagree), select
the one that best describes your opinion of the classroom assessment statement. These
descriptive ratings provide fairly distinct indications of the strength and direction of an
individual’s response to each provided item.
The five-descriptor format is the most common and recognized scale. However,
findings have been reported that greater reliability and measurement precision is gener-
ated when a seven-descriptor format (Absolutely Disagree, Strongly Disagree, Disagree,
Indifferent, Agree, Strongly Agree, and Completely Agree) is used (Dawes, 2008; Mun-
shi, 1990). With the five-descriptor format, sometimes the middle descriptor is omitted
(as in the math survey in Figure 14.2) and a four-descriptor option set is used, which
forces a respondent to provide either a supportive or nonsupportive response. Also, for
younger respondents, three (Agree, Not Sure, Disagree) or even two indicators (Yes or
No) may be appropriate.
In regards to the assessment statement listed in the previous paragraph (“Classroom
assessment is one of the most important skills that a teacher can possess”), what response
option did you select when you read the item, and why did you give it that response?
CONSTRUCTING A LIKERT
SELF-REPORT MEASURE
Let’s examine the basic steps (and major considerations) that are important to follow
when constructing a Likert self-report measure (see the box “Construction Steps for a
Likert Self-Report Measure”).
Question 1
Question 1 deals with a possible perception or viewpoint. However, there is not a clear
personal reference to the respondent (e.g., “I”), so a connection is not firmly established
and can only be assumed. It could be that a student actually believes social studies is
boring (i.e., provides an “Agree” response) but may really like it, which would not be
assumed based on the existing statement. A better statement would be, “I think social
studies is boring,” as this statement is directly linked to the respondent.
Question 2
Question 2 deals directly with student interest. A personal connection can be assumed
with the statement and a distinct learning event has been reported in the item. One clear
thought and expression is evident. This would be a good statement to potentially use in
an inventory.
Question 3
Question 3 deals with an affective response, but two conf licting emotions are listed
within one single statement. The personal reference is there, yet due to the two oppos-
ing reactions it is a potentially confusing item for a student to answer. The student
could agree with the first part of the statement and disagree with the second part of the
same statement and not know what option rating to mark given this dilemma. A better
approach would be to break the original statement into two separate and independent
statements.
The number of constructed items can vary, but several items (seven or more depend-
ing on the topic and the audience) will likely be needed to ensure adequate coverage of
an area within the scale. As a rule, more and finer selection items can be used with older
students while fewer and more general items often work better with younger respon-
dents. Also, it’s better to have a larger number of items than to have too few, and rec-
ognize the time requirements that will be necessary in order to complete the inventory.
The reading level of the students must also be taken into consideration, as minimal read-
ing requirements should exist with all selected items. The creation or selection of the
items can be completed by yourself, if you have the interest and background, or by a
group of supportive colleagues. If done by committee, group interest and participation
may prove helpful in the later administration of the instrument.
External Review
Once the items are generated or selected, you may believe that the inventory is ready to
be administered. However, having the items reviewed by a colleague in order to check
for accuracy and continuity is important. This could be a fellow teacher or a member of
the support staff. In particular, seek out the assistance of the school psychologist, as he
or she possesses an assessment background and will have experience and perhaps even a
high level of expertise in constructing inventories and surveys. Ideally, a group of inde-
pendent judges would be used to determine the acceptability of each item relative to the
identified concept. Also, inter-correlations among the items, based on the independent
ratings of the judges, are typically completed in order to select the most effective items
for the scale. However, most classroom instruments do not have this level of construction
support. Therefore, obtaining colleague feedback is essential in identifying the items that
“hold together” and address the intended domain.
Field Testing
The measure should also be field tested, preferably with a comparable set of students.
For instance, if this inventory is going to be given to your third-grade classroom, hav-
ing another third-grade class in a different elementary school in the district complete
and provide feedback would be ideal. Procedural problems, unclear questions, or other
kinds of difficulties could be caught and corrected. Also, student reactions and comments
should be sought out, as they are often extremely helpful in building a better instrument
and refining the administration of it.
Respondents must also be familiar and comfortable with how to process and respond
to these types of questions. Learning to listen and measure one’s feelings about a subject
is a personal and sometimes complicated journey. Trying to attach a measuring stick to
the provided descriptor options can also be tricky. The mere fact that students are asked
not to identify themselves when completing the survey, so that their responses cannot be
singled out, can be a strange requirement. As with most things, practice makes all the
difference. Having students complete Likert-based items as part of exit slips or other class
activities helps to establish a general comfort and acceptance level.
Comstock/PictureQuest
Future Modifications
After the data is collected, analyzed, and integrated into the educational process, it is
necessary to ref lect on the entire assessment process to determine what, if any, changes
are warranted. For example, it may be found that a statement turns out to be confusing
to the students, or perhaps a better ordering of the items is possible. Comments from
your students after the inventory is completed may confirm that a statement was confus-
ing and was consequently misinterpreted by the students. All of these occurrences are
possible, and so are many more. It’s just good practice to review your inventory and the
statements that compose it so that the measure is even better the next time it is used.
FOCUS ON ASSESSMENT
CONSTRUCTING AN AFFECTIVE
ASSESSMENT MEASURE
Imagine that you are helping a colleague, Ms. Welch, a third-grade teacher who has
some personal and professional concerns about how her students view reading and its
educational importance. Given the range of daily distractions such as the Internet and
gaming programs, she is worried that minimal interest in reading exists in her classroom
(as well as others). Consequently, she wants to find out her students’ attitudes and feelings
on this subject. She has asked for your help in developing an inventory that can help
address that concern.
Investigation Area
The first order of business is clearly determining the topic and assessment focus of the
inventory. Based on the teacher’s comments and description, it is reasonable to conclude
that “attitudes about reading” is the area of investigation. However, this should be con-
firmed and clearly stated before further steps are taken.
Selection of Statements
Once the area of investigation has been identified, the development of the statements that
will be used in the inventory needs to be completed. Consider and review the following
items that the teacher created or found for potential inclusion.
1. I believe reading books is fun.
2. I don’t like reading.
3. Reading in school is hard.
4. I like reading books by myself.
5. Reading is fundamental to a good education.
6. I like my teacher.
7. My teacher gives me good stories to read.
8. I really don’t like to write.
9. I have a lot of friends in my class.
10. I believe reading at home is important.
Would you use all of these? Would you use some of them? From the provided list, indi-
cate whether you would use each item and provide a reason for and against its selection.
Then compare your list with your classmates and to the examination of the statements
below.
Item 1: Is an acceptable item that addresses the investigation area.
Item 2: Is an acceptable item that addresses the investigation area.
Item 3: Currently is an unacceptable item (an opinion not directly connected to the
respondent). The statement “I believe reading in school is hard” would be a better
alternative.
Item 4: Is an acceptable item that addresses the investigation area.
Item 5: Is an unacceptable item. It is not an affective-directed question and it
includes wording (“fundamental”) that requires a strong vocabulary and makes the
interpretation of the statement confusing.
Item 6: Is an unacceptable item since it does not address the investigation area. It
may provide useful information, but not in regards to the present assessment.
How does your review match up with this one? Do you believe the justification for each
item is appropriate and reasonable?
Through this analysis (and the item justifications), you help your colleague organize
the items for the inventory, along with the directions and general format of the inven-
tory, which are presented below.
External Review
Since you have been directly involved in the construction of the inventory, you are also
assisting in the external review of the items and the general inventory procedures. This
work could certainly be reviewed by someone outside of your current dyad, but that may
not be possible, especially if a quick implementation schedule is being followed. Direc-
tions and the general layout of the inventory should also be developed at this time.
Pilot/Field Test
A small group of third graders (five to six would do) who are in Ms. Adam’s class (the
other third-grade classroom) are available to read and respond to the statements, and this
pilot would identify any potential problems that could be eliminated before the inven-
tory is actually given in Ms. Welch’s class. Since the administration of this interest inven-
tory is directly connected to the instructional process, approval and permission is likely
not needed from the students’ parents. Most school districts follow policies that do not
require parental permission if students are participating in school or classroom activities
that are:
♦ instructionally relevant and appropriate
♦ principal- and/or building-approved
♦ not research-focused or directed
♦ relevant to school-wide or district-wide decision making
However, it is always wise to check your school policy to make sure that prior permission
is not required, even with a pilot testing.
The pilot is completed and the following changes were warranted based on student
feedback and observation of the assessment.
♦ State in the directions that only one rating is to be selected for each statement, as
some students were confused with that.
♦ Verbally remind everyone that they are not to put their name anywhere on the
form. Due to habit, some students automatically wrote their names on the forms.
♦ Make sure that only pencils with erasers are used, in case changes to answers need
to be made.
2 students are not sure about the statement, and 30 percent or 6 students agree or strongly
agree with that statement. Collectively, a majority of students like reading but almost a
third of the class does not.
Given that only six items make up the reading interest scale, the overall cumulative
findings are restricted to this narrow item set. Nevertheless, the ratings across each of the
questions generated the following mean ratings: Question 1 (3.35); Question 2 (3.55—
reversal item); Question 3 (3.4—reversal item); Question 4 (3.1); Question 5 (3.75); and
Question 6 (3.25), with a grand mean (mean across all the means for Questions 1–6) of
3.4 and uniform variance for the ratings across the questions. Overall, the results indicate
relatively strong and comparable response strength with the individual questions, relative
to the composite mean; a particular strong and positive rating, in fact the highest, was
demonstrated for Question 5 (“My teacher gives me good stories to read”).
Upon further qualitative review, Ms. Welch merged questions into strands or
related combinations. For instance, Questions 5 and 6 can be combined into an “adult-
inf luenced” or external reading factor, which generates a strong positive reading rating.
Questions 2 and 4 fall into an “individual-inf luenced” or internal reading factor, and
a comparably strong reading rating was demonstrated. Finally, Questions 1 and 3 can
represent a “reading process” factor, and the strength of this factor is comparable to the
other two.
Educational Implications
As a teacher, what are the educational implications of this information, assuming the data
are valid and reliable? If you were Ms. Welch, would you be relieved or dismayed by the
findings? If possible, get into small groups and talk about the results and what kind of
data picture they create. What kind of message are you receiving from this inventory?
What insights could you share with Ms. Welch? Hopefully, several possible ideas have
emerged based on the collected information. In particular, based on the student ratings,
examine the following considerations and see how they coincide with your observations
and decisions.
♦ A majority of students, approximately half the class, like and enjoy reading. They
also believe that reading is not a difficult academic skill to execute and complete in
school.
♦ A smaller group of students, approximately a third of the class, do not like or enjoy
reading. Relative to the entire classroom, this is a large number of students who are
on the negative side of the reading issue. In regards to instructional and learning
dynamic issues, something needs to be done (and quickly) to help reach and change
the perceptions of these “detached” students. Moreover, they view reading as hard,
so their motivation to approach and complete reading activities will be adversely
affected. If something is not done to counteract this viewpoint, those perceptions
are likely going to become more solidified over time.
♦ Based on the student ratings, it would be prudent to carefully examine the use of
independent reading as an instructional practice since the class is split in regards to
the enjoyment connected with this activity. Similar results are demonstrated with
reading at home, so this also needs to be examined. Perhaps the reading require-
ments or the level of material being completed at home is discouraging the stu-
dents, particularly those who already hold negative views of reading.
♦ Ms. Welch clearly provides good and interesting stories to read, based on the stu-
dent responses. She should feel confident in her abilities to select stimulating stories
that her students find interesting and engaging.
These findings are interesting and should give Ms. Welch much to think about as she
examines and potentially rethinks her reading instruction. Just being aware that reading
is not a positive experience for several of her students and having evidence to support
that claim is important, because without that awareness nothing would likely change and
these students would continue on the same course. Moreover, this attitude exists in spite
of the fact that Ms. Welch provides interesting content for the students to read.
With this information, Ms. Welch can modify the reading practices in her classroom.
For example, she may want to have the students participate in specific individual and
group reading activities where she can identify particular issues (e.g., weak vocabulary
base, poor comprehension of what is read, little to no connection to reader’s life and
experiences) that may exist for the students who are struggling with reading. If com-
mon skill deficiencies emerge, specific instruction can be provided as part of her class-
room instruction. Also, in order to increase confidence and interest in reading, students
could select their own free-reading topics and be provided materials that are at or slightly
below their reading level. This would help to ensure motivation and success with this
reading activity and may help change the negative perception of reading that the strug-
gling readers possess.
Particular issues and problems were noted with independent and home reading, and
specific modifications can be considered for those activities. However, with follow-
up questions about those and other reading activities, even more information can be
obtained and used to improve the reading experience in Ms. Welch’s class. From the
present information alone, it is clear that a more developed and organized connection
between the school and home in the area of reading needs to exist. It is also important
to remember that at least half of the class feels good about reading, and this enthusiasm
needs to be maintained for those students. As you no doubt have learned from your
program and experiences, at times some students may require the same help or feedback
while other students will require other kinds of assistance. However, every student at one
time or another will require his or her own individualized help in the classroom.
Findings from this survey, as well as others, could also be shared with parents during
school conferences or as part of e-mail newsletters or blogs. For example, sharing the
results from Question 6 (“I believe reading at home is important”) can provide opportu-
nities to discuss the role and importance of reading at home—everything from how to
do it to useful resources, supplemental books, book clubs, and so on. This could certainly
serve as a valuable resource during the school year for general public issues and parent
education opportunities.
Measure Revision
Upon review of the inventory and the statements that compose it, Ms. Welch wants to
examine the “f low” of the items and their presentation order. What, if any, suggestions
would you have for her regarding this issue? After some discussion and review, she is now
considering the following order of the statements: (1) I don’t like reading, (2) I believe
reading books is fun, (3) I like reading books by myself, (4) I believe reading in school is
hard, (5) My teacher gives good stories to read, (6) I believe reading at home is impor-
tant. The new statement order follows a general to specific approach and ref lects a logical
f low, moving from reading in general to reading in specific locations such as in school
and at home. Ms. Welch is also considering adding some more negative statements to
help provide a more equal balance of positive and negative statements.
CHAPTER REFLECTION
In the Chapter Activity, you were asked to evaluate the classroom climate of your cur-
rent assessment/measurement course. In fact, three questions had to be answered. How
did your results compare with your peers? Was there general agreement, and if so, in
what areas?
It is important to recognize that receiving student feedback on classroom climate
through affective assessment is essential to providing and maintaining an effective
instructional practice. That is why university classroom evaluations are completed and
usually mandated at the end of the semester or quarter. However, if collected only at the
end of instructional periods, needed modifications are only possible for the next class.
This kind of information and feedback is really needed throughout the delivery of a
course, and many instructors do just that; they administer course evaluations once or
maybe twice before the final one is completed.
Periodic instructional feedback is important, whether you are teaching at the uni-
versity or P–12 level. At both levels, however, most educators have limited time and
resources to complete extensive reviews, but there are still ways to obtain such informa-
tion. One relatively easy way of incorporating this kind of data collection is to have class-
room climate questions periodically added to exit slips that are routinely administered
at the end of class. In this way a teacher can gain student data and monitor the learning
climate, as perceived by his or her students, as part of the standard classroom assessment
system.
CHAPTER REVIEW
Conclusion
In today’s classrooms, students must acquire and demonstrate proficiency across a number
of academic standards; consequently, day-to-day classroom instruction is focused on this
major goal, and that will not likely change. However, as mentioned earlier in the chap-
ter, students are much more than the academic work they produce. They are dynamic
learning beings that possess attitudes, feelings, and values that directly impact the educa-
tional accomplishments and skills that are expected with those mandated standards. For a
teacher, being able to measure these attributes can be just as valuable as the actual content
that is covered in a lesson.
Affective assessment has the potential of providing a teacher with useful information
about the internal workings and personal views of his or her students. There is the recog-
nized mission of educating the “whole” child, but a teacher has to know that child first,
and affective assessments can provide information to help educators see the entire learner,
not just the academic side. They allow for a more intimate awareness of the learning
audience, which can be used to improve the general learning climate of the classroom
as well as the quality of instruction that is provided. In addition, the opportunity for the
learners themselves to see and examine their own personal views, beliefs, and attitudes
is possible, maybe for the very first time. Every student deserves the opportunity to be
involved in a self-discovery process that can provide invaluable insight into and under-
standing of the affective side of their learning experiences.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. How can affective assessment be used to help you, as the teacher, better under-
stand your students and their learning needs in the classroom?
2. How will you integrate this kind of assessment into your instructional process?
3. How comfortable are you with developing a Likert-based affective assessment
inventory? Do you see this kind of measure fitting into your overall classroom
assessment system?
RESOURCES
The Alliance for the Study of School Climate (ASSC) (www.calstatela.edu/centers/
schoolclimate/index.html) at California State University provides valuable informa-
tion, research, and resources, including classroom climate surveys that can be used to
improve learning conditions and environments for students. Click on the “assessment”
button for the classroom climate surveys.
SurveyBuilder at surveybuilder.edmin.com/main.cfm is a useful Web site where free
online surveys can be constructed.
REFERENCES
Bloom, F. E., Lazerson, A., & Hofstadter, L. (1985). Brain, mind, and behavior. New York:
W. H. Freeman and Company.
Dawes, J. (2008). Do data characteristics change according to the number of scale points
used?: An experiment using 5-point, 7-point and 10-point scales. International Journal of
Market Research, 50(1), 61–77.
Fraser, B. J. (1998). Classroom environment instruments: Development, validity, and
applications. Learning Environments Research, 1, 7–33.
Freiberg, H. J. (1999). School climate: Measuring, improving, and sustaining healthy learning
environments. London: Falmer Press.
Griffith, J. (1995). An empirical examination of a model of social climate in elementary
schools. Basic and Applied Social Psychology, 17, 97–117.
Likert, R. (1932). A technique for the measurement of attitudes. Archives of Psychology, 140,
1–55.
Madaus, G. F., Airasian, P. W., & Kellaghan, T. (1980). School effectiveness: A reassessment of
the evidence. New York: McGraw-Hill.
Munshi, J. (1990). A method for constructing Likert scales. Retrieved from http://chaam
jamal.com/papers/likert.html
Sousa, D. A. (2001). How the brain learns (2nd ed.). Thousand Oaks, CA: Corwin Press,
Inc.
KEY FOCUS
The central focus of this chapter is to examine the fundamental question:
As a teacher, how do I start to make assessment an integral part of my teaching?
CHAPTER OBJECTIVES
After completing this chapter, you should be able to:
♦ Distinguish the assessment realities
that must be addressed before a class-
room assessment system is constructed
and implemented.
♦ Describe the importance of a self-
designed classroom assessment system
in enhancing instructional effectiveness
and student learning in the classroom.
♦ Identify the steps and required actions
of the classroom assessment cycle.
♦ Design a classroom assessment system
complete with identified instructional
questions, potential assessment mea-
sures, and a rationale for data review
and decision procedures.
♦ Explain the assumptions that will
guide the assessment process in your
classroom.
Introduction
It is hoped that through the journey of reading and reviewing the chapters of this book
you have gained some useful ideas and procedures to consider as part of your class-
room assessment system. As you continue your assessment journey there are additional
resources, including Web sites, professional associations, and books, that you will want to
examine and consider as you put your ideas together into concrete actions and products.
Assessment is just like teaching in that it requires continued development and refinement
of your professional skills. As you begin your teaching career, start out with a firm assess-
ment foundation and framework, and your expertise in this area will only grow with
time.
When you begin your teaching career, it is highly unlikely that you will receive direct
and continuous assistance in designing and implementing your classroom assessment
system. Currently, many school districts do not devote significant time and resources
toward that goal (McMillan, Myran, & Workman, 2002). What you will get from your
building administrator is a crystal-clear clarification on what your students must know
and demonstrate, relative to state academic standards, by the spring of the school year,
along with the clear directive to “get it done.” What happens to the students during the
year is your call and responsibility. School supports and resources are hopefully available
to help in the process. However, the responsibility and operation of the assessment system
in facilitating the learning of your students falls directly onto you as the teacher in the
classroom.
Therefore, the importance of constructing a prototype of a working classroom system
cannot be overstated. Although the opportunities to field test your system will not exist
until student teaching at the earliest, creating the foundation, framework, and operating
procedures that will exist within the system is important now. The assessment “skeleton”
needs to be in place and well connected so that further refinements can be made when
it is field tested under actual classroom conditions. With that in mind, this chapter is
designed to help set up that assessment foundation and framework. Completing the pro-
vided exercises represents an early investment opportunity, where your time and effort
constructing the system can provide you with long-term educational dividends through-
out your entire professional career.
CHAPTER ACTIVITY
It is assumed that you have started to develop a growing assessment knowledge and skill
base that can and will be used to provide the most effective instruction in your classroom.
As you ref lect back on your readings, discussions, experiences, and interactions, what ele-
ments stand out in your mind as the most significant assessment findings and important
issues for you as a future teacher? Take a moment and write them down. Once you have
created that list, consider this important question: How will these elements, issues, and
procedures be addressed and incorporated within your classroom assessment system?
FOCUS ON ASSESSMENT
dynamicgraphics/JupiterImages
can provide support in regards to assessment procedures and issues. But as a teacher
you are held accountable for accurately measuring and monitoring the learning
progress and accomplishments of your students. That is why you need to have a
firm assessment knowledge base in order to design and implement an effective
classroom assessment system.
Beyond the accountability pressure, a classroom assessment system provides
evidence that can directly address whether your teaching is truly effective and
whether your students are benefiting from that instruction. Always remember that
assessment and teaching are f lip sides of the same instructional coin, and the effec-
tiveness of your teaching is directly connected to your capacity to accurately assess
the learning progress of your students.
♦ The purpose and function of classroom assessment must be clearly defined.
The purpose and function of classroom assessment must be evident and clear as
the “why” of assessment is addressed. As the teacher, assessment must provide you
with essential services and must meet specific needs. In particular, there must be
important instructional targets connected to and measured by the assessment pro-
cess. Otherwise, assessment activities become misdirected events that take up valu-
able time in the classroom. Recognizing that the purposes of assessment systems
will vary, a classroom assessment system must be able to provide the systematic
collection of student learning evidence. That evidence must then be analyzed and
CLASSROOM ASSESSMENT
FRAMEWORK EXERCISE PART 1
Determining What Your Assessment System Will Do
Before the actual components of your classroom assessment system can be selected, you
need to identify the questions that your system will address and answer. As you try to
determine what your questions will be, start with the big ones: What does your assess-
ment system need to be able to do? What needs will it address for you?
Make sure your answers to these questions are as detailed as possible and include spe-
cific outcomes. For the moment, imagine that you are a fourth-grade teacher and con-
sider this as the primary assessment question you need to be able to answer: Are my
students making demonstrable progress toward meeting their expected fourth-grade aca-
demic content and skill standards in reading, math, and written language?
This is a reasonable question for a fourth-grade teacher to ask, assuming grade-level
skill standards exist within the district and instruction is matched to those standards. In
fact, this is a reasonable question for any teacher, regardless of grade level, since all teach-
ers need to know the progress points of their students and a classroom assessment system
should be designed to collect that kind of information. However, if you also need to be
able to formally evaluate the quality of their work (e.g., grading), which is very likely,
then that is an additional question that your system must be able to address. Recognize
that as the number of questions increases, so does the complexity and the time and effort
required to operate and maintain the system.
By completing this activity, you come to recognize the important needs that you have
identified for your assessment system. At this time, record the questions and needs that
you believe your assessment system must address; this information is necessary as you
continue through the framework exercises.
The questions/needs of my assessment system are or will be . . .
As the operator of this assessment system, you will need to determine what questions
regarding student learning need to be answered in your classroom. For example, a com-
mon question that confronts teachers is “What is the progress of my students in regards
to their expected, and typically required, academic learning targets and outcomes?”
This corresponds directly to the educational destination question (i.e., “Where do my
students need to be?”) that was identified as part of the instruction-learning-assessment
model discussed in Chapter 1. Since this fundamental question needs to be addressed in
every classroom, it is important to embed within your instructional delivery model a set
of procedures whereby student evidence can be collected that will answer that question.
Classroom assessment models or cycles typically require key elements within the problem-
solving process (Figure 15.1). Those basic elements include at least the following:
1. clear and identified learning targets
2. instruction directed and focused on those learning targets and related student
outcomes
CLASSROOM ASSESSMENT
FRAMEWORK EXERCISE PART 2
Determining What Assessment Measures Will Be Used in Your
Classroom
Given the critical importance that student learning evidence plays in the instructional
process, you need to determine what kinds of student learning evidence will be collected
(and when) within your instructional delivery model (e.g., the progression of instruc-
tional activities and general sequence of your daily lesson plans). At a minimum, the
learning status of your students at three key points within the instructional sequence
should be known:
♦ Midstream formative assessment evidence should document and confirm the prog-
ress of your students at that point in their instruction in order to verify the contin-
ued trajectory of their learning.
♦ Post-assessment or summative assessment evidence serves the important function
of documenting acquired skills and knowledge; it should serve as the confirmation
that the desired goals have been learned and retained by your students.
Student learning evidence should be collected through multiple pathways. For example,
progress feedback certainly needs to be part of the daily instructional regime. In addi-
tion, opportunities to develop and practice self-assessments should be incorporated into
the format of lesson plans whenever possible. Established and well-defined summative
assessments must also be developed, as these measures are recognized as the formal evalu-
ation and measurement of student learning and achievement. The selection and use of
various data collection methods (e.g., rubrics, portfolios, observations, checklists, video
clips, journals, diagrams, tests, interviews, performances) depends on how the informa-
tion is to be used. All of the aforementioned techniques can serve a formative, summa-
tive, or self-assessment function. It depends on how you choose to use these measures as
part of your overall classroom assessment system.
A key element in any assessment system is collecting meaningful evidence in a con-
sistent manner. Once it is known where students are, relative to the content that is to
be taught and their prerequisite knowledge and skill base, predetermined checks and
periodic monitoring of progress need to be conducted at various points throughout
the instructional lesson. Only in this way can a teacher be sure of the actual learning
progress of the students. As the instructor, you determine the actual number of checks
that are needed. However, at a minimum, a pre-assessment of student knowledge and
skill is needed along with a midstream “progress check” before an instructional unit is
completed.
The essential reason for the minimal halfway check is to confirm learning accuracy
before the lesson is completed. If inaccurate learning or minimal progress are demonstrated,
appropriate adjustments and review of needed material can take place before moving on to
the rest of the material and completing the lesson(s). Obviously, no instructor wants to
arrive at the end of a lesson to find out that his or her students have not learned the mate-
rial that has been taught. A midstream assessment activity helps to eliminate that scenario.
The ultimate focus of the assessment cycle is to collect student evidence so that data-
informed decisions, designed to promote continued student learning and progress, can be
made. Possessing student learning evidence allows teachers to make more relevant deci-
sions regarding the learning needs of their students. If collected data supports the con-
tention that students are making appropriate progress, instructional plans would likely
be maintained with minimal changes. On the other hand, if students do not evidence
expected learning gains and progress, instructional modifications would be necessary in
order to address the current learning issues.
CLASSROOM ASSESSMENT
FRAMEWORK EXERCISE PART 3
Aligning Your Classroom Assessments with Instructional
Learning Impact
Now it is time to examine and review the measures and procedures that have been
selected for your classroom assessment system. In the first chapter of this text, the essen-
tial assessment qualities of instruction as identified by the Assessment Reform Group
(1999) were examined. As you review your measures and procedures, make sure that
direct connections with the assessment practices of your classroom assessment system
exist, and can be identified, with each one of these practices.
1. Effective and meaningful learning feedback is provided to students during the
instructional process.
2. Efforts are made to ensure students are a part of their learning and are actively con-
nected to it.
3. Assessment information is used by the teacher to examine the learning progress of
students and to adjust the instructional process when and if necessary.
4. Assessment information is used to support and motivate students throughout the
instructional process.
5. Students self-assess their own learning progress and make the necessary adjustments
and modifications needed in order to reach the desired educational outcomes.
What specific opportunities exist so that these practices are active and f lourish in your
classroom?
By collecting student performance evidence over a period of time, the teacher is placed
in the prime position of being able to monitor and review both the cumulative progress
and the current learning status of his or her students. This information can be used to
improve student learning as well as evaluate general teaching effectiveness. Refinements,
reteaching, or other instructional changes can be completed in order to reach students
based on what they still need to learn relative to the content that is being taught.
Student data can also be shared with students to help them better understand their
learning progress and what they need to continue to work on in order to reach their
educational goals. Helping students to examine their learning progress through self-
monitoring is an invaluable life skill that can also reduce the workload of data collection
on the teacher’s part.
CLASSROOM ASSESSMENT
FRAMEWORK EXERCISE PART 4
Collecting Confirming Evidence of Student Learning
After the pre-assessment(s) but before the summative assessment(s) have been admin-
istered, the monitoring and facilitation of student progress is necessary as each teacher
strives to ensure that everyone stays on the intended instructional path. The need to
review student learning evidence is as important as the delivery of the content material
itself. Therefore, formative assessments and student performance feedback must exist as
the instructional experience is actually taking place. But how should this be done and
what should it look like?
As the instructor, you need to decide what kind of data will be collected, how often
it will be collected (e.g., daily, every other day, weekly), and how collection will take
place within the daily classroom routine. Another important consideration is whether
you or your students will collect and monitor the information obtained from the selected
assessment measures. The more objective and documentable the evidence is, the better.
For instance, in order to monitor reading f luency, reading checks (e.g., passage read-
ings) could be completed weekly, and the words read correctly through part or all of the
passage would stand as the progress standard. Memorization of math families and facts
could be done in a similar fashion. However, if more perceptual data is desired from the
learners (e.g., “I really understand the meaning of the words that I read” or “The mate-
rial that was covered today made sense to me”), Likert-based items on exit slips may be
more appropriate.
How data collection is done depends on what evidence needs to be collected. For
example, if a knowledge and skill base is being developed, periodic checks will be neces-
sary and observation-based methods and checklists could be used to confirm progress.
As the instructor, you want to possess multiple measuring points that can be tracked
over time showing general progression. The same checklists or step-by-step listings can
also be used by your students to help monitor their progress. Relying on more than one
data source is an excellent strategy when trying to measure progress, especially when
complex, highly sophisticated skill sets are being developed. Data collection can have a
number of different “looks” (e.g., written products, observational documentation, skill
demonstrations, video recordings), but there must be a consistent and reliable collection
mechanism in place so that the obtained information is reliable and (more importantly)
valid for its intended purpose.
As part of this exercise, provide a description of how you expect formative assessment
to work in your future classroom. Include in that description a listing of specific mea-
sures that you believe will provide you with the desired information and student learning
evidence. Also include a data collection timeline, general steps that will be followed, and
how your students may be included in this assessment process.
Consider these important formative assessment questions in this construction process.
1. What measures or procedures will be routinely used as part of the formative assess-
ment practices in your classroom?
2. What data collection/feedback timeline will you follow? Why?
3. How will your students be directly involved in the operation of your classroom
assessment system?
Generating an assessment system may present as a challenging exercise, especially when
you don’t have an actual classroom in which to implement your assessment procedures.
At this point it’s all still theoretical, with no real experiences, practical effects, or unique
daily phenomena connected to it.
However, as a pre-service teacher you are in the unique position of being able to
construct and critically review your system (i.e., a functional prototype) and the revi-
sions of your system before it is actually used in your classroom. In short, you can derive
the benefits of having it critiqued by faculty and peers and evaluated and field tested
during student teaching. With a functional prototype to rely on, you won’t have to con-
struct an assessment system on the spot or have to use someone else’s system because you
don’t have anything in place. Successful professionals, regardless of their area of specialty,
know that in order to meet an identified goal, a clear plan of action must exist. Educators
are no different when it comes to classroom assessment. Therefore, as a future teacher it is
important to have a model assessment plan for your classroom, one that meets your par-
ticular needs and can be explained to and understood by students and parents.
CLASSROOM ASSESSMENT
FRAMEWORK EXERCISE PART 5
Making Informed and Effective Decisions
Regardless of the specific foundation elements and measures that are utilized, an assess-
ment system must provide the teacher and his or her students with useful information so
that effective educational decisions can be made.
As a teacher goes through the process of collecting, reviewing, and using data and evi-
dence to direct instruction and learning activities, he or she becomes a prominent model
to his or her students. Data is seen as something to be used and valued. As the teacher you
need to be fully informed of student progress, but so do the students themselves, as well
as their parents. In fact, if students are involved in the collection and monitoring of their
own data, then they can serve as direct informants to the system.
Classroom assessment shouldn’t be a restricted “teacher-only” process. Students need
to be connected to this process so that they (1) become a contributing part of it, (2) are
ASSESSMENT IN ACTION
motivated by its effects and results, and (3) learn that they are directly responsible for the
learning outcomes that they achieve. With that in mind, examine all the components of
your proposed assessment system. As you complete the review, make sure you can answer
this question: Does your assessment system provide you (and your students) with the information
that you (and they) need in order to make the most effective educational decisions regarding learning
progress and student outcomes in the classroom?
Since the term “most effective” is open to interpretation, let’s agree to define it as
“in the direct and most beneficial interests of your students and their academic achieve-
ments.” Does the system you have constructed and integrated within your teaching
model answer that question? If you can honestly answer in the affirmative, then you
know you have something of real practical value that needs to be kept and refined during
field-based learning experiences such as student teaching. If you are not sure or if your
system falls short of answering this question, it is important to go back and reexamine
your exercise responses. Always remind yourself that you are only as good as the evi-
dence you collect and the decisions that are made based on that information.
SELF-ASSESSMENT OF YOUR
ASSESSMENT FOUNDATION
In Chapter 2, you were asked to complete the following exercise.
Based on the following assessment competencies (selected from the Ohio, Hawaii,
and Wisconsin educator assessment standards), rate your current assessment knowl-
edge and skills on a scale of 1 (extremely weak) to 5 (very strong) relative to these
statements:
a. Knowledge of various assessment types, their appropriate uses, and the data they
can generate. (1 2 3 4 5)
b. Ability to identify, develop, and utilize a variety of assessments (e.g., formative,
summative, self ). (1 2 3 4 5)
c. Ability to use data to review and monitor student progress and learning, and to
plan or differentiate instruction based on collected data. (1 2 3 4 5)
d. Ability to collaborate with and effectively communicate academic progress to stu-
dents, parents, and colleagues. (1 2 3 4 5)
e. Ability to involve students in self-assessment and goal setting in relation to their
learning goals. (1 2 3 4 5)
As mentioned previously, these assessment competencies represent a broad range of
expected skills that include the following:
♦ general knowledge of assessment and its purpose in the classroom
♦ ability to construct and implement assessments in the classroom
♦ ability to adjust instruction based on the data collected from assessments
CHAPTER REFLECTION
In your list of significant assessment findings from the Chapter Activity, did you list
your growing knowledge and skill base? What is your plan to continue developing and
refining your assessment skills? Do you see this area as a high priority for your ongoing
professional development? Why or why not? Where do you see yourself relative to the
practice of classroom assessment now and where do you want to be five years from now?
All of these questions relate back to the essential questions in the instruction-learning-
assessment model that were discussed in the first chapter:
1. Where are you?
2. Where do you need to be?
3. How do you get there?
As the teacher, you will be in the best position to accurately determine your current
assessment practice level and compare it to your desired future level. Getting there is a
matter of seeking out resources, refining your current procedures, and continuing to
develop in this professional area. Just as exercise and training are necessary in order to
build and strengthen the human body, you need to continue to develop the “assessment
muscles” that must be used on a daily basis.
CHAPTER REVIEW
Conclusion
The need to accurately measure student learning and progress has moved the topic of
classroom assessment to center stage for educators across the country. Teachers are now
expected to possess a classroom assessment plan, complete with various assessment tech-
niques and strategies for student evidence collection, self-assessment procedures, and data
review. Along with assessment skills, teachers must also possess the understanding to
conduct their assessment actions and decisions in the most professional and ethical man-
ner, which requires a strong ethical framework and structure.
Because assessment has emerged as a new professional expectation, you may not dem-
onstrate the same amount of confidence and comfort as you would in other aspects of
your training. Personal apprehension can be experienced, as assessment is a learned skill
that requires time and repeated practice. It’s not always about the tools and techniques;
sometimes it’s the personal issues and perceptions connected to the instructional chal-
lenges that matter most. However, as you travel down the assessment path, realize that
it will take time and practice to demonstrate effective assessment practices and that you
can’t or won’t know everything about assessment instantly. And that’s all right. Class-
room assessment is a professional journey that requires ongoing effort, time, and practice.
ASSESSMENT CHECK
It is important to make sure that the major concepts and issues that have been covered
become a part of your expanding knowledge and skill base. With that in mind, answer
the following questions in order to provide your own self-check or progress review on
the learning impact of this chapter.
1. What should a classroom assessment system provide for you and for your students?
2. A colleague approaches you and wants to know how your classroom assessment
system (i.e., identified instructional questions, assessment measures, data review
procedures) operates. What would you tell her and why?
3. Why should assessment be considered a continuing professional skill area for
teachers? Where does it rank (high or low priority) on your professional develop-
ment list? Why?
RESOURCES
The National Education Association (NEA) Web site at www.nea.org/index.html pro-
vides K–12 teachers, administrators, and parents with valuable resources and updates
on current educational practice issues, including assessment, legislation, and research
in the field of education.
The Alberta Assessment Consortium (AAC) at www.aac.ab.ca provides a variety of assess-
ment tools and resources for teachers, including continuing professional development
opportunities in a variety of assessment areas.
REFERENCES
Assessment Reform Group. (1999). Assessment for learning: Beyond the black box. Cambridge,
UK: University of Cambridge School of Education.
McMillan, J. H., Myran, S., & Workman, D. (2002). Elementary teachers’ classroom
assessment and grading practices. The Journal of Educational Research, 95(4), 203–213.
407
they may be misused or not used. To communicate effectively with others on matters of
student assessment, teachers must be able to use assessment terminology appropriately
and must be able to articulate the meaning, limitations, and implications of assessment
results. Furthermore, teachers will sometimes be in a position that will require them to
defend their own assessment procedures and their interpretations of them. At other times,
teachers may need to help the public to interpret assessment results appropriately.
Teachers who meet this standard will have the conceptual and application skills that
follow.
Teachers will understand and be able to give appropriate explanations of how the
interpretation of student assessments must be moderated by the student’s socioeconomic,
cultural, language, and other background factors. Teachers will be able to explain that
assessment results do not imply that such background factors limit a student’s ultimate
educational development. They will be able to communicate to students and to their
parents or guardians how they may assess the student’s educational progress. Teachers
will understand and be able to explain the importance of taking measurement errors into
account when using assessments to make decisions about individual students. Teachers
will be able to explain the limitations of different informal and formal assessment meth-
ods. They will be able to explain printed reports of the results of pupil assessments at the
classroom, school district, state, and national levels.
adequate yearly progress (AYP) A mandated aca- binary choice item A test item that has only two
demic requirement necessitating that all students score possible response options (e.g., true or false), com-
at prof icient or higher levels on state-generated monly used in classroom tests.
achievement tests.
causation The direct relationship of a cause to a spe-
age equivalent (AE) A generated score that is based cific effect.
on age rather than grade-level performance.
checklist The listing of essential components or ele-
alternate form reliability An aspect of reliability ments of a behavior or procedure.
designed to generate two parallel or equivalent forms
of a test. classroom test A collection of items that are
designed to measure and evaluate a learner’s skills,
analytical rubric A rubric that is designed to rate or
knowledge, general performance, or other capabilities
score each identified criterion of a required project.
as they relate to content that is covered as part of the
aptitude comparison grading A grading process instructional process.
that involves comparing each student’s learning prog-
College Entrance Examination Board (CEEB)
ress against his or her own estimated ability.
standard score A standard score with a mean of 500
Asperger' s syndrome A condition identified within and a standard deviation of 100.
the autism spectrum that ref lects social and interper-
sonal isolation but does not include the impaired speech concept map A diagram or other visual image that
and cognitive issues that are often associated with can be used to examine various concepts as well as pro-
autism. vide visual links to subordinate details and facts.
assessment bias A form of bias that is directly con- concurrent validity A measurement concept that
nected to the materials and/or procedures that an involves the completion of two measurement events
examiner may use with an assessment. simultaneously, such as a screening test and a criterion
evaluation (e.g., current academic or behavior perfor-
bar graph A visual representation of data in which
mance), where the strength of the relationship between
the height of the rectangular columns represents the
the test and the criterion measure is evaluated.
frequency of the scores that falls within a score inter-
val. Also known as a histogram. conductive hearing impairment The interference
benchmark A specific learning goal that is expected of sound waves through the outer ear to the inner ear
to be accomplished by a specified time, usually set at area.
specific grade levels. consensus gram A chart that contains a low-to-
best pieces portfolio A portfolio that emphasizes high scale that can be marked to indicate a learner’s
the best work and constructions of various assignments perceived understanding, skills, and performance level
of a student over a specified instructional period. relative to a specific statement or question.
bias An inaccurate view or opinion that interferes constructed response A type of test response
with and adversely impacts the collection and evalua- where the respondent must generate the answer mate-
tion of student learning. rial to a question.
G-1
ity in organizing and constructing a written response grouped frequency distribution A distribution
to a question. based on the construction of predetermined bands or
intervals where individual scores are placed.
fill-in-blank (FIB) question A form of test ques-
tion that requires the respondent to complete a state- group observation The collection of specif ic
ment or item by providing a missing word or words information on an entire set of individuals within a
that are intentionally left out of the item. particular setting or event.
f lowchart A geometric outline complete with growth portfolio Another term used to describe a
directional arrows to demonstrate the sequence of an developmental portfolio.
event or process. histogram A visual diagram of rectangular col-
formal observation The systematic collection and umns with the height of the columns representing the
recording of behavior. frequency of the scores that reside within a score
interval. Also known as a bar graph.
formative assessment A type of assessment that
provides constructive feedback to a learner regarding holistic rubric A scoring measure where a single
his or her performance in order to directly improve or score or rating, based on listed performance criteria, is
enhance that performance in the future. used to represent a student’s entire work.
individualized education plan (IEP) A required
frequency The number of times a specific score is
educational plan generated by a school committee that
obtained or a behavior is observed within a specified
serves as a learning contract for a student who quali-
period of time.
fies for special education services.
frequency distribution The listing and tally of
Indiv iduals w ith Disabilities Education
how often certain scores or behaviors are demon-
Improvement Act (IDEIA) Federal legislation
strated within a specific distribution of scores.
that reauthorized and refined the requirements of the
frequency polygon A line graph that connects the Education for All Handicapped Children Act in 2004.
listed frequencies of intervals together to form a visual The reauthorization ensured that students who are
image and shape of a distribution. identified with special needs received instruction with
their nondisabled peers in general education class-
frequency recording The documentation of a
rooms whenever possible and to the fullest extent pos-
repeated occurrence of an event or behavior within a
sible.
specific time period.
informal observation The viewing of an individ-
grade equivalent A generated score that describes ual or group that may occur at random times without
a student’s performance based on a grade-level com- preset notions of specific behaviors to examine and
parison. record.
grading The formal evaluation of student work instructional delivery The use of particular peda-
based on a set performance level and evidenced in the gogical approaches or methods to guide instruction in
form of a grade, number, or other form of reporting the classroom.
mark.
instructional preview A classroom procedure
grading adaptation approach A grading approach where an instructional overview is provided to learn-
where grading adjustments are made to an existing ers prior to a lesson or instructional activity.
grading system for students with disabilities.
instruction-learning-assessment alignment The
Graduate Record Exam (GRE) A standardized critical linkage of the content that is taught in the
aptitude test required by some graduate programs. classroom with the accurate assessment of the student
Results from this test are reported as College Entrance learning and proficiency based on the presented mate-
Examination Board (CEEB) standard scores. rial.
graph A diagram consisting of lines that describe internal consistency reliability A type of reli-
data or the relationships between the data. ability that examines how consistently items within a
test correspond to each other and uniformly measure a mini-interview An abbreviated exchange with an
concept or attribute. individual where specific and in-depth information
can be obtained.
inter-rater reliability The consistency of an
obtained score or measure that is demonstrated mixed hearing impairment A specific hearing
between two or more raters that independently review impairment or loss that involves the combination of
a certain event (e.g., test, observation, performance). both conductive and sensorineural hearing problems.
interval recording The observational recording of mode The score that occurs most frequently in a
specific preidentified behaviors based on set time seg- distribution.
ments or intervals. monitor The checking of a student’s progress
regarding a learning task or event; this process can be
interview A form of student evidence that involves
conducted externally (by the teacher) or internally (by
the collection of information obtained from an inter-
the learner).
viewee by way of direct questions from an inter-
viewer. multiple-choice item A selected-response test
question format that provides several potential answers
item discrimination index A ratio that compares from which the respondent must select the correct
a student’s complete performance on a test to his or response.
her accuracy on particular items on that test.
narrative An observational recording method that
journal A form of data collection where informa- involves the chronological recording of an event as it
tion is collected and recorded over a period of time as is taking place.
part of a personal daily record. narrative grading A qualitative grading system
learning outcome An expected student effect as a that involves an individualized written review of a
result of a specific educational experience. student’s performance that focuses on specific learning
accomplishments as well as ongoing needs and learn-
learning target A statement of desired student per-
ing issues.
formance that describes what a student should know
and be able to do after a specific lesson or instructional narrative rating scale A scale that uses descriptive
activity is completed. adjectives to gauge or rate a student’s behavior.
negative discrimination item A type of item dis-
letter grade A mark in the form of a capital letter
crimination index that is more likely answered cor-
that is assigned to student work based on preset qual-
rectly by students who perform poorly overall on a
ity and performance expectations.
test.
listing question A form of test question where negatively skewed distribution A distribution
context is presented in the question to generate the where most of the scores are located on the right side
desired response from the learner. of the distribution (and the mean) with few scores on
matching item A selected-response test question the left side.
that requires combining a word or statement from one No Child Left Behind (NCLB) A federal law
column with its corresponding answer from a second signed in 2002 that mandates high accountability,
column. clear learning standards, and measurable learning
matrices Visual frameworks designed to organize outcomes for all students. In particular, the law stip-
information into data components or sections dis- ulates that teachers must be highly qualif ied, that
played across selected rows and columns. “scientifically-based research” be used and followed
in schools, that students have greater public school
mean The numeric average, calculated by summing choice when schools do not demonstrate appropriate
all scores and dividing by the total number of scores. academic progress or improvement, and that all stu-
median The score that separates a distribution into dents’ academic progress must be measured on a
two equal halves, where 50 percent of the scores are yearly basis.
above the median and 50 percent of the scores fall nondiscriminating item A type of item discrimi-
below it. nation index where there is no difference in accuracy
between the test performances of high- and low-rank- according to predetermined standards or performance
ing groups. criteria.
normal curve A smooth, symmetrical curve or portfolio An assessment method that typically
shape that ref lects a balanced distribution of collected involves the collection of student work, such as writ-
data. ten evidence as ref lected in a writing portfolio.
normal curve equivalent (NCE) A standard positive behavior support The use of evidence-
score that has a mean of 50 and a standard deviation of based research and programs designed to promote
21.06. socially appropriate behavior in the school setting.
normal distribution A symmetrical distribution of positive discrimination item A type of item dis-
data that possesses characteristics that are useful in crimination index that is generated when a test item is
describing the ways in which scores vary from the more likely answered correctly by students who per-
mean of the distribution. form better overall on the entire test.
norm group An identified reference group that positively skewed distribution A distribution
serves as a comparison for students who complete the where most of the collected data or scores are located
same test or performance measure. on the left side of the distribution (below the mean)
norm-referenced A test or performance compari- with few on the right side.
son of a student to a comparable group, usually at a
power test A test that is administered to provide
specific age or grade level, that can be based on local,
adequate time so that most, if not all, students are able
state, or national classifications.
to complete the test.
norm-referenced grading A grading approach
pre-assessment A pre-instructional activity that
that involves comparing a student’s performance to a
specific group, such as a specific classroom or an entire assesses the specific content knowledge and skills of
grade level. the prospective learners before a lesson or learning
activity is initiated.
numerical grade A grade that uses number ranges
or percentage bands that are typically paired with spe- predictive validity The extent to which a score
cific letter grades. from a particular measure (e.g., a test or scale) predicts
future performance. It is used to establish criterion-
numerical rating scale The use of number or related evidence of validity by examining the predic-
point values to rate a student’s performance. tive or prognostic power of one measure (e.g., a test)
obtained score The score generated and recorded with that of a future desired criterion (e.g., passing a
on a test. See raw score. course).
odd-even reliability An internal reliability indica- premise A word or statement that is presented as
tor often completed on standardized achievement tests part of a matching test question.
that compares the consistency of all the even-num-
process A component of grading where informa-
bered test items with all the odd-numbered test items.
tion representing nonachievement factors such as
parent-teacher conference A scheduled meeting effort, participation, attendance, or other related vari-
where a teacher meets with parents about their child ables is included in the formal review and evaluation
regarding his or her school performance or experi- of a student.
ence.
product A component of grading where informa-
pass-fail A binary grading system that acknowl- tion on the learning accomplishments of a student is
edges only two possible grading outcomes: the passage included in the formal review and evaluation of that
or failure of a particular task, event, or class. student.
percentile The percentage of individuals scoring at progress A component of grading where informa-
or below a given point within a distribution. tion on the extent of a student’s learning during a
performance event Student-generated work that specified time period is collected and included in the
is presented before others and is assessed and evaluated formal review and evaluation of that student.
progress checklist A checklist that provides learn- response to intervention (RTI) The use of
ers with the ability to document the completion of an evidence-based intervention for students at all levels
assignment and/or a corresponding knowledge and of instructional need.
skill set.
restricted response The specific limitations on a
Public Law (PL) 94± 142 Federal legislation, also student’s response to a question, particularly as it
known as the Education for All Handicapped relates to the focus and potential length of response.
Children Act, that set up the basic structure, proce- rubric An assessment measure consisting of selected
dures, and overall process for the identification and criteria and a performance scale that allows for the
provision of special education services. review of specific student performances and products.
Public Law (PL) 99± 457 The federal law that SAT Reasoning Test A standardized test used for
extended Individuals with Disabilities Education Act college admission that reports results as College
coverage to preschoolers (ages three to f ive) and Entrance Examination Board (CEEB) standard scores.
ensured that students who are identified with special
needs receive instruction with their nondisabled peers scaffolding An instructional practice that directs
in general education classrooms whenever possible the instruction to the present learning level of students
and to the fullest extent possible. with the intent of moving them to the next or higher
levels of understanding and proficiency.
p value An item accuracy value that is generated by
dividing the number of correct responses to an item scatterplot The graphic construction and presenta-
by the total number of students that respond to that tion of data that is used to help determine whether a
item. potential relationship exists or not; involves the plot-
ting of scores from two different tests or measures
range A basic variance measure that is obtained along an x- and y-axis.
when the lowest score from a group of scores is sub-
tracted from the highest score. script An internal strategy, usually involving an
internal memorized listing of steps or procedures, that
rating The examination of identif ied behaviors a learner may use to independently examine and mon-
based on a predetermined scale, usually involving a itor his or her learning progress.
range of point values.
selected response A test response format where a
rating label A narrative description used with student is required to select the correct answer from a
assignments to differentiate the highest from the low- set of potential answers that are already provided.
est performance level. Descriptive labels help students
identify and understand what qualities their work self-assessment A type of assessment that involves
must possess in order to obtain a desired rating or a review system (either internal or external) that
focuses on selecting or prioritizing individual learning
level.
goals or outcomes and monitoring progress toward
raw score The score that represents the number of those learning outcomes, as well as determining what
items that are correctly answered on a test or other individual adjustments, if any, are needed throughout
measure. See obtained score. the instructional experience.
reliability A measurement requirement that must self-referenced grading A grading model that
exist for any assessment process that involves the involves comparing each student’s learning progress
obtainment of the same or similar results each time a against his or her own estimated ability and progress.
measurement takes place.
sensorineural hearing impairment A hearing
reliability coeff icient A number in the form of a impairment or loss that occurs due to the impaired
decimal (ranging from +1.00 to –1.00) that is used to detection of sound waves and vibrations by the hair
represent the strength and direction of a linear rela- cells of the cochlea of the inner ear.
tionship between two events.
short-answer question A test response format that
response The correct corresponding answer to a requires a minimal written response, ranging from a
stem in a matching test question. few words to a full sentence or more.
I-1
Binary options, grading by, 227 Clarke, S., 353 Communication Curriculum-based measurement
Black, P., 6, 7, 8, 27, 35, Classroom assessment deaf/blindness and, 315 (CBM), 351, 352
168, 217 assessment practices electronic, 301
Blindness. See Deaf/blindness
Bloom, B. S., 6, 35, 36,
for, 14–20
definition of, 7–8
grading and, 227
manual, 315
D
Data
58, 59, 60, 114 effective, 14 oral, 315 classroom assessment
Bloom, F. E., 365 importance of, 2–3 self-assessment and, 170 system and, 394, 397,
Bloom’s taxonomy, 29, 58–60 ineffective, 14 Communication 399, 400, 402
Bloom’s revised taxonomy, in instructional process, 3 development, 315 evidence and, 66
29, 60–62, 77, 196 instruction-learning- Competency, assessment in Likert self-report
Blueprint for Reform: The assessment alignment and, 2, 18, 20 measure, 378
Reauthorization of in, 12–14, 17 Comprehension RTI and, 342, 343,
Elementary and Secondary instruction-learning- in Bloom’s taxonomy, 344, 345, 353, 355
Education Act, A, 5 assessment model in, 8–12 58, 59, 60, 77 for statistics, 256–265
Bohanon, H., 352 introduction to, 2 tests and, 123 Data collection, 394,
Bond, L. A., 137 purpose of, 4–6, 8 validity and, 92 398, 400, 401
Borg, W. R., 90, 95, reality of, 4–6 Concept maps, defi nition Data collection methods, 398
100, 224, 274 RTI and (See Response to of, 214 De Graaff, S., 34
Borich, G., 100, 265, 271 intervention (RTI)) Conceptual knowledge, 60, 61 Deaf/blindness, 313,
Bosman, A., 34 in student lives, 6–7 Concurrent criterion-related 315, 318, 321
Boston, C., 6, 7, 11, 115 Classroom assessment validity, 87, 88 Deluca, D., 310
Boud, D., 168, 169, 171 framework exercise, Conductive hearing Deno, S. L., 6, 351
Braaksma, A. D., 352, 353 394–403 impairment, 316 Descriptive feedback, 38, 119
Braden, J. P., 281, 283, Classroom assessment Conferences, parent- Descriptive ratings, 374
313, 331 models/cycles, 395 teacher, 300–304 Descriptive statistics, 255,
Brain, affective assessment Classroom assessment system Confidentiality, 18 377. See also Statistics
and, 365 collecting evidence of Confi rming questions, 77 Descriptors, 228, 230
Brain injury, 314, 321 learning in, 400–401 Congenital abnormalities, 318 Developmental delay, 314, 315
Brain Injury Association determining classroom Consensus-gram, 31 Developmental portfolio,
of America, 321 assessment measures Constructed response, 80, 200, 145, 214
Bramble, W. J., 262, 271 for, 396–399 206–211, 213–217, 299 Deviation IQs, 268, 289
Braun, Henry, 295, 296 effective, 392 Construct-related evidence Deviation score method, 264
Breivik, P. S., 49 focus of, 399 of validity, 85, 90–91 Diagnostic and Statistical Manual
Brennan, R. L., 137 function of, 395–396 Contemporary five-point of Mental Disorders—Fourth
Brice, A., 310 instructional learning letter grade system, 225 Edition (DSM-IV), 317, 319
Broadfoot, P., 171 impact in, 399–400 Content Diagrams
Brookhart, S. M., 10, 35, introduction to, 390 affective assessment and, 370 classroom assessment
223, 225, 234, 245 key element in, 398 classroom assessment system and, 398
Bruce, C. D., 167 making decisions in, 401–403 system and, 394, 398 visual construction
Bruns, D., 310 personal, 391–394 instructional, 34 performances and, 214
Buckendahl, C. W., 138, 154 purpose of, 392–393, 394 rubrics and, 155 Dianda, Marcella, 18
Burling, K. S., 115, 119 self-assessment of assessment self-assessment and, 170 DIBELS. See Dynamic
Burns, M., 340 foundation for, 403–404 Content-related evidence of Indicators of Basic Early
Bursuck, W. D., 245, 247, 248 Classroom portfolios, 145–148 validity, 85, 87–88, 92 Literacy Skills (DIBELS)
Butler, S. M., 66, 145, 225 Classroom teaching, Cooperation, group Differentiated instruction,
assessment and, 27–41 observation and, 71 34–35, 115
C Closed head injury, 321 Cooperative learning groups, 34 Direct questioning, 76–77, 78
Calfee, R., 145 Clubfoot, 318 Core instruction, 344, 345–347 Directions, assessment
Callahan, C. M., 138 Coe, R., 37 Coronado, V. G., 321 and, 109
Cambron-McCabe, N., 49 Cognition, affective Corrective feedback, 38, 119 Disability categories,
Canter, A., 340 assessment and, 365 Correlation coefficient, special education.
Carmines, E. G., 87 Cognitive development, 315 97, 268–270 See Special education
Carney, K. L., 352 Cognitive disability, Creation, in Bloom’s revised disability categories
Causation, defi nition of, 270 314, 317–318 taxonomy, 63, 77 Disability/disabilities. See
CBM. See Curriculum-based Cognitive skills, 317 Criteria, rubrics and, 156–159 also Learning needs
measurement (CBM) Cohen, L. G., 137 Criterion-referenced cognitive, 314, 317–318
Central tendency Cohen, S. A., 12 grading, 241–242 intellectual, 314, 317–318
definition of, 261–264 Collaboration Criterion-referenced learning, 357
Cerebral palsy, 318 group observation and, 71 standardized test, 286 multiple, 357
Chappuis, J., 6, 7, 8, 10, 114, self-assessment and, 170 Criterion-related evidence specific learning,
119, 165, 167, 180, 234 Collection phase, 147 of validity, 85, 87, 88 314, 319–320
Chappuis, S., 8, 10, 114, 119, College Entrance Examination Cronbach, L. J., 100 testing and, 246, 247
165, 167, 180, 234 Board (CEEB) Cross, L. H., 225 Disease-related
Checklists, 74–76, 398 standard scores, 268 Cummings, K. D., 349 impairments, 318
Cirino, P. T., 340 Collins, V. L., 340 Curriculum guide, 50, Distracters, defi nition
Clarity, self-assessment and, 170 Colvin, G., 340 51, 52–53, 54 of, 201
issues involving, 227–235 Hearing impairment, 314, Informal observation, 67–69 Integrity
narrative, 232 315, 316–317 Inquiry-based approach, 114 procedural, 96
norm-referenced, 240–241 Heritage, M., 35, 36, 132 Inquiry-based instruction, 34 reliability and, 95
as part of educational Herman, J. L., 119 Instruction Intellectual disability,
system, 224–227 High-level thinking skills, 150 affective assessment and, 370 314, 317–318
percentage, 225, 227 High-stakes testing, 281, 283 classroom assessment Intelligence quotient
percentage-based criteria Hill, W. H., 58 system and, 399 (IQ), 317
for, 241–242 Hills, John R., 18 core, 344, 345–347 Intensive instruction, 344, 347
performance-based Histogram, 259 differentiated, 34–35, 115 Interests, affective assessment
criteria for, 241 Hofstadter, L., 365 individualized, 35 (See also and, 369–370, 375
problems with, 225 Hogaboam-Gray, A., 167 Differentiated instruction) Internal consistency, 96
professional consensus Holistic rubric, 151, 152 inquiry-based, 34 Internal consistency
for, 239–240 Hopkins, B. R., 274 intensive, 344, 347 reliability, 99–100
purpose and, 224 Hopkins, K. D., 264, 269, 274 multimodal, 314 Internal consistency/split halves
self-referenced, 242–243 Horn, S. P., 295 online, 34 or odd-even, 96–101
special needs students Horner, R. H., 353 self-assessment and, 170 Internal progress audit, 180
and, 243–248 Hosp, J. L., 340 supplemental, 344, 346–347 Internal review of a test, 272
standards-based, 229–232 Hyman, J. S., 12 Instruction delivery model, 396 Internal review process, 181
Grading adaptation Hypothetical constructs, Instructional alignment, Inter-rater reliability, 101–105
approach, 247–248 90–91 35–36 Interstate New Teacher
Grading based on effort, 246 Instructional approach Assessment and Support
I formative assessment Consortium (INTASC), 2
Grading based on progress,
242, 247 and, 114 Interval recording, 73–74
IDEA. See Individuals with teaching techniques Intervals, 257–258, 259
Grading consistency, 239 Disabilities Education and, 34–35 Intervention, 320, 344,
“Grading on a curve,” 240 Act (IDEA) Instructional content, 34 346–347. See also Response
Grading practices/ IDEA-97. See Individuals with Instructional learning to intervention (RTI)
realities, 248–250 Disabilities Education impact, 399–400 Intervention services, 315, 321
Grading standards, 245–246 Act Amendments of Instructional material, Interview/interviewing,
Grading synchronicity, 239 1997 (IDEA-97) preview of, 33–34 79, 398
Grading uniformity, 240 IDEIA. See Individuals with Instructional preview, 33–34 Inventory
Grading variation, 239 Disabilities Education Instructional process, in affective assessment
Graduate Record Exam Improvement Act (IDEIA) assessment in, 3 measure, 380–382,
(GRE), 268 IEP. See Individualized Instruction-learning- 384, 385
Grant, M., 35 education plan (IEP) assessment alignment, in Likert self-report
Graphs, 214–215, 259 Illustrations, 214 12–14, 17, 143, 154, 156, measure, 376, 377, 378
Green, B. F., 137 Impaired articulation, 320 157, 194, 272, 394 self-report, 372 (See also
Grimes, J., 340 Impara, J. C., 138, 154 Instruction-learning- Likert self-report measure)
Griswold, P. A., 234 Inclusion, 311. See also assessment model Investigation area, in affective
Gronlund, N. E., 95, Integration classroom assessment assessment measure, 380
96, 192, 268 Inclusion-focused model, 310 and, 8–12, 27–41 IQ. See Intelligence
Gross motor skills, 315 Individualized education classroom teaching quotient (IQ)
Group achievement tests, 53, plan (IEP) and, 27–41 Item discrimination,
281, 283, 296, 300 CBM and, 357 components of, 28 274–276
Group observation, 69, 71, 72 components of, 312 elements of, 25–41 Item discrimination
Grouped frequency definition of, 245, 312 formative assessment index, 274
distributions, 257, 258, 259 grading and, 245, and, 35–36, 41, 114
Growth
self-assessment and, 171
246, 247, 248
learning needs and,
instructional approach
and, 34–35
J
Jayanthi, M., 247, 248
standardized tests and, 295 311–313, 326, 327 instructional preview
Jensen, E., 32
Growth portfolio, 214. See also Individualized instruction, 35. and, 33–34
Jimenez-Silva, M., 150
Developmental portfolio See also Differentiated introduction to, 26
Johnson, C., 35
Guskey, T. R., 6, 10, 14, 35, 36, instruction learning feedback and, 37–38
Johnson, E. G., 137
223, 225, 228, 230, 234, Individualized learning learning targets and, 27–30
Journals
238, 239, 245, 247, 248 plans, 327 pre-assessment and, 30–32
classroom assessment
Individualized questions, 77 self-assessment and,
system and, 398
H Individuals with Disabilities
Education Act
38–40, 41
summative assessment
definition of, 213–214
Hamilton, L. S., 295 Joyce, L. B., 313, 331
Amendments of 1997 and, 40–41
Handler, M. W., 353 (IDEA-97), 243 teaching techniques
Harrison, C., 27, 35 Individuals with Disabilities and, 34–35 K
Harrow, A., 59 Education Act (IDEA), 310 INTASC. See Interstate New Kame’enui, E. J., 340, 349
Hasselman, F., 34 Individuals with Disabilities Teacher Assessment and Kaminski, R., 349, 350
Hastings, J. T., 6, 58 Education Improvement Support Consortium Kartub, D. T., 353
Hattie, J., 6, 37 Act (IDEIA), 301, 310, (INTASC) Katz, J., 311
Haugen, L., 353 311, 313, 318, 319, Integration, 315. See Kleiner, A., 49
Healy, K., 340 324, 340, 357 also Inclusion Klin, A., 315
Knowledge Language/voice performance event and, 143 Marston, D., 340, 351
affective assessment and, 378 impairment, 320 RTI and, 340 Marzano, R. J., 10, 14,
in Bloom’s taxonomy, Lau, M., 340 tests and, 123 50, 225, 239
58, 59, 60, 77 Lawrence-Brown, D., 35 Learning performance Masia, B. B., 59
conceptual, 60, 61 Lazerson, A., 365 feedback, 35 Mason, E. J., 262, 271
factual, 60, 61 Learner control, self- Learning plans, 327 “Mastery learning”
learning and, 49–51, 53–54 assessment and, 181 Learning status, 396–397 approach, 35–36
meta-cognitive, 60 Learning Learning strategies, 181 Matching
of performance Bloom’s taxonomy Learning targets, 27–30, definition of, 200
assessment, 161 and, 58–64 54–58, 124–125 as selected response,
procedural, 60 classroom assessment system Learning tasks, 181 200, 204–206
self-assessment and, and, 393, 394, 397, Least restrictive setting, 311 Math Applications
178, 181, 183 398, 399, 400–401 Lee, C., 27, 35 CBM, 351
tests and, 123 demonstration of, 58–64 Lentini, R., 353 Math portfolio, 145
validity and, 92 emotional disturbance Lesson plans Mathematics Computation
Knowledge and skill and, 316 learning targets and, 28 CBM, 351
areas, 60, 64 evidence of, 65–69, 71 planning, 64–65 Matrices, defi nition of, 214
formative assessment Letter grades, 227–228, 229 Maze CBM, 351
Knowledge and skill base, 2, 58,
and, 398 Lichtenstein, R., 320, 357 McCaff rey, D. F., 295
78, 306, 310, 391, 398, 400
interviewing and, 79 Lifelong learners, 168 McConeghy, J. L., 247, 248
Knowledge and skill
introduction to, 49 Likert, Rensis, 372 McCurdy, B. L., 353
expectations, 51
knowledge and, 49–51, Likert self-report measure. See McDonald, B., 168, 169, 171
Knowledge and skill gains, 80
53–54 also Self-report inventory McGuire, J., 331
Knowledge and skill levels, mental retardation McKevitt, B. C., 352, 353
17, 312, 397 administering
and, 318 inventory in, 377 McMillan, J. H., 151, 225, 390
Knowledge and skill sets, NCLB and, 283 McMunn, N. D., 66, 145, 225
127, 145, 161, 365 analyzing responses in, 377
performance assessment constructing, 374–378 McTighe, J., 6, 7, 10, 14,
Knowledge and skill status, 397 and, 139, 145 33, 151, 154, 170
Knowledge and skills construction of items
performance-based, 149 in, 375–376 Mean
assessment and, 4, 9, 10, planning lessons for, 64–65 affective assessment and, 377
domain selection in, 375
11, 18 project-based, 148–151 aptitude scores and, 292
educational implications
classroom assessment system questioning and, 76–79 definition of, 261–262
in, 378
and, 394, 399, 403 standardized tests and, 287 Deviation IQs and, 268, 289
external review in, 376
formative assessment student products and, 79–80 dispersion from, 263–264
field testing in, 376–377
and, 123, 129 Learning connections/ SAT and, 268
modifications in, 378
grading and, 244 outcomes, 147–148 standard scores and, 290
scoring in, 377
instruction-learning- Learning disabilities, topic in, 375 standardized tests and, 289
assessment model and, 36 RTI and, 357 T scores and, 267, 289
using data in, 378
learning needs and, 326, 330 Learning environment, 170 Likert-based items, 400 z-score and, 265
performance assessment Learning experiences, 114 Linan-Thompson, S., 340 Measure revision, in affective
and, 138, 139, 143, 147 Learning feedback, 37–38 Linn, R. L., 137, 138 assessment measure, 385
self-assessment and, 180 Learning goals, 50, 51, Listing Median, defi nition of, 262
standards and, 54 54–58, 67 as constructed response, Mental retardation,
statistics and, 272 Learning groups, 200, 206 314, 317–318
summative assessment cooperative, 34 definition of, 200 Merwin, Jack C., 18
and, 201, 202, 216 Learning impact, instructional, Listing questions, Meta-cognitive knowledge, 60
testing and, 292, 296 399–400 defi nition of, 206 Meyers, J. L., 115, 119
validity and, 88 Learning modalities, 80, 144 Livingston, R. B., 86, Midpoint assessment, 397
Knowledge dimensions, 60 Learning needs. See also 98, 234, 288 Midstream assessment
Koretz, D. M., 295 Disability/disabilities Lockwood, J. R., 295 activity, 399
Kovaleski, J. F., 340 assessment modifications Long, B., 10, 35 Miller, D. D., 300, 301
Kraft, N. P., 300, 301 and, 324–331 Longworth, N., 49 Miller, L., 324
Krathwohl, D. R., 29, 58, 59, 60 IEP and, 311–313 Lucas, T., 49 Miller, M. D., 138
Kubiszyn, T., 100, 264, 271 introduction to, 310–311 Luiselli, J. K., 353 Mini-interview, 79
Kuder-Richardson special education disability Luria, A. R., 61 Minnis-Kim, M. J., 352
formulas, 100 categories and (See Mirenda, P., 311
Special education Mirkin, P. K., 6, 351
disability categories) M Mitchell, S., 150
L special education Madaus, G. F., 6, 58 Mixed hearing impairment, 317
Language disorder, 320 legislation and, 311 Mainstreaming, 311 Mode, defi nition of, 263
Language impairment. Learning objectives, 54–58 Manley, M., 234 Modeling, 186, 187
See Speech/language Learning outcome Mannella, M. C., 353 Modifications, 328,
impairment assessment modifications Manual communication, 315 378, 399
Language skills. See Speech/ and, 327 March, R. E., 353 Moon, T. R., 138
language skills classroom assessment Mark, defi nition of, Moroz, K. B., 352
Language-based learning system and, 393, 403 226–227 Moss, C., 10, 35
problem, 319 definition of, 56 Marshall, B., 27, 35 Moss, P. A., 137
Motivation 283–284, 294, 301, PBS. See Positive behavior definition of, 137
assessment and, 7, 94, 365 310, 327, 340, 354 support (PBS) evaluating, 139
reliability and, 98 Nondiscriminating Pearson Product-Moment features of, 138
Motor impairment, 329 item, 274, 275 Correlation, 269 focus of, 138
Motor performances, 144 Norm group, 285 Peer reviews, 150 knowledge of, 161
Motor skills, 80, 144 Normal curve, 260, 265, Percentage grading, 225, 227 student production
Multimodal instruction, 314 267, 268, 289 Percentage-based criteria, and, 142–143
Multiple choice Normal curve 241–242 Performance levels, 147,
definition of, 200 characteristics, 266 Percentiles 154, 155, 157–159
as selected response, Normal curve distribution, definition of, 265 Performance review, 177
200, 201–203 260, 289 standardized tests Performance-based
Multiple disabilities, 314, 318 Normal curve equivalent and, 288, 303 criteria, 241
Munk, D. D., 245, 247 (NCE), 267 Perceptions, affective assessment Performance-based
Normal distribution, 264–265, and, 365, 371, 375 learning, 149
Muscular dystrophy, 318
268, 288, 289, 303 Performance Perkins, C., 310
Muyskens, P., 340
Norman, G. R., 79 assessment modifications Personal beliefs, 370–371
Myran, S., 225, 390 and, 329 Personal perceptions,
Norm-referenced grading,
240–241 classroom assessment 365, 371, 375
N Norm-referenced standardized system and, 393, 398 Personality assessment, 366
Narrative, 72–73 tests, 284–287 correlation coefficients Personality testing, 366
Narrative format, Numerical grades, 228–229 and, 268–270 Physical development, 315
grading by, 227 Numerical ratings, 230 curriculum guides and, 54 Pilot, 381–382
Narrative grading, 232 data and, 256 Plante, L., 247
O
Narrative rating scales, 75 educational objectives and, 57 Polloway, E. A., 247, 248
Narrative responses, 237 evidence and, 66 Polman, J. L., 150
Oberg, C., 30 fine/gross motor Popham, W. J., 105, 201, 223
National Assessment of
Observations construction, 216 Portfolio
Education Progress
checklists as, 74–76 grading and (See Grading) advantages of, 146
program, 283
classroom assessment learning target and, 54 best-pieces, 145, 214
National Association of State
system and, 398 motor, 144 classroom, 145–148
Directors of Special
Education, 340 event recording as, 73–74 normal distribution and, 265 classroom assessment
formal, 69 reviewing, 216–217, 275 system and, 398
National Board for Professional
Teaching Standards group, 69, 71, 72 RTI and, 340, 342–343, definition of, 145, 214
(NBPTS), 2 informal, 67–69 344, 351 developmental, 145, 214
National Center on Educational interval recording as, 73–74 on standardized tests (See disadvantages of, 146
Outcomes, 331 narrative as, 72–73 Standardized tests/testing) growth, 214 (See also
National Council on ratings as, 74–76 test, 80, 260, 268, 271, 303 Developmental portfolio)
Measurement in student products and, 79 testing, 271–272 math, 145
Education, 18 types of, 71–76 verbal, 79–80, 144 process-oriented, 145
National Education Association Obtained scores, 265 verbal construction, 215–216 product, 145
Code of Ethics, 16–17 O’Connor, K., 6, 7, 10, visual, 144 skill-focused, 145
14, 33, 170, 234 visual construction, 214–215 strengths of, 146
National Education Association
Odd-even reliability, 100 written, 144 theme-focused, 145
(NEA), 15, 18
Ohio Department of written construction, weaknesses of, 146
National Research Council, 32
Education, 28 213–214 writing, 145
NBPTS. See National Board written expression, 80
Ohio Resource Center, 28 Portfolio construction
for Professional Teaching
One-on-one tutoring, 34 Performance assessment process, 147–148
Standards (NBPTS)
Online instruction, 34 checklists as, 159–161 Positive behavior support
NCE. See Normal curve classroom portfolios (PBS), 345, 351–353
equivalent (NCE) Open head injury, 321
Oral communication, 315 as, 145–148 Positive discrimination
NCLB. See No Child Left definition of, 137–139 item, 274
Behind (NCLB) Orthopedic impairment,
314, 318 focus of, 138, 139, 142–143 Positive skew, defi nition
NEA. See National Education knowledge of, 161 of, 261
Association (NEA) Other health impairment,
314, 318–319 motor performances for, 144 Positively skewed
Negative discrimination PBL as, 148–151 distribution, 263
item, 274 Outcome measurement, 327
rubrics as, 139, 140–141, Post-assessment, 397, 398
Negative skew, defi nition 151–159 Post-instruction assessment, 397.
of, 261 P verbal performances See also Post-assessment
Negatively skewed P value, 272, 273, 274 for, 144 Powell-Smith, K. A., 349
distribution, 263 P-12 education, 6, 51 visual performances Power tests, 100, 329
Negatives, 299 Parents, grading and, 237 for, 144 Prasse, D., 340
Newbill, C., 324 Parent-teacher conferences, written performances Prater, K., 340
Nichols, P. D., 115, 119 300–304 as, 144 Pre-assessment, 9–10,
Nine-point scale, 228 Pass-fail grades, 229 Performance comparison, 289 30–32, 399, 400
Nitko, A. J., 18, 151, 223 Pass-fail system, 225 Performance event Predictive validity, 88, 90
No Child Left Behind (NCLB), PBL. See Project-based assessment modifications Pre-instruction assessment, 397.
4–5, 34, 41, 243, 281, learning (PBL) and, 329 See also Pre-assessment
Premises, 204
Preparation, for standardized
Q Research-supported
practices, 349–353
Roderique, T. W., 247, 248
Rolheiser, C., 167
Questioning
tests, 296–297 direct, 76–77, 78 Respect, affective Rosenfeld, J., 79
Presentation assessment and, 370 Ross, J. A., 167
focused, 78, 123
assessment change in, 328 Response modifications, 330 RTI. See Response to
learning and, 76–79
visual, 259 Response to Intervention: intervention (RTI)
Questionnaire, 367, 372.
Probes, 351 Policy Considerations and Rubrics, 139, 140–141
See also Survey
Probing questions, 77 Implementation, 340 analytical, 151, 152
Questions
Problem-solving/classroom Response to intervention (RTI) classroom assessment
assessment and, CBM and, 351
intervention cycle, 343 system and, 398
57–58, 78, 109
Procedural error, 95 in classroom, 342–349 effective, 153–155
classroom assessment definition of, 340
Procedural integrity, 96 focused, 151
system and, 395 introduction to, 340–341
Procedural knowledge, 60 general, 151
confirming, 77 learning needs and, 320
Process holistic, 151, 152
assessment, 326–327 essay, 208 PBS and, 351–353
individualized, 77 making, 155–159
definition of, 234–235 research-supported qualities of, 154
rubrics and, 155 lesson plans and, 64 practices and, 349–353
listing, 206 summative assessment
Process-oriented portfolio, 145 teachers as problem solver/ and, 193
Product mini-interview and, 79 researcher and, 353–357
probing, 77 task-specific, 151
definition of, 234–235 Responses
student, 79–80, 144 spontaneous, 77 affective, 376
Product portfolio, 145 standard, 77 constructed, 80, 200, 206– S
Production, student, 142–143 teacher-directed, 78 211, 213–217, 299–300 Sanders, James R., 18
Professional responsibility, 171 teacher-generated, 77 definition of, 204 Sanders, W. L., 294, 295
Progress extended, 209–210 SAT Reasoning Test, 88, 268
assessment modifications R in Likert self-report Sattler, J. M., 69, 73, 314, 321
and, 331 Range, defi nition of, 257 measure, 377 Scaffolding, defi nition of, 10
classroom assessment system Rathkey, C., 150 narrative, 237 Scatterplot, 269–270
and, 392, 393, 397, Rating labels, 158 restricted, 208–209 Schneider, Jeff rey, 18
398, 399, 400, 401 Ratings, 74–76 selected, 80, 200–206, 297 Score report, 287, 294
definition of, 234–235 Raw scores standardized tests and, Scores
developmental delay and, 315 definition of, 265 297–300 aptitude, 292
emotional disturbance standard deviation and, 264 student, 329–330 gain, 294, 295
and, 316 standardized tests and, 288 Responsibility obtained, 265
evidence and, 65–67, 69 z-scores and, 267 affective assessment and, 370 raw (See Raw scores)
formative assessment Readability, defi nition of, 109 classroom assessment standard (See Standard scores)
and, 114, 398 system and, 397 T, 267, 289
Reading CBM, 351
grading based on, 242, 247 professional, 171 test (See Test scores)
Reading literacy program, 349
IEP and, 312 Restricted response, 208–209
Reading research, 349 Scoring
NCLB and, 283 Re-teaching, 36
Reading skills, 349 in Likert self-report
performance assessment Reviews, 34, 117–119
Reflection, 147 measure, 377
and, 139, 147 in affective assessment
Rehabilitation Act of 1973, 322 NCLB and, 283
RTI and, 345, 347, 351 measure, 381
Reiter, H. I., 79 summative assessment
self-assessment and, 180, 183 checklists as, 160
Reliability and, 200, 205, 206,
standardized tests and, 287 classroom assessment
alternate form, 99 207, 209, 210, 211
summative assessment system and, 403
assessment and, of disability categories, 313 Scott, S., 331
and, 41
Progress checklists, 174–175 92–94, 94–105 formative assessment Scott, T. M., 352
Progress manager, 180–185 definition of, 94–95 and, 119–122 Script, 178, 179–180, 186
Progress monitoring, 170 forms of, 96–101 in Likert self-report Seclusion model, 310
Progress review script, 179–180 internal consistency, 99–100 measure, 376, 378 Section 504, 322, 324, 326, 327
Project-based learning inter-rater, 101–105 peer, 150 Selected response, 80,
(PBL), 148–151 odd-even, 100 performance, 177 200–206, 297
Project-oriented learning split halves, 100 performance assessment Selection phase, 147
event, 150 stability, 95, 96, 97, 98 and, 147, 150, 151, Self-assessment, 38–40, 41,
Provost, M. C., 354, 355 test-retest, 97 154, 160, 161 118, 119, 147, 150
Public Law 94-142, 311 types of, 96, 97 responses to test items of assessment foundation,
Public Law 99-457, 311 Reliability coefficient, 97 for, 276 403–404
Purpose Reschly, D. J., 340 rubrics and, 151, 154 classroom assessment system
of assessment measure, 193 Research self-assessment and, 181 and, 394, 397, 398
of classroom assessment, on academic feedback, 37 standardized tests and, 300 elements of, 170–171
4–6, 8 on assessment, 6 test, 198 introduction to, 167–168
of classroom assessment on formative assessment, 35 test completion, 199–200 progress manager
system, 392–393, 394 reading, 349 Revision, in affective and, 180–185
of grading, 224, 235, Research portfolio project, 145 assessment measure, 385 teacher and, 186–187
236, 238, 239 Research-based Reynolds, C. R., 86, 98, teaching and, 169–171
Putnam, R. F., 353 intervention, 320 234, 288 Self-assessment process, 178
Self-assessment techniques, test-taking, 296 Stability reliability, 95, 96, Standards-based tests, 283
171–180 visual, 80 97, 98 Stanine, 268, 288
Self-care skills, 315 Small groups, 347 Stability/test-retest, 96–101 Stanley, J. C., 274
Self-esteem, assessment and, 7 Smith, B., 49 Standard deviation Statistics
Self-monitoring, 400 Social development, 315 affective assessment and, 377 data for, 256–265
Self-referenced grading, Sousa, D. A., 31, 365 aptitude scores and, 292 descriptive, 255, 377
242–243 Spandel, J., 145 computing, 263–264 difficulty level of test
Self-reflective process, 187 Spearman-Brown formula, 100 definition of, 263 items and, 272–273
Self-regulation model, 170 Special education disability Deviation IQs and, 268, 289 introduction to, 255
Self-report inventory, 372. categories normal distribution and, item discrimination
See also Likert self- autism as, 313, 314–315 264–265 for, 274–276
report measure cognitive disability as, SAT and, 268 in reviewing student
Self-review, 184 314, 317–318 SEM and, 271 performances, 275
Self-review process, 175 deaf/blindness as, 313, standard scores and, 290 standard scores for, 265–271
Self-review skills, 187 315, 318, 321 standardized tests and, 289 testing performance
Self-review steps, 171 developmental delay as, T scores and, 267 for, 271–272
SEM. See Standard error of 314, 315 z-score and, 265 uses for, 255–256
measurement (SEM) emotional disturbance as, Standard error of measurement Stehouwer, J. D., 340
Senge, P., 49 314, 315–316 (SEM), 270–271 Stem, defi nition of, 201
Senn, J. A., 49 gifted and talented Standard questions, 77 Stiggins, R. J., 6, 7, 8, 10, 35,
Sensorineural hearing students as, 322 Standard scores 115, 132, 165, 167, 234
impairment, 317 hearing impairment as, aptitude scores and, 292 Stoner, G., 319
Sensory impairments, 315 314, 315, 316–317 CEEB, 268 Student attitudes, 365, 367, 369
Sensory information, 365 intellectual disability as, definition of, 265 Student interests, 369–370, 375
Setting, assessment change 314, 317–318 parent-teacher conferences Student outcomes,
in, 328–329 mental retardation as, and, 303 defi nition of, 27
Seven-descriptor format, 374 314, 317–318 for statistics, 265–271 Student production, 142–143
Shaw, S., 331 multiple disabilities as, test profile and, 289, 290 Student products, 79–80, 144
Shin, J., 351 314, 318 Standardized tests/testing Student response, assessment
Shinn, M. R., 340, 351 orthopedic impairment as, assessment modifications change in, 329–330
Short answer 314, 318 and, 327 Student teaching,
as constructed response, criterion-referenced, 286 assessment and, 27
other health impairment as,
200, 207–208 definition of, 284 Sudweeks, R. R., 96, 263,
314, 318–319
definition of, 200, 207 IEP and, 313 268, 288
specific learning disability
Silva, M., 247 interpreting test scores Sugai, G., 340
as, 314, 319–320
Simmons, D. C., 349 on, 287–294 Summative assessment,
speech/language impairment
Simpson, E. J., 59 introduction to, 281 40–41, 117, 119
as, 314, 320–321
Skewed distributions, 260, 261 NCLB and, 281, 283–284 classroom assessment system
TBI as, 314, 321
Skill areas, 60, 64 norm-referenced, 284–287 and, 394, 397, 398, 400
visual impairment as,
Skill base, 2, 58, 78, 306, parent-teacher conferences constructed response items
314, 321
310, 391, 398, 400 and, 300–304 in, 200, 206–211, 213–217
Skill expectations, 51 Special education
preparation for, 296–297 definition of, 10–11, 114, 192
Skill gains, 80 legislation, 311
results of, 300–304 introduction to, 192
Skill levels, 17, 312, 397 Special education services,
test-taking strategies selected response items
Skill mastery, 286–287 356, 357
for, 297–300 in, 200–206
Skill sets, 127, 145, 161, 365 Special education teachers, 310 VAMs and, 294–296
Special needs table of specifications
Skill status, 397 Standards and, 211–213
Skill targets, 122–123 in classroom, 325 classroom assessment
grading students Summative assessment
Skill-focused portfolio, 145 system and, 393, 394 construction, 193–200
Skills with, 243–248 definition of, 51
IEP and, 313 Supplemental instruction/
auditory, 80 ethical, 15
NCLB and, 327 intervention, 344,
Basic Early Literacy, 349 grading, 245–246 346–347
cognitive, 317 summative assessment and, 41 group achievement
Specific learning disability, Survey, 367, 370, 371, 372,
Dynamic Indicators of Basic tests and, 53
314, 319–320 374, 375, 376, 377, 385
Early Literacy, 345, 349 reliability and, 98
Speech/language impairment, Symbol-based methods, 228
fine motor, 315 science, 55
314, 320–321 Symbols, visual construction
gross motor, 315 teachers and, 54
Speech/language skills, 315 performances and, 214
high-level thinking, 150 Standards for Educational
Speed test, 100, 329 Synthesis, in Bloom’s
knowledge and (See and Psychological
Spelling, 207 taxonomy, 58, 59, 60
Knowledge and skills) Testing, 86, 327
language (See Speech/ Spelling CBM, 351 Standards for Teacher Competence
language skills) Spenciner, L. J., 137 in Educational Assessment T
motor, 80 Spina bifida, 318 of Students, 18 T scores, 267, 289
reading, 349 Spinal cord injuries, 318 Standards-based grading, Table of specifications,
self-care, 315 Split halves reliability, 100 229–232 defi nition of, 88
self-review, 187 Spontaneous questions, 77 Standards-based performance Tables, defi nition of, 214
speech/language, 315 Spriggs, M., 353 indicators, 227 Task-specific rubric, 151
Taxonomy table, 60, 63 Test-retest reliability, 97 Topic, in Likert self-report Viewpoints, affective
Taylor-Greene, S., 353 Test-taking skills, 296 measure, 375 assessment and, 371, 375
Tayrose, M. P., 281, 283 Test-taking strategies, 297–300 Torgesen, J. K., 340 Visscher, A. J., 37
TBI. See Traumatic brain Testing Towler, L., 171 Visual construction
injury (TBI) field (See Field testing) Traumatic brain injury performances, 214–215
Teacher-directed questions, 78 high-stakes, 281, 283 (TBI), 314, 321 Visual impairment, 314,
Teacher-generated questions, 77 P-12 education and, 51 “Triangulation” effect, 123 315, 321
Teachers personality, 366 Trice, Carolyn, 18 Visual performances, 144
competency standards reliability and, 98 True-false Visual presentations, 259
for, 18, 20 summative assessment definition of, 200 Visual skills, 80
formative assessment and, 40–41 as selected response, Voice impairment. See
and, 131–132 Tests. See also Assessment 200, 203–204 Language/voice
grading issues for, 227–235 achievement, 292–294 Tutoring, one-on-one, 34 impairment
knowledge and, 54 classroom assessment Volkmar, F. R., 315
as problem solver, 353–357 system and, 394, 398 U
questioning and, 78
reliability for, 92–94
constructed response and,
80, 200, 206–211
Uniform grading standard, W
245–246 Wagner, M. M., 246
as researcher, 353–357 determining difficulty Universal Design for Wald, M. M., 321
self-assessment and, 186–187 level of, 272–273 Instruction, 331 Walker, H. M., 340
as self-assessment model, 39 ethics and, 18 Walters, R. H., 40
special education, 310 formative assessment
validity for, 92 and, 122–123 V Weber, L. J., 225
Wechsler IQs, 268
Teaching group achievement, 53, Valdes, K. A., 246
Wesson, C., 351
classroom, 27–41 281, 283, 296, 300 Valencia, S. W., 145
Wetzel, K., 150
self-assessment and, 169–171 internal review of, 272 Validity
White, K. R., 90, 95, 96, 100,
student, 27 NCLB and, 5 assessment and, 85–94
224, 263, 268, 274, 288
Teaching techniques, power, 100, 329 categories of, 86
Wiggins, G., 114, 137
instructional approach selected response items classroom assessment
Wiliam, D., 6, 7, 8, 27, 35,
and, 34–35 for, 200–206 system and, 394
119, 168, 217
10-point grading scale, 228, 229 speed, 100, 329 concurrent criterion-
Williamson, C. L., 246
Test anxiety, 297 standardized (See related, 87, 88
Willson, V., 86, 98, 234, 288
Test awareness, 296 Standardized tests/testing) construct-related evidence
Wininger, S. R., 217
Test completion review, standards-based, 283 of, 85, 90–91
Winters, L., 119
199–200 Theme-focused portfolio, 145 content-related evidence
Witte, R., 185, 321
Test construction, 198, 199 Thornton, H., 246 of, 85, 87–88, 92
criterion-related evidence Workman, D., 225, 390
Test construction rules, 197–200 Three-point scale, 225 Worthen, B. R., 90, 95, 96, 100,
Test development, 197–198 Three-tiered intervention of, 85, 87, 88
definition of, 86, 87 224, 263, 268, 274, 288
Test environment, 297 model, 345 Wright, P., 294
Test implementation, 199 Tilly, W. D. III, 340, predictive, 88, 90
Value-added models (VAMS), Wright, R. G., 238
Test items, student 344, 346, 347 Writing portfolio, 145
responses to, 276 Time 294–296, 300
Written construction
Test preparation, 296–297 affective assessment and, 372 Values, affective assessment
performances, 213–214
Test profi le, 288–292 assessment change and, 370–371
Written Expression CBM, 351
Test results, 300–304 in, 328–329 Van Houten, R., 37
Vanderwood, M., 340 Written expression
Test review, 198 classroom assessment performances, 80
Test scores system and, 390, 400 Variability, defi nition of,
263–264 Written performances, 144
interpreting, 287–294 essay questions and, 210
normal curve characteristics PBL and, 150 Variance, defi nition of,
and, 266 performance assessment 263–264 X
normal distribution and, 265 and, 138 Vaughn, S., 340 Xu, L., 321
parent-teacher conferences standardized tests and, 300 Vellutino, F. R., 340
and, 303
as part of educational
test implementation and, 199
Time sampling, 73. See also
Verbal construction
performances, 215–216
Z
Zeller, R. A., 87
experience, 304 Interval recording Verbal performances, Zigmond, N., 246
SEM and, 271 Timperley, H., 6, 37 79–80, 144 z-scores, 265, 267
standard deviation and, Tindal, G., 351 Verbs, educational
263, 264 Todd, A., 353 objectives and, 57
standardized tests and, 295 Tomlinson, C. A., 34, 35 Verhoeven, L., 34