Download as pdf or txt
Download as pdf or txt
You are on page 1of 12

Dawn Ro g ie r

P h i l i pp i n e s

Assessment Literacy: Building


a Base for Better Teaching
and Learning

A
re you assessment literate? able about the key concepts of testing
What does that mean—to be and how they can inform the design
assessment literate? Assess- of assessments and decisions sur-
ment is something that we as teachers rounding their usage. This article will
must do all the time, but many of start you on the path to being assess-
us feel unprepared or uncomfortable ment literate. You will learn about the
when it comes to testing our students. terms that make up the cornerstones
Teachers often reuse tests without of testing, how to plan your courses
analyzing or revising them and sel- with assessments in mind, and how
dom use statistical procedures to see to make a test blueprint. Knowing
how a test—or a test item—is actu- more about assessment will not only
ally performing. Assessing students help you to assess your students more
often means reaching for a test or effectively, but it will also provide you
quiz that is already prepared, whether with a means of evaluating your own
it be a test included with a textbook, teaching and help you to produce
something another teacher prepared, tests that will actually motivate your
students to learn. Let’s begin by learn-
or a standardized test produced by
ing more about the words testing and
a major testing organization or our
assessment.
institution. These aren’t necessarily
bad choices (and sometimes it may Testing vs. assessing
not be our choice at all), but to make The word test can make people
sure they are good choices, we must nervous. It has semantic qualities that
be knowledgeable about the prin- make us think of being judged or
ciples and practices of assessment. measured by someone or something.
In order for assessment to be Many people have an emotive reaction
effective, classroom teachers need to to testing and associate it with negative
be assessment literate—knowledge- experiences that they may have had as

2 2014 N u m b e r 3 | E n g l i s h T e a c h i n g F o r u m
students. In an educational context, the terms Teacher-education programs are also at fault
testing and assessment are often used inter- for not making sure teachers are adequately
changeably to indicate the measurement of trained before entering the classroom (Mertler
student learning. However, although a test is a 2004). As Taylor (2009) points out,
type of assessment—usually thought of in the language education programs at
traditional sense of an exam or quiz—assess- graduate level typically devote little
ment is a more comprehensive term. It often time or attention to assessment
indicates the collection of information about theory and practice, perhaps just
student learning that might include not only a short (often optional) module;
tests but also a variety of techniques such as and although there is no shortage
performance tasks, portfolios, and observation. of books on language testing and
While tests are thought of as a means to assessment available today, many of
give grades to students, assessments offer these are perceived to be (and often
diagnostic information for both students and are) highly technical or too special-
teachers. The ultimate purpose of assessment ized for language educators seeking
is to improve student learning, as opposed to to understand basic principles and
just being able to give a mark for the amount practice in assessment. (23)
of course content a student has mastered.
During our time in school and teacher-train-
Today teachers tend to talk about assessing
ing courses, we take many tests, but how often
(rather than testing) their students because we
are we actually given practice creating them,
see the ongoing evaluation of student learn-
marking them, and interpreting the results?
ing as more than just testing knowledge and
Developing these skills is part of becoming
skills in a particular area at one point in time
assessment literate.
for grading purposes. Thus, throughout this
article, references to tests will be made with Assessment literacy
the ultimate goal of using them as assessment An essential element of assessment literacy
tools and not purely as testing instruments. is the ability to connect student assessment to
Importance of assessment
the learning and teaching process. Teachers
can make this link by first matching test items
As we all know, assessment plays an impor- to instructional objectives, then using the test
tant role in teaching and learning. It affects results to provide feedback on both student per-
decisions related to instruction, determines formance and how well the instructional objec-
the extent to which instructional objectives tives were met. An assessment-literate teacher
are met, and provides information for admin- is able to interpret data generated from a test
istrative decisions. It has been estimated that to make useful modifications to teaching and
teachers spend as much as 50 percent of their to use assessments as a tool to improve student
time in assessment-related activities (Stiggins learning. Assessment-literate teachers are also
1991), and that when assessment is imple- able to discuss assessments with others in terms
mented effectively, student achievement is of key concepts in testing. With this in mind,
improved (Campbell and Collins 2007). we can explore common terms associated with
Yet many teachers feel assessment and test- tests, along with their practical application.
ing are not relevant to their classroom practice
and report that they feel unprepared to under- Key concepts and considerations
take assessment-related activities. Popham Seven key concepts—usefulness, reliability,
(2004) reports that most public school edu- validity, practicality, washback, authenticity,
cators in the United States tend to think of and transparency—are cornerstones in testing
assessment as “a complex, quantitative arena that help to ensure that a test is solid (i.e., that
well beyond the comprehension of mere mor- it will consistently measure what you want it
tals” (82). Some of these feelings may come to measure in an efficient manner, and that
from the anxiety that teachers felt when they both teacher and student will see it as a valu-
were students taking tests, especially if they able source of information regarding learn-
didn’t understand how the tests were graded ing). Understanding these concepts and being
or if the objectives of the tests weren’t clear. able to improve practices related to them are

E n g l i s h T e a c h i n g F o r u m | Number 3 2014 3
important in developing assessment literacy. administering it. If you choose a pre-made test
Each is discussed separately below, but as you and it does not match your students’ needs or
will notice, they are connected to and support your purpose, then it will not be an adequate
one another; together, they form the basis for assessment of your students and will not pro-
building solid assessments. vide the information that you need in order to
make informed decisions about the teaching
Usefulness and purpose and learning taking place in the classroom.
According to Bachman and Palmer (1996), For example, if you wanted to measure the
usefulness is the most important consid- reading ability of your students to see if they
eration when choosing or designing a test. would be able to order from a menu when
Teachers must consider what the purpose of visiting the United States on an exchange
a particular assessment is and whether this trip, you couldn’t just use any reading test you
purpose is congruent with the students they find in a textbook or online. You would need
are testing and the course they are teaching. to find one (or better yet, make one) that is
All language tests must be developed with a specific to the skills taught in class, that meets
specific purpose, a particular group of test the vocabulary needs of the situation the stu-
takers, and a specific language use in mind. dents would be immersed in, and that uses
Even tests with the general purpose of test- an appropriate text style that matches what
ing English language ability (proficiency) are you expect the students to encounter. Having
designed with a specific group of test takers in them read a passage from a newspaper or a
mind. Take, for example, three standardized short story and then answer questions would
tests used globally for the purpose of measur- not adequately measure their ability to read
ing language ability: the Test of English as a and order from a menu at a restaurant. So
Foreign Language (TOEFL), the Internation- when you choose or design a test, consider the
al English Language Testing System (IELTS), purpose of the test, the group of test takers it
and the Michigan English Test (MET). Each is designed for, and the specific language use
of these has been developed with very specific you want to evaluate.
audiences and purposes (see Figure 1).
The examples in Figure 1 illustrate that Reliability
tests are designed with very specific audi- Your assessments not only need to be use-
ences and purposes in mind. This specificity ful for the intended purpose, they also need to
is what allows them to effectively measure be reliable. Reliability refers to the consistency
what they are designed to measure and makes of test scores. If you were to test a student
them useful for a specific purpose. You must more than once using the same test, the results
carefully consider the purpose of a test before should be the same, assuming that nothing

Test Information about the Purpose

TOEFL Measures the ability “to use and understand English at the university level,” and
evaluates how well the test taker can “combine listening, reading, speaking and
writing skills to perform academic tasks” (Educational Testing Service 2014).

IELTS Has an academic and a general-training version. The academic version is for
those who want to study in an English-speaking university; the general version
focuses on basic “survival skills in broad social and workplace contexts” (IELTS
2013).

MET “Intended for adults and adolescents at or above a secondary level of education
who want to evaluate their general English language proficiency” in social, edu-
cational, and workplace contexts. It is “not an admissions test for students apply-
ing to universities and colleges in the United States, Canada, and the United
Kingdom” (Modern Language Center 2010).

Figure 1. Examples of standardized proficiency tests and their purposes

4 2014 Number 3 | E n g l i s h T e a c h i n g F o r u m
else had changed. Reliability can be threat- written questions are not beyond the level
ened by fluctuations in the learner, in scoring, of the students. Otherwise, you will be test-
or in test administration. Fluctuations in the ing them on more than just their listening
learner are out of the testing administrator’s comprehension skills and thus decreasing the
control; we cannot control whether a student validity of the test as a measure for listening
is sick, tired, or under emotional stress at the ability.
time of a test. But we as teachers can limit the A number of factors can have an adverse
fluctuations in scoring and test administra- effect on validity, including the following:
tion. The guidelines for how a test is adminis- • unclear directions
tered, the length of time allotted to complete • test items that ask students to perform
the test, and the conditions for testing should at a skill level that is not part of the
be established in advance and written in a course objectives
test-specifications document. (See the Validity • test items that are poorly written
section for an example of test specifications.) • test length that doesn’t allow for ade-
As much as possible, there should be consis- quate sampling or coverage of content
tency in testing conditions and in how a test • complexity and subjectivity of scor-
is administered each time it is given. Teachers ing that may inaccurately rank some
can minimize fluctuations in score by prepar- students
ing answer keys and scoring rubrics, and by The best way to ensure validity and reli-
holding norming sessions with those who will ability is to create test specifications and exam
be scoring the test. blueprints. These will help ensure that tests
You can take steps to improve the reli- created and used match what is intended for
ability of your tests. You need to make sure the course and the students. Figure 2 shows
that the test is long enough to sample the an example of general information for the
content that students are being tested on and test specifications of a final exam for a high-
that there is enough time for most of the er-education pre-academic English-language
students to finish taking the test. The items program course. For each of the subtests (lis-
should not be too easy or too difficult, the tening, reading, and writing), specifications
questions should not be tricky or ambiguous, would also be written and would include the
the directions should be clear, and the score type of skills being assessed, level of vocabu-
range should be wide. Before you administer lary, grammar structure, and length of text to
the test, you might want to have someone else be used.
take it to see whether he or she encounters
problems with directions or content. Use that Practicality
person’s feedback to see where the test might Practicality refers to how “teacher friendly”
need to be improved. a given test is. Practicality issues include the
cost of test development and maintenance,
Validity time needed to administer and mark the test,
One thing to keep in mind is that a test ease of marking, availability of suitably trained
may be highly reliable, but not valid. That is, markers, and administration logistics. If the
it might produce similar scores consistently, test you want to give requires computers, and
but that does not mean it is measuring what these are not available or connectivity is unre-
you would like it to. A test has validity when liable, there will obviously be a practicality
it measures what you want it to measure. The issue with the delivery of the test. For many
most important aspect of validity is the appro- teachers, the amount of time required to
priateness for the context and the audience of mark a test is an important practicality issue.
the test. Think about what is to be gained by You can overcome this issue by weighing how
administering a test and how the information important a particular assessment is in terms
will be used. Suppose your goal is to measure of overall course mark and determining how
students’ listening ability, and you give a test much time you want to spend marking it.
in which students answer questions in written For example, if a vocabulary quiz will not
format about a lecture they hear. In that case, be worth much in the overall course mark,
you need to make sure that the vocabulary, you might consider having students exchange
sentence structure, and grammar usage in the papers and mark them instead of marking

E n g l i s h T e a c h i n g F o r u m | Number 3 2014 5
General Test Information — Final Exam
Purpose (Why are you testing?) To test student mastery of listening, reading, and writ-
ing curricular objectives for Level 1
Intended population (Who are you Students in university pre-academic intensive English
testing?) program – Level 1
Intended decisions/stakes (How High stakes – weighted as 40% of the final grade
important is the test for the course
grade?)

Response format (What type of Listening: multiple choice, short answer, matching, gap
questions will you use? How will the fill, and information transfer
test taker show mastery of the
objective?) Reading: multiple choice, short answer, matching, gap
fill, and information transfer
Writing: one-paragraph response to a prompt (input is
a picture or personal knowledge)
Number of examiners (How many One test supervisor per 20 students; two markers per
people are needed to administer the exam (cannot be the class teacher)
test? Are there any restrictions for
test supervisors?)
Number and weighting of items/ Listening: approximately 20 items (33%)
tasks (How many questions will Reading: approximately 20 items (33%)
there be on each part? How much Writing: 1 task (34 %)
will each part be worth for the
overall grade of the test?)
Examination length (How much Maximum of 2 hours
time will the assessment take overall? Listening: 30 minutes
Is there a time length per section?) Reading: 40 minutes
Writing: 40 minutes
Order of tasks (In what order will 1. Listening
the sections be tested?) 2. Reading
3. Writing

Rating scale type (Conditions Reading and Listening: Answer key agreed to before
necessary for marking the exam) the test; Writing: Two markers, analytical criteria, third
marker if necessary

Reporting type (How will the score Single score (maximum 100%; pass mark 70%)
be reported? As a whole score, or per
section? What is the passing grade?)

Figure 2. Example of general test specifications

each one yourself. This arrangement also Washback


allows students to review the materials at the Washback refers to the effects of testing on
same time. For marking writing, it might be students, teachers, and the overall program. It
more practical to have students review each can be positive or negative. Positive washback
other’s work and peer edit the first draft than occurs most often when testing and curriculum
to have the teacher make comments on each design are based on clear course outcomes that
initial draft. are known to all students and teachers. On

6 2014 Number 3 | E n g l i s h T e a c h i n g F o r u m
the other hand, exams that require extensive aware of the skills, vocabulary, and grammar
preparation can have negative washback and be that they will be expected to learn, and they
harmful to the teaching and learning process; if should receive a clear explanation of how
instruction solely focuses on helping students these will be assessed. Transparency makes
pass the test, other learning activities may be students part of the testing process by ensur-
neglected. To make sure washback is positive, ing that they understand what the course
teachers should link teaching and testing to objectives are and what will be tested, as well
instructional objectives. Tests should reflect the as the format of tests and how they will be
goals and objectives of the course along with the used and graded. Students should have the
types of activities used to teach the content. That chance beforehand to practice question types
underscores the importance of planning assess- that will be used in a test. Using a new test
ments at the same time you plan the course. format, one that students are unfamiliar with,
Another way to bring about positive wash- could affect the test’s reliability. When stu-
back is through feedback. Providing feedback dents do not perform well on a test, it should
in a timely manner is important if you want be because they have not learned the material,
students to learn and benefit from the assess- not because they didn’t understand the direc-
ment process. In the above example of practi- tions to complete a task.
cality, having students mark their classmates’ Increasing transparency will also reduce stu-
papers provides timely feedback to the stu- dents’ test anxiety and allow them the chance
dents and helps them understand where they to perform better. To increase transparency,
might need further practice or review. Using many schools and educational institutions pub-
short quizzes that are graded immediately by lish their test specifications. For example, the
the students throughout the course may let Oregon Department of Education publishes
students know where they need to study more; test specifications for the English Language
it may also redirect teacher energy toward the Proficiency Assessments by grade level on its
areas that need more instruction time. website (Oregon Department of Education
Involving students in the marking process 2014). These documents list not only content
is one way to create positive washback from to be tested, but also in what ratio, along with
testing. Other ways are to use authentic testing appropriate test-item types.
materials and to make the assessment procedures Planning your assessments
transparent—the topics of the next two sections.
Now that you are familiar with the seven
Authenticity cornerstones of assessment, let’s examine how
Tasks that reflect real-world situations and you would go about planning an assessment
contexts in which the language will be used that is useful, valid, reliable, practical, authen-
provide motivation for learners to perform tic, and transparent, and that has positive
well. Assessment tasks should be relevant to washback.
real-life contexts in which the language will Planning your assessments goes hand in
be used. For example, if a course is designed hand with developing your course learning
for students who will be answering phones objectives and should start when you begin
in English in a call center, an oral exam that planning the course. How you will assess stu-
mimics a telephone-call format would be dent learning will affect how you present mate-
more authentic than a test in which students rials and teach the course. There are several
listen to an academic lecture and respond to phases in the assessment process. One of the
questions related to the lecture, or one where most important is the initial planning stage.
the students write the correct forms of verbs in When you plan an exam, begin by describing
sentence blanks. The assessments should relate your assessment context. Think about what
to the purpose of the course, which in turn the purpose of the course is, which resources
relates to course objectives, which are then you have available, and how the instructional
tested on the assessments. setting and larger educational context influ-
ence the course. This is the information that
Transparency you will put in the test specifications, discussed
Transparency refers to the availability of above in the Validity section, in the categories
information to students. Students should be for purpose and intended population.

E n g l i s h T e a c h i n g F o r u m | Number 3 2014 7
The next step is to identify students’ learning and as a means of assessing the course
needs and develop course learning objectives. materials and instructional practices.
Learning objectives are determined by what
you want your students to know—and may Creating an exam blueprint
be mandated by institutional or national Having an exam blueprint increases the
priorities for education within your context. likelihood that you will actually test what you
You should specify what you want your stu- set out to test (i.e., the test will have validity).
dents to learn or be able to do after taking Test blueprints help you avoid overemphasizing
the course. This will guide you in develop- one area or completely missing another area
ing not only lessons and curriculum, but that needs to be tested. A blueprint is a tool to
also in deciding how you will assess whether determine what is important for the students
students have learned what you want them to know and the relative weight of each area
to. Identifying course learning objectives in relation to other areas or skills being tested;
will give you and the students goals to work at the same time, a blueprint ensures that the
toward during the course. Each of these content being taught is properly represented on
objectives can then be divided into the skills the test. The blueprint can also help a teacher
needed to accomplish the objectives, whether see that the method used for assessing matches
they relate to vocabulary, structure, or flu- the cognitive demand that is intended.
ency skills. With these learning objectives in Begin creating a blueprint by listing the
hand, you will be able to design a test and learning objectives you want to measure,
check that the test you hope to use will accu- the way they will be tested, and how much
rately measure these objectives. of the total exam will cover each area. There
The best way to do this is to create a may be several items on the exam related to
blueprint of the assessment, matching course each objective, but by first mapping out what
objectives to the test questions. By using the you hope to test, you can be sure to include
course learning objectives to guide the content questions that assess all your objectives. An
and the purpose of your exam, you can make example of a simplified blueprint is given in
sure that your assessments serve both as a Figure 3. It lists the skill area (in this case, the
tool for providing information about student skills to be tested are reading, listening, writ-

Skill Learning Objective How Objective(s) Will Be Tested Total


Reading • Can scan to find specific Read a paragraph and answer 30%
information questions related to a reading
• Can recognize main idea of a passage using multiple-choice,
paragraph short-answer, matching, gap-fill,
• Can understand pronoun and information-transfer items
references

Listening • Can recognize main idea of a Listen and answer questions 30%
section using multiple-choice,
• Can listen for specific short-answer, matching, gap-fill,
information and information-transfer items
• Can listen for numbers
Writing Can use pronouns to show Write sentences related to a 15%
cohesion between sentences personal topic
Grammar • Present simple Multiple-choice and fill-in-the- 25%
• Adjectives blank items
• Subject/object pronouns

Total 100%
Figure 3. Simplified exam blueprint

8 2014 Number 3 | E n g l i s h T e a c h i n g F o r u m
ing, and grammar), the learning objective, the what content is being tested and whether it is
item/question type, and overall percentage of tested in the correct proportion to what you
importance in the context of this assessment. hope the students are learning. Figure 4 shows
You can also select a test that is already sample questions that might be on a reading
made and map it backward to see if it will fit test (in this case, the topic of the passage was
your purposes or if items need to be added, dogs); the questions have been analyzed to
adjusted, or replaced. To map backward, you determine the objectives being assessed and
would list each question, what it tests, and the mix of item types being used.
the number of points it is worth. At the end One way to develop items for a test is to
of this exercise, you should be able to see write them on notecards (or if you have a

Item Learning Objective Item Type


What is the main idea of the reading? Reading: Can Multiple-choice
A. How to care for a dog understand main idea question
B. The many different breeds of dogs
C. The many ways that dogs are important
to people
In Paragraph 1, what does they refer to? Reading: Can Multiple-choice
A. veterinarians understand pronoun question
B. dog trainers references
C. dogs
What is the meaning of the word flush in Reading: Can Multiple-choice
Paragraph 3? understand vocabulary question
A. raise in context
B. remove
C. even with
What is the main idea of Paragraph 3? Reading: Can Multiple-choice
A. Dog showing is a popular sport. recognize main idea of question
B. There are several hundred breeds of dogs. a paragraph
C. Obedience training for dogs is important.
According to the reading, which is considered a Reading: Can scan Multiple-choice
sporting dog? to find specific question
A. collie information
B. fox terrier
C. pointer
Match the type of dog with its description, Reading: Can scan Matching
according to the information in the reading. to find specific
information
basset hound _____
poodle _____
Chihuahua _____
terrier _____

A. is trained to pull sleds


B. has long ears
C. is the smallest pure-bred dog
D. sheds very little hair
E. is used to herd animals

Figure 4. Analysis of exam questions relative to learning objectives

E n g l i s h T e a c h i n g F o r u m | Number 3 2014 9
computer, add them to a spreadsheet) so that to use the questions again and which learn-
you can then sort them according to skill, ing objectives might need to be reviewed or
question type, or objective covered. This sys- taught in a different manner in future lessons.
tem will let you know whether you have too In this way, an assessment can provide both
many questions of the same type or need to teacher and student feedback.
add more in a certain category. Writing ques-
tions at the end of each lesson you teach is also Providing feedback
an effective way to reflect on what you taught An important process in assessment devel-
during the lesson and how you could assess opment and revision is feedback. By providing
that content in a manner the students will be feedback on an actual test after administer-
familiar with. Working with a group of teach- ing it, teachers ensure constant improvement
ers to make new test items, or even working with regard to the reliability and validity of all
with your students, is a way to add variety to assessments administered within a program.
your item bank (your collection of possible Information gathered from the test can be
test questions) and to get other opinions on useful to determine final marks of an exam
what might be appropriate. and to decide whether any questions should
Having students become part of the assess- not be counted in the grading process. All
ment process can increase transparency, help information that is generated through testing
them recognize what is important to know should be kept with that test and reviewed
from the content being taught, and motivate before the test is administered again. Teach-
them to review and study together. If students ers who use an assessment should be asked to
have a clear understanding of what they are complete a feedback form. Sample questions
being tested on and how they will be tested, and the areas you might consider for feedback
positive washback is promoted. If you are are given in the following section.
working at an institution that uses very high-
stakes testing and standardized exams for all Creating an exam file
students, you might consider having the stu- Reflecting on the exam is an important
dents create assessments for review or quizzes. part of the test development process. After
For example, you could ask each student to you administer each test that might be reused,
write questions for a particular grammar item create an exam file. The file should contain
or vocabulary list. This method will also help information about the actual tasks in the test,
you informally assess student knowledge of its administration, and its level of difficulty.
the topics being taught as you review students’ You could create a feedback form that has
questions and answers to decide whether they questions related to these topics, then fill it
are suitable for a small-scale assessment. out when developing and reviewing the test,
After questions are written, label each with and reflect on it after the administration.
the learning objective that it covers and the Questions that might be useful in gathering
item type. You can then look at the questions, feedback on an exam are listed below.
make selections, and put them together for
the quiz. You can look at item cards your- Setting: Task and Administration
self or with other teachers to determine the • Is it clear how the student is expected to
strength of each question for testing a given respond to each item?
objective. Looking at questions with others is • Does the exam test the curriculum
a useful practice that can also generate more content?
items or improve the ones you have. You • Does the exam contain formats familiar
can make stacks of questions relating to the to the students?
same curriculum objective and determine how • Are the exam tasks authentic and
many of each you will need for the appropri- meaningful?
ate weighting on the test or quiz. Once the • Is there more than one logical option/
assessment has been administered, you will answer possible for any blank/question?
also want to add information to each question • Is the layout user-friendly? Does the
about how it performed (see Basic Statistics formatting match specifications?
for Testing, below). These testing or item sta- • Are there any typos, misprints, or other
tistics will be useful in determining whether errors in the production of the test?

10 2014 Number 3 | E n g l i s h T e a c h i n g F o r u m
• Are the listening and reading texts at an who took the test. You can compute a mean
appropriate level? score for the whole test and for each section.
• Are the texts clear in their organization The mean score provides information about
and content? the reliability of a test over time, especially if
• Is the quality (e.g., pace, sound clarity, the delivery of the course remains the same.
voice quality) of the recording good? If a test is reliable, you would expect a similar
• Is the time allotment appropriate for result among students at different times or
the length of the exam? between groups of similar students. If you
have a test in which various skills are tested,
Demands: Task such as the one in the example of the test
• Do the test tasks adequately sample the specifications with a listening, reading, and
skills and strategies needed for [listen- writing section (Figure 2), computing the
ing/reading/writing]? mean for each section would give you an idea
• Are any questions answerable by stu- of which skill areas the students have more
dents who haven’t heard the lecture or difficulty with, as indicated by a lower mean
read the text? score.
• Do the task types reflect student class- To determine the pass/fail rate, meanwhile,
room experience? you must first decide what a passing score is
• Are the task types adequate for measur- on the exam. I have been in places where 60
ing the course objectives? percent is passing at the university level for
• Is this test an accurate measuring tool some courses, depending on their difficulty,
for progress for this course? and other places where 75 percent or even
Along with the feedback form, an exam higher represents the passing score. It depends
file would also include the test-specifications on the level at which students are expected to
document indicating the purpose and general master the course objectives, and on whether
audience for the test, the test blueprint indi- the purpose of testing is just to measure learn-
cating how each question matches a learning ing objectives or to move only the top per-
objective, and basic statistics. Changes to formers on to the next course or level of study.
an exam or scoring process should be based Once you have determined a passing score,
on the analyses and feedback. Making notes add the number of students who reached the
about the test immediately after grading it passing score or higher and divide by the total
offers an opportunity to update an exam number of test takers. This number represents
while it is fresh in your mind. You might even the pass rate. The same can be done for the
think about having your students answer a fail rate, counting instead the number of stu-
few questions about the test as well. These dents who did not receive at least the passing
could include how difficult they thought it mark and dividing by the total number of test
was, whether there were questions they found takers.
confusing, and whether the content covered The pass/fail rate can be calculated for the
was what they expected would be on the test. test as a whole, for each section or skill area,
Valuable information about your assessment and even for individual questions. Looking at
can also be gathered through basic statistical individual-item pass rates gives you an idea of
analysis. the difficulty of each question and can help
you find items that may need to be rewrit-
Basic statistics for testing ten or areas that will need further instruction
After you administer an assessment, basic or review. If you add this information to the
statistics give you an idea of how your stu- questions that you created on notecards or
dents are performing and information about a computer spreadsheet, in the future you
the test in general. Two useful statistics to can also put together tests with questions
measure are the mean score and the pass/fail that have a variety of difficulty levels. This
rate. The mean is the mathematical average of information will also tell you which skills
all the test takers’ scores; you can compute the you might not need to spend so much time
mean by adding all the test takers’ scores and reviewing in the future and which ones may
dividing the total by the number of students need more instruction time.

E n g l i s h T e a c h i n g F o r u m | Number 3 2014 11
Practice makes perfect that “teacher professional development is
Just as learning a language takes practice, more effective when it is school embed-
so does developing assessment skills. By ded, cooperative and sustained over time
thinking about the assessment process and … organized within and/or across schools
analyzing the tests we use, we can all become and focuses on improving practice over time
better teachers and test developers. Teachers through the sharing of knowledge, experience
often report that assessment literacy is not and expertise” (O’Leary 2008, 112).
something that they acquired by taking a So, with that in mind, make sure to share
course, but instead is something they learned with your colleagues what you have learned
on the job through the ongoing experience from this article (or share this article with
of working with and developing assessments. them), talk about your assessments, work
When I asked teachers how they had gained together to improve them, and remember
the knowledge that they have about assess- that assessment plays an important role in the
ment, almost all mentioned post-formal- teaching and learning processes. By reflect-
education experience (“through discussion ing on the tests used in your classroom, how
of the effectiveness of assessments with they are developed, and the results obtained
colleagues,” “trial and error,” “being part of from them, you will become more assessment
a group making tests,” “on-the-job experi- literate and a better teacher. Testing will not
ence”). Teachers seem to feel that assessment seem so complex or beyond comprehen-
literacy is not something you are taught in a sion, as many teachers feel it is. You, in turn,
course, though you can get the fundamentals will provide students with a better experi-
and theories there; it is basically something ence, empowering them as learners instead
that must be developed through ongoing of instilling fear of tests. With assessment
experience and practice. Assessment literacy literacy comes the understanding of the valu-
is a skill that you have to learn by doing, not able role assessment plays in the classroom for
just by reading about it. Real-life teaching in both teachers and students.
the classroom provides the perfect opportu-
References
nity to apply assessment fundamentals and
to develop assessments that will motivate Bachman, L. F., and A. S. Palmer. 1996. Language
testing in practice. Oxford: Oxford University
your students to learn—and that will moti- Press.
vate you to develop lessons that will facilitate Boyles, P. 2006. Assessment literacy. In New visions
learning. in action: National assessment summit papers,
A lot of research has focused on what ed. M. Rosenbusch, 18–23. Ames: Iowa State
information about assessment should be University.
Brookhart, S. M. 2001. The standards and class-
taught in education courses, but maybe
room assessment research. Paper presented at
instead we should be looking at the oppor- the Annual Meeting of the American Associa-
tunities for development that present them- tion of Colleges for Teacher Education.
selves throughout our teaching careers. As Campbell, C., and V. L. Collins. 2007. Identifying
Boyles (2006) suggests, teachers can use the essential topics in general and special education
introductory assessment textbooks. Educational
information from assessment to adjust teach-
Measurement: Issues and Practice 26 (1): 9–18.
ing practices, to provide evidence of student Educational Testing Service. 2014. “About the
performance, and to guide the curriculum- TOEFL iBT® Test.” www.ets.org/toefl/ibt/
review process. The information we gather about?WT.ac=toeflhome_ibtabout2_121127
from assessments should provide feedback to IELTS. 2013. “Test Takers—Academic or General
students, but it should also be used as a means Training?” www.ielts.org/test_takers_informa-
tion/what_is_ielts/academic_or_general_train-
for teachers and program administrators to ing.aspx
review instructional practices and curricular Mertler, C. A. 2004. Secondary teachers’ assess-
objectives. Brookhart (2001) notes that in ment literacy: Does classroom experience make
order for teachers to “own the goal of high- a difference? American Secondary Education 33
quality assessment, they must be convinced (1): 49–64.
Modern Language Center. 2010. “MET: Michigan
that assessment is just as important to student English Test.” www.mlc.edu.jo/met-michigan-
learning as … effective classroom manage- english-test
ment or lesson design.” Research indicates O’Leary, M. 2008. Towards an agenda for profes-

12 2014 Number 3 | E n g l i s h T e a c h i n g F o r u m
sional development in assessment. Journal of
In-service Education 34 (1): 109–114.
Oregon Department of Education. 2014. “Test-
ing—Student Assessment.” www.ode.state.or.us/
search/page/?id=496
Popham, W. J. 2004. Why assessment illiteracy is
professional suicide. Educational Leadership 62
(1): 82–83.
Stiggins, R. J. 1991. Relevant classroom assessment
training for teachers. Educational Measurement:
Issues and Practice 10 (1): 7–12.
Taylor, L. 2009. Developing assessment literacy.
Annual Review of Applied Linguistics 29: 21–36.

Dawn Rogier is the Regional English


Language Officer for Brunei, Malaysia,
the Pacific Islands, Papua New Guinea,
and the Philippines. In her free time, she
enjoys analyzing test items and discussing
statistical procedures.

E n g l i s h T e a c h i n g F o r u m | Number 3 2014 13

You might also like