Professional Documents
Culture Documents
Assessment in Learning 1 Module
Assessment in Learning 1 Module
CHAPTER 1
BASIC CONCEPTS OF
ASSESSMENT
Teachers make professional judgements on learners’ performance in every
teaching and learning session undertaken, whether consciously or subconsciously.
Teachers, therefore, need training and support to enable them to make valuable
assessment decisions, to provide quality feedback to learners, and to teach learners
to receive feedback positively and use the information contained within it effectively
to improve their work. Assessment for Learning and quality feedback can and do
promote increased learner progress. However, assessment of learning rather than
assessment for learning has preoccupied the minds of the profession for many years
in an effort to meet awarding body requirements. Assessment of learning can detract
from effective classroom practice and prevent feeding back assessment decisions to
learners on their performance with the view to improving their work.
In this chapter with the change of focus in instruction from content to learning
outcomes came the need to redefine and clarify the terms used to determine the
progress of students towards attainment of the desired learning outcomes. These
are measurement, evaluation and assessment.
LEARNING CONTENT
PURPOSES OF ASSESSMENT
1. TEACHING AND LEARNING
What makes assessment for learning effective is how well the information
is used.
2. SYSTEM IMPROVEMENT
Students and teachers can use the information gained from assessment to
determine their next teaching and learning steps.
Parents and families can be kept informed of next plans for teaching and
learning and the progress being made, so they can play an active role in their
children‘s learning.
EVALUATION
Evaluation is a broader term that refers to all of the methods used to find out
what happens as a result of using a specific intervention or practice. Evaluation is
the systematic assessment of the worth or merit of some object. It is the systematic
acquisition and assessment of information to provide useful feedback about some
object. Evaluation is a process designed to provide information that will help us
make a judgment about a particular situation. The end result of evaluation is to
adopt, reject or revise what has been evaluated.
Though the terms assessment and evaluation are often used interchangeably
(Cooper, 1999), many writers differentiate between them. Assessment is defined as
gathering information or evidence, and evaluation is the use of that information or
evidence to make judgments (Snowman, McCown, and Biehler, 2012).
Measurement involves assigning numbers or scores to an "attribute or characteristic
of a person in such a way that the numbers describe the degree to which the person
possesses the attribute" (Nitco and Brookhart, 2011, p. 507). Assigning grade
equivalents to scores on a standardized achievement test is an example of
measurement.
B. List down three (3) activities or processes involved in each of the following:
1. Measurement
a) ______________________________________________________
b) ______________________________________________________
c) ______________________________________________________
2. Assessment
a) ______________________________________________________
b) ______________________________________________________
c) ______________________________________________________
3. Evaluation
a) ______________________________________________________
b) ______________________________________________________
c) ______________________________________________________
ROLES OF ASSESSMENT IN
MAKING INSTRUCTIONAL
DECISION
In this chapter we will learn that aassessment plays an important role in
the process of learning and motivation. The types of assessment tasks that we
ask our students to do determine how students will approach the learning task
and what study behaviours they will use. In the words of higher education
scholar John Biggs, “What and how students learn depends to a major extent
on how they think they will be assessed.” (1999, p. 141).
Given the importance of assessment for student learning, it is important
to consider how to best measure the learning that you want your students to
achieve. Assessment should integrate grading, learning, and motivation for
your students. Well-designed assessment methods provide valuable
information about student learning. They tell us what students learned, how
well they learned it, and where they struggled. Good assessments allow you
to answer the question.
LEARNING OBJECTIVES
CLASSIFICATION OF ASSESSMENT
There are four types of assessment: placement, diagnostic, formative and
summative. Although these four are generally referred to simply as assessment,
there are distinct differences between the three.
LEARNING THEORY
Learning theories are conceptual frameworks describing how information is
absorbed, processed and retained during learning. Cognitive, emotional, and
environmental influences, as well as prior experience, all play a part in how
understanding, or a world view, is acquired or changed and knowledge and skills
retained.
Behaviorists look at learning as an aspect of conditioning and will advocate a
system of rewards and targets in education. Educators who embrace cognitive
theory believe that the definition of learning as a change in behavior is too narrow
ISABELA STATE UNIVERSITY|ASSESSMENT IN LEARNING 1 9
and prefer to study the learner rather than their environment and in particular the
complexities of human memory. Those who advocate constructivism believe that a
learner's ability to learn relies to a large extent on what he already knows and
understands, and the acquisition of knowledge should be an individually tailored
process of construction. Transformative learning theory focuses upon the
oftennecessary change that is required in a learner's preconceptions and world view.
1. Behaviorism - Behaviorism is a philosophy of learning that only focuses on
objectively observable behaviors and discounts mental activities. Behavior
theorists define learning as nothing more than the acquisition of new behavior.
Experiments by behaviorists identify conditioning as a universal learning
process. There are two different types of conditioning, each yielding a
different behavioral pattern:
Classic conditioning occurs when a natural reflex responds to a stimulus.
The most popular example is Pavlov's observation that dogs salivate when
they eat or even see food. Essentially, animals and people are biologically
"wired" so that a certain stimulus will produce a specific response.
Behavioral or operant conditioning occurs when a response to a stimulus is
reinforced. Basically, operant conditioning is a simple feedback system: If a
reward or reinforcement follows the response to a stimulus, then the response
becomes more probable in the future. For example, leading behaviorist B.F.
Skinner used reinforcement techniques to teach pigeons to dance and bowl a
ball in a mini-alley.
How Behaviorism impacts learning?
LEARNING OBJECTIVES
LEARNING CONTENT
3. BALANCE
Assessment methods should be able to assess all domains of learning and
hierarchy of objectives.
The cognitive domain involves the development of our mental skills and the
acquisition of knowledge. The six categories under this domain are:
B. Affective Domain
The affective domain involves our feelings, emotions and attitudes. This domain is
categorized into 5 subdomains, which include:
3. Valuing: the ability to see the worth of something and express it.
Example: An activist shares his ideas on the increase in salary of laborers.
4. Organization: ability to prioritize a value over another and create a unique value
system.
Example: A teenager spends more time in her studies than with her boyfriend.
5. Characterization: the ability to internalize values and let them control the
person`s behaviour.
Example: A man marries a woman not for her looks but for what she is.
3. Guided Response: the ability to imitate a displayed behavior or to utilize trial and
error.
Example: A person follows the manual in operating a machine.
4. Mechanism: the ability to convert learned responses into habitual actions with
proficiency and confidence.
Example: A mother was able to cook a delicious meal after practicing how to cook
it.
HIERARCHY OF OBJECTIVES
(BLOOM) (ANDERSON)
Evaluation Create
Synthesis Evaluate
Analysis Apply
Application Analyze
Comprehension Understand
Knowledge Remember
1. VALIDITY
Assessment should be valid. There are several types of validity that are to be
established.
Types of Validity
Construct validity: Does the test measure the concept that it’s intended to
measure?
Content validity: Is the test fully representative of what it aims to measure?
Face validity: Does the content of the test appear to be suitable to its aims?
Criterion validity: Do the results correspond to a different test of the same thing?
Note that this article deals with types of test validity, which determine the
accuracy of the actual components of a measure. If you are doing experimental
research, you also need to consider internal and external validity, which deal with
the experimental design and the generalizability of results.
1. Construct validity
Construct validity evaluates whether a measurement tool really represents the
thing we are interested in measuring. It’s central to establishing the overall
validity of a method.
What is a construct?
A construct refers to a concept or characteristic that can’t be directly observed,
but can be measured by observing other indicators that are associated with it.
Example
To achieve construct validity, you have to ensure that your indicators and
measurements are carefully developed based on relevant existing knowledge. The
questionnaire must include only relevant questions that measure known indicators of
depression.
The other types of validity described below can all be considered as forms of
evidence for construct validity.
2. Content validity
Content validity assesses whether a test is representative of all aspects of the
construct.
Example
3. Face validity
Face validity considers how suitable the content of a test seems to be on the
surface. It’s similar to content validity, but face validity is a more informal and
subjective assessment.
Example
You create a survey to measure the regularity of people’s dietary habits. You
review the survey items, which ask questions about every meal of the day and
snacks eaten in between for every day of the week. On its surface, the survey seems
like a good representation of what you want to test, so you consider it to have high
face validity.
4. Criterion validity
Criterion validity evaluates how closely the results of your test correspond to the
results of a different test.
What is a criterion?
The criterion is an external measurement of the same thing. It is usually an
established or widely-used test that is already considered valid.
Example
5. RELIABILITY
Assessment should show consistent and stable results. There are methods which
can be used to measure and establish reliability.
7. AUTHENTICITY
Assessment should touch real life situations and should emphasize practicability.
ASSESSMENT FORMS:
Assessment should not be used to derogate the students. One example of this is the
right to confidentiality.
Assessment should have a positive effect. It should motivate students to learn and
do more and should give way to improve the teacher's
instruction.
learning task
DEVELOPMENT OF CLASSROOM
ASSESSMENT FOR MEASURING KNOWLEDGE
AND LEARNING
Teachers teach content then test students. Tests seek to see what students
such have learned. Such tests are used to evaluate student learning, skill level of
growth and academic achievements at the end of the course.
LEARNING OBJECTIVES
Design a test that are appropriate to measure the student’s learning and
thinking skills..
Match the different learning targets and their appropriate assessment
techniques
Construct classroom assessment tools in measuring knowledge and thinking
skills.
Demonstrate understanding of the basic definitions used in item analysis and
some of its key concepts
Analyze test item
What you want to assess should be related to your learning outcomes for the
course.
Knowledge or how it is used. You can design your test questions to assess
students’ knowledge or ability to apply material taught in class.
Process or product. You can test students’ reasoning skills and evaluate the
process by focusing the marks and other feedback on the process they follow
to arrive at a solution. Alternatively, you can evaluate the end product.
The communication of ideas. You can evaluate students’ communication
skills their ability to express themselves - whether this is by writing a cogent
argument, or creating an elegant mathematical proof.
Convergent thinking or divergent thinking. You can test your students’
ability to draw a single conclusion from different inputs (convergent thinking).
Or you may alternatively want them to come up with different possible
answers (divergent thinking). Do you expect different answers from students,
or do you expect all of them to provide the same answer?
Absolute or relative standards. Is student success defined by learning a set
amount of material or demonstrating certain skills, or is student success
measured by assessing the amount of progress the students make over the
duration of the course?
The overall exam should be consistent with your learning outcomes for the course.
There are a number of ways to review and prioritize the skills and concepts taught in
a course. You could:
“To be able to prepare a GOOD TEST, one has to have a mastery of the subject
matter, knowledge of the pupils to be tested, skill in verbal expression and the use of
the different test format.”
Validity – the extent to which the test measures what it intends to measure
Reliability – the consistency with which a test measures what it is supposed
to measure.
Usability – the test can be administered with ease, clarity, and uniformity.
Scorability – easy to score
Interpretability – test results can be properly interpreted and is a major basis
in makings educational decisions
Economical – the test can be reused without compromising its validity and
reliability
PRODUCE A
TABLE OF
SPECIFICATION DRAFT
ORDER
TEST ANALYZE
SUBMISSION
TABLE OF SPECIFICATIONS
A two-way chart that relates the learning outcomes to the course content.
It is a test map that guides the teacher in constructing a test.
It ensures that there is a balance between items that test lower level thinking skills
and those which test higher order thinking skills (HOTS).
It ensures balance between easy and difficult items •It enables the teacher to
prepare a test containing a representative sample of student knowledge in each of
the areas tested.
An objective test is a test that has right or wrong answers and so can be marked
objectively. It can be compared with a subjective test, which is evaluated by giving an
opinion, usually based on agreed criteria. Objective tests are popular because they are
easy to prepare and take, quick to mark, and provide a quantifiable and concrete result.
Analyzes a statement
When/how to use:
Critical content should be readily apparent and identified for analysis, avoiding cleverness,
trickery, and verbal complexity.
Avoid..
Variations in answers:
Test instructions:
Before the test, give clear, proactive instructions on what content is covered, level of
detail, and what type of questions will be asked:
Encourage comprehension: cause and effect, if/then, sequences
Avoid memorization
Detail exactly what must be exactly memorized: dates, locations, proper names,
sequences
Be consistent in test administration over time
Have students indicate their answers by circling complete words of "true" "false"
(not "t" "f")
Do not have students write their response of t/f or true/false to (avoids
distinguishing/problems of hand writing and sloppiness)
Avoid plus or minus signs "+" of "-"
Indicate how the test is scored: total right, or total right minus wrong?
Limitations:
Scoring tends to be high since guessing yields a 50-50 score (half right half wrong)
as a base. i.e. if there are 100 items, and the student knows the correct answer to
50, and guesses on the other half, the score will be 75 knowing only half the
material.
Since the stem can cue a correct answer, guessing is enhanced without really
understanding the question
The format does not provide diagnostic information on why a student got it wrong
It may be easy to cheat
Content can be simplistic and/or trivial
GUIDELINES:
1. Use negatively stated stems sparingly and when using negatives such as NOT,
underline or bold the print.
2. Use none of the above and all of the above sparingly, and when you do use them,
don’t always make them the right answer.
3. Only one option should be correct or clearly best.
4. All options should be homogenous and nearly equal in length.
5. The stem (question) should contain only one main idea.
6. Keep all options either singular or plural.
7. Have four or five responses per stem (question).
8. When using incomplete statements place the blank space at the end.
1. Definition – terms
2. Historical events – dates
3. Achievements – people
4. Statements – postulates
5. Descriptions – principles
They can also assess student’s ability to apply knowledge by requiring a test-taker
to match the following:
1. Examples – terms
2. Functions – parts
3. Classifications – structures
4. Applications – postulates
5. Problems – principles
Matching questions are really a variation of multiple choice format. If you find that
you are writing multiple choice questions which share the same answer choices,
you may consider grouping the questions into a matching item.
This test consists of a series of items which requires the students to fill a word
or phrase on the blanks. This is also called filling the blank type of test.
Direction: Fill in the blank the correct word or phrase. Write your answer at the right
column.
- J.P. Guilford
Difficulty index(DF) refers to the proportion of the number of students n the upper
and lower groups who answered an item correctly. In a classroom achievement test,
the desired indices of difficulty not lower than 0.20 nor higher than 0.80. The average
index of difficulty from 0.30 or 0.40 to a maximum of 0.60.
DF =PUG+ PLG
2
PUG= proportion of the upper group who got an item right
PLG= proportion of the lower group who got an item right
Index of discrimination
Discrimination Index is the degree which the item discriminates between high
performing groups in relation of scores on the total test. Index of discrimination is
classified into positive discrimination, negative discrimination and zero
discrimination.
Positive Discrimination if the proportion of students who got an item right in the
upper performing group is greater than the proportion of the lower performing group.
Zero Discrimination if the proportion of the students who got an item right in the
upper performing group and low performing group are equal.
Maximum Discrimination is the sum of the proportion of the upper and lower
groups who answered the item correctly. Possible maximum discrimination will occur
if the half or less of the sum of the upper and lower groups answered an item
correctly.
Notations:
Formula:
Di = PUG – PLG
DE = Di
DM
DM =PUG + PLG
Given:
Number of students took the exam =80
27% of 80 = 21.6 or 22, which means that are 22 students in the upper performing
group and 22 students in the lower performing group.
PUG = 6 = 27%
PLG = 4 = 18%
22
Di = PUG – PLG
= 27% - 18%
Di = 9%
DM = 45%
DE = Di
DM
.
= 09
.45
DE = 0.20 0r 20%
This can be interpreted as on the average, the item is discriminating at 20% of the
potential of an item of its difficulty.
Measures of Attractiveness
Difficulty Level
Discriminating
Power
Overall
Evaluation of
the Item
Highly
Acceptable highly acceptable very good item
Highly
Acceptable/
Acceptable Acceptable good item
Distractor Analysis
Good Distractor
Poor Distractor
Completely Ineffective
Example
1 b 1 7 Mo Mo Poor
c 3 1 HA HA CI
d 0 0
a 6 10 .12 .07 Good Poor Item.
c 0 0 Un Un CI
*d 2 1
Estimate the item difficulty (percentage of students who got the item right)
RU RL
ID
T
Where :
Estimate the discriminating power of each item (difference between the number of
pupils in the upper group and lower group who got the item right)
RU RL
DP
1
T
2
Where :
Ru- The number in the upper group who answered the item correctly
Good distractor – attracts more students from lower than from the upper
group
Poor distractor – attracts more students from upper group than from the
lower group
Completely Ineffective – attracts no one
Difficulty Level
Discriminating Power
ESSAYS
The following are rules of thumb which facilitate the scoring of essays:
Rule 1: Phrase the direction in such a way that students are guided on the key
concepts to be included.
Rule 2: inform the students on the criteria to be used for grading their essays. This
rule allows students to focus on the relevance or substantive materials rather than
on peripheral and unnecessary facts of information.
Rule 4: Decide on your essay grading system prior to getting the essays of the
students.
Rule 5: Evaluate all the students’ answers before proceeding to the next question.
Rule 6: Evaluate answers to essay questions without knowing the identity of the
writer.
Rule 7: Whenever possible, have two or more persons grade each answer.
Calculate the Difficulty Index (p) and the Discrimination Index (D).
Student Total
Score
Questions
(%)
Name 1 2 3 4 5 6 7 8 9 10
Bob 100 1 1 1 1 1 1 1 1 1 1
Julie 80 1 0 1 1 1 1 0 1 1 1
Toli 70 0 1 1 1 1 1 0 1 0 1
Vasili 70 1 1 1 0 1 1 0 1 0 1
Nora 60 1 1 1 0 1 1 0 1 0 0
Jaana 50 0 0 1 0 1 1 1 1 0 0
Zella 40 1 1 1 0 0 0 1 0 0 1
Kate 20 0 1 0 0 0 1 0 0 1 0
Calculate the Difficulty Index (p) and the Discrimination Index (D).
Assessment Tasks
Assessment Task 1:
Assessment Task 2:
Request for a set of major examination result from any of your teacher and conduct
item analysis.
LEARNING OBJECTIVES
PERFORMANCE-BASED TEST
Intended to assess what it is that students know and can do with the
emphasis on doing.
Have a high degree of realism about them.
Involve:
(a) activities for which there is no correct answer,
Should assess higher level cognitive skills rather than narrow and lower level
discreet skills.
Direct measures of skills of interest.
a) work, often days and weeks. Students perform, create, construct, produce, or
do something.
b) Deep understanding and/or reasoning skills are needed and assessed.
c) Involvement is sustained
d) Calls on students to explain, justify, and defend.
e) Performance is directly observable.
f) Involves engaging in ideas of importance and substance.
g) Relies on trained assessor’s judgments for scoring
h) Multiple criteria and standards are pre-specified and public
i) There is no single correct answer.
Learning Targets
SKILLS
Ex: Speaking
PSYCHOMOTOR SKILLS
Fine motor: cutting papers with scissors, drawing a line tracing, penmanship,
coloring drawing, connecting dots
Gross motor: Walking, jumping, balancing, throwing, skipping, kicking
Complex: Perform a swing golf, operate a computer, drive a car, operate a
microscope
Visual: Copying, finding letters, finding embedded figures, identifying shapes,
discrimination
Verbal and auditory: identify and discriminate sounds, imitate sounds,
pronounce carefully, blend vowels
PRODUCTS
Variation of authenticity
Explain what would you te Show how to perfor Play a basketball game.
ach to students learning b m basketball skills in
asketball practice
2. Develop descriptions of the task and the context in which the performance is to be
conducted.
3. Write the specific question, prompt, or problem that the student will receive.
Complexity of task
Restricted-type task
Extended-type task
Examples:
Characteristics of Tasks
1. Should integrate the most essential aspects of the content being assessed with
the most essential skills.
2. Should be authentic
Realistic
Require judgment and innovation
Ask the student to do the subject
Replicates or stimulates
Assess the students ability to efficiently and effectively use a repertoire of
knowledge and skill to negotiate a complex task
Allows opportunities to rehearse, practice, consult resources, and get
feedback and refine performances and products.
3. Structure the task to assess multiple learning targets
4. Structure the task so that you can help students succeed.
5. Think through what students will do to be sure that the task is feasible
6. The task should allow for multiple solutions
Performance Criteria
What you look for in student responses to evaluate their progress toward meeting
the learning target.
Example of Criteria
Learning target:
Students will be able to write a persuasive paper to encourage the reader to accept
a specific course of action or point of view.
Criteria:
Rating Scales
1. Numerical
2. qualitative
3. combined qualitative/quantitative
Numerical Scale
Example:
No understanding
Clear organization
Fluent reader
Qualitative scale
Type B: Checklist
Example of Type A:
Holistic scale
The category of the scale contains several criteria, yielding a single score that gives
an overall impression or rating
ISABELA STATE UNIVERSITY|ASSESSMENT IN LEARNING 1 57
Example
level 4: Sophisticated understanding of text indicated with constructed
meaning
Analytic Scale
Example:
5 4 3 2 1
Creative ideas
Logical organization
Relevance of detail
Variety in words and sentences
Vivid images
Rubrics
When scoring criteria are combined with a rating scale, a complete scoring
guideline is produced or rubric.
A scoring guide that uses criteria to differentiate between levels of student
proficiency.
Example:
1. Teacher-Centered vs Learner-Centered
Paradigms. (Huba & Freed, 2000)
4. On reflective assessment.
Choose one:
Assessment Task 1:
1. Learner-centered assessment
What is constructivism?
Assessment Task 2:
CHAPTER 6
When teachers have rich information about what their students know, can do
and need to do next, they are able to involve students as active participants in their
learning and assessment of their own learning. They are also in a position to consult
parents and the school’s communities about students’ progress.
LEARNING OBJECTIVES
LEARNING CONTENT
NORM-REFERENCED TEST
Norm-referenced refers to standardized tests that are designed to compare
and rank test takers in relation to one another. Norm-referenced tests report whether
test takers performed better or worse than a hypothetical average student, which is
determined by comparing scores against the performance results of a statistically
selected group of test takers, typically of the same age or grade level, who have
already taken the exam.
Calculating norm-referenced scores is called the “norming process,” and the
comparison group is known as the “norming group.” Norming groups typically
comprise only a small subset of previous test takers, not all or even most previous
test takers. Test developers use a variety of statistical methods to select norming
groups, interpret raw scores, and determine performance levels.
Norm-referenced scores are generally reported as a percentage or percentile
ranking. For example, a student who scores in the seventieth percentile performed
as well or better than seventy percent of other test takers of the same age or grade
level, and thirty percent of students performed better (as determined by norming-
group scores).
Norm-referenced tests often use a multiple-choice format, though some include
open-ended, short-answer questions. They are usually based on some form of
national standards, not locally determined standards or curricula. IQ tests are
NORMAL DISTRIBUTIONS
TYPES OF SCORES
ACHIEVEMENT TESTS
DIAGNOSTIC TESTS
APTITUDE TESTS
• Norm-referenced • Criterion-referenced
– General ability – Mastery
– Range of ability – Basic skills
– Large groups – Prerequisites
– Compares people to – Affective
people-comparison groups – Psychomotor
– Selecting top candidates – Grouping for instruction
learning task
LEARNING TASK 1: MULTIPLE CHOICE: Encircle the letter of the correct answer.
a. Individual Students
b. Teachers
c. A group of students
ISABELA STATE UNIVERSITY|ASSESSMENT IN LEARNING 1 69
d. A group of teachers
6. All but which of the following are negatives of Norm-Referenced state testing?
a. State Assessments
b. Local curriculum
c. Nationally used textbooks
d. None of the above
10. Test takers design questions that most students will get wrong. In addition, If
too many students get a question right, it is thrown out to achieve a normal
curve.
a. True c. Slightly true
b. False d. Slightly false
LEARNING OBJECTIVES
LEARNING CONTENT
A. Feedback
The term feedback is often used to describe all kinds of comments made after the
fact, including advice, praise, and evaluation. But none of these are feedback, strictly
speaking.
Basically, feedback is information about how we are doing in our efforts to reach
a goal.
Feedback Essentials
Goal-Referenced
Effective feedback requires that a person has a goal, takes action to achieve the
goal, and receives goal-related information about his or her actions.
Any useful feedback system involves not only a clear goal, but also tangible results
related to the goal.
Actionable
User-Friendly
Ongoing
Adjusting our performance depends on not only receiving feedback but also
having opportunities to use it. What makes any assessment in education formative is
not merely that it precedes summative assessments, but that the performer has
opportunities, if results are less than optimal, to reshape the performance to better
achieve the goal. In summative assessment, the feedback comes too late; the
performance is over.
Thus, the more feedback I can receive in real time, the better my ultimate
performance will be.
Consistent
Grading
In such a system:
a. Grade depends on what group you are in, not just your own performance
b. Typical grade may be shifted down or up, depending on group’s ability
c. It is the widely used system
Example:
Consider the following two sets of scores in an English 1 class for two sections ot ten
students each:
A={30,40,50,55,60,65,70,75,80,85}
B={60,65,70,75,80,85,90,90,95, 100}
In the first class , the student who got a raw score of 75 woulg get a grade of
80% while in the second class , the same grade of 80% would correspond to a raw
score of 90. Indeed , if the test used for the two classes are the same , it would be a
rather “unfair” system of grading . A wise student would opt to enroll in the class A
since it is easier to get higher grades in that class than in the other class (class B).
In norm-referenced grading, the students, while they may work individually , are
actually in competition to achieve a standard performance that will classify them into
the desired grade range.
Criterion-Referenced Grading
a) Test Validity
b) Test Reliability
c) Test Difficulty level and other
The K - 12 Grading System: Reproduced from the DepEd Order No. 31 , s. 2012
Levels of Proficiency
At the end of the quarter, the performance of students shall be described in the report
card, based on the following levels of proficiency:
What shall appear in the report card is not the numerical value, but the equivalent
level of proficiency, abbreviated as follows:
B for Beginning;
D for Developing;
A for Advanced.
At the end of the four quarters, the Final Grade for each learning area shall be
reported as the average of the four quarterly ratings, expressed in terms of the levels
of proficiency. The general average shall be the average of the final grades of the
different learning areas, also expressed in terms of levels of proficiency with the
numerical equivalent in parenthesis.
As a matter of policy every learning deficiency should be bridged even for those
students whose level of proficiency is above the Beginning level. The guidelines for
bridging gaps in learning are in a separate DepEd Order.
Honor students shall be drawn from among those who performed at the Advanced
Level. Subsequent guidelines shall be issued as basis for ranking of honors.
At the end of the every quarter, schools are encouraged to put up exhibits of
student products across subjects as culminating activity. Students may also do an
exhibition of their performance in different subjects as evidence of their learning or
attainment of performance standards. Thus, when parents receive the report card
and confer with teachers they will actually be witnessing what students are learning
in school.
Pass-fail System -System that utilizes a dichotomous grade system. Other colleges
and universities , faculties, schools, and institutions use pass-fail grading systemin
the Philippines, especially when the student work to be evaluated is highly
subjective( as in fine arts and music), there are no generally accepted standard
gradations (as with independent studies).
Non-Graded Evaluation- Non- Graded Evaluation does not assign numeric or letter
grades as a matter of policy. This practice is usually based on a belief that grades
introduce an inappropriate and distracting element of competition into the learning
process, or that they are not meaningful as a measures of intellectual growth and
development as are carefully crafted faculty evaluations.
learning task
Learning Task 1:
1. What is feedback? Who are the recipients of feedback? Who are the
sources of feedback information?
Learning Task 2:
1. What is grading? How can grades be used effectively?
2. What are some issues about grading? In what ways have these be
approached? What can you recommend in the context of your own
academic setting?
Grade and Assessment: What’s the diffenrence? The crucial differences between
grading and assessment
https://education.cu-portland.edu/blog/classroom-resources/grading-vs-assessment-
whats-the-differnce/
How am I doing? This article describes how effective feedback helps students see what
they know and what they need to keep working on.
https://ascd.org/publications/educational-leadership/sept12/vol70/num01/%c2%how-
am-i-doing%c2%a2%c2%a3.aspx
https://www.edutopia.org/blog/tips-providing-students-meaningful-feedback-
marianne-stenger
[1] Navaro, R.L., Santos, R.G., and Corpuz, B.B. (2019). Assessment of Learning, OBE &
PPST Based, Fourth Edition. Lorimar Publishing, Inc.
[2] Bartlett, J. (2015). Outstanding Assessment for Learning in the Classroom. Routledge,
Taylor & Francis Group
[3] Frey, N. and Fisher, D. (2011). The Formative Assessment Action Plan. USA: ASCD
[5] Ecclestone, K. et. al. (2010) Transforming Formative Assessment in Lifelong Learning.
UK: McGraw-Hill Open University Press
[6] Airasian, Peter W. (2005). Classroom Assessment Concept and Applications. New York:
McGraw-Hill Companies, Inc.
[7] Shermis, Mark D. and Di Vesta, Francis J. (2011) Classroom assessment in action.
Rowman & Littlefield Publishers, Inc.
[8] Anderson, Lorin W. (2003). Classroom Assessment: Enhancing the Quality of Teaching
Decision Making. Lawrence Erlbaum Associates, Publishers
[9] Lovely Professional University. (2012). Educational Measurement and Evaluation. USI
Publications
[10] Mehrens, Williams, A and Lehmann, Irvin J. (1991). Measurement and Evaluation in
Education and Psychology. Wadsworth/Thomson Learning
Websites
https://dictionary.apa.org/subjective-test
https://en.wikipedia.org/wiki/Educational_assessment#:~:text=Placement%20assessment%
20%E2%80%93%20Placement%20evaluation%20is,i.e.%20the%20tests%20that%20colleges
https://www.queensu.ca/teachingandlearning/modules/assessments/04_s1_01_intro_secti
on.html
https://www.scribbr.com/methodology/types-of-validity/
https://thesecondprinciple.com/instructional-design/threedomainsoflearning/
Images
https://www.google.com/url?sa=i&url=https%3A%2F%2Fwww.visualpharm.com%2Ffree-
icons%2Ftest%2520passed-
595b40b85ba036ed117da493&psig=AOvVaw1UObbP1_t_ypEYDa-6D-
_0&ust=1597732317792000&source=images&cd=vfe&ved=0CAIQjRxqFwoTCJCEiZzeoesCFQ
AAAAAdAAAAABAK
https://www.google.com/url?sa=i&url=https%3A%2F%2Ficon-
library.com%2Ficon%2Fintroduction-icon-
0.html&psig=AOvVaw1mYlPy2srDA0N11ZvMTUF4&ust=1597731498262000&source=images
&cd=vfe&ved=0CAIQjRxqFwoTCPjM2ZHboesCFQAAAAAdAAAAABAD
https://www.google.com/url?sa=i&url=https%3A%2F%2Fwww.clipartmax.com%2Fmiddle%
2Fm2H7d3Z5G6i8H7A0_self-assessment-
icon%2F&psig=AOvVaw1jxv1bBZTNbSSv45gXbh5P&ust=1597730763047000&source=images
&cd=vfe&ved=0CAIQjRxqFwoTCJDJv77YoesCFQAAAAAdAAAAABAF
https://www.google.com/url?sa=i&url=https%3A%2F%2Ficons8.com%2Ficons%2Fset%2Fobj
ectives&psig=AOvVaw3bARz-62W-
UobmsSUGgDUS&ust=1597728360197000&source=images&cd=vfe&ved=0CAIQjRxqFwoTC
NDvobzPoesCFQAAAAAdAAAAABAD
https://www.google.com/url?sa=i&url=https%3A%2F%2Ficons8.com%2Ficons%2Fset%2Fass
essment&psig=AOvVaw1nAFt8FkSZYqBOKY2gdMkf&ust=1597728164659000&source=image
s&cd=vfe&ved=0CAIQjRxqFwoTCICSxuXOoesCFQAAAAAdAAAAABAD
https://www.google.com/search?q=assessment+of+learning+1+graphics&tbm=isch&ved=2a
hUKEwixvI-U_qDrAhVNbJQKHSpBBYMQ2-
cCegQIABAA&oq=assessment+of+learning+1+graphics&gs_lcp=CgNpbWcQA1DonAFYmqsBY
PCtAWgAcAB4AIABAIgBAJIBAJgBAKABAaoBC2d3cy13aXotaW1nwAEB&sclient=img&ei=xM8
5X7HKDc3Y0QSqgpWYCA&rlz=1C1CHBD_enPH791PH791
Books