Professional Documents
Culture Documents
Reddy (2011) - Design and Development of Rubrics To Improve Assesment Outcomes PDF
Reddy (2011) - Design and Development of Rubrics To Improve Assesment Outcomes PDF
Reddy (2011) - Design and Development of Rubrics To Improve Assesment Outcomes PDF
www.emeraldinsight.com/0968-4883.htm
QAE
19,1 Design and development of
rubrics to improve assessment
outcomes
84
A pilot study in a Master’s level business
Received January 2010 program in India
Revised July 2010
Accepted November 2010 Malini Y. Reddy
Department of Marketing and Strategy, IBS Hyderabad,
Icfai Foundation for Higher Education, Andhra Pradesh, India
Abstract
Purpose – This paper seeks to discuss the characteristics that describe a rubric. It aims to propose a
systematic method for developing curriculum wide rubrics and to discuss their potential utility for
program quality assessment.
Design/methodology/approach – Implementation of rubrics is a recent phenomenon in higher
education. Prior research and theoretical issues related to rubric design and development are
discussed. The proposed method for rubric development is illustrated by deriving generic analytic
scoring rubrics for two assessment methods, namely projects and cases in a Master’s level business
program in India. Aspects related to the validity of the rubrics developed are investigated and results
of reliability study conducted using intraclass correlation coefficient (ICC) are reported.
Findings – Upon testing, the rubrics were found to be reliable and valid grading tools. Results of
inter-rater and intra-rater reliability analyses illustrated that the rubrics supported standardization of
the manner in which cases and projects could be evaluated in different business courses.
Practical implications – Whereas previous studies on rubric construction have largely
concentrated on task specific rubrics, this study focuses on development of curriculum wide rubrics
that can be employed for assessment of students’ learning at both course and program level.
Originality/value – To date there has not been any published work on issues of assessment of
student learning through project and case analysis rubrics within diverse courses in a business
program. The method detailed in the study can guide the development of generic rubrics for
alternative assessment methods employed in business programs as well as in other disciplines.
Keywords Assessment, Learning, Higher education, Business schools, Students, India
Paper type Research paper
Introduction
A “rubric” in education literature is commonly understood as an assessment tool that is
used to describe and score observable qualitative differences in performances. It
captures the essence of performance in academic tasks by “listing the criteria, of what
counts, and describing levels of quality from excellent to poor” (Andrade and Du, 2005,
p. 1). In assessment methods such as projects, case analysis, essays, portfolio, where
Quality Assurance in Education
Vol. 19 No. 1, 2011
the constructed responses given by students cannot be evaluated with complete
pp. 84-104 objectivity, rubrics are considered an effective approach for achieving reliable
q Emerald Group Publishing Limited
0968-4883
(consistent) and valid (accurate) professional judgment of students’ performances
DOI 10.1108/09684881111107771 (Pellegrino et al., 1999).
The development and use of rubrics for assessment of learning outcomes has Design and
become a popular and recognizable trend since the early 1990s. While they have been
extensively used at the school level in the USA, their application in higher education, as
development of
well as their usage in other countries has gathered momentum only in the last one rubrics
decade (Montgomery, 2002). Several business schools in the USA are currently using
rubrics for assessment of students’ learning at course and program level[1]. This
interest in use of rubrics in business programs is growing rapidly in other countries as 85
well (Martell and Calderon, 2005). The departments of education and quality assurance
agencies in several countries (e.g. the USA, the United Kingdom, Australia, France,
Turkey) as well as globally recognized accreditation bodies (e.g. AACSB, EFMD,
AMBA) which have adopted assessment of learning outcomes as the cornerstone of
their assurance and accreditation process, have contributed to the growing awareness
and importance of rubrics in business education. They have been proposed as a viable
means for scrutinizing actual samples of student work produced in the programs for
assessing and assuring the attainment of the specified learning outcomes.
Having said this, it is puzzling to note the miniscule literature on the development and
use of rubrics in business programs. A majority of the limited samples of rubrics designed
for use in post secondary level business courses are task specific. There are no published
studies on development and use of a rubric for an assessment method such that it could be
easily implemented for different assignments/tasks in a variety of business courses (hence
forth, referred as curriculum wide generic rubrics). This paper, attempts to fill this gap by
demonstrating a systematic process for developing rubrics for two assessment methods,
namely, written analysis of cases and course based projects. The process illustrated can
be used for developing rubrics that are generic enough to suitably serve a variety of
assignment objectives in a course; can be implemented in different courses across
academic terms; and can possibly be applicable at multiple institutions.
Discussions on the inconsistencies in the scores given during the norming process help
in reconciling the divergence in understanding of the rubric (Maki, 2001). Additionally,
it serves the purpose of enhancing the reliability of subsequent scoring of student
performance between different raters (interrater reliability) and between the same rater
over a period of time (intrarater reliability). Testing the rubric on a pilot sample of
student work or on “colleagues willing to role-play as students” (Bresciani et al., 2004,
p. 35) and using the knowledge so gained to revise the rubric have been suggested as
the final step involved in rubric development.
Establishing the reliability and validity of the rubrics developed has not received
enough attention. Several empirical studies mention conducting pilot and reliability
tests, however very few (such as Dunbar et al., 2006; Oakleaf, 2006; Boulet et al., 2004)
report the results. Information about the procedures, analyses, and results would
enable readers to better understand claims made about validity and reliability. Studies
need to report how the validity of a rubric was established, scoring reliability,
including rater training and its contribution toward achieving inter-rater reliability.
Further, applicability of a rubric at multiple sites necessitates attention to rater
training as well as instrument use and bias.
From the literature reviewed it can be seen that the process of development of
rubrics requires tremendous investments of time and effort by instructors. Also, rubric
design poses some serious limitations. While it is possible to attain a consensus on the
process to be applied for rubric design and development, agreement on the criteria and
performance levels is difficult to achieve. Selection of criteria and drafting the
descriptors is highly subjective and contextual. It is dependent upon a variety of
factors including level of higher education (bachelor’s, master’s, doctoral); emphasis of
the program and course; quality of the students; instructor’s experience and expertise
etc. Even in the same context, different instructors may view the importance of criteria
for inclusion and their descriptions differently. As a result a rubric constructed for the
same assessment method or task could be drafted differently by different instructors.
Majority of the empirical studies do not elucidate the process used for rubric
development; rarely involve students in the development process; and hardly ever
provide information on validity and reliability of the rubrics developed. Further,
literature does not provide an approach for designing and developing such rubrics that:
.
are acceptable to several instructors from multiple disciplines; and
. can be used across courses and academic terms. This study proposes an
approach for addressing these aspects.
Study Design and
Based upon a review of literature on rubric design and construction, this study
proposed and tested a multi-step approach for construction of curriculum wide rubrics.
development of
This is presented in Table I, followed by a description of the implementation processes. rubrics
The rubrics in this study were developed by involving 35 instructors and 95
students from two reputed business schools in Hyderabad, India. Four instructors, who
have researched and published on the development and utility of rubrics and whose 89
work has been extensively cited were also involved as experts in the process of
development of rubrics. The rationale behind the procedure was to bring together the
views of a group of interested colleagues in the schools to build upon their common
knowledge, as this has a greater impact than any one individual may have upon
definition of learning goals and expectations, course design and course delivery
(Suskie, 2004). The composition of the multidisciplinary team of instructors was: 61
percent male; average age 35 years; 56 percent with more than five years of teaching
experience; specialization in the disciplines of finance and accounting (8 percent);
economics (18 percent); management (18 percent); human resources (14 percent);
operations (8 percent); marketing (26 percent); education (8 percent). Table II provides
further demographic details of the participants involved in rubrics development.
Steps
1 Identification of the learning objectives to be served by the use of the assessment method and
leading to the identification of qualities (criteria) that need to be displayed in a student’s work
to demonstrate proficient performance
2 Identification of levels of performance for each of the criteria
3 Development of separate descriptive scoring schemes for each evaluation level and criteria
4 Obtaining feedback on the rubrics developed
5 Revision of rubrics based on feedback from primary stakeholders Table I.
6 Testing the reliability and validity of the rubrics Step-by-step procedure
7 Pilot testing of the rubrics used for designing
8 Using the results of the pilot test to improve the rubrics generic analytic rubrics
QAE
Demographic item Participants (%)
19,1
Instructors (n ¼ 39)
Gender
Female 39
Male 61
90 Years of teaching experience
, 2 years 8
2-5 years 36
5-10 years 26
. 10 years 30
Area of specialization
Finance and accounting 8
Economics 18
Management 18
Human resources 14
Operations 8
Marketing 26
Education 8
Students (n ¼ 95)
Gender
Female 40
Male 60
Average age (years) 23
Table II. Work experience
Demographic data on , 1 years 48
participants involved in 1-2 years 34
development of rubrics . 2 years 16
Issues Identifies and differentiates Identifies and differentiates Identifies the issues in the case Not able to clearly identify the
identification between the crucial and non- between the crucial and non- however differentiation between issues presented in the case
crucial issues while clearly crucial issues while partially crucial and non crucial is either
describing the multiple describing the multiple superficial or not presented
perspectives of different perspectives of different
characters in the case. characters in the case
Analysis Reasoning is logical, well Examination of the facts Examination of the facts Examination of facts presented
articulated and based upon: presented in the case is detailed presented in the case is at times in the case is superficial and
detailed and correct and correct however the same is incorrect and does not draw incorrect leading to
examination of the facts not suitably articulated to upon relevant theories/concepts unsupported ideas with flawed
presented in the case; sound explain how the facts or relevant leading to inconsistent reasoning
theoretical knowledge; and theories/concepts support the reasoning
personal experience assertions/reasoning
Recommendations Vigorously explores multiple Proposes practical solutions for Proposes solutions which have a Proposes solutions which are
practical solutions and carefully each of the issues identified theoretical basis and are related either impractical or unrelated
details the consequences of each however only partially to the issues but are too general to the issues identified
solution examines the consequences of to implement
each action proposed
Communication Logical organization of write up Minor problems in organization Errors in organization of write Frequent errors in organization
presented in a professional of write up presented in a up presented with some lapses of write up and in presentation
manner with well connected professional manner with in title page, referencing, section distract the reader and interfere
sections and proper title page, proper title page, referencing, headings, word choice, with meaning/comprehension
referencing, section headings, section headings, word choice, grammar and spelling which
word choice, grammar and grammar and spelling distract the reader
spelling
Total score /16
Table IV.
development of
94
19,1
QAE
Table V.
Rubric developed for
course based projects
Levels
Exemplary Accomplished Developing Beginning
Criteria 4 3 2 1 Score
Statement of purpose Presents focused and novel Presents focused project Presents focused project Presents research objective(s)
project objective(s) that reveal objective(s) that enable objective(s) that enable which are either unfocussed or
a new line of enquiry and involvement in meaningful involvement in meaningful, lend themselves to readily
enable involvement in and challenging work though not challenging work available answers
meaningful and challenging
work
Information/literature Presents a critical review of Partial attempt at presenting a Presents only a summary of Presents a summary of either
review relevant and current literature/ critical review of relevant and relevant and largely current irrelevant or outdated
information to provide a clear current literature/information literature/information which literature/information
context and rationale for the to provide a clear context and does not provide a clear
proposed project rationale for the proposed context and/or rationale for the
project proposed project
Methodology Takes an appropriate Takes an appropriate The approach to examine the Takes an inappropriate
approach to study the project approach to study the project project objective(s) though approach to examine the
objective(s) and provides objective(s) and provide appropriate, appears to be pre- project objective(s)
logical explanation for explanations which partially selected with no attempt to
selection of the same justify the selection of the explain its selection
same
Computation and Absence of any errors in Presence of such minor errors Presence of such errors in Presence of such errors in
reporting of results computation and reporting of in computation and/or computation and/or reporting computation and/or reporting
results reporting of results that have of results that have a marginal of results that have a
no impact the analysis negative impact on the substantial negative impact on
analysis the analysis
Findings and analysis Presents factual findings as Presents factual findings as Presents factual findings as Presents only factual findings
well as analysis of connections well as analysis of connections well as analysis of connections without even making an
between them along with between them, however between them but synthesis attempt to examine the
synthesis with relevant synthesis with relevant with relevant theory/ connections between them or
theory/principles/literature theory/principles/literature principles/literature and synthesis with literature and
and personal experience and personal experience is personal experience is either personal experience
limited incorrect or absent
(continued)
Levels
Exemplary Accomplished Developing Beginning
Criteria 4 3 2 1 Score
Conclusions and Draws logical conclusions Draws logical conclusions Draws some of the conclusions Draws conclusions based on
recommendations from the evidence/analysis to from the evidence/analysis to from evidence/analysis and personal opinion to provide
provide action oriented provide recommendations some based on personal recommendations which are
recommendations while which are mostly action opinion to provide either unrelated to the problem
clearly articulating their pros oriented but articulation of the recommendations which are or are not action oriented with
and cons, thereby pros and cons do not either unrelated to the problem articulation of pros and cons of
demonstrating an demonstrate either an or are not action oriented with the recommendations being
understanding of the understanding of the articulation of pros and cons either absent or sketchy
complexity of the problem and complexity of the problem or being vague.
practicality of the practicality of the
recommendations recommendations
Documentation Format is professional with Format is professional with Lapses in organization, Lapses in organization,
proper organization, indexing, minor lapses in organization, indexing, cover page, indexing, cover page,
cover page, referencing, titles, indexing, cover page, referencing, titles, section referencing, titles, section
section headings, grammar referencing, titles, section headings, grammar and headings, grammar and
and spelling headings, grammar and spelling distract the reader spelling distract the reader and
spelling though not interfering with interfering with meaning
meaning
Total score /28
Table V.
rubrics
Design and
95
development of
QAE reliability study and two of these raters participated in the intra-rater reliability study
as well (average age 37 years; 64 percent male). Reliability of the rubrics developed was
19,1 tested by intraclass correlation coefficient (ICC), a measure of correlation, consistency
or conformity for a data set with multiple groups. ICC falls within the framework of
analysis of variance (ANOVA) and there are six formulas for calculating the ICC
depending upon the purpose of the study, the design of the study and type of
96 measurements taken. This study used a two-way random effects model of ICC. In this
model, both raters and samples are randomly selected from a larger pool. Each of the
raters scores all the samples and the individual rater’s score constitutes the unit of
analysis. Instead of an average measure, a single measure was selected as eventually,
the rubrics were intended to be used in a manner that the scores given by individual
raters (and not average score of raters) would be used to assess student work. This
version of ICC is interpreted as being generalizable to all possible judges (Shrout and
Fleiss, 1979) and is appropriate when the purpose is to establish that the scales can be
used effectively by several other raters. The analysis was carried out in SPSS version
12 and a 95 percent confidence interval was defined.
The selection of the raters was from the pool of instructors who had participated in
the previous steps. All the raters had taught the particular course whose samples they
were evaluating at least twice. With an aim to achieve high inter-rater reliability,
extensive discussions were held separately with each one of the raters involved in the
study. Anchor papers were used to clarify the criteria and the levels of performance. A
total of 100 student artifacts of written case analysis (43) and projects (57) were
collected from a cross section of core introductory courses to represent a spread of
qualitative and quantitative orientation. The courses were Marketing management
(MM); Financial management (FM); Macro economics and business environment
(MEBE); Business research methods (BRM); Management control and information
systems (MCIS); and Organizational behavior (OB). Along with the samples, the
assignment briefs were collected and shared with the raters to bring in clarity on the
objectives and expectations of the assignments.
Literature on inter-rater reliability states that for classroom assessments two raters
are sufficient to produce acceptable levels of inter-rater agreement. In this study two
raters independently scored each of the student artifacts. The raters were blind to the
other raters and also to the origination of the samples. The sample size in this study
was in the acceptable range of six-12 samples per course and assessment method
(Bonett, 2002; Walter et al., 1998). For each of the courses in which samples were
collected, absolute agreement and consistency were calculated for both total scores
given by the raters and for scores given against criterion in the rubric.
The results presented in Figure 1 show that the absolute agreement for the total
scores of case analysis across subjects ranged between 0.78-0.92 and consistency ranged
between 0.90-0.95. The same for the project rubric stood at 0.61-0.99 (agreement) and
0.71-0.99 (consistency). This is an acceptable range for initial evidence of score reliability
involving rubrics (Shrout and Fleiss, 1979; Wainer and Thissen, 1996).
Results of criteria wise ICC were also calculated to identify specific areas of low
reliability. In the case analysis rubric, relatively lower agreement as well as
consistency (0.64) was found in the MM course, in criteria 1. The same in the
course-based projects rubrics were seen in MEBE course in criteria 1, 2, 3, and 6
(agreement and consistency both reflected 0.615). Discussions were held with the
concerned raters to probe the reasons for the discrepant scores. It was found that one of
the raters for MEBE had difficulties in interpreting the levels correctly, while one of the
Design and
development of
rubrics
97
Figure 1.
Aggregate inter-rater
reliability (agreement and
consistency) of case and
project rubrics
References
Andrade, H. and Du, Y. (2005), “Student perspectives on rubric-referenced assessment”, Practical
Assessment, Research & Evaluation, Vol. 10 No. 5, available at: http://PAREonline.net/
getvn.asp?v¼10&n¼3 (accessed December 14, 2006).
Arter, J. and McTighe, J. (2001), Scoring Rubrics in the Classroom, Corwin, Thousand Oaks, CA.
Bigelow, J.D. (2004), “Using problem-based learning to develop skills in solving unstructured
problems”, Journal of Management Education, Vol. 28 No. 5, pp. 591-609.
Bonett, D.G. (2002), “Sample size requirements for estimating intraclass correlations with desired
precision”, Statistics in Medicine, Vol. 21, pp. 1331-5.
Boulet, J.R., Rebbecchi, T.A., Denton, E.C., Mckinley, D. and Whelan, G.P. (2004), “Assessing the
written communication skills of medical school graduates”, Advances in Health Sciences
Education, Vol. 9, pp. 47-60.
Bresciani, M.J., Zelna, C.L. and Anderson, J.A. (2004), Assessing Student Learning and
Development: A Handbook for Practitioners, National Association of Student Personnel
Administrators (NASPA), Washington, DC.
Brown, G., Bull, J. and Pendlebury, M. (1997), Assessing Student Learning in Higher Education,
Routledge, Oxon and New York, NY.
Callison, D. (2000), “Rubrics”, School Library Media Activities Monthly, Vol. 17 No. 2, pp. 4-36, 42.
Campbell, A. (2005), “Application of ICT and rubrics to the assessment process where
professional judgement is involved: the features of an e-marking tool”, Assessment &
Evaluation in Higher Education, Vol. 30 No. 5, pp. 529-37.
Corey, E. (1980), Case Method Teaching, Harvard Business School, Boston, MA.
Council for Higher Education Accreditation (CHEA) (2002), “Student learning outcomes
workshop”, CHEA Chronicle, Vol. 5 No. 2, pp. 1-3.
Crittenden, W.F. (2005), “A social learning theory of cross-functional case education”, Journal of
Business Research, Vol. 58, pp. 960-6.
Dunbar, N.E., Brooks, C.F. and Kubicka-Miller, T. (2006), “Oral communication skills in higher
education: using a performance-based evaluation rubric to assess communication skills”,
Innovative Higher Education, Vol. 31 No. 2, pp. 115-28.
Fraser, L., Harich, K., Norby, J., Brzovic, K., Rizkallah, T. and Loewy, D. (2005), “Diagnostic and
value-added assessment of business writing”, Business Communication Quarterly, Vol. 68
No. 3, pp. 290-305, available at: http://bcq.sagepub.com/cgi/content/abstract/68/3/290
(accessed May 2, 2007).
Glickman-Bond, J. and Rose, K. (2006), Creating and Using Rubrics in Today’s Classrooms,
Christopher-Gordon Publishers Inc., Norwood, MA.
Gopinath, C. (2004), “Exploring effects of criteria and multiple graders on case grading”, Journal
of Education for Business, Vol. 79 No. 6, pp. 317-22.
Green, R. and Bowser, M. (2006), “Observations from the field: sharing a literature review rubric”, Design and
Journal of Library Administration, Vol. 45 Nos 1/2, pp. 185-202.
development of
Greenhalgh, A.M. (2007), “Case method teaching as science and art: a metaphoric approach and
curricular application”, Journal of Management Education, Vol. 31 No. 20, pp. 181-94. rubrics
Huba, M.E. and Freed, J.E. (2000), Learner-Centered Assessment on College Campuses: Shifting
the Focus from Teaching to Learning, Allyn and Bacon, Boston, MA.
Lapsley, R. and Moody, R. (2007), “Teaching tip: structuring a rubric for online course 103
discussions to assess both traditional and non-traditional students”, Journal of American
Academy of Business, Vol. 12 No. 1, pp. 167-72.
Lundberg, C.C. and Enz, C. (1993), “A framework for student case preparation”, Case Research
Journal, Vol. 13, pp. 133-44.
Maki, P.L. (2001), “From standardized tests to alternative methods: assessing learning in
education”, Change, Vol. 33 No. 2, pp. 28-32.
Martell, K. and Calderon, T. (2005), Assessment of Student Learning in Business Schools: Best
Practices Each Step of the Way, Vol. 1, Nos 1/2, The Association for Institutional Research,
Tallahassee, FL.
Mertler, C.A. (2001), “Designing scoring rubrics for your classroom”, Practical Assessment,
Research and Evaluation, Vol. 7 No. 25, available at: http://PAREonline.net/getvn.
asp?v¼7andn¼25 (accessed December 17, 2005).
Montgomery, K. (2002), “Authentic tasks and rubrics: going beyond traditional assessments in
college teaching”, College Teaching, Vol. 50 No. 1, pp. 34-40.
Moskal, B.M. (2000), “Scoring rubrics: what, when, and how”, Practical Assessment, Research,
and Evaluation, Vol. 7 No. 3, available at: http://ericae.net/pare/getvn.asp?v¼7andn¼3
(accessed December 9, 2006).
Moskal, B.M. (2003), “Recommendations for developing classroom performance assessments and
scoring rubrics”, Practical Assessment, Research & Evaluation, Vol. 8 No. 14, available at:
http://PAREonline.net/getvn.asp?v¼8&n¼14 (accessed January 22, 2007).
Moskal, B.M. and Leydens, J.A. (2000), “Scoring rubric development: validity and reliability”,
Practical Assessment, Research & Evaluation, Vol. 7, pp. 71-81, available at: http://
pareonline.net/getvn.asp?v¼7&n¼10 (accessed December 10, 2005).
Oakleaf, M.J. (2006), Assessing Information Literacy Skills: A Rubric Approach, UMI No. 3207346,
PhD dissertation, University of North Carolina, Chapel Hill, NC.
Parke, C.S. (2001), “An approach that examines sources of misfit to improve performance
assessment items and rubrics”, Educational Assessment, Vol. 7 No. 3, pp. 201-25.
Parkes, K.A. (2006), The Effect of Performance Rubrics on College Level Applied Studio Grading,
UMI No. 3215237, PhD dissertation, University of Miami, Coral Gables, FL.
Pellegrino, J.W., Baxter, G.P. and Glaser, R. (1999), “Addressing the two disciplines problem:
linking theories of cognition and learning with assessment and instructional practice”,
Review of Research in Education, Vol. 24, pp. 307-53.
Petkov, D. and Petkova, O. (2006), “Development of scoring rubrics for IS projects as an
assessment tool”, Issues in Informing Science and Information Technology, Vol. 3,
pp. 499-510.
Popham, W.J. (1997), “What’s wrong and what’s right with rubrics”, Educational Leadership,
Vol. 55 No. 2, pp. 72-5.
Popham, W.J. (2003), Test Better, Teach Better: The Instructional Role of Assessment,
Association for Supervision and Curriculum Development, Alexandria, VA.
QAE Reitmeier, C.A., Svendsen, L.K. and Vrchota, D.A. (2004), “Improving oral communication skills
of students in food science courses”, Journal of Food Science Education, Vol. 3, pp. 15-20.
19,1 Schrock, K. (2000), Kathy Schrock’s Guide for Educators, available at: http://school.discovery.
com/schrockguide/assess.html (accessed March 10, 2007).
Schroeder, H. and Fitzgerald, P. (1984), “Peer evaluation in case analysis”, Journal of Business
Education, Vol. 60, pp. 73-7.
104 Shrout, P.E. and Fleiss, J.L. (1979), “Intraclass correlations: uses in assessing rater reliability”,
Psychological Bulletin, Vol. 86 No. 2, pp. 420-8.
Song, K.H. (2006), “A conceptual model of assessing teaching performance and intellectual
development of teacher candidates: a pilot study in the US”, Teaching in Higher Education,
Vol. 11 No. 2, pp. 175-/190.
Stefl-Mabry, J. (2004), “Building rubrics into powerful learning assessment tools”, Knowledge
Quest, Vol. 32 No. 5, pp. 20-5.
Stevens, D.D. and Levi, A.J. (2005), Introduction to Rubrics, Stylus, Sterling, VA.
Suskie, L. (2004), Assessing Student Learning: A Common Sense Guide, Anker Publishing
Company, Bolton, MA.
Tierney, R. and Simon, M. (2004), “What’s still wrong with rubrics: focusing on the consistency of
performance criteria across scale levels”, Practical Assessment, Research and Evaluation,
Vol. 9 No. 2, available at: http://PAREonline.net/getvn.asp?v¼9andn¼2 (accessed
September 19, 2005).
Wainer, H. and Thissen, D. (1996), “How is reliability related to the quality of test scores? What is
the effect of local dependence on reliability?”, Educational Measurement: Issues and
Practice, Vol. 15, pp. 22-9.
Walter, S.D., Eliasziw, M. and Donner, A. (1998), “Sample size and optimal designs for reliability
studies”, Statistics in Medicine, Vol. 17 No. 1, pp. 101-10.
Woolf, H. (2004), “Assessment criteria: reflections on current practices”, Assessment &
Evaluation in Higher Education, Vol. 29 No. 4, pp. 479-93.
Further reading
Marzano, R.J., Brandt, R.S., Hughes, C.S., Jones, B.F., Presseisen, B.Z., Rankin, S.C. and Suthor, C.
(1988), Dimension of Thinking: A Framework for Curriculum and Instruction, Association
for Supervision and Curriculum Development, Alexandria, VA.