Professional Documents
Culture Documents
Azu Etd 14700 Sip1 M
Azu Etd 14700 Sip1 M
by
Adnan Alobaid
__________________________
DOCTOR OF PHILOSOPHY
2016
THE UNIVERSITY OF ARIZONA
GRADUATE COLLEGE
As members of the Dissertation Committee, we certify that we have read the dissertation
prepared by Adnan Alobaid, titled Testing, Assessment, and Evaluation in Language
Programs and recommend that it be accepted as fulfilling the dissertation requirement for the
Degree of Doctor of Philosophy.
Final approval and acceptance of this dissertation is contingent upon the candidate’s
submission of the final copies of the dissertation to the Graduate College.
I hereby certify that I have read this dissertation prepared under my direction and recommend
that it be accepted as fulfilling the dissertation requirement.
2
STATEMENT BY AUTHOR
This dissertation has been submitted in partial fulfillment of the requirements for an
advanced degree at the University of Arizona and is deposited in the University Library to
Brief quotations from this dissertation are allowable without special permission, provided
that an accurate acknowledgement of the source is made. Requests for permission for
3
ACKNOWLEDGEMENTS
This dissertation could not have been accomplished without the assistance and
support of many individuals. First and foremost, I would like to express my deep and sincere
gratitude to my supervisor Suzanne Panferov, who has been more than generous with her
assistance, patience, and expertise in addition to guiding me through the entire process of
Peter Ecke, and Edmond White who have been substantially supportive, encouraging, and a
great source of inspiration. I also would like to thank Dr. Nicholas Ferdinandt with whom I
took program evaluation class. Dr. Ferdinandt has been very supportive and helpful in
providing me with all necessary advice, references, and information pertaining to the program
evaluation industry.
has been very instrumental in providing me with thoughtful insights in addition to discussing
crucial issues relevant to my dissertation topics. Furthermore, I would like to thank Ms.
Amani Suleiman Al-Samhan who helped me collect data from the study site. She has been
very generous with her time and assistance carrying out the interview part of one of this
dissertation articles.
Last but not least important, my deepest debt of gratitude goes to my family members,
especially my parents, Othman and Lulwa, for their prayers, support, and encouragement. I
also would like to thank my wife, Reem, and my two children, Wissam and Tala, for their
support, patience, and assistance in that without whom I could not have been able to complete
this dissertation.
4
TABLE OF CONTENTS
ACKNOWLEDGEMENTS ............................................................................................................ 4
ABSTRACT .................................................................................................................................. 13
ABSTRACT .............................................................................................................................. 20
INTRODUCTION .................................................................................................................... 21
A Historical Review on ESL Placement Tests: Traditional ESL Placement Tests ..................... 23
METHODOLOGY ................................................................................................................... 31
5
Significance of the Study ............................................................................................................. 35
Participants .................................................................................................................................. 37
Procedures ................................................................................................................................... 38
FINDINGS ................................................................................................................................ 39
LIMITATIONS ......................................................................................................................... 52
IMPLICATIONS ...................................................................................................................... 54
ABSTRACT .............................................................................................................................. 59
INTRODUCTION .................................................................................................................... 60
Definition ..................................................................................................................................... 61
Learner-centeredness ................................................................................................................... 63
6
Self-directed Learning ................................................................................................................. 63
Self-assessment ............................................................................................................................ 64
Correlations ................................................................................................................................. 65
Discrepancies ............................................................................................................................... 65
METHODOLOGY ................................................................................................................... 69
Research Tools............................................................................................................................. 70
Participants .................................................................................................................................. 70
Procedures ................................................................................................................................... 72
FINDINGS ................................................................................................................................ 74
7
ABSTRACT .............................................................................................................................. 88
INTRODUCTION .................................................................................................................... 89
8
Data Analysis ............................................................................................................................. 112
Evaluating the Goals and Objectives of the Two EFL Programs .............................................. 116
Student Satisfaction with the Practices of the Two Departments .............................................. 128
9
CHAPTER 5: CONCLUSION ................................................................................................... 137
Appendix J: Color-coded CEFR, TOEFL, and IELTS Equivalency Table (Tannenbaum &
Figure 2. Scatterplot (BIVAR)=TOEFL with Ratio by Level of Proficiency for Male and
Female. ..................................................................................................................................... 80
Figure 3. Student learning outcomes cover all language skills. ............................................ 119
Figure 4. Teaching strategies are proper for various learning styles. .................................... 120
Figure 6. You are engaged presentations & leading discussions .......................................... 120
Figure 10. Students’ responses to how often their work is graded and returned to them
Figure 11. Students’ responses to the extent to which SLOs meet their expectations. ......... 129
11
LIST OF TABLES
Table 1 ..................................................................................................................................... 38
Table 2 ..................................................................................................................................... 40
Table 3 ..................................................................................................................................... 41
Table 4 ..................................................................................................................................... 42
Table 5 ..................................................................................................................................... 44
Table 6 ..................................................................................................................................... 71
Table 7 ..................................................................................................................................... 73
Table 8 ..................................................................................................................................... 75
Table 9 ..................................................................................................................................... 76
Table 10 ................................................................................................................................... 77
Table 11 ................................................................................................................................... 82
12
ABSTRACT
language testing, assessment, and evaluation. The first article (Saudi Student Placement into
ESL Program Levels: Issues beyond Test Criteria) addresses a crucial yet understudied issue
Saudi students. Based on data obtained from different stakeholders, the findings revealed that
assessment rubric compared to students’ TOEFL scores, and the extent to which gender and
levels of language proficiency cause any potential score underestimation. By obtaining data
from 21 ESL students attending the Center for English as a Second Language at University of
self-assessed scores and their TOEFL scores. However, the participants reported that the
On the other hand, the third article (Quality Assurance and Accreditation as Forms for
of the NCAAA (the National Commission for Academic Accreditation and Assessment) and
CEA (the Commission on English Language Program Accreditation). The findings indicated
that the standards of the mission, curriculum, student learning outcomes, and program
development, planning, and review, were partially met, whereas the standards of teaching
Program
Evaluation
Classroom
Alternative
Assessment
Placement Tests
14
CHAPTER 1 INTRODUCTION
(Fulcher & Davidson, 2007), and they play a fundamental role in fulfilling a number of
functions in the educational process at different levels. Research stresses the important of
these three concepts in planning, guiding, and implementing modern educational processes.
For example, Fulcher (2010) asserts that tests are “vehicles by which society can implement
equality of opportunity or learner empowerment” (p. 1). They function as gatekeepers for
education through a process of evaluating student performance to ensure they satisfy or meet
a set of standards or requirements, respectively. Furthermore, tests are tools, which are used
to verify the extent to which an individual has mastered a specific skill(s). Assessment, on the
other hand, is concerned with documenting the target knowledge, skills, and attitudes in that a
These three concepts are an indispensable part of L2 learning. For example, Carroll
(1961) indicates that language tests are key to formal language learning as they “render
information to aid in making intelligent decisions about possible courses of action” (p. 314).
This suggests that verifying the extent to which L2 student learning outcomes are achieved is
unlikely to be made without any form of testing. Moreover, language assessment plays a vital
role not only in L2 learning but also in applied linguistics, by “operationalizing its theories
and supplying its researchers with data for their analysis of language knowledge or use”
(Clapham, 2000, p. 148). Furthermore, program evaluation has become key to many L2
language programs as “no curriculum should be considered complete without some form of
program evaluation” that corresponds with the program emerging developments (Brow, 2007,
p. 158). This three-article dissertation addresses language testing, assessment, and evaluation.
15
The first article, “Saudi Student Placement into ESL Program Levels: Issues beyond
Test Criteria”, attempts to account for why some Saudi ESL students studying in the United
States are placed into lower than-expected levels. In other words, one of the IEP programs in
the Southwestern United States has received some complaints raised by the Saudi Arabia
Cultural Mission (SACM), a Saudi student sponsor in the US, about Saudi students being
placed into lower ESL levels than expected. Of course, this placement may be attributed to
numerous factors including students’ low level of language proficiency, fatigue, lack of
adequate background on ESL placement tests, and so forth. Nevertheless, based on initial
anecdotal evidence found by the researcher, it was found that there are other reasons causing
this issue. That is, many Saudi ESL students indicated that they purposefully underperform
on ESL placement tests for several reasons including but not limited to having more time to
learn English, adapting to the target educational and environmental contexts, preparing well
for admission-related tests (TOEFL, IELTS, and GRE), and so on. To examine whether this
issue exists and therefore represents a dilemma for some or all stakeholders (language
programs, scholarship agencies, students, teachers), a set of different data were collected
from a larger, more diverse group of Saudi and Gulf Cooperation Council (GCC) students.
The article will help solve this issue, to whatever extent it exists, by first obtaining the
opinions of various stakeholders. Then, based on their responses and attitudes, implications
will be provided to help mitigate the issue. The findings of this article; however, are not
intended only for Saudi or GCC students; they are also intended for ESL students of other
nationalities. Thus, the article will be a unique contribution to possibly avoiding any form of
negative impact caused by this issue on stakeholders. That is, having a classroom with a large
number of self-misplaced students may pose greater issues for ESL teachers in particular, for
levels of language proficiency. To do so, self-assessed scores of a group of ESL students are
compared against their TOEFL/IELTS scores in order to identify the extent to which this self-
assessment measure can be used in L2 contexts for proficiency purposes. The reason for
twofold. First, as a previous EFL student, the researcher never used any self-assessment
measures during his entire undergraduate studies. This lack of experience led him to
encounter several difficulties when he first pursued his higher studies in the United States.
student-centeredness (Nunan, 1988), enhancing student learning (Boud 1995; Taras, 2010),
traditional assessment techniques. That is, it proposes endeavors be made to use CEFR or any
other valid self-assessment rubrics not only in L2 programs in Saudi Arabia, but also in many
ESL programs in general. More importantly, the findings of this article are hoped to
encourage university students to develop their autonomy through self-assessment. At the end
of the paper, EFL departments in Saudi Arabia in particular and other L2 practitioners in
from the National Commission for Academic Accreditation and Assessment (NCAAA), this
article provides a simulated evaluation of two EFL departments at a Saudi University to serve
as an effective model and benchmark for other language programs that seek to obtain
accreditation. Another salient gap that needs to be addressed is that through its Ninth
aspect of its policies (Ministry of Economy and Planning, 2015). Hence, this article
Another main gap that this article attempts to fill is that program evaluation does not
techniques”, for it is deemed “as of a lower priority than the more obviously immediate
activities associated with design and planning” (MacKay, 1998, p. 33). In other words, as
MacKay notes, some program administrators argue that once a program is planned, designed,
and implemented, program evaluation processes will become natural and spontaneous as a
evaluation and review, and that program evaluation is not, under any circumstances, “an ad
hoc, unprincipled, or arbitrary activity” but rather it is a planned, organized, and guided
process (MacKay, 1988, p. 34). Therefore, by evaluating two university English language
programs in Saudi Arabia through a set of integrated standards of the NCAAA and CEA, this
article explores program evaluation in general yet simulating through other lenses.
18
CHAPTER 2 SAUDI STUDENT PLACEMENT INTO ESL PROGRAM LEVELS:
19
ABSTRACT
ESL Placement tests are used “to place students at the stage of the teaching program
most appropriate to their abilities” (Hughes, 1989, p. 14). Placement decisions should thus
reflect students’ actual levels of language proficiency; otherwise, improper placements may
affect their learning achievement (Bachman & Palmer, 1996). Although most ESL placement
tests adhere to testing theoretical underpinnings, another issue may exist. That is, some Saudi
ESL students studying in the United States are placed into lower than-expected levels, which
might be attributed to several factors (e.g., students’ low level of language proficiency); yet,
there are other reasons for this issue. To gain initial insight into this issue, a piloted survey
was circulated to 27 Saudi students studying in the U.S. The findings revealed that 40% of
them purposefully underperformed on university ESL placement tests for various reasons.
To examine the extent to which this issue exists, a survey was circulated to a
randomly selected group of Saudi ESL students attending three ESL programs in the U.S.
This was followed by conducting semi-structured interviews with a randomly selected sample
underperformed on ESL placement tests. Then, the same survey was circulated to GCC (Gulf
Cooperation Council) students studying in the U.S. The findings showed that 18% of them
also deliberately underperformed on ESL placement tests. Finally, some ESL administrators
were surveyed about this issue and provided contradicting percentages of students
purposefully underperforming on ESL placement tests. This article, therefore, provides some
Key Words: Test validity, test reliability, test authenticity, test practicality,
teaching, for it addresses key issues pertaining to language evaluation and assessment such as
test design, test administration, scores interpretation and so forth (Bachman & Palmer, 1996;
Brown, 2010; Hughes & Scott-Clayton, 2011; Kunnan, 1998; McNamara, 2006). ESL
placement tests are widely used by universities, community colleges, and language programs
for the purpose of placing prospective students into homogeneous groups based on accurate
measures of their levels of language proficiency (Bachman, 1990; Bachman & Palmer, 1996;
Fulcher, 1997; Wall, Clapham, & Alderson, 1994). The main purpose of ESL placement tests
is to provide accurate measures of students’ language abilities in order to place them in the
appropriate level classes (Kunnan, 2000). However, failure to report accurate scores that
reflect students’ actual language proficiency can lead to placing them into incorrect levels,
which can impact their current and future learning (Hughes, 1989).
Research has approached language placement testing from different perspectives. For
example, Fulcher (1997) points out that early placement testing publications (e.g., Goodbody,
1993; Schmitz & DelMas, 1990; Schwartz, 1985; Truman, 1992) were “concerned with the
placement of linguistic minority students” into language programs (p. 113). Before the 1990s,
researchers rarely examined the validity of ESL placement tests compared with other studies
conducted on first language and standardized foreign language tests because placement tests
are often low stakes as opposed to other language tests such as exit tests (Brantmeier, 2006).
This paper investigates another, but as of yet understudied, issue concerning the reasons why
many Saudi ESL students are placed into lower-than-expected levels in intensive university
ESL programs abroad and the extent to which this behavior affects program placement
systems.
21
In this paper, it is assumed that the placements of many Saudi ESL students are lower
than expected for numerous reasons. First, the Saudi Arabian Cultural Mission (SACM) has
long complained about placing Saudi ESL students into lower levels (ESL program
administrator, personal communication, March 12, 2015). SACM is concerned with the
reasons why lower-level ESL classes typically contain a disproportionate number of Saudi
students, and SACM ESL coordinators have attempted to identify the factors causing this
issue. They wonder if it is a language issue, a testing issue, a lack of experience in taking
placement tests, or a combination of issues. Second, research has reported that Saudi ESL
students’ oral skills are better than their writing skills. For example, Johnson (2015) found
that Saudi students perform better on the University of Illinois entry interviews compared
with their written placement scores. Saba’s (2014) longitudinal study also found that within a
short period of time, some Saudi students achieved TOEFL scores higher than their scores on
the Virginia Tech Language and Cultural Institute (VTLCI) placement test.
The discrepancies between Saudi ESL students’ placement scores and their actual
level of English proficiency can be ascribed to numerous factors such as lack of testing skills,
test-takers’ varied abilities in taking computerized test, test anxiety, and so on. Nevertheless,
more tangible evidence is needed to support this extrapolation. Thus, the researcher collected
anecdotal evidence from a randomly selected group of Saudi ESL students, all who indicated
that they purposefully performed poorly on ESL placement tests for various reasons, which
will be discussed below. This finding encouraged the researcher to review literature
addressing this issue in order to identify if some ESL students intentionally underperformed
on ESL placement tests, and if so, to what extent? Unfortunately, the researcher was unable
to find adequate studies addressing the issue. Hence, a historical review on ESL placement
tests and an in-depth discussion about test criteria will be provided in detail instead.
22
A Historical Review on ESL Placement Tests: Traditional ESL Placement Tests
Due to the dominance of SLA theories that emerged during the 70’s, 80’s and 90’s,
ESL placement tests have undergone profound changes (Green & Weimer, 2004). For
example, ESL placement tests of earlier years were greatly impacted by grammar-translation
approaches, resulting in tests that focused mainly on language structures (Brown, 1989).
Hughes, Weir, and Porter (1995) argue that traditional ESL placement tests merely test
homogeneous grouping (p. 13). Later, the interactionalist paradigms emerged (Gass &
Varonis, 1994; Long, 1981; Mackey, 1999), which emphasized the importance of presenting
test items with rich contextual clues through which “interactional adjustments” could be
triggered (Gass, 2010, p. 221). This contextualized ESL placement tests items, “whose recall
ESL placement tests completely neglected other language skills such as listening and
speaking.
Communicative language teaching (CLT), which emerged in the early 1970s (Hymes,
1972) and 1980s (Wesche, 1983), stressed that oral skills be included in ESL placement tests
(Brown, 1989). Consequently, this shifted testing from focusing on language structure (e.g.,
This paradigm shift resulted in the inclusion of more authentic texts (reading), interactive
tasks (speaking), lecture note-taking tasks (listening), and integrative-based tasks (writing).
The CLT approach has also introduced performance- and task-based test items, which made
ESL placement tests move away from solely testing L2 learners’ grammatical knowledge to
23
ESL Placement Test Item Formats
true/false, error correction, and so forth (Bachman, 1990). Although multiple-choice formats
have been used for quite some time, they were initially integrated into ESL placement tests
during the early 1990s, primarily as a reaction to the widespread and popular item response
theory research (IRT) (Bock, 1997). Researchers have addressed the issue of test format and
the extent to which it affects test-takers’ performance. For example, Bridgeman and Lewis
(1994) argue that some test-takers are “relatively strong on essays and weak on multiple-
choice question and vice versa” (p. 133), implying that the nature of the test construct has a
great impact on test-takers’ scores. This suggests that test format should take into careful
Another issue that test designers should bear in mind when designing a placement test
is the extent to which ESL placement test format affects test-takers’ performance. Bridgeman
and Morgan (1996) conducted a study to validate an ESL placement test designed with two
sample of ESL students in two separate sessions. The results indicated that some students,
who achieved higher scores in the essay-based test, obtained lower scores in the multiple-
choice format and vice versa. In the same vein, in validating the Lancaster University EAP
test, Wall et al. (1994) found evidence that multiple-choice items were fairly easy for many
test-takers, with a total mean scores ranging from 70% to 76%, whereas the mean scores of
the essay-based test was lower (57.91%). According to Lancaster University’s policies,
students who achieved higher scores met the targeted program’s requirements, while those
who achieved lower scores were required to take pre-sessional or remedial courses to help
beginning the early 1980s, ESL placement tests were delivered and graded more and more by
computers, known as computer-based tests (CBTs), which slowly replaced human raters.
Brown (1997) lists three key benefits of CBTs: 1) they can be “individually administered,
even on a walk-in basis”, 2) they do not involve many proctors, and 3) they can be kept for
future use, review, and adaptation (p. 45). Despite these benefits, it was questioned whether
the two formats would report equivalent results. Based on a Cronbach’s alpha analysis,
Fulcher (1999) found that the scores of a CBT correlated significantly with those of a PBT,
where the former’s reliability was 0.90, while the latter’s was 0.95. This implies that both
formats can report consistent scores provided test-takers are familiar with using computers.
Currently, many ESL placement tests are being delivered online, a dramatic change in
the placement test industry compared with their PBT and CBT counterparts (Roever, 2001).
That is, the Internet provides impetus opportunities for “test-taking at the learner’s
convenience and providing immediate and specific feedback to learners” (Chapelle, Jamieson
& Hegelheimer, 2003, p. 409). In fact, many ESL placement tests today are effectively
delivered and graded online despite the fact that some technical glitches may occur during
taking the test. One question that arises is to what extent do different delivery test formats
yield consistent scores? Research has revealed conflicting results; however, regardless of
delivery or design formats, the consensus is that all test-takers should have equal
opportunities in taking the test (Brantmeier, 2006; Brown, 1997; Fulcher, 1999; Kirsch,
Jamieson, Taylor & Eignor, 1998). The next section sheds some light on test criteria that
should be taken into account during the design process of ESL placement tests.
25
Test-takers Intentionally Failing Exams: Myth or Reality?
identify similar cases, from any discipline, that show people making conscious attempts to
purposefully perform poorly on an exam. In other words, do people have legitimate motives
to intentionally perform poorly on a test? Literature from various disciplines has reported that
some people fail a test to achieve desired goals. For example, based on neuropsychological
data, Coppel (2011) found that some high school “athletes have been known to deliberately
perform poorly on baseline to create a low comparison point to evaluate change on post-
concussion follow-up” (p. 658). This suggests that purposefully performing poorly on a
baseline exam will enable high school athletes to be excused from practice, take longer
leaves, receive lighter assignments, and be exempt from doing their homework.
malingering. That is, malingers have been found to deliberately perform poorly during
clinical examinations by feigning memory problems caused by traumatic events hoping that
the results would lead them to achieve specific external incentives (Sekely, 2014). For
obtaining drugs” (Diagnostic & Statistical Manual of Mental Disorders 4th ed. American
Psychiatric Association, 2000, p. 739). In academic contexts, Ahlgren Reddy and Harper
(2013) reported that some prospective mathematics students often “intentionally perform
easy way to control for this” (p. 688). Based on the examples above, one may realize that
Only one research article was found that addressed the motives of ESL students who
deliberately perform poorly on ESL placement tests. Kahn, Butler, Weigle and Sato (1994)
found that some ESL students tend to “intentionally perform poorly in oral interviews in
order to remain at a lower level...to avoid missing anything, or when they want to stay with
friends or an instructor at a lower level” (p. 38). Yet, no other evidence pertinent to this issue
has been reported, which can be ascribed to the assumption that ESL students might find no
benefit of revealing their intentions to throw any test. Moreover, as a Saudi, I assume that
many Saudi ESL students might be reluctant to acknowledge such behavior due to the fear of
losing their scholarship. Thus, this leads us to seek additional language-related cases of ESL
diagnostic tests, so that “some students may actually be motivated to fail or at least to get a
low score on criterion-referenced tests” (p. 287). This indicates that in some cases, students
may intentionally underperform on tests for different reasons. Brown and Hudson (2002)
provide an interesting yet inappropriate choice of students’ attempts to fail a test on purpose
in that some “cynical students may try to outsmart the criterion-referenced testing process
when they understand what is going on” (p. 287). That is, taking the test at the beginning of
the course for diagnostic purposes “and are told by the teacher (or guess)” that they will
retake it at the end of the course for achievement purposes, students tend to deliberately
underperform on the pre-test, and then take the post-test seriously (p. 87). By doing so, they
will be able to make perceived learning progress due to the difference between their pre- and
post-test scores without any learning actually occurring. Such deliberate failures have been
failing a test. Similarly, although the Modern Language Aptitude Test (MLAT) is often used
to predict students’ ability to learn a foreign language (Carroll & Sapon, 2002), it has been
recently used to diagnose students who claim that they have learning disabilities that may
impede them from learning a foreign language. The findings indicated that some students
intentionally failed the test, so they could waive or postpone a foreign language course. To
address this issue, some schools hire counselors who conduct interviews with test-takers to
ensure that their MLAT reported scores are accurate. Despite all the efforts exerted by
schools to overcome or at least alleviate the aforementioned behavior, many cases of students
failing the MLAT test still occur (Carroll & Sapon, 2002).
who was one of the attendees, reported that one of the presenters suggested that a new teacher
evaluation system be imposed as students’ reported scores are inaccurate. That is, since
students know that they will go through a pre- and post-test evaluation process to identify
their progress in a course, they deliberately fail a test to reflect a rapid achievement based on
their post scores. However, their motives for failing a test intentionally are not always for
their own benefit but rather for their teachers’. For example, knowing that their teachers are
evaluated based on their achievement, these high school students intentionally fail a
diagnostic pre-test so that the teacher(s) they like can obtain higher scores on teacher
evaluation (Leonardatos, 2012). This assumes that the students would achieve higher scores
on the achievement post-test. Conversely, students might perform well on the diagnostic pre-
test so that the teacher(s) they dislike can gain lower scores on teacher evaluation.
28
In the previous sections, a number of significant issues pertaining to ESL placement
tests were discussed in detail including how these tests have developed over time, with
particular focus on item formats, delivery modes, theoretical underpinnings, and so forth. It
was also been emphasized that regardless whether ESL placement tests have diagnostic
ultimately reflect test-takers’ actual levels of English proficiency. This is due to the fact that
they are originally developed “to assess students’ level of language ability so that they can be
placed in the appropriate course or class” (Alderson, Clapham & Wall, 1995, p. 11). Such
purposes can be achieved by ensuring a set of test criteria in order to make the test as useful
as possible. In some cases, however, despite being valid, reliable, practical, and even
authentic, some placement tests may fail to report test-takers’ actual levels of language
As reported by the study’s anecdotal evidence and pilot study, it was discovered that
some Saudi ESL students purposefully underperformed on ESL placement tests for certain
planned or personal reasons. This paper examines the accuracy of these findings by surveying
a larger number of Saudi and GCC ESL students. The reason for choosing Saudi ESL
students in particular is two-fold. First, as a Saudi, the researcher has witnessed a wide range
of Saudis intentionally performing poorly on ESL placement tests. That is, whenever he
attends Saudi gatherings, meets with Saudi friends, or even has some discussions with other
Saudi students, he notices that some of them encourage each other to fail placement tests on
purpose for many reasons, most importantly to have more time to seek university admission.
These students believe that underperforming on a placement test will help them stay longer in
the U.S., for if they are placed into higher ESL levels and finish the ESL program in a short
period of time, they are required to obtain admission sooner or be forced to leave the country.
29
Saudi Students Studying Abroad
The first study abroad scholarships began in 1927, when six students were sent to
Egypt to pursue higher degrees at the Kingdom’s expenses (MOHE1, 2014). Since that time,
study abroad scholarships have grown exponentially. According to Alqahtani (2014), one of
the most prosperous eras of these scholarships was the launch of King Abdullah Scholarship
Program in 2005, which was established based on an agreement between King Abdullah and
George W. Bush (Taylor & Albasri, 2014). The program sends qualified Saudi students to
“the best universities worldwide for further studies leading to academic degrees (Alqahtani,
2014, p. 33). As of 2014, the number of Saudi students studying abroad dramatically
increased to 200,000 students, approximately 111,000 of them in the U.S. (MOHE, 2014).
The Saudi Arabian Cultural Mission (SACM) is a Saudi agency in the U.S. that
oversees the King Abdullah Scholarship Program, represents Saudi universities, and provides
scholarship Saudi students with academic and financial support (MOHE, 2014). The ESL
Department at SACM functions as an intermediary between ESL programs and ESL Saudi
students (SACM, 2013). There is no required level of English proficiency for obtaining a
requirements (MOHE, 2014). Thus, in most cases, Saudi students’ levels of English
proficiency are generally based on their high school English backgrounds, which range from
beginner to intermediate with some exceptions. Some students, like those from ARAMCO
who are taught in English in KSA, generally have higher levels of English proficiency, while
those with very limited exposure to English are at much lower levels (Al Murshidi, 2014).
1
Ministry of Higher Education
30
METHODOLOGY
The key motivation for conducting this study was to identify the reasons why many
Saudi ESL students are placed into lower than-expected ESL levels. This requires carrying
out further investigations, recruiting a large number of participants, or even triangulating the
data from various resources. Therefore, data collection involved several stages. First, an
anecdotal confirmation of this behavior was obtained from some Saudi ESL students. Second,
based on the anecdotal findings, a pilot study was circulated to many Saudi ESL students.
Third, more data was collected from Saudi ESL students attending three ESL programs in the
southwestern United States. Finally, the study survey was circulated to a large number of
GCC students studying in the United States. The findings suggested numerous reasons
accounting for why many Saudi ESL students are placed into lower than-expected levels.
Being uncertain about why many Saudi ESL students are placed into lower than-
expected ESL levels, the researcher decided to collect some anecdotal evidence from a
randomly selected sample of Saudi students studying English in the United States. To
accomplish this, he conducted face-to-face interviews with seven of them and communicated
online via chat and email with nine of them. During the interviews, the researcher noticed
consistent themes about the issue at hand. Some Saudi students indicated that they
deliberately performed poorly on ESL placement tests. Based on these themes, a pilot study
consisting of a twelve-item survey was then administered online through Twitter (See
Appendix A). Then, it was sent via Whatsapp to a group of Saudis who fit into the targeted
category. After asking them about some introductory questions about their experience in
taking ESL placement tests, the participants were then directly asked if they have, in one way
indicated that they had indeed failed an ESL placement test on purpose for several reasons,
which will be discussed thoroughly in the Findings section. Although this is a pilot study with
a limited number of subjects, the findings are still stunning. That is, almost half of the
randomly selected ESL learners made a conscious attempt to fail a placement test, a problem
worth further investigation, especially given that the results are consistent with those of the
anecdotal evidence. To further explore the issue, in the subsequent data collection stages, a
more developed survey was circulated to a larger number of Saudi and GCC ESL students in
addition to interviewing a random sample of them in order to allow them to expand more on
The IEP programs examined in this study are located in the southwestern United
States. In order to maintain anonymity, they are labeled Program A, Program B, and Program
C throughout this article. Program A consists of six ESL levels: Basic (1 & 2), Intermediate
(1 & 2), and Advanced (1 & 2). If students obtain a GPA of 3.0 and above in Advanced 2,
then they meet the university language requirement. Program B, on the other hand, consists
of seven levels often ending with an optional Bridge Program. Each level takes eight weeks
to complete unless a student fails two or more classes, in which case they must repeat the
level. In the Bridge Program, academic credit-bearing courses are combined with some ESL
courses to develop students’ English skills and to engage them in a university course in their
areas of specialization. If students complete level 7 successfully, they will fulfill the
consists of six levels. Students can obtain an endorsement for university entrance if they earn
reasons. Unfortunately, the researcher did not have access to Program A’s placement test. He
believes that the placement tests of Programs B and C are valid since the former’s was
abstracted from a previously validated test, the International Test of English Proficiency or
iTEP, and students are reassigned a level whenever misplacements are detected (Program B
its levels were compared against the TOEFL IBT proficiency descriptors on which placement
decisions are based (Program C Handbook). Hence, prospective students are placed into a
level of instruction based on their TOEFL or in-house placement scores. Moreover, the first
two weeks are a period of provisional placement for all students. If a student is misplaced,
s/he is then reassigned to the proper level. Given that two of the study placement tests are
valid, it would be unusual if some Saudi ESL students are placed into lower than-expected
levels unless there are other external factors causing these persistent misplacements.
but not limited to test-takers’ potential low levels of language proficiency, lack of adequate
experience in taking language placement tests, and even fatigue when taking the test. In cases
when the classes are too easy or frustration when they are too challenging. Sometimes, these
misplaced students show neither boredom nor frustration to reflect their dissatisfaction with
misplacement. In these cases, one may assume that there are other factors causing
in placement tests, or any other test-related factors. To examine this issue prior to conducting
this study, anecdotal evidence was obtained and a pilot study was conducted from which
between a researcher and another” (Bourke, 2014, p. 2). It mainly deals with a researcher’s
interrelated fields. For purposes of full disclosure, the researcher’s nationality is the same as
the participants, Saudi Arabian. A researcher is expected to collect and analyze data as an
participants towards certain desired findings. The researcher’s physical, language, and
cultural access to the participants provided him with an emic position. Both the researcher
and participants are scholarship students sponsored by SACM to pursue their desired degrees.
This enabled him to meet with several Saudi students at the Saudi Student Club, weekend
The researcher’s emic advantage helped him obtain much information from the
participants, which might not have been accessible by researchers of other nationalities.
Although the participants appeared to be comfortable discussing this issue with him as
opposed discussing it with their ESL teachers or classmates, his objectivity was challenged
during some stages of data collection process. For example, during one interview, the
researcher discussed with a participant some issues about ESL placement tests in order to
allow the issue at focus, purposefully underperforming on ESL placement tests, to enter into
the conversation naturally. However, the participant kept discussing the accuracy of ESL
placement tests in detail, and the researcher repeatedly attempted to redirect the discussion to
the study issue. This may have impacted some of the participants’ responses and led them to
This study is a substantial endeavor in accounting for why some Middle Eastern,
especially Saudi, ESL students are placed into lower than-expected levels. In addition, based
on the literature, it examines students’ suggestions and ESL administrators’ views to provide
implications to address this issue effectively. According to the anecdotal evidence and pilot
study findings, failing an ESL placement test on purpose, for any reason, is likely to lead to
further negative consequences. Assuming that 40% of the approximately 10,000 Saudi
students, who are granted governmental scholarships on an annually basis to study abroad,
intentionally underscore on placement exams, then 4,000 students will fail their ESL
placement test. This will not only affect the outcomes of Saudi students studying abroad, but
A large number of misplaced students can pose greater problems for ESL teachers, as
there might be severe discrepancies among students (Hughes, 1989). Furthermore, although
we do not have accurate statistics of scholarship ESL students who were unable to obtain
university admission and thus returned to Saudi Arabia, this study will be helpful to
encourage not only Saudi but all ESL students to take the ESL placement tests seriously to
benefit from the ESL levels into which they are placed. As a consequence, the significance of
this study centers upon the hope that many Saudi scholarship administrators, ESL
administrators and teachers, and Saudi ESL students will find the results very important in
effects, and general language pedagogy. Therefore, this study aims to answer the following
key question, do ESL Saudi students purposefully underperform on ESL placement tests?
Based on the findings of this question, the study also investigates any negative consequences
2. What are some reasons that make students deliberately underperform on the English
placement tests?
3. To what extent does this behavior affect the placement system of ESL programs?
5. What are some possible implications that can contribute to resolving or mitigating this
issue/phenomenon?
Research Instruments
A mixed-methodology approach was used to collect data from the three programs
studied because an understudied issue is being investigated. The research tools are a 20-item
survey (Appendix B) and semi-structured interviews (Appendix C). The survey was designed
based on the findings of the anecdotal evidence and pilot study. The last item of the survey
asked the participants to provide their emails if they wished to participate in the interviews.
On the other hand, the semi-structured interviews, which consisted of ten questions, were
designed in a way that helps the participants expand more fully on their responses to the
survey in order to identify their motives behind deliberately underperforming on their ESL
placement tests. Moreover, the same survey was circulated, with some modifications, to a
wide range of GCC students studying in the U.S. through several platforms to test the initial
findings of the study. The last research tool used in this study was a six-item survey
(Appendix D), which was designed and sent out randomly to some ESL administrators and
assessment coordinators (n= 17) in order to gain their views on the study’s issue.
36
Participants
The participants of this study were divided into three categories (See Appendix E).
The first group consisted of 71 randomly selected Saudi ESL students (50 male and 21
female) attending three university ESL language programs in the southwestern United States.
Their levels of language proficiency were identified in the survey, which first asked students
about their current ESL program levels, range from beginner to advanced (Figure 1).
Although 127 students participated in the survey, only 71 of them completed the
entire survey, so the remaining participants were excluded from analysis. Regarding their
distribution across the three study programs, 25 participants are attending Program A, 36
attending Program B, 13 attending Program C, and one attending another (Table 1).
37
Table 1
Lower-
Upper- Advance Passed all
Answer Beginner intermediat Intermediate
intermediate d ESL Levels
e
Program A 3 0 5 5 8 4
Program B 5 2 7 8 8 3
Program B 3 1 1 1 1 6
Total 71 11 3 13 14 17 13
The second group of participants consisted of 216 GCC students studying in the
United States who received the same survey with some modifications. The third group
included 17 ESL administrators and assessment coordinators to obtain their attitudes towards
Procedures
Saudi students attending the target ESL programs. Unfortunately, the initial number of
participants was too small to draw any conclusions about the issue. Therefore, the researcher
asked some teachers from the study’s programs to encourage Saudi students to participate in
the survey. He also asked many Saudi students with whom he is acquainted to encourage
their friends or any other Saudi students they knew of who satisfied the parameters of the
focus population to take the survey. After three months, 127 survey responses were received,
71 of which were completed in their entirety. The responses were organized into three
themes: 1) general information about the participants, 2) the study’s main question (Have you
ever intentionally performed poorly on an ESL placement test?), and 3), suggestions for
38
As for the second round of data collection, those who provided their emails in the
survey were contacted and given the choice to take part in face-to-face, phone, or other voice
protocol interviews. Program B students were interviewed at their program site, whereas the
other students were interviewed via phone, Skype, and Tango2. Each interview session lasted
20 to 25 minutes. Due to cultural constraints, the researcher was unable to interview female
participants face-to-face; instead, the interviews were done via Whatsapp. All interviews
were carried out in English unless some participants found difficulty conveying their
messages, then Arabic was used. To make the findings generalizable to more ESL students
beyond the study population, the survey was sent to a number of GCC students studying in
the U.S. via Student Clubs, Whatsapp, and Twitter. Finally, a six-item survey was sent to a
random group of ESL program administrators to ask them about any impact that purposefully
underperforming on ESL placement tests may have on their program placement system.
FINDINGS
After being asked some introductory questions concerning their experiences in taking
ESL placement tests in order to avoid leading them to any desired findings (i.e., being
influenced by the wording of the study’s main question), the participants were directly asked
the following question: “Did you intentionally, in one way or another, perform poorly on the
ESL Placement Test to be placed in lower levels?” One noteworthy issue is that this key
question was introduced in both English and Arabic to ensure that all participants understood
it fully. The findings showed that 14 out of 71 (20%) of these Saudi participants reported that
they did deliberately perform poorly on an ESL placement test for various reasons (Table 2).
2
A social media platform
39
Table 2
# Answer Response %
20%
1 Yes 14
80%
2 No 57
Total 71 100%
Although 20% represents only half that of the pilot study, it is still high. However, it
on placement tests exist at this percentage. In other words, 20% of participants is likely to be
evidence, such as observing participants’ classroom performance right after taking the
placement tests. Another salient issue is that the researcher had no access to the placement
systems used by the three programs that participated in the study in order to examine these
findings. As a consequence, the researcher decided to support these findings with other data
collected from a larger number population that meets the parameters of the target population.
Therefore, the same survey was circulated to numerous GCC students studying in the U.S..
In this round of data collection, almost 336 GCC students responded to the survey, yet
only 216 of them completed the survey. Although the survey targeted GCC students across
the United States, the vast majority of the participants are Saudis (202), with only (6)
Emiratis, (7) Kuwaitis, and (1) Omani as indicated in Table 3 below. This can be attributed to
the researcher’s limited access to GCC students, except for Saudis. In other words, the
40
researcher contacted Saudi students across the Unites States with whom he is acquainted and
Table 3
2 Emirati 4 2
3 Kuwaiti 4 3
4 Omani 1 0
5 Qatari 0 0
6 Saudi 153 49
Total 162 54
The findings revealed that 39 out of 216, or 18% of the GCC participants indicated
that they did intentionally underperform on ESL placement tests (Table 4). Although these
percentages (20% & 18%) represent around half of that of the pilot study (40%), they are still
high and concerning for scholarship agencies. Given that 10,000 Saudi scholarship students
are annually sent to study abroad, this means that 1800 to 2000 of them are likely to
deliberately perform poorly on ESL placement tests. These percentages can cost scholarship
agencies tens of millions of dollars, as students will take unnecessarily additional courses.
41
Table 4
Did you intentionally, in one way or another, perform poorly on the ESL Placement Test to be placed in lower levels?
Answer Response %
18%
Yes 39
82%
No 177
One may argue that these data are insufficient; hence, more evidence is still needed to
gain further insight into the issue. Therefore, a group of ESL administrators and assessment
coordinators were surveyed. They were first asked if they have ever noticed any case of
indicated that they had encountered some form of conscious attempts made by students to
perform poorly on ESL placement tests, yet with a very low percentage (<5%). More
specifically, five of these participants stated that the percentage of students intentionally
underperforming on the placement test is very low, probably around 1%, while two of them
indicated that it is likely less than 5%. In contrast, one of them indicated that she has
personally witnessed this behavior only twice in her entire 25-year ESL teaching experience.
Another participant stated that although she found some students deliberately
underperforming on ESL placement tests, she maintained that it is difficult to tell the rate.
Interestingly, only six of these participants indicated that they have never witnessed a student
42
Clearly, there is a large discrepancy between the responses of ESL administrators and
assessment coordinators and those of students. This poses another issue whether this behavior
exists or not. It is possible that some ESL administrators might be unaware of their students’
intentions to underperform on ESL placement tests, which could account for this discrepancy.
Of course, this argument makes sense if the study had a large pool of data collected through a
long-term period, or perhaps data were triangulated. Otherwise, the discrepancy between
these two sets of percentages may question the accuracy of these findings, especially since no
concrete evidence of students’ work was collected such as writing samples, classroom
observations, and subsequent tests. Nevertheless, even with these two sets of data, one may
wonder why these students would still intentionally perform poorly on ESL placement tests.
tests were asked about their motives. That is, if a participant selected no, indicating that s/he
has never intentionally performed poorly on ESL placement tests, the survey automatically
skipped to the next section. The participants’ provided different, yet interrelated, reasons
Table (5). Six of the 14 participants indicated that they had failed placements test
purposefully in order to a have longer time to practice English, as the more ESL classes they
take, the more exposure to English they will have. Moreover, five of them indicated that they
had placed themselves into lower ESL levels so that they could have less effort in carrying
out ESL tasks, which would allow them to focus more on preparing for admission-required
43
Table 5
Responses
Answers
6
To have more time for learning English.
Other reasons. 1
Moreover, two participants justified their reasons by stating that this would help them
spend extra time in the ESL program to have more time to become fully familiar with the
American educational system. However, only one participant ascribed his intentions to “other
reasons”, without providing additional information in the subsequent question. Having seen
these findings, one may consider students’ justifications logical to some extent, provided this
behavior does not affect their current or future learning. However, some qualitative data is
placement test argued that if they were placed into classes reflecting their actual levels of
complete the English language program in a shorter period of time. In such cases, they are
44
required per SACM policies to obtain immediate admission to a university or return to Saudi
Arabia. These two options are mandatory because once a student completes his/her ESL
classes, their I20 for language studies is cancelled. If a student decides to intentionally
underperform on an ESL the placement test to be placed into a lower level, this will provide
them more time to: 1) learn English, 2) prepare for standardized tests, and 3) correspond with
or even visit several universities before choosing one. Moreover, three of the participants
contend that being placed into a lower ESL level allowed them to start learning English from
scratch.
test? Surprisingly, eight of them indicated that their friends who had taken ESL placement
tests prior encouraged them to fail the test, suggesting that more focus be given to obtaining
university admission. That is, their friends told them that the higher the ESL level, the more
demanding the class assignments will be. This may not allow them to have sufficient free
time to prepare for the TOEFL, IELTS, or GRE. On the other hand, four participants stated
that no one told them to purposefully underperform on the test; rather, this decision was
based on personal motives. Another interesting finding is that two of these participants
placement test; however, shortly before taking the test, some students with whom they took
the test encouraged them to fail it. As for GCC students, many indicated that they paid more
This brings the researcher to another question. If those students had no scholarships,
would they still have intentionally performed poorly on their ESL placement tests? Twelve
out of 14 participants indicated that they would not, under any circumstances, have
45
purposefully underperformed on the ESL placement tests if they had not been granted a
scholarship, as they are unable to afford paying the tuition fees. Nonetheless, only two
participants insisted that they would still fail the test without having a scholarship. Hence,
this suggests that since their tuition fees are paid by SACM, these students are not concerned
with any financial costs resulting from taking unnecessary, additional ESL classes. In
contrast, the findings of GCC participants were considerably consistent with the above
mentioned in that 26 out of the 39 participants, or 67%, indicated they would not
intentionally underperform on a placement test if they were responsible for paying tuition
fees.
ESL placement tests, one may wonder to what extent this issue affects the placement system
replacement rates or any other similar statistics for the three programs involved in the study.
Alternatively, some ESL administrators and assessment coordinators were asked about the
potential impact of having 18 to 20% of their students misplaced in their placement systems.
The findings revealed that only one participant indicated that these percentages would not
impact the entire program but would only impact these particular students’ placements. She
argued that her program has five different measures for placement, including a face-to-face
interview, which makes it difficult for students to consistently underperform on all of the
assessments deliberately.
On the other hand, the remaining 16 ESL administrators and assessment coordinators
expressed their concerns with such high percentages. For example, 13 pointed out that these
46
percentages would be problematic especially if there were several scholarship students
intentionally underperforming on their ESL placement tests. For example, one of these ESL
administrators said, “We would need to move them to the appropriate section and would
possibly need to add a class/take away a class; this would cause changes in the teachers'
assignments, groups of students in each class, etc.” (personal communication, February 11,
2016). This would therefore cause additional work and frustration for all involved.
Furthermore, two participants argued that such high percentages would lead to decreasing
student motivations for the self-misplaced students would not display the proper motivation
and work ethic needed in the classes they attended. In other words, studying with misplaced
Another participant indicated that such behavior would change the entire classroom
dynamic and could ultimately affect other students' perception of the seriousness of the class.
In addition, another participant stated that if the content validity of their placement test was
very high, having scholarship students deliberately underperforming on ESL placement tests
would be a serious issue. For example, she pointed out that her program’s placement test was
mostly abstracted from their program curriculum in order to ensure high content validity.
poorly on ESL placement tests, this would “create level drift in our program” (personal
communication, February 10, 2016). In addition, the remaining participants argued that if
they had 18% self-misplaced students, they would take immediate action to reexamine the
that if this issue exists at these rates, this will create a great challenge for ESL practitioners.
47
Purposefully Underperforming on ESL Placement Tests: A Major or Minor Issue?
One of the key research questions of this study is the extent to which intentionally
underperforming on ESL placement tests is a serious or minor issue. Now that multiple sets
of evidence (i.e., anecdotes, pilot study, Saudi students, GCC students, and ESL
administrators) have been provided, one can, at least based on this study’s population, draw
some conclusions about the study’s issues. One issue is the discrepancy between students’
reported rates (18 to 20%) and those provided by ESL administrators (5% at most). That is,
one may argue that having 20% out of 71 and 18% out of 216 participants claim that they had
position are likely to consider student assertion to be strong evidence that this behavior exists,
collected two weeks past the placement test date, observing students’ performance, or
monitoring the study program placement systems, it is difficult to consider this behavior as
either widespread or detrimental to the students and/or the programs in which they are
enrolled. In fact, both arguments can be valid for two reasons. First, the discrepancy between
the students’ reported rates and those of ESL administrators does not necessarily mean that
the behavior is a minor issue simply because students are unlikely to inform their teachers of
their intentions to fail a test. In other words, these participants find no plausible reasons for
telling ESL administrators that they deliberately underperformed on ESL placement tests.
48
Second, the other argument seems also logical since drawing generalizable conclusions of
As a compromise between these two positions, the Saudi researcher, who was a
previous ESL student and is currently an EFL teacher, contends that although the issue at
hand exists based on the study data, such data are inadequate to consider it a common
phenomenon. This requires a much larger number of participants to corroborate the data even
though the issue still occurs in several different contexts. For example, during the writing of
this study, the researcher discovered four other cases of Saudi students intentionally
underperforming on ESL placement tests that were not included into this data set; however,
documenting them poses another perplexing problem, student dishonesty. For example, one
of the study’s three ESL programs students stated that although he had completed an
intensive English course at ARAMCO and had obtained a TOEFL score that meets admission
requirements, he decided to fail the test unlike his classmates, who took it seriously and were
subsequently placed in inter-mediate ESL levels. He did this because he plans to prepare for
Intensive English Programs (IEP), the total number of Saudi IEP students enrolled in U.S.
programs reached 32,557, ranking them top on the list of places of origin of IEP students in
the country in 2014. Despite this large number, some Saudi ESL students are placed into
lower than-expected levels. This can be attributed to numerous factors such as students’ low
levels of language proficiency, students’ lack of adequate exposure to the target language,
and so forth. However, based initial anecdotal evidence and pilot study, the findings indicated
that 11 out of 27 (40%) Saudi ESL students pointed out that they had intentionally
49
underperformed on university ESL placement tests for several reasons. Both the anecdotal
evidence and pilot study helped the researcher to rationalize the study issue.
Surprised by these findings, the researcher decided to explore more tangible evidence
encountered difficulty locating resources addressing the topic, so he decided to collect more
data from a larger population. The first set of data showed that 14 out of 71 (20%) of Saudi
participants reported that they did purposefully perform poorly on ESL placement tests. To
examine the accuracy of this percentage, 216 GCC ESL students were surveyed about the
same issue of which 39 of them (18%) indicated that they had intentionally underperformed
on ESL placement tests. In order to gain further insight into this issue, 17 ESL administrators
and assessment coordinators were surveyed. Ten out of 14 participants indicated that they had
seen some form of conscious attempts made by some ESL students to deliberately
by some ESL administrators and assessment coordinators. In other words, although the
findings of the pilot study, Saudi ESL students, and GCC students were high (40%, 20%, and
18% respectively), the ESL administrators reported significantly lower percentages. This
discrepancy between the percentages can be ascribed to numerous reasons. First, the behavior
may indeed exist at these high percentages, but the self-misplaced students did not inform
ESL administrators of their motives. Second, these percentages might be exaggerated due to
impacts caused by some of the survey’s statement wording, which may have led students to
desired findings. Such discrepancies suggest that further research studies be conducted in
order to recruit more participants, obtain concrete evidence of participants’ work collected
50
two weeks past the placement test date, and monitor placement systems in order to make data
more valid.
Although the study does not have evidence besides students’ reports to have
purposefully underperformed on ESL placement tests, such findings suggest that immediate
action be taken by SACM for several reasons. First, if such behavior exists at these rates, it
would cost scholarship agencies additional money without gaining the desired outcomes.
Second, this behavior is likely to lead students to not take ESL classes seriously because they
are taking ESL levels far below their English proficiency level in order to focus more time
and energy for obtaining university admission. Third, these students might not improve their
English skills by attending lower ESL levels. Fourth, some ESL programs offer English
language endorsements by which the minimum TOEFL score requirements for university
entry can be waived. In such cases, these students might delay this chance, as they have
placed themselves in lower levels, and it will take longer to reach the higher levels that allow
can be concluded that deliberately underperforming on ESL placement tests indeed exists and
needs to be addressed. Therefore, this study proposes a number of concrete implications that
will help mitigate or even eradicate the problem. Prior to implementing any of the subsequent
implications, it is highly recommended that SACM, the Saudi Ministry of Education, or any
other scholarship agencies survey their students to identify their language needs upon which
the implications can be implemented. In other words, it might not be effective to take any
action regarding this problem without carefully examining students’ motives behind
deliberately underperforming on ESL placement tests. For example, students should be first
asked about their experiences in taking ESL classes in general and ESL placement tests in
51
particular in addition to any issues they confronted during their ESL experiences. This will
certainly help provide concrete implications based on students’ real issues pertaining to ESL
placement tests.
LIMITATIONS
One of the primary limitations of this study is that it lacks supporting data with
concrete evidence of students’ language work. In other words, one may assume that it may be
insufficient to ask ESL students if they have ever deliberately underperformed on an ESL
placement test, but rather data should have also included samples of students’ classroom
work (e.g., writing samples). This would enable us to compare participants’ responses to their
responses to the survey and interview, provided that these samples were written within a
short period of time of taking the ESL placement tests. Unfortunately, the researcher was
unable to obtain students’ language work due to fear of placing substantial burden on
participants to take part in the study. As a consequence, this option was excluded from the
study. To overcome this limitation, a larger number of participants (GCC students) were
Another limitation of this study is concerned with the wording of the survey
questions, especially that of the study’s main question, “Did you intentionally, in one way or
another, perform poorly on the ESL Placement Test to be placed in lower levels?” One may
argue that participants should not have been asked directly if they have ever purposefully
underperformed on an ESL placement test since such wording may lead them to the
researcher’s desired findings. Alternatively, the researcher should have let the participants
talk about their experiences so that any naturally arising issues pertaining to the study’s
targeted behavior could be documented and further explored. However, this strategy, albeit
asking the participants about this behavior. Instead, s/he should find another effective strategy
Another limitation is that one of the survey questions asked participants about the
reasons why they intentionally performed poorly on the ESL placement tests. They were
given a close-ended format of options consisting of four main reasons accounting for their
Although the participants were allowed to select all choices that applied, the wording of this
question, in addition to the limited choices, could still make participants’ responses very
restrictive. Moreover, giving limited choices might implant some ideas that push the
participants towards certain preferred answers. To address this limitation, the participants
were provided with a subsequent question, “If you have any other reasons, please mention
them here,” in order to allow them to provide reasons other than those listed.
In addition to the aforementioned limitations, another limitation was found. That is,
the researcher’s objectivity was challenged at some of the data collection stages. For
example, in one of the interviews, the researcher asked a participant about some introductory
questions about ESL placement tests so that the participant could not be led to any desired
findings. However, the participant was off topic, discussing the accuracy of ESL placement
tests in detail. Thus, the researcher repeatedly attempted to redirect the discussion to the
study’s issue, which may have impacted some of his responses. For example, the researcher
53
asked the participant to assume he had no scholarship, would he still throw the test? The
participant seemed to be very hesitant to respond, and he did not mention that he had ever
deliberately underperformed any ESL placement test with having a scholarship. Hence, the
researcher felt that the participant was trying to find the answer that best fit the question.
IMPLICATIONS
Pedagogical Implications
Based on participants’ responses, one of the most intriguing implications of this study
is the notion that scholarship students should be enrolled in English programs in their home
country prior to studying abroad. There are several successful examples of this approach. For
example, through the College Degree Program for Non-Employees, that Saudi Aramco
Petroleum Company grants scholarships for qualified high school graduates to pursue
bachelor degrees in certain majors. Before studying abroad, the students are enrolled in a ten-
month language preparation program (Saudi Aramco College Preparatory Program), where
they are taught intensive English and some preparatory courses in their intended field of
study (Al Murshidi, 2014). After completing the program successfully, students are then
sponsored to study abroad. Despite the criticisms leveled at ARAMCO for having students
US” (Al Murshidi, 2014, p. 41), these preparatory programs have been very helpful for
This implication is consistent with the TESOL position statement that “It is important
that the sponsors either send students already at a high enough proficiency level to progress
sufficiently within the sponsor’s time limit or that the sponsors recognize that additional time
may be necessary for English language study” (TESOL, 2010, p. 2). Thus, it is recommended
that the Saudi Ministry of Education consider engaging scholarship students in intensive
54
English courses prior to studying abroad. If this implication is considered undesirable, Saudi
ESL students should then be given more time to learn English in the target country, which
will allow them adequate time to develop their English skills. Moreover, it will provide them
with sufficient time for preparing for admission-related tests such as the TOEFL, GRE, and
GMAT.
One of the survey statements asked students about what the Ministry of Education
should do to help their ESL learning experiences. The vast majority of the Saudi and GCC
students chose the abovementioned two suggestions, providing students with English
programs prior to studying abroad and extending ESL study period. However, one of the
proposed suggestions that many participants unexpectedly selected was that a certain TOEFL
or IELTS score should be required for obtaining a scholarship. At first, the researcher thought
that the participants had selected this suggestion because it had been introduced to them in a
item actually showed the contrary. That is, 12 out of 71 Saudis and 28 out of 216 GCC
students indicated since they found difficulty obtaining the minimum scores on the TOEFL or
IELTS required to meet admission requirements, it would be more useful if a specific score
was required prior to studying abroad. Although one may doubt these students’ willingness to
suggestion that could help them overcome some of the difficulties they face during their ESL
learning experiences. They suggested that before studying abroad, they should be provided
with preparation programs for standardized tests (e.g., TOEFL, IELTS, GRE, GMAT). For
example, some participants stated that they lack the basic test-taking skills needed for high-
55
stake standardized tests, as opposed to, for example, Japanese, Korean, and Singaporean high
school students who are required to take exit exams. Other participants stated that it is quite
difficult to focus on ESL program assignments and admission language requirements at the
same time. As a result, they suggested that one of them should be provided in Saudi Arabia
prior to studying abroad so that students can concentrate well on both of them upon arrival in
another, scholarship students should then be given mandatory exam preparation programs
about admission-related tests (e.g., TOEFL, GRE, GMAT). Although Saudi ESL students
have unlimited opportunities to take these tests at the expense of SACM, these exam prep
programs should be part of the scholarship requirements so that students will take them
seriously. Given that these programs are optional, some students indicated that they take them
to avoid visa issues rather than for developing their test taking skills. For example, one of the
participants pointed out that he had completed all ESL levels, so his I20 was about to expire.
To avoid having a hold placed on his scholarship, losing tuition fees and stipend, or worse,
being deported from the U.S., he took a GRE course to extend his I20.
Given that regular ESL placement decisions are not as accurate as desired (Hughes &
exacerbate this issue and cause dire consequences. To overcome this issue, substantive efforts
should be exerted to raise students’ awareness of the adverse impact of being placed into
improper ESL levels including but not limited to susceptibility to boredom, laziness,
goals (Banegas, 2013). The awareness-raising process should be systematic in a way that
56
informs students of the negative consequences of intentionally underperforming on placement
tests. For example, this issue can be included in scholarship regulations, addressed during
orientation programs in Saudi Arabia, or constantly reminded and sent to students’ emails by
their sponsors. In conclusion, future researchers are advised to make sure that students do not
report having purposefully underperformed on ESL placement tests because they do not want
to appear weak in their English skills. Moreover, they should take into account that some
students might refrain from participating in such a study, as it may question their honesty.
57
CHAPTER 3 INTEGRATING SELF-ASSESSMENT TECHNIQUES INTO L2
58
ABSTRACT
teacher” (Chen, 2008, p. 238). However, to promote students’ learning autonomy and
involvement in classroom activities, other alternative assessment forms have been introduced
assessment, and journal assessment (Brown, 2010; Shohamy, 2001). One of the more widely
Ekbatani & Pierson, 2000). Unfortunately, most EFL departments in Saudi Arabia lack
integrating alternative assessment techniques into assessment processes (Al Asmari, 2013;
As a result, this article aims to explore the use of alternative assessment forms in
English as a Second Language (ESL) and English as a Foreign Language (EFL) classroom
contexts. It examines the accuracy of ESL students’ self-rated scores using CEFR self-
assessment rubric compared to their recently obtained TOEFL scores. It also explores
whether gender and levels of language proficiency are major influential factors for causing
score underestimation. The participants of this study are 21 ESL students who are attending
the Intensive English Program (IEP) at the Center for English as a Second Language at the
University of Arizona in Tucson. Based on CESL’s class system, their levels of proficiency
are intermediate (n=5), upper-intermediate (n=8), and advanced (n=8). The findings revealed
no statistically significant correlation between the participants’ self-assessed scores and their
TOEFL scores. However, based on the qualitative data, the participants reported that they
find CEFR self-assessment rubric accurate in reflecting their levels of language proficiency.
The past three decades have witnessed a paradigm shift in L2 contexts from teacher-
to student-centered teaching (Benson, 2001, 2012; Brown, 2007; Ellis, 1994; Nunan, 1988;
Tarone & Yule, 1989; Tudor, 1993, 1996). This new pedagogical approach originated as a
reaction to several criticisms raised by some L2 researchers who had suggested that
substantive attention be paid to L2 learners rather to language forms and structures (Bachman
2000; Ellis, 1994; Huba & Freed, 2002; Nunan, 1988). This has led ESL practitioners and
involving L2 students in language learning processes (Benson, 2012, p. 33). Later, self-
directed learning was introduced into L2 learning contexts, as a stronger form of learner-
In parallel with this paradigm shift, other researchers have suggested that learner
involvement in language learning not be limited to learning aspects; rather, it should also
include engaging them in assessment processes (Ekbatani & Pierson, 2000). For example,
Nunan (1988) argues that to make a learner-centered curriculum more effective, “both
teachers and learners need to be involved in evaluation” (p.116). Thus, several alternative
assessment approaches have been widely used in ESL contexts. Among these approaches is
learner’s self-assessment, which refers to involving learners in making decisions about their
own learning. The effectiveness of self-assessment has long been discussed in the literature
(Cassidy, 2001; Ekbatani, 2011; Harris, 1997; LeBlanc & Painchaud, 1985; McNamara,
1995). Hence, this paper investigates the extent to which self-assessment, as a key element of
self-directed learning, can be an accurate and reliable measure for ESL and EFL learners to
identify and monitor their levels of language proficiency and serve as a supplementary
200). Similarly, O'Malley and Pierce (1996) define self-assessment as “an appraisal by an
individual of his or her own work or learning process” (p. 240). On the other hand, Andrade
a process of formative assessment during which students reflect on and evaluate the
quality of their work and learning, judge the degree to which they reflect explicitly
stated goals or criteria, identify strengths or weaknesses in their work and revise
Self-assessment can be used for placement (Bachman, 2000), achievement (McDonald &
Boud, 2003), and diagnostic (Andrade & Du, 2007) purposes. Moreover, it can be used “to
detect changes and patterns of development over time” (Dörnyei, 2001, p. 194).
(Ekbatani, 2011; Harris, 1997; McNamara & Deane, 1995). Moreover, Brown and Hudson
(1998) point out that self-assessment facilitates student involvement in their own learning,
promote their learning responsibility, and increases their motivation. Furthermore, it helps
prevent any form of cheating (LeBlanc & Painchaud, 1985) and enables students to monitor
their learning (Harris 1997). However, there are some limitations of self-assessment one of
which is subjectivity in that students are likely to be “either too harsh on themselves or too
self-flattering” (Brown, 2001, p. 145). Moreover, Tara (2008) argues that when students have
inadequate experience in self-assessment, they may “judge their own work within their own
learning contexts and how early teaching methodologies have significantly influenced its use.
Before the 1970s, most if not all earlier second language teaching methodologies were
primarily concerned with analyzing language forms and structures (Brown, 2007). For
memorizing isolated words, and translating literary texts (Larsen-Freeman, 1986; Mitchell &
Vidal, 2001). This classical method, in addition to the audio-lingual method, was deemed to
neglect the learners’ role in language learning and merely considered them to be a passive
entity. As a consequence, an alternative approach was needed to empower the learners’ role
in language learning.
Later, Communicative Language Teaching (CLT) emerged in the U.S. and across
Kumaravadivelu, 2006; Nunan, 1988; Widdowson, 1978). This dynamic approach focuses on
all components of language so that learners have the opportunity to demonstrate their
language abilities more effectively and be more active role players in language learning
(Brown, 2007; Lee & VanPatten, 2003; Nunan, 2003). Benefits of the communicative
language teaching approach led several L2 researchers and theorists to conduct an extensive
number of research studies aimed at promoting the vital role of L2 learners in language
learning processes (Ellis, 2003). Forms of CLT such as task- and process-based have both
indicated that learners are evidently capable of taking charge of their own learning, while
teachers can act as facilitators and model/scaffold providers (Johnson, 2003; Nunan, 2004).
62
Learner-centeredness
During the 1980s, research stressed the need for learner-focused teaching, a
pedagogical approach that places substantive emphasis on “learners and learning in language
teaching, as opposed to a focus on language and instruction” (Benson, 2012, p. 30). Learner-
focused teaching was initially proposed through the work of Nunan (1988), Tarone and Yule
(1989), and Tudor (1993). Their contributions were followed by numerous research studies
curriculum” (Nunan, 1989), “needs analysis” (Richards, 2001), “learner training” (Wenden,
1995), and “learning styles” (Brown, 2007). Among these types of learner-centeredness is
Self-directed Learning
into L2 learning, a key issue concerning potential misconceptions between self-centered and
or preferences” (Benson, 2012, p. 33); nevertheless, it does not necessarily involve giving
students some control on their learning nor does it include obtaining their consultation about
that learning (Brown, 2007). On the other hand, learner-directed learning involves learners’
planning, monitoring, and evaluating of their learning (Garrison, 1997). Benefits of self-
directed learning include empowering learner autonomy (Benson 1995; McNamara & Deane,
1995), increasing learner motivation (Dörnyei, 1994), and promoting learner self-confidence
(Taylor, 1995), which have introduced self-assessment into language learning (Brown, 2010).
63
Self-assessment
a principle that was first coined by Henri Holec (1981). This principle stresses the importance
of learners’ ability to be accountable for their own learning including setting their learning
goals, monitoring their performance, taking learning decisions, and developing their
motivation, which all reflect the characteristics and values of self-assessment (McNamara &
Deane, 1995). Thus, self-assessment has dramatically influenced second language assessment
in many ways including contributing to placing L2 learners into proper levels, reporting their
potential areas of strengths and weaknesses, providing them with feedback, and assessing
their attitudes (Saito, 2003). However, in its initial uses, as Brown (2010) noted, self-
assessment was deemed as “an absurd reversal of politically correct power of relationships”,
contending that some L2 learners, especially novice learners, are unlikely to be capable of
reporting an accurate assessment of their own performance (p. 144). On the contrary, research
learning, which will be discussed thoroughly in this paper. However, what best accounts for
the efficacy of self-assessment instruments depends on examining how accurate they are
Accuracy of Self-assessment
in second language learning (Andrade & Valtcheva, 2009; Bachman & Palmer, 1989;
Blanche & Merino, 1989; Butler, 2010; Cassidy, 2001; Garrison, 1997; Jassen-van Dieten,
1989; LeBlanc & Painchaud, 1985; McDonald & Boud, 2003; Oscarson, 1997; Taras, 1995,
2001). Most of these studies were correlational-oriented, through which two forms of
assessment are compared in order to yield any potential correlation or lack of it.
64
Correlations
Drawing on Raasch’s (1979) study, von Elek (1982) argued that the validity of self-
assessment can be similar to or at least not substantially lower than that of traditional
assessment instruments. In the same vein, LeBlanc and Painchaud (1985) found evidence of a
English proficiency exam (r = .53). Likewise, in her pilot study, Jassen-van Dieten (1989)
second language though she focused only on grammar and reading skills. Results
demonstrated a higher correlation between the two instruments (ranging from .60 to .79). She
concluded that, out of 25 students, 21 were able to place themselves in the same levels into
In the same manner, Bachman and Palmer (1989) suggested that self-assessment “can
be reliable and valid measures of communicative language abilities” (p. 22), indicating that
their techniques had demonstrated unexpected high reliability. Moreover, Cassidy (2001)
reported higher correlations (r = .87 to .97) between students’ self-reported scores and their
actual SAT scores. In addition, in his meta-analysis validation study, Ross (1998) found that
self-assessment techniques could provide high validity, suggesting that “the degree of
experience learners bring to the self-assessment context influences the accuracy of the
product” (p. 16). However, he also noticed that the participants were more accurate in
assessing their receptive skills (listening and reading) than those of their productive skills
Discrepancies
Other studies have reported conflicting conclusions. For example, unlike her
aforementioned study that was limited to assessing students’ grammar and reading skills,
Janssen-van Dieten (1989) later conducted a study on all four language skills - listening,
65
speaking, reading, and writing - to obtain a broader perspective of any potential correlation
between the two instruments. She reported little relationship between self-assessment and
previously validated proficiency tests. Moreover, Wesche, Morrison, Ready and Pawley
(1990) found that self-assessment did not show any statistically significant correlation with
traditional tests. Furthermore, Pierce, Swain and Hart (1993) found weak correlations
French immersion programs. In addition, Wesche (1993) found that “placement via self-
Reasons accounting for the discrepancies between self- and traditional assessment
tools vary depending on certain factors including but certainly not limited to “the linguistic
skills and materials involved in the evaluations” (Blanche & Merino, 1989, p. 315). Other
factors may include the type of language tasks used in both instruments, the degree of task
assessment (Blanche, 1988; Coombe, 1992; Oscarson, 1989; as cited in Wolochuk, 2009).
Despite the fact that self-assessment has received heated discussions in the literature,
many researchers have proposed effective implications that help validate self-assessment
techniques. Harris (1997), for example, suggests that self-assessment rubrics be clearly
specified in advance so that students can pinpoint their areas of strengths and weaknesses.
Moreover, students should receive adequate training on using self-assessment rubrics in order
for them to gain insight into the target evaluation criteria (Taras, 2003). However, Brown and
Hudson (1998) argue that self-assessment might be more effective when used for research
rather for placement or diagnostic purposes since the former is unlikely to make learners
In 1996, the Council of Europe published the first version of the Common European
language standard that describes language proficiency “through a group of scales composed
levels: A1 and A2, B1 and B2, C1 and C2 (Weir, 2005, p. 281). These six levels fall into
three broad levels of proficiency: Basic User (A1, A2), Independent User (B1, B2), and
Proficient User (C1, C2) (See Appendix F). Each of these levels “attempt[s] to specify as full
a range of language knowledge, skills and use as possible” through which CEFR users can
identify their levels of language proficiency (Council of Europe, 2001, p. 7). As of 2014, the
website of Council of Europe indicates that CEFR language proficiency descriptors have
the Association of Language Testers in Europe (ALTE) developed and validated user-
oriented and performance-related scales anchored to CEFR levels “to establish a framework
of ‘key levels’ of language performance, within which exams can be objectively described”
(Council of Europe, 2001, p. 244). This led to developing the CEFR self-assessment scales
with six different levels. Each level consists of Listening, Reading, Spoken Interaction,
Spoken Production, Strategies, Language Quality, and Writing. According to the CEFR self-
assessment rubric, in order for CEFR users to reach a given level, they would need to respond
as ‘I-can’ to more than 80% of the given level’s ‘Can-do’ statements. That is, after reading
CEFR self-assessment rubric language descriptive tasks (statements), users should assess
their abilities to perform these tasks by ticking either ‘I-can-do’ or ‘I-can-not-do’. If a user’s
overall number of ticked ‘I-can-dos’ are 80% or above, this places him/her in the given level.
67
Validity of the CEFR Self-assessment Rubric
The validity of the CEFR self-assessment rubric has long been investigated in the
literature. For example, Alderson et al. (2004) contend that although CEFR itself provides
rather than a theory of development in listening and reading activities” (p. 3). In addition,
Huhta et al., (2002) argue that “the theoretical dimensions of people’s skills and language use
which CEFR discusses are on a very high level of abstraction.” (p. 133). In investigating the
validity of CEFR self-assessment used by refugees in Ireland, Little, Lazenby Simpson, and
O’Connor (2002) found that “evidence of the difficulty that learners encounter in using the
CEFR to maintain on-going reflective self-assessment suggests a need for more detailed
Moreover, Jones (2002) points out that “different people tend to understand ‘Can-do’
somewhat differently” (p. 181), thereby creating discrepancies in their self-rated scores. All
of the above raise legitimate concerns about the validity of CEFR self-assessment rubric.
On the contrary, Little (2006) argues that one of the advantages of the CEFR is its
ability “to bring curriculum, pedagogy and assessment into much closer interdependence than
has usually been the case” (p. 382). Despite what was reported by Huhta et al. (2002) above,
this multi-tasking ability engages CEFR users in ‘action-oriented’ language scenarios through
which they can demonstrate their language abilities (Council of Europe, 2001). Moreover,
drawn upon the (Basic Interpersonal Communication Skills) and (Cognitive Academic
Language Proficiency), North (2007) argues that the CEFR distinguishes between productive
and receptive language use by dividing the former into interaction and production, resulting
in “34 illustrative scales for listening, reading, oral production, written production, spoken
university EFL department in Saudi Arabia, has first-hand experience regarding the lack of
using alternative assessment techniques in many EFL departments in Saudi Arabia. Thus, this
that substantive efforts be exerted to use alternative assessments by not only L2 programs in
Saudi Arabia, but also in many other L2 contexts. The study involved 21 ESL students; 18 of
them are Saudis who represent a sample of a larger population of Saudi L2 learners.
The ultimate purpose of conducting this research was to promote incorporating self-
assessment techniques into EFL classroom assessment processes. Because some EFL
assessment processes, this paper aims to promote the use of alternative assessment
techniques, in particular self-assessment, not only in EFL contexts in Saudi Arabia, but also
in other L2 contexts. This paper examines the accuracy of ESL students’ self-rated scores
obtained by CEFR rubric compared to recently obtained TOEFL scores. Finally, the paper
will provide EFL departments in Saudi Arabia and other L2 practitioners with implications
Research Questions
1. Do ESL students’ self-rated scores correlate with their TOEFL scores? If not, why?
2. Are gender and levels of proficiency major influential factors for causing any potential
score underestimation?
A mixed-methodology approach was used to collect data for this study using three
main research tools. First, there was a web-based CEFR self-assessment rubric (Appendix B),
which has already been mapped onto the TOEFL proficiency levels. It would, however, be
difficult to use the whole rubric, for it contains 227 ‘I-can-do’ and ‘I-cannot-do’ CEFR
statements, which can be time-consuming to use in its entirety and possibly inapplicable in
some situations. Thus, only levels B1, B2 and C1, which are equivalent to the TOEFL
selected. The second research tool was semi-structured interviews conducted with those who
provided their emails to participate in the second round of data collection (Appendix H). The
third research tool was participants’ most recent TOEFL scores, which were obtained based
on student consent.
Participants
The participants of this study were 24 (14 male and 10 female) ESL students
attending CESL (Center for English as a Second Language), University of Arizona. Their
(n = 8). In addition, they were attending CESL levels 3 through 7 in a seven-level program.
Their nationalities, as shown in Table 6, are Saudi (n=18), Chinese (n=3), Mexican (n=2),
and Qatari (n=1). For confidentiality purposes, their names were coded by their initials,
which they were requested at the beginning of the study rubric. Three of the participants were
excluded from the analyses because they did not provide their TOEFL scores, nor did they
provide their emails, leaving 21 participants. Moreover, after contacting those who provided
their emails, ten out of these participants agreed to participate in the interviews in order to
allow them to expand more on their attitudes towards using CEFR self-assessment rubric.
70
Table 6
Background on Site
The Center for English as a Second Language (CESL) is a nationally accredited IEP
located at The University of Arizona, in Tucson, Arizona, USA (CESL, 2014). Its IEP
consists of seven levels; each level is comprised of an eight-week session leading to optional
bridge programs for either undergraduate or graduate students It is also one of the ESL
71
programs highly recommended by SACM (SACM, 2015). CESL provides three types of
programs. First, there are the English language programs, which include eleven programs:
ESL graduate and undergraduate bridge programs, evening programs, intensive English (full-
and part-time) programs, local portable classes, online programs, skill classes, skill intensive
workshops, a teen English program, and tutoring. Second, it provides two different teacher
training programs: general teacher training and Content Area Teacher Training (CATT)
Procedures
A randomly selected group of CESL students, who were soon going to take or had
just taken the TOEFL, were recruited. To encourage them to participate in the study, they
were offered a free one-time proofreading for one of their class papers (10 pages or less). The
recruitment flyer (Appendix I) was sent to these students’ emails by CESL the IEP
coordinator. The flyer contained links to the self-assessment rubric, which correlated to the
three levels of English proficiency that the participants could select (intermediate, upper-
intermediate, or advanced) to self-determine their current CESL levels. The participants were
also told that they should use the rubric two weeks at most prior to or right after taking the
TOEFL test in order to avoid any potential effect of learning progress between the study’s
two sessions. Next, they were asked for their permission to provide their TOEFL scores. The
first part of the rubric explained in detail the entire self-rating process in order to ensure that
After that, participants’ self-rated scores were re-examined to ensure that they
satisfied the study’s conditions. For example, those who did not provide their TOEFL scores
were contacted via email a month later because two weeks were needed as a maximum
timeframe between the two sessions, and two other weeks were required for TOEFL scores to
72
be released by ETS. Fortunately, sixteen participants had taken the TOEFL test on January
30, 31, or February 7, 2016, before they used CEFR self-assessment rubric in mid- and late-
February. Hence, they provided their TOEFL scores before they performed the self-
assessment rubric. As for those who did not provide their TOEFL scores, two of them
responded and provided their TOEFL scores, whereas the remaining three participants did not
respond and were thereby excluded from the study. Finally, those who provided their emails
in the first session were asked via email to participate in the semi-structured interviews. Two
days later, 10 out of 17 participants agreed to participate in the interviews through Skype,
email, or Tango.
Before analyzing the data, the CEFR, TOEFL, and IELTS equivalency table
developed by (Tannenbaum & Wylie, 2007) was used to compare participants’ self-rated
scores against their TOEFL scores (See Appendix F) At first, the researcher attempted to
color-code each level of the equivalency table; nevertheless, this strategy appeared to be very
confusing and even seemingly misleading (see Appendix J). As a result, given that the study
has only three levels of English proficiency (intermediate, upper-intermediate, advanced), the
Basic and Proficient levels were excluded from the study equivalency table in order to make
the table readable and more meaningful. Table 7 illustrates how the range of scores of the
TOEFL and IELTS are constructed to correspond to CEFR different levels of language
proficiency.
Table 7
since intermediate and upper-intermediate levels have 42 statements, their ratio is (36.6). In
the same vein, the advanced level has 37 statements; thus, its ratio is (29.6). After that,
and the self-rated scores of advanced participants were divided by (29.6) so that they can be
entered into ANOVA as ratios. One crucial issue is that the participants were not told that if
they have achieved 80% of ‘I-can-do’ statements, this would indicate that they met the level.
This was done to avoid any potential effects of participants focusing on reaching 80% instead
of rating their actual level of proficiency, which might impact their overall self-rating
performance.
FINDINGS
1. Do ESL students’ self-rated scores correlate with their TOEFL scores? If not, why?
Data were analyzed first using a two-factor ANOVA with participants’ gender (male,
factors (see Table 8). Participants’ TOEFL scores were mainly used as a covariate in each of
the analyses described below. The findings indicate that the participants’ self-rated scores did
not correlate with their TOEFL scores or with their levels of proficiency. For example, the
analysis showed that the gender effect was not significant F(1, 14) = .027, p = .87, and the
level of English proficiency was not significant either: F(2, 14) = 1.81, p = .335. This is
because there were only three participants (two intermediate and one upper-intermediate)
who had over 80% of ‘I-can-do’ statements, suggesting that they had probably reached these
levels.
74
Table 8
This statistical analysis did not, however, reveal the extent to which each participant’s
self-rated score was accurate. In other words, although the findings concluded that there is no
correlation between participants’ self-rated and TOEFL scores, they did not show how many
participants reached, were very close, close, somewhat close, far, or very far from reaching
the assigned ratio of 80%. To obtain a fuller explanation of these findings, a descriptive
Therefore, data were examined descriptively in order to gain further insight into these
findings. As discussed earlier, in order for participants’ self-rated scores to correlate with
their TOEFL scores, the participants should have at least 80% or more of the responses
selected in ‘I-can-do’ column to reach the level in which they participated. As shown in Table
9, only three out of 21 participants were able to obtain over the assigned ratio, achieving
100% (42/42), 85% (36/42), and 95% (40/42) respectively, suggesting that their self-rated
75
Table 9
Data of the Participants Who Reached or Were Very Close to Reaching the Assigned Ratio
Furthermore, the data showed that only three participants (4, 5, and 6) were very close
to 80%: 78% (33/33.6), 78% (29/29.6), and 75% (28/29.6). On the other hand, how close or
far the remaining self-rated scores from the assigned ratio (80%) varied depending on
participants’ self-rated performance. For example, they ranged from close (73%=31/33.6 and
far (40%=15/29.6 and 28%=12/33.6)3. These percentages suggest that participants’ self-rated
scores were not normally distributed. In other words, they do not reflect participants’
determined levels of language proficiency, nor are they consistent with participants’ TOEFL
scores. In order to account for these discrepancies, a randomly selected sample of the
The interviews were designed in a way that could help provide further insight into the
factors that could have potentially affected participants’ self-rated scores were included in the
3
For full details of each participant’s self-rated scores, look at Appendix (K)
76
interview questions. This included how many times have participants taken the TOEFL test
(test familiarity), if they ever used a self-assessment rubric before (lack of exposure), and
other questions that could help answer the other part of the first question (If not, why?). It
was found that three out of 10 participants had taken the TOEFL test for the first time,
whereas the remaining seven participants had taken the test more than two times. This
suggests that most of the participants have background knowledge about the TOEFL test,
Moreover, none of these 10 participants had ever used any self-assessment measures
before. This calls into question the extent to which participants’ lack of adequate exposure to
self-assessment rubrics may have led to lack of correlation between their self-rated TOEFL
scores. This is supported by the findings that most of the participants’ TOEFL scores were
Table 10
Advanced 79 – 95
Upper-intermediate 65 – 78
Intermediate 30 – 64
To identify why the majority of participant’s self-rated scores are not consistent with
their levels of language proficiency or TOEFL scores, the participants were also asked which
measure scores, TOEFL or self-rated, they believe to be more accurate. All 10 participants
argued that their self-rated scores are more accurate in reflecting their actual levels of English
proficiency; nevertheless, this still does not answer the second part of the first question.
77
To account for why 18 out of 21 participants obtained lower self-assessed scores,
three measures: TOEFL, level of proficiency, and self-assessment. Since this study did not
intend to collect such empirical evidence, the participants who were interviewed were asked
to what extent they consider their self-rated scores accurate. All 10 participants indicated that
they feel that their self-rated scores reflect their actual levels of English proficiency. After
that, each participant was provided with a table showing his or her level of proficiency,
TOEFL scores, and self-rated scores and was asked what s/he thought of them. Although all
of the ten participants, except one, obtained lower self-rated scores compared to their TOEFL
However, seven out of 10 participants indicated that they did not expect their self-
rated scores would be very low. After being shown their different scores, three of the
participants attributed their low self-rated scores to the fact they often become uncertain to
they decided to underestimate their scores. For example, one of the participants indicated that
once he read the survey instructions, he decided to assess his language abilities accurately
without any form of bias. This, therefore, has led some participants to underrate themselves.
Moreover, another participant argued that when the CEFR self-assessment rubric asked her
about language tasks that she once performed, she found no difficulty choosing ‘I-can-do’ or
‘I-can-not-do’ statements based on her previous background. Nevertheless, when asked about
language tasks that she has never encountered before, she spent some time visualizing herself
performing this task, and then selected ‘I-can-not-do’ statements. Another participant asserted
that when he wanted to select ‘I-can-do’, he remembered his reluctance to interact with native
78
speakers and ended up choosing ‘I-can-not-do’ statements, a CEFR problem that Weir (2005)
Having indicated that none of them had ever used a self-assessment rubric for
proficiency purposes, the participants were asked about any other factors that have could
have potentially impacted their self-rated performance. Although the self-assessment task
was explained to them in writing, the participants pointed out that they encountered some
difficulties using the CEFR self-assessment rubric, especially during the first parts of the
rubric. This suggests that engaging students in performing self-assessment tasks might not be
as effective unless the tasks are demonstrated for them through modeling or scaffolding
(Tara, 2003). For example, McDonald and Boud (2003) divided their study participants into
experimental and treatment groups. The former received formal training on using self-
assessment rubrics, while the latter received no training. The findings revealed that the
experimental group “outperformed their peers who had been exposed to teaching without
such training in all curriculum areas” (McDonald & Boud, 2003, p. 217).
Another interesting finding accounting for participants’ low self-rated scores was the
ambiguity of some statements that some participants faced. For example, one participant
argued that during responding to some of the statements, she found difficulty understanding
the given language task fully. Another participant pointed out that some statements have
unknown or confusing words (especially verbs), which makes it difficult for them to respond
to the statement accurately. This is consistent with previous literature that urges CEFT self-
assessment rubric developers to take the synonyms they use into considerable account. For
example, Alderson et al., (2004) argue that B2 level has several different verbs are used
locate, monitor, identify, and so forth. The researchers express some concerns whether these
79
verbs are simply “stylistic synonyms” or they indicate “real differences in cognitive
processes” (p. 9). During the interviews, one intermediate and two upper-intermediate
participants indicated that they were unable to identify the meanings of some verbs.
2. Do gender and level of English proficiency cause any form of score underestimation?
Given that the number of participants was relatively small, the two main effects
(gender and levels of proficiency) were not significant, and that only three participants were
able to do more than 80% of the ‘Can-do’ statements, it would be rather difficult to run an
analysis of variance to identify whether gender and levels of proficiency have a substantial
impact on CEFR users’ score underestimation. However, the only statistical analysis that can
help provide some yet inadequate, explanation about any potential association between these
two factors and CEFR users’ score underestimation is by using a scatter plot.
Figure 2. Scatterplot (BIVAR)=TOEFL with Ratio by Level of Proficiency for Male and
Female.
As shown in Graph 1, the overall pattern of the plot shows that participants’ gender
and levels of proficiency scattered randomly, as the correlation is very weak. For example,
80
completely different from each other. On the other hand, although the correlation of advanced
participants’ scores is also weak, they are less randomly distributed, suggesting that the more
advanced participants obtain higher TOEFL scores, they more their self-rated scores will
increase. As for gender, the findings suggest that the more female participants obtain higher
TOEFL scores, the more their self-rated scores increase. However, this does not help draw a
clear conclusion about the impact of these two variables on score underestimation.
The other option was to examine data descriptively to identify any potential effect of
gender and levels of proficiency in causing score underestimation. To do so, a cutoff score
should be first established indicating that a user’s self-rated score is underrated compared to
his/her TOEFL scores. For example, if a participant placed herself in advanced level but her
what is the cutoff score that suggests that this user should be in a lower level (e.g. B2 or B1).
To the best of the researcher’s knowledge, no cutoff score has ever been identified to indicate
the lowest performance in CEFR self-assessment rubric as opposed to the 80%. Thus, due to
the lack of a previous established minimum cutoff score in addition to not having internal
reliability estimated by Cronbach's alpha, this option was not pursued further.
The following table contains the participants who achieved less than 60%, taking into
account that this is a random rather a cutoff percentage. It can be noticed that there are ten
participants who were either far (59% - 41%) or very far (40%28%) from the assigned ratio
(80%). Even if all external variables such as test anxiety or inadequate experience in using
self-assessment rubrics are controlled, the data in Table 11 still do not reveal any pattern
indicating that gender and levels of proficiency cause any form of score underestimation.
Therefore, it can be concluded that because the aforementioned approaches did not reveal
clear findings, further data are needed so that valid conclusions can be drawn.
81
Table 11
Finally, the participants were asked about their attitudes towards using the CEFR self-
assessment rubric. In addition to what have been discussed earlier, the participants provided a
wide range of interesting insights from which some implications can be drawn. All ten
participants consensually indicated that they found the CEFR self-assessment rubric more
effective in determining their levels of English proficiency. For example, two participants
argued that the rubric enabled them to pinpoint both their strengths and weaknesses. Another
participant asserted that the rubric allowed him to assess his abilities to demonstrate his
another participant pointed out that one of the advantages of the CEFR self-assessment rubric
82
is that it does not only focus on users’ four language skills, but it also includes other skills
including interaction, strategies, and language quality. Such skills, as another participant
noted, helped her visualize herself perform certain tasks inside and outside the classroom.
However, in addition to the limitations discussed at the beginning of this study, some
participants also provided some drawbacks of using self-assessment rubrics. For example,
one participant, as indicated earlier, argued that some of the rubric statements are
dichotomous, which makes her uncertain which ‘Can-do’ choice she should select. Moreover,
another participant indicated that using the rubric for purposes other than grades made her in
some cases do not take the assessment seriously. Furthermore, two intermediate participants
provided interesting comments in that although they can perform some of the rubric language
tasks, they often found themselves not motivated enough to select ‘I-can-do’ choice. In
addition and more importantly, the participants complained about the difficult level of some
words used in the rubric statements. For example, one of the participants argued that he
responded to some prompts without fully understanding the task. Yet, all of the participants
concluded that they are interested in using the CEFR self-assessment in classroom settings.
Self-assessment has long been used in L2 learning contexts for various different
purposes (Ekbatani, 2011; Cassidy, 2001; Harris, 1997; LeBlanc & Painchaud, 1985;
McNamara, 1995). However, the extent to which self-assessment rubrics correlate with
traditional measures is still controversial. This study explored the experience of 21 ESL
participants in using the CEFR self-assessment rubric in order to identify whether their self-
rated scores correlate with their TOEFL scores, whether gender and levels of proficiency play
an influential role in causing score underestimation, and the possibility of incorporating this
rubric into L2 classroom assessment procedures. The findings indicated that 18 out of 21
83
participants were unable to rate themselves to the level reflecting their TOEFL scores and
CESL levels, even though the latter were found to be consistent. Moreover, the small number
of participants, lack of significance of the two main effects, and the fact that only three
participants were able to rate themselves to the given levels did not help obtain adequate data
to conclude whether gender and level of proficiency have major effects in score
underestimation.
Nevertheless, based on their responses during the interviews, ten participants showed
great enthusiasm and high motivation to use self-assessment rubrics in L2 learning settings.
In addition, they provided valuable insights about their experience in using the CEFR
assessment rubric, which can be very beneficial for L2 practitioners for promoting the
that self-assessment rubrics can be very effective if certain conditions are met. For example,
students should be trained on using these rubrics efficiently in order to avoid any potential
discrepancies between their self-rated scores and their actual levels of proficiency. Moreover,
TOEFL-related Limitations
One of the main limitations of this study was that the participants were not asked if
their reported TOEFL scores were obtained the first time they took the test. This is important
because familiarity and more practice typically increase test scores, and we are not certain
about the correlation between the participants’ TOEFL scores and their level of proficiency.
Although seven out of 10 participants who were interviewed indicated that they have taken
the TOEFL test more than two times, this does not necessarily mean that their TOEFL scores
were not impacted by any external factors. For example, test anxiety is a critical factor that
84
affects test-takers’ performance on any test, especially high-stake tests. That is, when taking
the self-assessment rubric, it is unlikely that the subjects had high-test anxiety, as the task
they were performing was for research rather for grade purposes. Thus, concurrent validity is
needed.
Another limitation is concerned with the limited number of participants (21), which
calls into question the appropriateness of drawing valid and generalizable conclusions about
the accuracy and effectiveness of self-assessment rubrics. In addition to the limited number of
participants, only three out of 21 participants were able to select more than 80% of the ‘I-can-
do’ statements, a problematic issue that which poses critical questions. Moreover, in order to
gain valid and generalizable findings, the participants should have been divided into two
groups: an experimental group, which receives training on the CEFR self-assessment rubric,
and a treatment group, which receives no training. This would certainly help draw more
accurate conclusions. Furthermore, time constraints are also another limitation of this study.
That is, the participants were not given certain time constraints to complete the CEFR self-
assessment rubric. This is unlike many assessment measures where time allotment is one of
ESL program administrators should first compare their program levels to a previously
validated measure (e.g., TOEFL, IELTS, ITEP) against which students’ self-rated scores can
tools, self-assessment rubrics, achievement tests) can be observed and compared. It is also
procedures from early levels so that students’ progress throughout the subsequent levels can
85
be effectively monitored. This will help ESL practitioners and self-assessment researchers
identify the extent to which level of language proficiency impact users’ self-rated scores.
One of the key implications for ESL/EFL teachers is training students on using self-
assessment rubrics. Teachers can start by having students use a small version of any self-
assessment measure to rate their performance on a certain language task. As a result, with the
continuous use of self-assessment rubrics, students will be able to identify their areas of
language strengths and weaknesses in that the more they practice using self-assessment
rubrics the more they will become accurate in identifying their levels of proficiency.
Moreover, the CEFR ‘Can-do’ rubric is highly recommended to be used for a pre- post
Future researchers are advised to recruit a larger number of participants so that they
can be able to increase statistical power. Moreover, they should also divide participants into
two main groups: 1) an experimental group and 2) a treatment group in order to explore the
can look at participants’ language skill subsets (reading, listening, speaking, writing).
86
CHAPTER 4 QUALITY ASSURANCE AND ACCREDITATION AS FORMS FOR
87
ABSTRACT
In 2004, the Saudi Ministry of Higher Education, currently named the Ministry of
Education, established the National Commission for Academic Accreditation and Assessment
(NCAAA) to ensure that Saudi higher institutions adhere to predetermined national standards
education. As a result, many programs in Saudi Arabia seek quality in order to obtain
academic accreditation. As an initial step to assure quality, this paper provides a simulated
University throughout the paper. The evaluation processes were based on a review using a
combination of integrated standards of the NCAAA and CEA (the Commission on English
Language Program Accreditation). Data were collected qualitatively through surveys and
interviews with students, faculty, and quality assurance coordinators. Moreover, the mission,
The findings indicated that the two EFL departments appear to partially meet the
standards based on their the mission, curriculum, student learning outcomes (SLOs), and
program development, planning, and review, whereas the standards for teaching strategies,
assessment methods, and student achievement were not met. At the end of the paper, some
concrete suggestions for improvement are provided that should help the two departments
address areas of weaknesses in order to be ready to obtain academic accreditation from CEA
Key words: NCAAA, CEA, quality assurance, academic accreditation, program review,
88
INTRODUCTION
Brown (2007) points out that “no curriculum should be considered complete without
some form of program evaluation” that corresponds with its changing conditions and
emerging developments (p. 158). Research suggests that academic programs be evaluated on
a constant basis to ensure that their learning outcomes are achieved effectively (Ellis, 1993;
Halliday, 1994; Mercado, 2012; Royse et al., 2009). Pennington and Hoekje (2010) point out
that program evaluation can be undertaken on a formal and informal basis, where the former
may include external review and accreditation, and the latter may involve ongoing internal
quality control. Royse, Thyer, and Padgett (2009) argue that program evaluation is
attempt to build theory” (p. 2). In other words, it is a process of evaluating a program using
1998; Mercado, 2012; Stensaker & Rosa, 2007; Van Damme, 2004; Westerheijden) and
developing countries (Sallis, 2002; Shawer, 2013; Smith, & Abouammoh, 2013). The
relationship between program evaluation on one side and quality assurance and accreditation
on the other has long been recognized in higher education contexts as one of the key
impetuses for program evaluation that “takes the form of accreditation-mandated student
learning outcomes assessment” and quality control (Norris, 2016, p. 173). As a result and in a
response to the rapidly growing quality assurance and accreditation interests, this paper
evaluates quality assurance processes at two EFL departments in a Saudi Arabian university
89
using an integrated set of accreditation standards of NCAAA and CEA to obtain accreditation
program evaluation. From the early 1980s to the present, these definitions have revolved
around one central theme even though other considerations have been suggested for the term
program evaluation. For example, Palmer (1992, p. 144) defines program evaluation as a
process of finding out whether a program is “feasible” in terms of its curriculum (practicality
issue), “productive” in terms of producing the intended learning outcomes (validity issue),
success of a specific program” (p. 149). In a broader view, this investigation should involve
examining the efficiency of “the individual components of [any] program in relation to each
other and to contextual factors, goals, criteria of value” (Pennington & Hoekje, 2010, p. 262).
performance.
Allen (2004) defines program evaluation as an ongoing process “for focusing faculty
objectives, curricular organization, pedagogy, and student development” (p. 4). In this study,
the main focus is on the program evaluation form that leads to maintaining quality assurance
and thereby obtaining academic accreditation. This is because program evaluation processes
in Saudi Arabia have recently received substantial attention for accreditation purposes. As a
result, this paper adopted Pennington and Hoekje (2010)’s comprehensive definition of
LITERATURE REVIEW
Given that applied linguistics is a field that began in the 1940s (Kaplan, 2010),
program evaluation literature “within the field is quite scant” as well (Lynch, 1996, p. 12).
However, research has documented some evidence of early systematic evaluations in public
education. For example, the first documented program evaluation dates back to the 19th
century, when many federally-funded schools in the United States, Great Britain, and
Australia began evaluating school curricula (Rea-Dickins, 1994). The premise underpinning
the evaluation processes undertaken during that era was based on evaluating curricula
“through scrutiny of the competence and behavior of the teacher” (Kiely & Rea-Dickins,
2005, p. 18). Then, during the late 19th century, programs were evaluated quantitatively based
on data (e.g., student retention rates, learning outcomes) from a positivistic view (Gitlin &
Smyth, 1989). Nevertheless, program evaluation during these eras remained recognized as a
during the 1960s and early 1970s (Beretta, 1992; Cronbach, 1963, Keating, 1963; Lynch,
1990; Sherer & Wertheimer, 1964; Smith, 1970). According to Lynch (1996), the emergence
of program evaluation field was due to some researchers’ calls for more systematic
approaches to program evaluation that could provide us with “what counts as evidence” (p.
9). In other words, program evaluation should be used as an effective tool for providing
concrete evidence of the appropriateness and efficiency of teaching and learning of the
administrators to provide them with some statistical information as proof indicating that a
91
program indeed achieves the desired learning outcomes that will allow it to continue to
There was a plethora of studies on foreign language program evaluation from the
1960s through the 1990s (Brown, 1995). According to Beretta (1992), until the late 1980s,
most of the studies and discussions on program evaluation were limited to research papers,
and there were very few books addressing this area. This was, as Beretta (1992) puts it, due to
“the seemingly never-ending” quantitative versus qualitative research-method debate (p. 5).
Beretta (1992) reviewed most, if not all, of these studies, and the extent to which they
contributed to shaping the modern field of program evaluation. For example, one of the early
studies on foreign language program evaluation, conducted by Keating (1963), examined the
Similarly, Campbell and Stanely (1963) conducted the first true- and quasi-experimental
study on L2 program evaluation, which provoked the research-method debate for a long time.
During the 1970s, several program evaluation studies were conducted (e.g., Asher et
al. 1974; Bushman & Madsen, 1976; Gary, 1975; Levin, 1972; Postovsky, 1974; Smith,
1970; as cited in Beretta, 1992). These studies focused on manipulating the teacher variable
so that variations could be minimized. This variable was controlled by comparing teachers’
performance from different programs (Smith, 1970), comparing the performance of two
program taught by the same teacher (Bushman & Madsen, 1976; Postovsky, 1974), and
finally “[eliminating] teachers altogether and [replacing] them with tape-recorded lessons”
(Beretta, 1992, p. 11). During the 1980s, Tyler’s Model of Evaluation, developed in the
1940’s, was again used by program evaluators by comparing the desired outcomes with
92
achieved outcomes (Beretta & Davies, 1985; Prabhu, 1987; Wagner & Tilney, 1983; Wolfe
studies addressing many areas of program evaluation. For example, Beretta (1992) examined
‘program-fairness’ of evaluation and concluded that there was a gap between program
evaluation theory and practice that needed to be addressed. Drawing upon this, Lunch (1996)
suggested that data-gathering techniques “provide information that makes sense and counts as
evidence” of program performance (p. 155). To achieve these two implications, standards of
the Joint Committee on Standards for Educational Evaluation (JCSEE) were reviewed in the
mid-1990s, which led to innovative and standard-based program evaluation (Patton, 1997). In
the late 1990s, several researchers advocated the incorporation of political, affective, and
cultural aspects into program evaluation processes, for language learning involves all these
factors into program evaluation, which can be found in Pennington’s (1998) work that studied
dynamic and interactive dimensions of language program evaluation than solely focusing on
quantitative data.
At the beginning of the 21st century, the epistemological and methodological tenets of
program evaluation have changed due to the impact of a wide range of political, economic,
and technological changes (Kiely & Rea-Dickins, 2005). This has resulted in spurring the
demand for compliance with mandates as a basis for language program evaluation.
Evaluation approaches such as self-study, peer review, key performance indicators, and
external audits have become key to verifying program’s compliance with mandates (Kiely,
2001), which have led to conducting language program evaluation for meeting “internal and
albeit not new, has made language program evaluation key for accreditation purposes (Eaton
Evaluation Paradigms
makes it crucial at this point to understand how programs have been evaluated. Lynch (1996)
paradigmatic camps: 1) positivistic view and 2) naturalistic view. Debates between these two
paradigms are based on “the epistemological basis of research” (Lynch, 1996, p. 13).
approaches to conducting inquiry (Kiely & Rea-Dickins, 2005). This paradigm has generated
two key inquiry approaches: 1) true experiments, which assign participants randomly to
experimental and control groups to draw comparisons between them, and 2) quasi-
experiments, which compare the two groups without random assignment (Lynch, 1996).
Many researchers (Beretta, 1986; Long, 1983) criticized the positivistic paradigm, for it
merely focuses on “product or outcome rather than also attending to the process of how the
On the other hand, the naturalistic paradigm “requires an emergent or variable design”
where evaluation takes place as the “evaluator proceeds to investigate the program setting,
allowing new information and insights to change how and from whom data will be gathered”
(Lynch, 1996, p. 14). This form of evaluation does not manipulate any conditions or
variables; rather, it observes, describes, and interprets how a program performs in real-life
contexts. In other words, a naturalistic inquiry is carried out in a program’s natural settings,
which is consistent with the naturalistic second language learning approach (Krashen, 1982).
94
Such an approach includes “observations, interviews, questionnaires, tests, and program
approaches was applied to help us articulate “what counts as evidence for our evaluation”
Most early program evaluation studies have focused mainly on program design and
2003, p. 2). Later, a wide range of program evaluation approaches emerged, changing to
some extent the motives why language programs are evaluated. For example, during the mid-
and late-1990s, program evaluation shifted from merely improving program specific features
This impetus for accountability was a result of the necessity of gaining more funds from
funding agencies (Brindley, 1998). Program outcome-based evaluation can therefore have
several motives including but not limited to responding to greater accountability needs,
maintaining quality assurance and control, obtaining academic accreditation, gaining higher
rankings, and so forth. Regardless of the purpose of program evaluation, it is prudent realize
that program evaluation provides valuable conclusions about a program performance as well
as the extent to which strengths and weaknesses should be supported and addressed
respectively.
then decide how program evaluation processes will be planned, implemented, and even
evaluated in a systematic way. To help program evaluators make a decision with this regard,
Brown (1995) argues that prior to conducting program evaluation, evaluators should first
1) formative or summative
95
2) involves external experts or internal participants
These illustrate some of the key conceptual and practical approaches to program evaluation
that program evaluators, administrators, teachers, practitioners, and external reviewers should
Formative and summative evaluations are two distinct measures that have long been
used in the evaluation industry (Alderson, 1986; Bachman & Palmer, 1996; Black & William,
2009; Brown, 1995). Although formative evaluation was first introduced into educational
psychology literature by Scriven (1967), it was defined in detail by Bloom et al. (1971) as
“the use of systematic evaluation in the process of curriculum construction, teaching, and
learning for the purpose of improving any of these three processes” (p. 117). It is aimed at
process. In other words, it refers to the process of collecting information and data that
(Bachman & Palmer, 1996, p. 62). However, Bailey (2009) defines program level formative
evaluation as a form of appraisal that provides “feedback for program improvement” (p. 707).
It helps stakeholders gain periodic or ongoing guidance that helps them “adjust their activities
measure that “typically occurs at the end of a program or a funding period for a program”
96
(Bailey, 2009, p. 707). It is typically used to provide “evaluative conclusions for any other
reasons besides development” (Scriven, 1991, p. 21). Brown (1995) argues that summative
evaluations occur at the end of a program “to determine whether the program was successful”
(p. 228). This form of evaluation provides teaching practitioners with an inclusive summary
of the value of their program for future improvement. As for the study contexts, most of the
evaluations undertaken by the NCAAA are summative, while those conducted by program
participants are formative. A useful distinction of this dichotomy was suggested by Alderson
(1986) that “when the cook tastes the soup, that is formative. When the guests taste the soup,
both. Some program administrators invite external experts to carry out their program
evaluation processes. For example, during one interview, the quality assurance coordinator of
the male EFL department under study indicated that they once invited a quality assurance
expert from another Saudi university to evaluate their learning outcomes. The NCAAA
requires Saudi higher educational institutions to bring in a group of external experts to help
them conduct the developmental self-study or the mock program review. According to Brown
(1995), external reviewers can provide “a certain amount of impartiality and credibility to the
results” (p. 232). Nevertheless, Alderson and Scott (1992) argue that outside evaluators “are
perceived by insiders as at least threatening to themselves and the future of their project, and
at worst as irrelevant to the interests and perspectives of the project” (26-27). Hence, they
97
The Participatory Model includes involving stakeholders in program planning,
implementation, and evaluation processes (Alderson & Scott, 1992). This model helps
different perspectives about the effectiveness of their program. One of the main benefits of
this model is that it helps “develop greater insights not only about the roles that stakeholders
may play in evaluations, but also how learning may take place” (Kiely & Rea-Dickens, 2005,
p. 201). Nonetheless, to overcome any potential bias, Alderson and Scott (1992) suggest that
judgment. Moreover, Ross (1992) recommends engaging teachers in the evaluation processes
to reduce their anxiety of “being watched” and to promote their roles as “both practitioner
Another aspect that program evaluators should be aware of is whether the evaluation
processes will be undertaken through field or laboratory research. Beretta (1986, p. 296)
short-term evaluation that often “involves the testing of individual components of a theory in
an environment in which extraneous variables are artificially held constant” (Beretta, 1986, p.
296). Reliance on either form depends on the purpose of the evaluation (Brown, 1995).
Alderson and Scott (1992) argued that some program evaluators prefer a field evaluation to
obtain an overall image of a program, while others prefer to conduct a small-scale laboratory
evaluations depending on the evaluation purpose and context. According to Brown (1995),
previous studies of program evaluation were primarily longitudinal in that they were
conducted during the implementation of a program. On the other hand, short-term evaluations
(often called after-program evaluations) are carried out at the end of a program in an attempt
researchers pay substantive attention to the content of the program by observing student
performance constantly (on-going evaluation), while others seek to look for the final output
(short-term evaluation). Brown (1995) suggests that a combination of both ongoing and
short-term evaluations be integrated “during the program, immediately after it, and in a
follow-up as well” so that all potential aspects of the program can be covered effectively (p.
233).
questionnaires, and so forth (Lynch, 1996). Advocates of qualitative approaches adopt the
(Lynch 1990, 1992; McKay, 1991; Rea-Dickens, 1999). Qualitative approaches are useful for
providing “the best array of information types” allowing program evaluators to gain broader
and deeper insight into program processes (Brown, 1995, p. 234). However, such approaches
can be problematic, for data collection and analysis are time-consuming and may include
multi-faceted themes, an issue that encourages many program evaluators to rely more on
quantitative methods.
99
On the other hand, some researchers choose to collect and analyze data quantitatively
(positivist perspective). Brown (1989) defines quantitative data as any form of information
“gathered using those measures which lend themselves to being turned into numbers and
statistics” including course grades, test scores, accountability criteria, faculty qualifications,
ratio of students per teaching faculty and so forth (p. 231). One of the benefits of quantitative
data in program evaluation is that they have been seen as “easier to gather, and more
amenable to analysis and summary” (Alderson & Scott, p. 53). Another benefit is that they
can help alleviate as much stakeholders’ bias as possible, as rarely is there any form of
subjective analysis (Ross, 2003). Nevertheless, recent program evaluation studies have
(Alderson & Beretta, 1992). According to Ross (2003), this diversity of approaches can help
us “yield richer contextualized data about program processes as well as empirical data about
that focuses mainly “on what it is that is going on in the program (process) that helps to arrive
at those goals (product)” (p., 231). During the 1990s, as denoted earlier, program evaluation
was focused more on process-oriented methodologies (Lynch, 1996). That is, program
evaluation advocates called for examining how learning outcomes are achieved effectively
(Beretta, 1986b; Burstein & Guiton, 1984; Campbell, 1978; Elley, 1989; Lynch, 1990, 1996;
Ross, 2003). Process-oriented evaluations can provide internal stakeholders (e.g., students,
100
teachers, administrators) and external stakeholders (e.g., accreditation agencies, funding
agencies) with detailed information about their program to enhance its performance and to
identify the different motives of educational institutions for evaluating their programs. Over
time, according to Thomas (2003), program evaluation has changed dramatically to include
other considerations such as accreditation, quality assurance, external review, and so forth in
addition to curriculum development, accountability, the value and worth of programs, and
fund gaining (as cited in Kiely & Rea-Dickins, 2005). Hence, purposes for program
evaluation differ depending on the expectations of the stakeholders of the program being
evaluated. That is, program evaluation can be sought for different purposes including
program being evaluated (Posavac, 2015, p. 23). Most of these purposes lead to quality
component of program evaluation (Van Damme, 2004). The term quality assurance refers to
a systematic evaluation process through which a program’s learning outcomes are evaluated
to ensure that they meet specific predetermined standards (McNaught, 2009). According to
Halliday (1994), quality assurance was first sought in educational contexts during the 1980s
“as a response to political demands for institutional accountability” (p. 36). The term quality
assurance is “derived partly from manufacturing and service industry, partly from health
care” and then was pervasively integrated into educational contexts (Ellis, 1993, p. 3). Norris
101
(1998) listed six approaches to curriculum evaluation of social programs that help manage
quality assurance and comply with mandates, namely “1) Experimentation, 2) Measurement
Gosling and D'Andrea (2001) went beyond quality management to a higher phase that
they call Quality Development, which is “an integrated educational development model that
incorporates the enhancement of learning and teaching with the quality and standards
monitoring processes in the university” (p. 11). The premise of this model is to assure total
quality of program curriculum rather than solely focusing on achieving other goals, for
example, accreditation. Thus, various new initiatives have been introduced into quality
assurance literature including total quality, quality control, trust, and so forth (Sallis, 2002).
Research reveals many approaches to assuring quality such as self-report (Ellis, 1993), needs
analysis (Richard, 2001), SWOT analysis (Dyson, 2004), benchmarking (Ellis & Moore,
2006), and site visit (Norris, 2009). Moreover, one of the purposes of assuring quality is to
obtain accreditation, a process by which institutions, universities, and programs are validated
formal approval of an institution or program that has been found by a legitimate body to meet
predetermined and agreed upon standards” (Van Damme, 2004, p. 129). Accreditation is a
colleges, universities, and educational programs for quality assurance and quality
improvement” (Council for Higher Education Accreditation, 2002, p. 1). That is, it entails
evaluating the extent to which certain standards are met by an institution or program.
102
According to Eaton (2006), accreditation functions as “a reliable authority on academic
quality” whereby quality assurance, federal and state funds, accountability, etc. can all be
maintained effectively (p. 3). There are two types of accreditation in higher education:
with “a license to operate”, while the latter accredits programs “for their academic standing”
departments, and programs” to assure quality in their programs (Norris, 2009). The
postsecondary institutions in many developing countries, including Saudi Arabia, have begun
to seek “to participate and compete in the global economy” in order to ensure that their
programs meet predetermined accreditation standards (Smith & Abouammoh, 2013, p. 104).
In fact, some Saudi universities have imposed strict systems that lead their programs to meet
national accreditation standards. For example, in 2009, King Saud University imposed
probation on its programs, “and unless they satisfy the NCAAA standards by September
Education in Saudi Arabia has been undergoing drastic reforms that have resulted in
universities in 2007, and ending with sponsoring several private higher education institutions
in 2010 (Alamri, 2011). This has led the Saudi Ministry of Education to implement quality
103
assurance in higher education institutions. Nonetheless, according to Darandari et al. (2009),
“there was no quality system for higher education at the national level in Saudi Arabia”
before 2004 save some institutional endeavors (p. 40). Since that time, the NCAAA has
become the official government accreditation body that governs national quality assurance
and accreditation standards to which many public and private postsecondary institutions
NCAAA
was established in 2004 to oversee quality assurance and accreditation processes in Saudi
institutions (NCAAA, 2013). The mission of the NCAAA is “to encourage, support, and
evaluate the quality assurance processes of postsecondary institutions and the programs they
offer” (NCAAA, 2014). In 2005, it became a member of the International Network for
Quality Assurance Agencies in Higher Education (Darandari et al., 2009). The quality
assurance and accreditation trends in conjunction with the expansion of higher education
The Saudi higher education system has recently witnessed a significant expansion
(Smith & Abouammoh, 2013). As of 2015, there are 26 government universities, 18 male and
Colleges, and 49 Private Universities and Colleges (MOHE, 2015). Although obtaining
accreditation is not mandatory in Saudi Arabia, nor does it impact government-based funding,
many Saudi universities seek to obtain accreditation from the NCAAA. According to the
104
Saudi Ninth Development Plan (2014), accreditation is a national strategic dimension of the
policies of the country (Ministry of Economy and Planning, 2015). Hence, all universities are
required to restructure their academic programs by closing those that do not fulfill job market
needs and promoting those that do. Moreover, no university is allowed to establish new
programs unless high employability rates relating to these programs can be obtained. Thus,
many universities tend to customize their current programs to meet job market needs. One of
Quality assurance and accreditation are therefore a key trend to shaping the future of
Saudi higher education (Darandari et al., 2009). Since quality assurance processes, as
Mercado (2012) noted, are better performed by internal reviewers first (Emic perspective),
this paper serves as a simulated evaluation of two EFL departments in a Saudi university as
an initial step for assuring quality and thereby obtaining academic accreditation. Therefore, in
coordination with the vice-rector and the heads of the two departments, the researcher
conducted a site visit to the two departments to evaluate their quality assurance processes. It
is also hoped that this paper will be the first step toward promoting language program
evaluation in Saudi Arabia in order to assure quality and obtain academic national and
METHODOLOGY
The primary motive for conducting this study is to guide two EFL departments
accreditation standards of NCAAA and CEA commissions that meet the nature, needs, and
purposes of the 18 EFL departments. The quality assurance practices of two departments (one
for male students and the other for female students) were evaluated to serve as internal
105
benchmarks for the remaining 16 departments to help their administrators assure quality and
obtain accreditation from NCAAA and CEA. Any inapplicable or unnecessary standards
were excluded (e.g., marketing, housing, and others). To evaluate these two EFL departments
effectively, the researcher made a site visit to X University where these departments are
located. The researcher circulated surveys to male and female students, teaching staff, and
department heads, and he also conducted semi-structured interviews with a randomly selected
sample of participants.
Background on Site
X University has 18 male and female EFL undergraduate departments, which are
planning to obtain departmental accreditation from the NCAAA in the next two years and
perhaps from the CEA in the near future. Both the departments have a four-year, full-time
education program divided into eight levels and preceded by a one-semester pre-university
intensive English language program. The total credit hours required for graduation are 137,
22 of which are non-major courses. The departments have two main tracks: 1) EFL and
Translation and 2) EFL and Literature. These two departments were selected in particular
because they are the largest male and female EFL departments at X University, so they can
University.
Research Objectives
The research objectives of this paper vary but fall under one overarching goal. First, they
aim to obtain stakeholders’ (students, faculty, and department administrators) opinions about
the effectiveness of the two departments being studied. Second, they attempt to identify the
extent to which the two departments meet the integrated standards of NCAAA and CEA in
order to prepare them to obtain academic accreditation from these two commissions. Third,
106
they pinpoint both positive and potentially poor quality assurance practices in the two
departments. Fourth, they identify potential dilemmas and barriers that may delay the two
EFL departments from obtaining academic accreditation from the NCAAA and CEA. Fifth,
they provide some implications that will contribute to helping the two EFL departments
maintain quality assurance and thereby obtain academic accreditation. Finally, the researcher
will provide a concise report describing quality assurance processes of these two EFL
departments. It is hoped that the report will be useful not only for EFL departments at X
University, but also for any similar EFL departments that plan to obtain national or
Research Questions
1. To what extent do the two departments meet the study-integrated model of standards?
2. To what extent are students satisfied with the two departments’ curricula, teaching
4. What are poor quality assurance practices in the two departments that need to be
improved?
5. Are there any potential dilemmas and barriers that may delay the two EFL
Participants
The participants of this study are divided into two groups, male and female. Each
group is subdivided into three subgroups: teaching staff, students, and administrators. The
first group consists of a randomly selected sample of teaching faculty, students, and the
quality assurance coordinator of the male EFL department. The second group consists of a
randomly selected female group of the same categories from the female department. The first
107
subgroup of participants is male and female undergraduate students majoring in EFL. Their
levels of English proficiency, which were identified based on their current levels at the two
departments, range from beginner to advanced. Two hundred twenty seven students studying
in the EFL departments participated in the survey, and 15 of them (six male and nine female)
were interviewed. The second subgroup consists of two male and four female teaching
faculty. The third subgroup consists of one male quality assurance coordinator and one
Research Tools
The central research tools used in this study included a web-based survey (Appendix
M), semi-structured interviews (Appendix N), and the researcher’s evaluation checklist
(Appendix O). The survey consisted of six themes: general information (4 items), course
objective (10 items), teaching strategies (10 items), student learning outcomes (10 items), and
student assessment (10 items). There were also three different semi-structured interviews: 1)
student interviews, 2) teaching faculty interviews, and 3) administrator interviews. The third
research tool was the evaluation checklist (Appendix O), which was used to evaluate the
mission, curricula, objectives, and goals of the two departments. Survey items, semi-
structured questions, and evaluation checklist were all designed based on the standards and
substandards of the integrated model of NCAAA and CEA to ensure that the evaluation
It is relatively difficult and perhaps time-consuming to integrate all NCAAA and CEA
standards in Table 12 into one model; thus, only standards suiting the Saudi higher education
contexts were included. Moreover, this paper focuses on the standards that address student
learning experience and ignores those that deal with administrative and/or financial matters.
108
Hence, 1) Mission, Goals, and Objectives, 2) Learning and Teaching, and 3) five sub-
standards of Learning and Teaching standard were selected from NCAAA standards.
5) Student Achievement were selected from CEA standards. The integration process also
included the sub-standards of the selected standards, as it is shown in Table 13 below. One
noteworthy issue is that meeting the integrated model of standard does not mean in any case
that the two departments meet all of the CEA and NCAAA standards; rather, it means that
109
Table 12
NCAAA CEA
Table 13
NCAAA CEA
Procedures
110
Data were collected circulating surveys and by conducting semi-structured interviews
with a randomly selected sample of the two departments (i.e., students, teaching faculty,
program quality assurance coordinator). That is, having visited the male department, the
researcher attended some classes, where he was allowed to spend time with students to
collect data. Although the survey was delivered via Qualtrics, online survey software, the
participants did not have access to the Internet. At first, the researcher thought a technical
glitch occurred for some reason; nevertheless, it turned out that the department did not have
Internet access. Thus, the researcher had to print 300 copies of the survey. Then, students
were provided with a detailed description of the study’s purpose and procedure, and the
survey was explained to them in order to clarify any potential confusion. Nonetheless, for
ethical considerations, students were given the option of completing the survey via hard copy
After that, a randomly selected sample of male students was engaged in semi-
structured interviews to allow them to expand on their attitudes towards the department. Each
interview session took approximately 20-25 minutes. By randomly choosing some students
from each level, the researcher ensured that the interviewed students had different levels of
proficiency to obtain a broader insight into the department. Moreover, the mission, goals,
curriculums, teaching strategies, assessment methods, and learning outcomes of the two
departments under study were evaluated using the evaluation rubric that has been designed
based on the integrated standards model (See Appendix O). As for the female-section
department, the researcher was unable to interview female participants face-to-face due to
cultural constraints; thus, a female faculty was assigned by the department head to carry out
111
Data Analysis
Data were analyzed systematically and qualitatively as carried out by Mitchell (1989)
for her evaluation of the language program at the University of Stirling, Scotland. That is,
responses of all participants were used as evidence for exploring the extent to which the two
departments meet the target standards and then to provide a comprehensively descriptive and
responses, all potential themes, explicit or implicit patterns, or emerging trends were sought.
For example, all answers obtained in the second round of data collection (interviews) were
transcribed and then analyzed to identify themes using comparative techniques against CEA
and NCAAA standards. The findings of this study were divided into four major areas: 1) the
two departments’ adherence to the integrated model of standards, 2) student satisfaction with
the departments, 3) good and poor quality practices of the two departments, and finally 4) any
FINDINGS
The extent to which the practices of the two departments meet the CEA and NCAAA
standards was examined by evaluating them against the study-integrated model of standards.
defined by Allen (2004) as a “holistic vision of the values and philosophy” of an institution or
program (p. 28). According Kiely and Rea-Dickens (2005), articulating a mission is key to
statements are "ubiquitous in higher education” in that “accreditation agencies demand them,
strategic planning is predicated on their formulation, and virtually every college and
university has one available for review” (Morphew & Hartley, 2006, p. 456). Conway,
112
Mackay and Yorke (1994) provide six criteria that a mission statement should fulfill: 1)
program context, and 6) values sought by the program over other similar programs.
As discussed earlier, the mission statement plays an integral role in determining how a
program operates within a previously determined set of standards. Therefore, both the CEA
mission statement that can be accessed by stakeholders at anytime and from anywhere.
Moreover, the CEA mission standard states that a program should have “a written statement
of its mission and goals, which guides its activities, policies, and allocation of resources”
(CEA, 2015, p. 7). Surprisingly, the two EFL departments under investigation have the same
mission statement in writing, which also appears to have areas of weaknesses. For instance, it
was derived from the mission of the college under which the two departments operate (See
Appendix P).
college’s mission, which functions as the umbrella for the two departments’ mission.
Although the college’s mission is posted on the college’s website, which is adequate for
meeting one aspect of the NCAAA sub-standard, “Stakeholders should be kept informed
about the mission and any changes in it” (NCAAA, 2013, p. 8) and CEA main standard, “The
students, student sponsors, and the public” (CEA 2015, p. 7), it appears to be too wordy. This
may make it difficult to be understood by stakeholders and evaluators. Moreover, it does not
specify the college’s activities, policies, and allocation of resources. Rather, it only describes
Paradoxically, the Arabic version of the college’s mission is also posted on the same
the local, national, and international scientific and social fields. (college’s website,
2015)
This inconsistency may create some ambiguity for stakeholders. For example, when the
researcher first visited the college’s website, he was confused as to which mission, the
English or the Arabic one, guides the college’s goals, policies, and allocation of resources, as
each mission has a completely different purpose. Hence, one of these two mission statements
The mission of the two departments (Promoting the overall performance of learners of
language, literature, and translation to achieve the possibly highest quality of learning
outcomes) is posted on the college’s website as well. Nevertheless, having examined the
department specifications report, it was noticed that the two departments have a different
mission, namely, “Preparing highly qualified cadres with skills and expertise in English
language, translation, and literature, and well-trained educated researchers who follow the
scientific approach in thinking and in dealing with technology, multi-faceted thinking, and
problem solving” (Department specifications report, 2015). This mission seems to have been
written clearly as required by CEA and NCAAA. Moreover, it includes “the elements
114
commonly understood to form the basis for a higher education mission” (Morphew, &
The departments’ mission statement appears to fulfill only two criteria of those
mentioned by Conway et al. (1994); it highlights the departments’ goals and objectives, and it
includes the programs offered (i.e., literature, & translation). However, it neglects the
remaining four criteria: 1) relevance to the college mission (though it is almost aligned with
that of X University), 2) the target groups (EFL students), 3) the departments’ geographical
areas, and 4) the added value sought by the departments over other similar EFL departments.
Another drawback is that the mission statement does not have a controlling idea. That is,
when compared with the mission statement from a comparable department, for example, the
learning environment for the rehabilitation of graduates in English language and translation
able to meet society's expectations for a promising future” (Department website, 2015), it is
clear that the latter has a controlling idea (creating an interactive English learning
environment), whereas the former does not. Furthermore, unlike that of X University, King
Taking the above into account, the two departments’ mission statement fail to
provide a sense of the current identity of X University. In addition, based on data obtained
from the interviews, 11 out of 15 of the students asserted that they have no idea what the
mission statement of these two departments is. More importantly, data showed that this
mission statement was updated only once in the past four years. This may impede the mission
from reflecting “new goals or shifts in the focus of [their] educational programs or services
and whenever activities and policies are conceived, implemented, or revised” (CEA, 2015, p.
7). All of these factors suggest that this mission statement appears to only partially meet the
115
Mission Standard of CEA and NCAAA. A poorly phrased mission statement would make it
difficult to meet other standards such as ensuring alignment of the curriculum with the
Since CEA does not have a separate standard for goals and objectives, NCAAA
standards were used to identify whether the goals and objectives of the departments meet
these standards. The goals and objectives of the two departments are almost identical and can
translators, competent linguists, and excellent researchers. These are clear, attainable, and
“consistent with and support the mission” (NCAAA, 2013, p. 8), which mostly fulfill the
NCAAA sub-standard (Relationship between Mission, Goals & Objectives). However, the
them for seeking higher studies, and developing their presentation and rhetorical skills are not
derived from the mission of the departments. That is, the mission does not indicate a desire to
instill in the students a sense of providing community service, which seems to be irrelevant to
One of an instructor’s main duties is to ensure that students are fully aware of course
objectives (Allen, 2004). Out of the 202 participants who responded to this item, 13 (6%)
students indicated that they strongly agree that most of the departments’ course objectives are
communicated to them at the beginning of each course, while 95 (47%) of them agreed with
this item. On the other hand, 22 (11%) of the participants disagreed, and 12 (6%) strongly
disagreed with this statement, suggesting that only a small number of the students are
dissatisfied with the poor communication of course objectives or lack of them. However, 60
116
(30%) of the participants were uncertain (neither agree nor disagree) about the delivery of
course objectives at the beginning of the course. This high percentage poses serious questions
as to why these participants would be uncertain about whether course objectives were
communicated to them earlier. This can be attributed to several factors detailed in the
following discussion.
Moreover, some of these participants are likely to have “no clear idea of what the
intended outcomes of the course are,” and thus ending up being confused about the ultimate
goal of the course itself (Menges, Weimer, & Associates, 1996, p. 188). Another bewildering,
albeit reasonable, question is why are these students unsure about being informed of these
objectives in advance if, as Allen (2004) argues, articulating course objectives on a course
syllabus “allows students to make informed decision before enrolling, to monitor and direct
their own learning, and to communicate what they have learned to others, such as graduate
schools, employers, or transfer institutions” (p. 44)? To obtain further insight into this issue,
some of the participants were asked during interviews about their uncertainty. Seven out of
15 argued that they are rarely provided with written syllabi that highlight course objectives.
In fact, one of the female participants asserted that the she has never been given a syllabus in
writing.
The second and one of the key standards of CEA is curriculum, which refers to the
goals, methods, and assessment through which an academic program achieves its targeted
instructional objectives (Nation & Macalister, 2009). As for NCAAA, the word curriculum is
not mentioned in their handbook; rather, all curriculum components are distributed across
various standards and sub-standards. To better evaluate the curricula of the two EFL
outcomes, 4) processes for teaching and learning, and 5) means of assessment. Nonetheless,
the educational goal and purpose as well as the objectives for each course in the curriculum
have been discussed in the Mission, Goals, and Objectives standard, so we are left with the
The two departments do not have a curriculum in writing, per se; instead, they have a
Study Plan and Program Specifications from which adequate information about the
curriculum components can be obtained. Having compared these two documents against CEA
and NAAA standards, there were some interesting findings. The departments’ study plan is
fairly consistent with their mission. That is, it has “adequate and appropriate levels to meet
the assessed needs of the students through a continuity of learning experiences” (CEA, 2015,
p. 9). Moreover, it has several strengths in that it is designed with a logical progression from
one level to the next through eight levels. It also has consistent objectives across courses and
levels in a way that is appropriate for students’ needs. Nevertheless, one of the drawbacks of
this study plan is that it is not well-organized, nor does it allow evaluators and stakeholders to
monitor and document students’ performance. Despite this, the study plan of the two
departments appears to partially meet the standards with further improvements needed.
learning” (Adam, 2004, p. 2). They are “broad statements of what is achieved and assessed at
the end of a course of study” (Harden, 2002, p. 155). Although SLOs of the two departments
are specified, measurable, attainable, aligned with each other, and relatively consistent with
However, what concerns the researcher is the extent to which SLOs cover all
language skills. According to students’ responses (Fig. 3), out of 115 participants who
responded to this item, 12 strongly agreed that SLOs cover all language skills, 41 agreed, 26
were uncertain, 30 disagreed, and 6 strongly disagreed. This suggests that SLOs of the two
departments need to be reviewed. For example, three out of six teaching staff participating in
this study indicated that SLOs do not appear to cover all language skills. Based on these
findings, SLOs practices appear to partially meet the standards of both the CEA and NCAAA
Teaching Strategies
Among various NCAAA standards, Learning and Teaching standard is the largest
standard in length, suggesting that it plays a vital role in helping an institution or program
obtain academic accreditation. Therefore, both students and teaching faculty were asked
about the practices of this standard extensively. The graphs below (Figs. 4-7) show students’
119
responses to four key items of the Teaching Strategies section, which were derived from the
Figure 4. Teaching strategies are proper for various learning styles. Figure 5. Instructors communicate with you in English.
Figure 6. You are engaged presentations & leading discussions Figure 7. You are engaged in research projects.
and 12 strongly disagreed that teaching strategies used in the two departments are proper for
their different learning styles. Yet, the participants complained that rarely do teaching faculty
speak English. Moreover, they stated that due to the large number of students per class, they
are not given the opportunity to demonstrate their learning. On the other hand, their responses
to the extent to which they are engaged in research projects varied in that 10 strongly agreed,
38 agreed, 40 were unsure, seven strongly disagreed, and 21 disagreed. Based on these
results, it can be concluded that teaching strategies at the two departments appear to partially
120
Assessment Methods
objectives that faculty value” (Allen, 2004, p. 13). Thus, it is crucial to ensure that program
assessment methods are “for learning (i.e., formative in nature) rather than of learning (i.e.,
not summative, as in conventional tests) (Cumming, 2009, p. 93). Despite this, two out of six
teaching faculty indicated that they only use direct assessment tools (e.g., midterm/final
exams, homework). Lack of using other forms of assessment (e.g., performance assessment)
may not help identify students’ strengths and weaknesses (McNamara, 1996). When asked
about the infrequent use of alternative assessment methods, one faculty member noted that
they have “a set schedule to follow,” which does not give them the flexibility to diversify
One of the crucial good practices of the CEA and NCAAA assessment substandards is
to inform students of assessment methods in advance. The two departments have achieved a
noticeable progress in this respect. For example, out of 112 students who responded to this
item, 10 students strongly agreed and 34 agreed that assessment methods are communicated
to them in advance. Nevertheless, only 16 students disagreed and 12 strongly disagreed with
this statement, which can be considered a small percentage, yet it still needs to be addressed.
121
Moreover, course assessment methods are set out in student handbook distributed to students
Since assessment methods are closely related to CEA Student Achievement Standard,
it may be better to shed some light on the practices of the two departments relating to this
standard. First, CEA requires that assessment practices be undertaken consistently throughout
the program and aligned with its mission, curriculum, objectives, and student learning
outcomes. As shown in Figure 9, students revealed different, yet profound insights into
assessment methods used in the two departments. That is, 16 out of 186 students strongly
agreed that assessment practices are used consistently, 57 agreed, 38 disagreed, and 19
strongly disagreed. One perplexing finding is that almost 56 participants were unsure about
this, which has raised the researcher’s curiosity to know why. After being asked about their
uncertainty, eight of these students argued that even though assessment methods are
undertaken consistently, they are limited to quizzes, a midterm, and a final exam. They
suggested that assessment methods be diversified so that their understanding and mastery of a
As for NCAAA standards, students’ responses to the assessment methods section suggest that
assessment practices in the two departments do not appear to meet the standards. For
122
example, the majority of the participants indicated that they are either unsure (68), disagree
(32), or strongly disagree (10) that assessment techniques measure aspects on which they
expect to be measured, grade distribution is appropriate to course objectives, and exams and
assignments are graded and returned to them promptly, all of which imply that this standard
As program evaluation processes can spark investigation across the program, this
performance, taking into account other considerations such as “the social and political basis
and motivation for language learning and teaching (Lynch, 1996, p. 10). Both the CEA and
NCAAA place extensive emphasis on program evaluation. The last standard of CEA
form of two sub-standards of the fourth standard Learning and Teaching as Program
Development Processes and Program Evaluation & Review Processes. The departments’ data
about this standard obtained from teaching faculty did not reveal any findings worth-
mentioning because their responses focused on two main themes: 1) lack adequate training in
processes effectively due to their heavy teaching loads (23 hours per week).
In order to gain further insight into the two departments’ evaluation practices, one of
processes was interviewed apart from the department heads. Therefore, this section will be
allocated for discussing the interview responses obtained from the coordinator of the quality
assurance unit, John, with whom the researcher recently spoke using Line, a widely-used
123
communication application. The interview with John, who has been the quality assurance
coordinator for the male student department for three years, took nearly 30 minutes. He
shared his experience in taking over the department’s quality assurance processes. He
indicated that during the past three years, quality assurance evaluation and processes were
John indicated that the department does not have a plan in writing for planning,
stated, that the NCAAA handbook has been used as a plan for the department evaluation
processes. However, the NCAAA requires that all academic programs and/or departments
have a plan in writing for managing all “quality assurance processes across the [program],
and the adequacy of responses made to evaluations processes that are made” (NCAAA, 2013,
p. 18). The lack of such plan has resulted in not setting out a route for reviewing the
activities, policies, and services. To address this issue, a quality assurance unit was
established at the two departments in 2013. Although this unit takes on all responsibilities of
department evaluation and quality assurance processes, it still does not comply with the
Although all teaching and other staff are engaged in program evaluation processes,
there is a lack of quality assurance and program evaluation literacy. According to the male
processes were solely limited to filling in the NCAAA course reports without involvement in
complained about the lack of training teaching faculty on quality assurance processes.
Furthermore, John described a lack of effective cooperation with other departments affiliated
124
with the same college regarding exchanging experiences about quality assurance processes.
Another issue concerns the instructional staff’s teaching and administrative workloads. For
example, John is responsible for teaching 23 hours per week in addition to supervising
quality assurance and examination processes. This does not meet the NCAAA substandard
that states “Sufficient time should be given for a senior member to supervise quality
Unlike many language programs, where a program evaluation director can exercise
authority over all available or needed resources (Westerheijden, et al., 2007), John, along
with his team, was not given adequate authority that meets the standards of the NCAAA and
authority to obtain all necessary is a basic requirement of the NCAAA and CEA so that s/he
can close the loop of data feedback (Allen, 2004, p. 17). However, John indicated that he
finds difficulty obtaining program evaluation-related data from various sections such as the
program, the college, and other similar programs, for he is not authorized to do so. This has
resulted in poor practices of the two departments in evaluating, reporting on, and improving
quality assurance arrangements. Moreover, when John asked the college to provide him with
funding for conducting external benchmarking, he said, “the college refused to do so for
unrealistic reasons.”
To promote an effective practice of program evaluation, both the NCAAA and CEA
student department has had only one informal external evaluation, which was conducted by a
visiting professor from King Saud University. Although the external reviewer provided
professor of Arabic. According to John, despite the external reviewer’s great efforts, he was
125
unable to pinpoint areas of weaknesses from which language-related suggestions could be
made for further improvement and systematic review. On the other hand, the female
department has never been subjected to any evaluation. When asked about this, the
department head stated that they have no access to data needed for internal or external
benchmarking, as their role is only limited to filling in the NCAAA forms. Thus, it appears
that this standard is not being met for either the NCAAA or CEA.
Now that the extent to which the two departments meet the integrated model of
standards has been identified, one can understand the students’ level of satisfaction with the
departments. One noteworthy issue is that the evaluation findings of some standards may help
us identify students’ satisfaction with the two departments (e.g., curriculum, learning
outcomes, teaching strategies, assessment methods), while others may not provide us with
review). Having conversed with some male students, the researcher noticed that they seemed
dissatisfied with some of the department’s educational activities. For example, six out of 15
students who participated in the interviews indicated that they have not noticed dramatic
progress in their English skills. Specifically, one student argued that some of the courses do
not promote students’ critical thinking skills; instead, they primarily rely upon memorization.
Moreover, nine of the 15 participants expressed anxiety and frustration with their poor
speaking and writing skills in English. One of them argued that most of the courses are
designed in a way that improves their receptive skills (e.g., listening and reading), but
neglects improving their productive skills. Another student indicated that he had to engage
himself in communicative activities outside the classroom, which has helped him speak
specialization, the CEA standards; on the other hand, place substantial emphasis on
improving language skills as a whole. Having examined the male department’s objectives
(See Appendix Q), it can be noticed that instead of focusing on developing students’
language skills, they address other important skills such as technical, research, and
Furthermore, one of the students complained that although they are “placed in a safe,
clean, livable environment” (CEA, 2015, p. 29), many students complain that it is not a
motivating learning environment. He said, “Some of the department courses are taught in a
high-school manner wherein the instructor dominates the classroom discourse without
When asked about this issue, one of the female faculty members indicated that this could be
true to some extent, pointing out that teaching students who have drastic learning
backgrounds makes it difficult to diversify teaching strategies. Thus, some teaching faculty
members use merely a lecture-based teaching format. Another teaching faculty complained
that having a large number of students per class does not allow them to avoid lecturing and
One of the students complained that some non-native English speaking teaching
faculty members speak Arabic frequently. To determine if this was true, the researcher asked
some faculty members how often they spoke Arabic inside class. Two of them asserted that
they speak English most of the time; however, when they want to provide students with
important instructions (e.g., about exams), they often repeat what they have said in Arabic.
Another student complaint was that their assignments are not graded and returned in a timely
manner. According to the survey results, 20 students strongly agree that their work is not
127
graded and returned to them, 55 agree, 56 are uncertain, 30 disagree, and 25 strongly disagree
(Fig. 10).
Figure 10. Students’ responses to how often their work is graded and returned to them
promptly.
Although most of the practices discussed earlier were found inadequate to fully meet
the study-integrated model of standards, students seemed satisfied with some of these
practices. For example, it can be noted that the percentages of students who strongly agree
and agree with most of the statements outweigh those who strongly disagree or disagree. This
may suggest that the practices of these statements are relatively satisfying, with further
improvements needed. Another example is that one of the highest overall ratings of student
satisfaction of the practices was about student learning outcomes. When asked about the
extent to which SLOs meet their learning expectations, 95 out of 190 participants (50%)
seemed satisfied with SLOs, while only 37 of them (20%) were dissatisfied that SLOs did not
128
Figure 11. Students’ responses to the extent to which SLOs meet their expectations.
In addition, as a former student and current lecturer at the male EFL department, the
researcher has noticed progress in numerous aspects of the department performance. For
example, compared to its performance in 2008, when the department had its first review, the
male department currently operates according to predetermined standards and KPIs based on
which the achievement of the department outcomes are verified. Teaching strategies, SLOs,
and assessment methods are all evaluated on a semesterly basis and the evaluation results are
included into the NCAAA templates for improvement. For example, each course coordinator
is responsible for preparing course blueprints indicating the consistency between the intended
SLOs with teaching and assessment strategies (Male Department Annual Report, 2014).
Another satisfying practice of both the male and female departments, which was not
included into the survey but naturally arose during the interviews, was concerned with the
academic and career counseling and advice provided for students. Two out of 15 of the
participants commended the efforts of teaching faculty in providing students with effective
advising about their current academic performances and potential career opportunities. They
indicated that the department administrators have allocated separate offices for student
individual counseling services. Moreover, four of the female participants pointed out that
whenever they face any form of learning difficulties, the academic advisors to which they are
assigned provide them with all needed counseling and assistance. These good practices meet
one of the salient substandards of the NCAAA, namely, Educational Assistance for Students.
129
Generally, students appeared to be satisfied with the two departments’ newly
established facilities. In the interviews, three of the male students indicated that unlike the old
building, the environment of the new building is clean, attractive, and well maintained, which
meets Standard 7, Facilities & Equipment (NCAAA, 2013, p. 39). Despite their complaints
about the lack of Internet connectivity, they also noted that the main library provides them
with “efficient access to online databases, research, and journal material” relating to the
department courses, which meets Standard 6, Learning Resources (NCAAA 2013, p. 35). The
female department has also recently transferred to a new building equipped with adequate
technological and learning resources that facilitate the educational process. In the interviews,
having attended classes at both buildings, five out of nine participants indicated that the new
building is well-equipped and adapted to meet their learning requirements as opposed to the
old building.
Given that the ultimate purpose of this paper is to evaluate the readiness of the two
departments for obtaining academic accreditation from the NCAAA and CEA in the future, it
would be very useful to highlight both good and potentially poor practices so that they can be
considered carefully for future enhancement and improvement respectively. In general, the
two departments have numerous good practices that the researcher has identified in this
study. For example, although the two departments’ mission statement seems to partially meet
the Mission Standard of the CEA and NCAAA, it only needs few improvements with greater
emphasis placed on students’ language skills, and then it will fulfill most of the practices of
the mission standard of the NCAAA and CEA. Therefore, if the problematic issues raised
earlier about the departments’ mission statement are addressed effectively, the mission
standard will be one of the strongest standards in meeting the two commission standards.
130
Another good quality assurance practice of the two departments is collecting data on
student achievement of SLOs, advancement from one level to another, graduation, and so
forth (Male Department Annual Report, 2014), which is highlighted as a positive aspect in
the CEA Program Development, Planning, and Review standard. According to John, most
data is collected on a semester basis and included into the NCAAA evaluation templates for
future use and improvements. In each department, John continued, a coordinator is assigned
to oversee the extent to which course reports and specifications are prepared on a semester
basis according to the NCAAA templates. Another good quality practices was evident in
rating all practices on an annually basis. For example, according to the two departments’
recent annual reports, feedback from stakeholders (e.g., students, teaching staff, employers,
Despite the good practices discussed above, the two departments still have some poor
quality assurance practices that need to be addressed in order to improve their performance.
First, as stated in their annual report (2014), the two departments fail to conduct internal or
recommendations for improvement can be made. Second, most of the quality assurance
processes, as John noted, are based on individual initiatives even though there is a quality
assurance unit in both departments. In other words, since quality assurance processes at the
two departments are not institutionalized, this results in not involving all teaching and other
staff in quality assurance processes. This was noticed in John’s complaints about the lack of
having a plan in writing for quality assurance practices, lack of cooperation among quality
assurance coordinators at different departments, and lack of having access to data necessary
well. In other words, although the two departments lack necessary data for conducting
benchmarking, they can still overcome this issue by utilizing any data at hand. Unfortunately,
despite collecting some essential data about student performance (e.g., student achievement
and progression) as discussed earlier, the departments’ annual reports do not indicate that
these data are used for improvement (closing the loop). For example, no department has ever
conducted internal benchmarking against its performance from previous years. Another
observed poor practice of the two departments was that their quality assurance processes are
not themselves evaluated and reported on a regularly basis. This is one of the fundamental
quality assurance practices that the NCAAA highlights in its substandard Administration of
Assessment-related Dilemmas
One of the potential dilemmas that may delay the two EFL departments from
obtaining academic accreditation from the NCAAA and CEA is concerned with their
assessment practices. First, as indicated in their annual reports (2014), the departments lack
group-project assessment, portfolios and so forth. Second, the two departments fail to engage
students in a culminating experience course through which they can demonstrate their gained
skills throughout their learning experience in the departments. For example, the study plans
of the two departments lack a capstone or research course that verifies students’ overall
achievement. Third, although one of the two departments’ main goals is to graduate EFL
teachers, their study plans lack a teaching practicum course, which is a basic requirement of
their teaching and other staff on quality assurance and program evaluation processes. As
noted earlier, this poses a barrier for obtaining academic accreditation especially from the
NCAAA. In other words, one of the NCAAA substandards requires that “all teaching and
other staff participate in self-evaluations and cooperate with reporting and improvement
processes” (p. 16). Thus, it is likely that teaching and other staff find difficulty participating
in self-study evaluations without having adequate training programs about quality assurance
processes. Moreover, another potential dilemma of obtaining accreditation from the CEA is
the lack of having a program evaluation plan in writing that “guides the review of curricular
elements, student assessment practices, and student services policies and activities” (CEA,
2015, p. 46).
In general, the two EFL programs are to be commended for their efforts in applying
academic accreditation standards. Nonetheless, it is clear that their efforts lack adequate
words, the researcher noticed that most of the practices lack essential compliance with the
standards of both CEA and NCAAA. For example, the programs’ mission is lacking for
several reasons. First, the programs have two different missions: one stated in program
specifications and another one posted on the university’s website. This does not meet the
NCAAA sub-standard “The mission should be publicized widely within the institution and
action taken to ensure that it is known about and supported by teaching and other staff and
students” (NCAAA, 2013, p. 8). In addition, neither department has a curriculum in writing,
which made it difficult for me to document evidence as they are scattered in various
133
documents. Nevertheless, the curriculum of the two programs appears to partially meet the
As for SLOs, they seem to meet the standards partially as they do not “represent
significant progress or accomplishment” (CEA, 2015, p. 10). Another SLO issue is that they
do not reflect various levels of students’ cognitive, affective, and interpersonal skills as
required by the NCAAA. Furthermore, teaching strategies were also found to only partially
meet the standard, for they are not appropriate for various learning styles. In addition, some
students were unsatisfied with why some teaching faculty members persistently speak Arabic.
Furthermore, both assessment practices and student achievement do not meet the standards
for the considerations discussed before. Therefore, this standard requires immediate action
leading to further improvements. Finally, based on the data obtained, the Program
Development, Planning, and Review standard appears to only partially meet the standards of
Implications
have two different mission statements, only one of them should be chosen and posted on the
university website. It is also recommended that the mission statement fulfill all six criteria
proposed by Conway et al. (1994). More specifically, it lacks relevance to the college
mission, the target groups (EFL students), the departments’ geographical areas, and the added
value sought by the departments similar to other EFL departments. In addition, in order to
develop a mission statement that meets the NCAAA and CEA standards, the mission should
have a controlling idea that guides the two departments’ activities, policies, and allocation of
resources. Finally, a plan for regularly reviewing, evaluating and reaffirming the mission
recommended that a separate curriculum be written in a guide and be made clear for all
stakeholders and evaluators. Moreover, SLOs should focus on all learning domains as
the standards. In addition, teaching strategies used in the two departments should be proper
for various types of learning outcomes. Performance assessment techniques should also be
integrated into assessment practices. Furthermore, the departments’ objectives should focus
on developing students’ language skills. Finally, teaching faculty should be trained on “the
assessment standard is the CEA Student Achievement Standard, which needs further
improvement. For example, it is highly recommended that systems be established “for central
recording and analysis of course completion and program progression and completion rates”
(NCAAA, 2013, p. 23) by “timely reporting to students of their progress through a level
and/or completion of the course” (CEA, 2015, p. 41). Another significant implication is
concerned with ensuring that all assessment practices are aligned with program-level
outcomes to ensure that they meet the professions of teaching, translating and interpreting.
Finally, it is highly recommended that an assessment expert be hired at the two departments
as an assessment coordinator to ensure that assessment practices meet the targeted standards.
One of the essential implications for promoting the two departments’ development,
be trained on program evaluation and quality assurance processes so that they can obtain
adequate knowledge and skills needed for undertaking these processes. Moreover, to make
quality assurance and program evaluation processes consistent and involve all teaching and
include it into the job descriptions of teaching and other staff to ensure that quality assurance
practices are integrated into normal planning and development strategies. Furthermore, all
available data should be utilized for self-evaluation purposes instead of relying on difficult-
to-obtain data. Finally, quality assurance coordinators should be given delegated authority to
136
CHAPTER 5: CONCLUSION
assessment, and evaluation even though some of them still need further research. For
example, although the findings of article one, “Saudi Student Placement into ESL Program
Levels: Issues beyond Test Criteria”, revealed that 11 out of 27 of the pilot study participants
(40%), 14 out of 71 of the Saudi ESL students (20%), and 57 out of 216 of GCC students
(18%) reported that they deliberately underperformed on ESL placement tests, these cannot
placement and replacement rates, Saudi ESL students’ placement rates across different
programs (obtained from SACM), and participants’ language performance were not obtained
as concrete evidence to support the findings. Moreover, the substantial discrepancies between
participating students bring into question the validity of the latter. That is, if the issue at hand
really exists in these percentages, one may wonder why the vast majority of ESL
As a consequence, given that there are 32,557 Saudi IEP students enrolled in US
programs, and perhaps double this number of other IEP scholarship students, the findings of
article one suggest an urgent need for further research studies that better provide more
More specifically, future researchers should avoid asking students directly if they have ever
intentionally performed poorly on ESL placement tests. It is recommended that they allow
such findings to occur naturally in order to make the data more valid and reliable. In sum,
regardless of the differences between the percentages of the study’s participants, it can be
concluded that this study’s issue indeed exists and is worth addressing. Therefore, this study
137
proposes some concrete suggestions that may contribute to eradicating or at least mitigating
this issue.
assessment techniques into their assessment procedures. That is, although self-assessment
techniques have long been utilized in many L2 classrooms (Ekbatani, 2011), several language
programs, especially in Saudi Arabia, still lack using not only self-assessment rubrics but also
many other alternative assessment techniques. As a result, this study suggests that much more
(Boud 1995; Taras, 2010), and augment learner autonomy (Benson, 2012). Nonetheless, the
study concluded that a larger number of participants should be recruited in order to identify
any potential correlation between participants’ self-assessed scores and their TOEFL/IELTS
scores.
More specifically, this study investigated the experience of only 21 ESL participants
using the CEFR self-assessment rubric, which was inadequate to draw valid and
generalizable conclusions that answer the research questions and accurately reflect students’
attitudes and experiences in using self-assessment rubrics. However, although the study did
not reveal a statistically significant correlation between the two measures, ten participants
reported that they found self-assessment rubrics very accurate in reflecting their actual levels
of language proficiency. For example, they provided valuable insights about their experience
in using CEFR assessment rubric including but not limited to pinpointing their areas of
strength and weakness, engaging them in higher language skills such as interaction,
strategies, and language quality, and involving them more in their own learning. Generally,
138
self-assessment rubrics can be very effective especially if they are first used as a supplement
As for the third article, “Quality Assurance and Accreditation as Forms for Language
hoped that it will contribute to raising language program stakeholders’ awareness of the
general, the article concluded that the two EFL programs that participated in the study should
be commended for utilizing many good quality assurance practices. However, they still lack
adequate background regarding program evaluation. In other words, it was found that most of
their practices lack compliance with the standards of the CEA and NCAAA due to lacking
adherence to the two agencies guidelines. For example, the two departments lack a program
evaluation plan in writing, which resulted in dissatisfaction of students with some essential
standards and substandards. Moreover, lack of having a curriculum in writing is another issue
common to the two departments, which may make it difficult to document evidence on their
current performance. However, student achievement standards are not being met due to the
It is important to note that this study is not solely intended for EFL programs, but it
also provides all language programs with a simulated evaluation that can be used as a
reference for future program evaluation projects. In other words, in addition to making this
example, the findings here can serve any program that seeks to obtain programmatic
accreditation from the NCAAA since the study integrated a set of standards can be applicable
139
to any post-secondary programs in Saudi Arabia. In conclusion, administrators of any
program in general but language programs in particular are highly advised to make their
program evaluation processes as systematic as they can in order to meet the standards of the
140
APPENDIX A: PILOT STUDY FINDINGS
141
8. If you have other reasons, please mention them here.
142
11. Do you think the Ministry of Education should establish effective mechanisms to
143
APPENDIX B: SURVEY QUESTIONS
If you are a Saudi ESL student (who has once taken an ESL placement test), would
you please participate in this survey to help me collect some data for my dissertation research
Acquisition and Teaching (SLAT), University of Arizona, Tucson, USA. For my doctoral
dissertation, I am conducting a research study on the placement of Saudi ESL students into
ESL levels. If you decide to participate in this study, you would have to respond to a 20-item
online survey, which will take you approximately 10-15 minutes. Then, I will, if you provide
me with your email address, interview you either at your ESL program, at any other place
you prefer, over Skype, or via any other online platforms. To participate in the survey, please
respond to the following items. However, at the end of the survey, you will be asked to
provide your email address in case you decide to participate in the interview part. Please
share this email with your friends, classmates, or any other categories that you think fit this
study. I will not collect any video recordings for this research; however, I will, with your
permission, audiotape and take notes during the interview. I might also need to collect one of
All survey responses, interview transcriptions, audio recordings, and writing samples
will be stored on a password-protected drive. The drive will be kept in a locked departmental
office at the University of Arizona for six years past the date of project completion. However,
no one other than (the primary investigator) will have access to the data. Thus, all of your
information and responses will be kept confidential. Also, your name will not be recorded on
the tape as well as will not be associated with any part of the written reports or oral
presentations. Your participation is voluntary. You may refuse to participate in this study.
You may also leave the study at any time. No matter what decision you make, it will not
144
affect your grades or future benefits. Participation Consent: I have read (or someone has read
to me) this form, and I am aware that I am being asked to participate in a research study. I
have had the opportunity to ask questions and have had them answered to my
satisfaction.
- Yes
- No
Q2. Do you allow the researcher to publish your responses for a research project and
potential publication? (Note that your name will not appear in the survey).
- Yes
- No
Q3. Gender?
- Male
- Female
Q4. Which of the following ESL programs are you currently or have previously
attended?
- (Program A).
- (Program B).
- (Program C).
- Others.
Q5. How long have you been studying English abroad in ESL Classes?
145
- 0 – 3 months.
- 4 – 6 months.
- 7 – 12 months.
- Beginner.
- Lower-intermediate.
- Intermediate.
- Upper-intermediate.
- Advanced.
- Others.
Q8. Do you think your placement scores reflected your actual level of language
proficiency?
- Definitely yes
- Probably yes
146
- Probably not
- Definitely not.
Q9. Do you think the content of ESL Placement Test reflected real-life language uses?
- Definitely yes
- Probably yes
- Probably not
- Definitely not.
- Too long.
- Relatively long.
- Relatively short.
Q11. Did you intentionally, in one way or another, perform poorly on the ESL
- Yes.
- No.
Q12. If yes, what are the reasons that made you intentionally place yourself into lower
Q13. If you have any other reasons, please mention them here:
- Myself.
Q15. If you did not have a scholarship, would you still internationally perform poorly?
- Yes.
- No.
- Maybe.
- I don't know.
Q16. Do you think the Saudi Ministry of Education should establish effective
- Definitely yes
- Probably yes
- Probably not
- Definitely not.
Q17. If yes, what are some suggestions that help establish such mechanisms? Please
Q18. If not mentioned above, what are other suggestions do you have?
Q19. Briefly, provide us with some comments about your experience in taking ESL
Q20. Please provide your email if you would you like the researcher to interview you
look forward to meeting with you (either face-to-face or online) to obtain your opinions
about ESL Placement Tests. (Note that your information will be kept confidential).
149
APPENDIX C: INTERVIEW QUESTIONS
1. What are the challenges you faced when you took the ESL placement test?
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
2. Do you think that the ESL placement tests that you have taken are accurate in reporting
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
3. Do you think that the ESL placement test that you had taken placed you in a level
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
4. Did you prefer to be placed into a lower, higher, or the same level of that of your level of
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
150
5. If you had intentionally performed poorly in the ESL placement test, did you benefit from
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
6. After you were placed into lower than-expected levels, were you able to improve your
English better more than if you were placed into a class that fits your level of proficiency?
Please explain.
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
7. If you had not intentionally performed poorly on ESL placement tests, can you tell me why
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
8. If the Saudi Ministry of Education demands you to take a six-month ESL course in Saudi
Arabia prior to coming to the targeted country, would you commit to attending that class?
151
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
9. If the Saudi Ministry of Education demands you to obtain a score of (4.5) in the IELTS or
a score of (40) in the TOEFL prior to studying abroad, would you be willing to do so? Why
___________________________________________________________________________
___________________________________________________________________________
___________________________________________________________________________
10. Do you have any comments that you have not mentioned in the survey?
___________________________________________________________________________
___________________________________________________________________________
152
APPENDIX D: ESL ADMINISTRATORS & ASSESSMENT COORDINATORS
conducting a research study on the placement of Saudi scholarship students into ESL levels.
Based on my data, I found that some ESL Saudi students intentionally perform poorly on
ESL placement tests to be placed into lower ESL classes. This is because some of them need
to have more time learning English, preparing for standardized tests, applying for
universities, or avoiding returning to Saudi Arabia (as they did not obtain unconditional
his findings, Stewart (1999) for example reported that 22% of ESL placement test-takers
were misplaced. This rate has resulted in complaints among test-takers. Moreover, Crusan
(2002) argued that if misplaced, students would have to pay additional financial expenses. In
addition, Kokhan (2014) pointed out that cases of misplacement are likely to make students
fail to significantly improve their English. In my data, I am missing an important part, which
is concerned with the extent to which such behavior (students throwing a test in purpose) can
affect the whole placement system of an ESL program. Therefore, I would like to obtain your
valuable opinions on this issue. Moreover, if you, based on your permission, provide your
email, I will send you the findings of this study. Note: No names of participants or ESL
Yes (1)
No (2)
153
1. Have you ever noticed any incident of students throwing an ESL placement test in
2. How do you identify cases of misplaced students? Please check all that apply.
Teacher feedback.
Others.
3. If you have other means of identifying cases of misplacement, please state them here:
4. How do you usually deal with significant misplacements? Please check all that apply.
Teacher evaluation.
Others.
5. If you have other ways for alleviating misplacements, please state them here:
intentionally perform poorly on the placement test, how would this affect the whole
placement system?
7. If you wish to receive the findings of this study, please provide your email address:
154
APPENDIX E: STUDY PARTICIPANTS
Total
Type of Participants Male Female
50 21 71
Saudi ESL Students
GCC Students
162 54 216
155
APPENDIX F: CEFR, TOEFL, AND IELTS EQUIVALENCY TABLE
0-8 0 - 1.0
A1 19 - 29 2.0 - 2.5
A2 30 - 40 3.0 - 3.5
B1 41 - 52 4.0
B2
B2 65 - 78 5.5 - 6.0
156
APPENDIX G: CEFR SELF-ASSESSMENT RUBRIC DESCRIPTORS
Taken from the Bank of descriptors for self assessment in European Language Portfolios
1. Your First Name and Middle Initial (Again, your name will not appear in the study).
2. Gender:
m Male (1)
m Female (2)
3. What is your current ESL Level? (Note that your level of proficiency must be
m Intermediate (1)
m Upper-intermediate (2)
m Advanced (3)
4. What was your most recently obtained TOEFL score? And when did you take the
test? If you are planning to take the TOEFL very soon, please provide the test date.
157
Welcome ESL Intermediate Students (B1)
Instructions:
I can do this
I cannot do this
a. Please try to assess your level of language proficiency by choosing the appropriate (I can do this or I cannot do
b. If you think you can efficiently do the skill/task that each statement describes, choose the (I can do this) choice.
However, if you think you can't, in a way or another, do it, please select the (I cannot do this) choice.
Listening 1 2
I can follow clearly articulated speech directed at me in everyday conversation, though I sometimes have to ask for repetition of
I can generally follow the main points of extended discussion around me, provided speech is clearly articulated in standard dialect.
I can listen to a short narrative and form hypotheses about what will happen next.
I can understand the main points of radio news bulletins and simpler recorded material on topics of personal interest delivered
I can catch the main points in TV programmes on familiar topics when the delivery is relatively slow and clear.
I can understand simple technical information, such as operating instructions for everyday equipment.
Reading 1 2
I can understand the main points in short newspaper articles about current and familiar topics.
I can read columns or interviews in newspapers and magazines in which someone takes a stand on a current topic or event and
I can guess the meaning of single unknown words from the context thus deducing the meaning of expressions if the topic is familiar.
I can skim short texts (for example news summaries) and find relevant facts and information (for example who has done what and
where).
I can understand the most important information in short simple everyday information brochures.
I can understand simple messages and standard letters (for example from businesses, clubs or authorities).
In private letters I can understand those parts dealing with events, feelings and wishes well enough to correspond regularly with a
pen friend.
I can understand the plot of a clearly structured story and recognise what the most important episodes and events are and what is
I can start, maintain and close simple face-to-face conversation on topics that are familiar or of personal interest.
I can maintain a conversation or discussion but may sometimes be difficult to follow when trying to say exactly what I would like
to.
I can deal with most situations likely to arise when making travel arrangements through an agent or when actually travelling.
I can express and respond to feelings such as surprise, happiness, sadness, interest and indifference.
I can give or seek personal views and opinions in an informal discussion with friends.
Spoken Production 1 2
I can explain and give reasons for my plans, intentions and actions.
I can paraphrase short written passages orally in a simple fashion, using the original text wording and ordering.
Strategies 1 2
I can repeat back part of what someone has said to confirm that we understand each other.
I can ask someone to clarify or elaborate what they have just said.
When I can’t think of the word I want, I can use a simple word meaning something similar and invite ”correction”.
Language Quality 1 2
I can keep a conversation going comprehensibly, but have to pause to plan and correct what I am saying – especially when I talk
I can convey simple information of immediate relevance, getting across which point I feel is most important.
I have a sufficient vocabulary to express myself with some circumlocutions on most topics pertinent to my everyday life such as
Writing 1 2
I can write simple connected texts on a range of topics within my field of interest and can express personal views and opinions.
I can write simple texts about experiences or events, for example about a trip, for a school newspaper or a club newsletter.
159
I can write personal letters to friends or acquaintances asking for or giving them news and narrating events.
I can describe in a personal letter the plot of a film or a book or give an account of a concert.
In a letter I can express feelings such as grief, happiness, interest, regret and sympathy.
I can reply in written form to advertisements and ask for more complete or more specific information about products (for example a
I can convey – via fax, e-mail or a circular – short simple factual information to friends or colleagues or ask for information in such
a way.
160
Welcome ESL Upper-intermediate Students (B2)
Instructions: Students
I can do this
I cannot do this
a. Please try to assess your level of language proficiency by choosing the appropriate (I can do this or I cannot do
b. If you think you can efficiently do the skill/task that each statement describes, choose the (I can do this) choice.
However, if you think you can't, in a way or another, do it, please select the (I cannot do this) choice.
Listening 1 2
I can understand in detail what is said to me in standard spoken language even in a noisy environment.
I can follow a lecture or talk within my own field, provided the subject matter is familiar and the presentation straightforward and
clearly structured.
I can understand most radio documentaries delivered in standard language and can identify the speaker’s mood, tone, etc.
I can understand TV documentaries, live interviews, talk shows, plays and the majority of films in standard dialect.
I can understand the main ideas complex speech on both concrete and abstract topics delivered in a standard dialect, including
I can use a variety of strategies to achieve comprehension, including listening for main points; checking comprehension by using
contextual clues.
Reading 1 2
I can rapidly grasp the content and the significance of news, articles and reports on topics connected with my interests or my job,
I can read and understand articles and reports on current problems in which the writers express specific attitudes and points of
view.
I can understand in detail texts within my field of interest or the area of my academic or professional speciality.
I can understand specialised articles outside my own field if I can occasionally check with a dictionary.
I can read reviews dealing with the content and criticism of cultural topics (films, theatre, books, concerts) and summarise the
main points.
I can read letters on topics within my areas of academic or professional speciality or interest and grasp the most important points.
I can quickly look through a manual (for example for a computer program) and find and understand the relevant explanations and
plot.
Spoken Interaction 1 2
I can initiate, maintain and end discourse naturally with effective turn-taking.
I can exchange considerable quantities of detailed factual information on matters within my fields of interest.
I can convey degrees of emotion and highlight the personal significance of events and experiences.
I can engage in extended conversation in a clearly participatory fashion on most general topics.
I can account for and sustain my opinions in discussion by providing relevant explanations, arguments and comments.
I can help a discussion along on familiar ground confirming comprehension, inviting others in, etc.
I can carry out a prepared interview, checking and confirming information, following up interesting replies.
Spoken Production 1 2
I can give clear, detailed descriptions on a wide range of subjects related to my fields of interest.
I can understand and summarise orally short extracts from news items, interviews or documentaries containing opinions,
I can understand and summarise orally the plot and sequence of events in an extract from a film or play.
I can explain a viewpoint on a topical issue giving the advantages and disadvantages of various options.
Strategies 1 2
I can use standard phrases like ”That’s a difficult question to answer” to gain time and keep the turn while formulating what to
say.
I can make a note of ”favourite mistakes” and consciously monitor speech for them.
I can generally correct slips and errors if I become conscious of them or if they have led to misunderstandings.
Language Quality 1 2
I can produce stretches of language with a fairly even tempo; although I can be hesitant as I search for patterns and expressions,
I have sufficient vocabulary to express myself on matters concerned to my field and on most general topics.
I can communicate with reasonable accuracy and can correct mistakes if they have led to misunderstandings.
Writing 1 2
162
I can write clear and detailed texts (compositions, reports or texts of presentations) on various topics related to my field of
interest.
I can discuss a topic in a composition or “letter to the editor”, giving reasons for or against a specific point of view.
I can develop an argument systematically in a composition or report, emphasising decisive points and including supporting
details.
I can write about events and real or fictional experiences in a detailed and easily readable way.
I can express in a personal letter different feelings and attitudes and can report the news of the day making clear what – in my
163
Welcome ESL Advanced Students (C1)
Instructions:
I can do this
I cannot do this
a. Please try to assess your level of language proficiency by choosing the appropriate (I can do this or I cannot do
b. If you think you can efficiently do the skill/task that each statement describes, choose the (I can do this) choice.
However, if you think you can't, in a way or another, do it, please select the (I cannot do this) choice.
Listening 1 2
I can follow extended speech even when it is not clearly structured and when relationships are only implied and not signalled
explicitly.
I can understand a wide range of idiomatic expressions and colloquialisms, appreciating shifts in style and register.
I can extract specific information from even poor quality, audibly distorted public announcements, e.g. in a station, sports
stadium etc.
I can understand complex technical information, such as operating instructions, specifications for familiar products and services.
I can understand lectures, talks and reports in my field of professional or academic interest even when they are propositionally
I can without too much effort understand films employing a considerable degree of slang and idiomatic usage.
Reading 1 2
I can understand fairly long demanding texts and summarise them orally.
I can read complex reports, analyses and commentaries where opinions, viewpoints and connections are discussed.
I can extract information, ideas and opinions from highly specialised texts in my own field, for example research reports.
I can understand long complex instructions, for example for the use of a new piece of equipment, even if these are not related to
164
I can read contemporary literary texts with ease.
I can go beyond the concrete plot of a narrative and grasp implicit meanings, ideas and connections.
Spoken Interaction 1 2
I can use the language fluently, accurately and effectively on a wide range of general, professional or academic topics.
I can use language flexibly and effectively for social purposes, including emotional, allusive and joking usage.
I can express my ideas and opinions clearly and precisely, and can present and respond to complex lines of reasoning
convincingly.
Spoken Production 1 2
I can give an extended description or account of something, integrating themes, developing particular points and concluding
appropriately.
I can give a clearly developed presentation on a subject in my fields of personal or professional interest, departing when
necessary from the prepared text and following up spontaneously points raised by members of the audience.
Strategies 1 2
I can use fluently a variety of appropriate expressions to preface my remarks in order to get the floor, or to gain time and keep
I can substitute an equivalent term for a word I can’t recall without distracting the listener.
Language Quality 1 2
I can express myself fluently and spontaneously, almost effortlessly. Only a conceptually difficult subject can hinder a natural,
I can produce clear, smoothly-flowing, well-structured speech, showing control over ways of developing what I want to say in
order to link both my ideas and my expression of them into coherent text.
I have a good command of a broad vocabulary allowing gaps to be readily overcome with circumlocutions; I rarely have to
search obviously for expressions or compromise on saying exactly what I want to.
I can consistently maintain a high degree of grammatical accuracy; errors are rare and difficult to spot.
165
Writing 1 2
I can express myself in writing on a wide range of general or professional topics in a clear and user-friendly manner.
I can present a complex topic in a clear and well-structured way, highlighting the most important points, for example in a
composition or a report.
I can present points of view in a comment on a topic or an event, underlining the main ideas and supporting my reasoning with
detailed examples.
I can put together information from different sources and relate it in a coherent summary.
I can give a detailed description of experiences, feelings and events in a personal letter.
I can write formally correct letters, for example to complain or to take a stand in favour of or against something.
I can write texts, which show a high degree of grammatical correctness and vary my vocabulary and style according to the
166
APPENDIX H: INTERVIEW QUESTIONS
3. During learning English, what was/were the most useful/accurate technique(s) that have helped
you identify your levels of proficiency?
4. In this study, you have obtained two different scores (self-rated and TOEFL), which one do you
think is more accurate in reporting your actual levels of proficiency?
5. Have you ever used a self-assessment technique before this study? If yes, how was it?
6. Tell me about your experience in using the self-assessment rubric in this study?
7. Would you like to use this assessment rubric (or a similar one) in classrooms? Why and why not?
8. Do you think you can assess your English skills using self-assessment tools? Why or why not?
9. When you used CEFR self-assessment rubric in this study, were you able to detect your areas of
strengths and weaknesses? How?
10. When you used the CEFR self-assessment rubric in this study, did you find it difficult to use?
11. When you used CEFR self-assessment rubric in this study, did you find any difficulty assessing
yourself in one particular area/skill? Would you explain it?
12. Please provide some advantages and disadvantages of using self-assessment techniques?
13. Finally, if you support using self-assessment techniques in classrooms, do think it should be used
alone? Or it should be supported with other assessment techniques? Why?
167
APPENDIX I
My name is Adnan Alobaid, a doctoral candidate, in the Second Language Acquisition and Teaching
I am offering you a one-time free proofreading of one of your CESL papers (10 pages or less) if you
participate in this study. To participate in this study, you would have to:
1) Assess your level of English proficiency using an online CEFR rubric, which will take you
2) After you assess yourself, you will, upon your permission, provide me with your TOEFL scores
once you obtain them (Note that you have the right to refuse to do so).
https://goo.gl/nEFoWY
https://goo.gl/0K153y
168
If you are in Level (7) or Bridge Program, please click on this link:
https://goo.gl/EietQm
Important Notes:
1. Your name, TOEFL scores, or any other private information will not appear in the study.
2. Do not forget to include your email so that we can arrange for proofreading.
3. Keep in mind that proofreading will be done for the first 50 participants ONLY. Therefore, it is highly
adnanalobaid@email.arizona.edu
Yours,
Adnan Alobaid.
169
APPENDIX J: COLOR-CODED CEFR, TOEFL, AND IELTS EQUIVALENCY TABLE
170
APPENDIX K: PARTICIPANTS’ SELF-RATED SCORES
171
APPENDIX L: CONSENT FORM
My name is Adnan Alobaid, a doctoral candidate, in the Second Language Acquisition and Teaching
program at the University of Arizona, Tucson. I am conducting a research study for using alternative
assessment approaches for proficiency purposes. To participate in this study, you would have to:
1) Assess your level of English proficiency using an online CEFR rubric, which will take you
approximately 15-20 minutes.
2) Upon your permission, provide me with your TOEFL scores once you obtain them (Note that you have
the right to refuse to do so.
3) And finally be engaged in a face-to-face or online interview at your discretion to allow you to expand
more on your experience in and attitudes towards using self-assessment rubrics.
All self-rated scores, interview transcriptions, and audio recordings will be stored on a password-
protected drive that will be kept in a locked departmental office at UA for six years past the date of
project completion (May. 2016). However, no one other than the primary investigator (me) will have
access to the data. In case we are unable to arrange for a face-to-face meeting, we can carry out the
interview over Skype. However, The interview will be audio recorded after obtaining your consents. In
case you choose not to be recorded, you can still participate in the interview part without being recorded
under any circumstances. Moreover, if you refuse to participate in the interviews, your participation in the
first round of data collection is still highly appreciated and adequate for the study.
Your participation is voluntary. You may refuse to participate in this study. You may also leave the study
at any time. No matter what decision you make, it will not affect your grades or future benefits.
Participation Consent
I have read and I am aware that I am being asked to participate in a research study. If you have any
questions about your rights are a research participant, please contact the Human Subjects Protection
Program at 520-626-6721 or at www.orcr.arizona.edu/hspp or you can contact me via: (818) 614-4522,
adnanalobaid@email.arizona.edu
By taking part in the survey, you are agreeing to have your responses used as part of research.
172
APPENDIX M: WEB-BASED STUDENT SURVEY
Statements4
Practices
Course Objectives
objectives clearly.
requirements in advance.
4
The statements have been developed from the NCAAA and CEA standards.
173
assignments help you master the targeted
objectives.
measurable.
Teaching Strategies
for you.
performance constantly.
174
Student Learning Outcomes
the course
learning outcomes.
skills.
175
have emphases on real-life language uses
Student Assessment
the course
courses.
objectives?
course objectives.
176
APPENDIX N: INTERVIEWS5
a. Curriculum:
1. To what extent does the curriculum fulfill your language needs? Please explain?
2. How well does the curriculum materials reflect real-life language uses and situations?
3. How well is the curriculum structured with regard to students’ level-to-level progression?
4. How well does the curriculum prepare you for your future career as stated in course goals?
b. Teaching Strategies:
2. How often are you engaged in activities such as leading discussions, presentations etc.?
1. To what extent are you provided with professional counseling whenever needed?
2. To what extent are you provided with academic advising during your study?
1. How accurate are course assessment methods in assessing the course-learning objective?
5
All interview questions have been developed based on the NCAAA and CEA standards.
177
3. To what extent does your instructor use a variety of different assessment methods?
4. Are you given the opportunities to demonstrate your English skills well? If yes, How?
5. Do the assessment results provide you with areas of strengths and weaknesses?
6. Are you provided with a proficiency scale that shows your levels of language proficiency?
a. Learning Outcomes
1. How are student intended learning outcomes defined? Please explain it in detail?
2. When defining intended learning outcomes, do you ensure that they are aligned with the Saudi
4. Are intended learning outcomes communicated to students at the beginning of the course?
6. How often do you review and update student learning outcomes of your courses
b. Student Assessment
3. How often do you integrate alternative assessment techniques into assessment processes?
4. How often do you provide students with informative feedback to their performance?
5. What mechanisms do you use to respond to students’ complaints about their grades?
6. Do you use direct and indirect assessment tools to assess student learning outcomes?
c. Teaching Strategies:
1. What mechanisms do you use to ensure that teaching strategies are appropriate for various
learning styles?
178
3. What strategies do you use to facilitate learning activities for students (e.g. modeling,
scaffolding etc.)?
collaborative/cooperative learning?
d. Faculty Development
1. How often does the Program offer you general training programs (e.g. teaching skills)?
2. How often do you integrate technologies into your classroom teaching activities?
3. Do you maintain a portfolio that shows evidence of the course you teach?
4. Do you think the Program administrators provide you with adequate time and opportunity for
5. Are you engaged in heavy teaching and/or administrative workloads? If yes, to what extent do
6. Generally, what are some challenges you face in teaching the Program courses?
a. Program Mission:
1. Has the program mission been approved by the University Senior Administration?
2. Has the Program mission been developed after obtaining consultation of stakeholders?
179
1. How do you undertake faculty recruitment processes?
2. What mechanisms do you use to ensure that prospective faculty members have appropriate
3. What strategies do you undertake to encourage teaching faculty members to develop strategies
6. Are faculty members given the opportunities to attend and participate in conferences?
1. Do you have a plan for the Program development and evaluation? Please explain it?
2. What mechanisms do you use to evaluate the Program (e.g. surveys, employment data)?
3. How often do you do so for the Program courses, learning outcomes, and study plan?
4. Do you invite professional people, from fields relevant to the Program areas of specialization,
5. Does the program have KPIs that include learning outcomes? How are they determined?
180
APPENDIX O: RESEARCHER’S EVALUATION CHECKLIST
College Mission:
__________________________________________________________________________________
__________________________________________________________________________________
College Goals
College Objectives
181
Program Mission
__________________________________________________________________________________
__________________________________________________________________________________
Program Goals
Program Objectives
182
RESEARCHER’S EVALUATION CHECKLIST CONT.
Program Mission
affiliated?
emphasis on research?
Program Goals
6
The statements have been developed from the NCAAA and CEA standards.
183
mission?
the curriculum?
Program Objectives
specific?
related?
Program’s goals?
assessable?
Curriculum
program’s mission?
184
from one level to the next?
needs of students?
techniques?
measurable
objectives?
Teaching Strategies
185
course specifications?
activities?
for students?
groupwork?
students’ performance?
written?
186
measurable?
attainable?
cognitive skills?
affective skills?
Communication, Information
result of learning?
other?
187
Assessment Methods
methods used?
language skills?
188
APPENDIX P: COLLEGE MISSION
The college aims at preparing the prospective teacher who is religiously committed; the one is
attached strongly to his homeland. This teacher should be a good model for his students in his
work and behavior. He should be at a high level of professionalism. Learning instinct and love
for the career should be part and parcel of him. He should be completely experienced and fully
aware of his role in facilitating learning approaches. He should be continuously developing in his
field of specialty as well as his teaching styles. He should have the traits of the strong leader who
has the ability to convince others and to prove that his opinion is correct. He should attain the
capacity of making a decision and of taking on responsibility. He should be able to plan well and
to put into consideration individual differences among students. He should be a good thinker
possessing all types of thinking skills. He should be able to develop these skills in his students.
He should have the traits of a social pioneer who has the ability to communicate effectively with
his society and to cooperate in solving its problems. He should be ever renewing in how to deal
with modern technology and how to function it properly in all instructional settings. He should
be a guide taking charge of directing the instructional process in fulfilling its targets and solving
its problems.
189
APPENDIX Q: MALE DEPARTMENT OBJECTIVES
190
REFERENCES
Adam, S. (2004). Using learning outcomes. In Report for United Kingdom Bologna Seminar (pp.
1-2).
Ahlgren Reddy, A., & Harper, M. (2013). Mathematics placement at the University of Illinois.
Alamri, M. (2011). Higher education in Saudi Arabia. Journal of Higher Education Theory and
Alderson, J. C., & Scott, M. (1992). Insiders, outsiders and participatory evaluation. Evaluating
Alderson, J. C., Clapham, C., & Wall, D. (1995). Language test construction and evaluation.
Alderson, J. C., Figueras, N., Kuijper, H., Nold, G., Takala, S., & Tardieu, C. (2004). The
Assessment: Reading and Listening: Final report of The Dutch CEF Construct Project.
Andrade, H. & Valtcheva, A. (2009). Promoting learning and achievement through self-
191
universities. International Journal of Research Studies in Language Learning, 3(5).
University Press.
Bachman, L. F., Davidson, F., & Ryan, K. (1995). An investigation into the comparability of
two tests of English as a foreign language (Vol. 1). Cambridge University Press.
Bachman, L. & Palmer, A. (1996). Language testing in practice. New York: Oxford
University Press.
(Eds.), Learner-directed assessment in ESL (pp. ix-xii). New Jersey: Lawrence Erlbaum
Associates, Inc.
Bachman, L. F. (2005). Building and supporting a case for test use. Language Assessment
teaching, 706.
Publishing, http://search.proquest.com/docview/1369866651.
192
Barker, R. (2003). The social work dictionary. Washington, DC: NASW Press.
Benson, P. (1995) 'A critical view of learner training'. Learning: JALT Learner
Benson (2012) in Burns, A., & Richards, J. C. (Eds.). (2012). The Cambridge guide to
pedagogy and practice in second language teaching. New York: Cambridge University
Press.
Beretta, A. (1986). Program‐Fair Language Teaching Evaluation. TESOL Quarterly, 20(3), 431-
444.
Black, P., & D. Wiliam. (2009). Developing the theory of formative assessment. Educational
Bloom, B. S., Madaus, G. F., & Hastings, J. T. (1981). Evaluation to improve learning. New
York: McGraw-Hill.
Bock, R. D. (1997). A brief history of item theory response. Educational Measurement: Issues
Bourke, B. (2014). Positionality: Reflecting on the research process. The Qualitative Report,
19(33), 1-9.
193
programs: A review of the issues. Language Testing, 15(1), 45-85.
Johnson (ed.) The second language curriculum. Cambridge, UK: Cambridge University
Press, 222-241.
Brown, J. D. (1989). Improving ESL placement test using two perspectives. TESOL
Brown, D. (1995). Language Program Evaluation: Decisions, Problems and Solutions. Annual
Brown, J. D. (1996). Testing in language programs. New Jersey: Prentice Hall Regents.
Brown, D. (1997). Computers in language testing: present research and some future
Brown, J. D., & Hudson, T. (2002). Criterion-referenced language testing. UK: Cambridge
University Press.
Butler, Y. G., & Lee, J. (2010). The effects of self-assessment among young learners of
Cargan, L. (2007). Doing social research. Maryland: Rowman & Littlefield Publishers.
Carmines, E. G., & Zeller, R. A. (1979). Reliability and validity assessment. Beverly Hills, CA:
194
Sage publications.
CESL (2014). About Center for English a Second Language. Retrieved March 17th, 2016,
from http://www.cesl.arizona.edu/content/about-cesl
19, 254-272.
Chapelle, C. A., Jamieson, J., & Hegelheimer, V. (2003). Validation of a web-based ESL test.
Chapelle, C. A., Chapelle, C. A., Enright, M. K., & Jamieson, J. M. (Eds.). (2011). Building a
validity argument for the Test of English as a Foreign LanguageTM. New York:
Routledge.
Clapham, C. (2000). Assessment and testing. Annual Review of Applied Linguistics, 20, 147-
161.
Conway, T., Mackay, S., & Yorke, D. (1994). Strategic planning in higher education: Who are
Council of Europe.
195
Council of Europe (2001). Common European Framework of Reference for Languages:
Council of Europe (2015). Education and Languages, Language Policy. Retrieved March
Cronbach, L. (1963). Course improvements through evaluation. The Teachers College Record,
64(8), 672-672.
Cumming, A. (2009). Language assessment in education: Tests, curricula, and teaching. Annual
Darandari, E. Z., Al‐Qahtani, S. A., Allen, I. D., Al‐Yafi, W. A., Al‐Sudairi, A. A., &
Catapang, J. (2009). The quality assurance system for post‐secondary education in Saudi
Dornyei, Z. (2001). Teaching and researching motivation. London: Pearson Education Ltd.
Education Accreditation.
Eckes, T., & Grotjahn, R. (2006). A closer look at the construct validity of C-tests. Language
Educational Testing Service. (2010). Linking TOEFL iBT scores to IELTS scores: A
196
research report. Princeton, NJ: ETS. Retrieved February 21st, 2016, from
https://www.ets.org/s/toefl/pdf/linking_toefl_ibt_scores_to_ielts_scores.pdf
Ekbatani, G., & Pierson, H. D. (Eds.). (2000). Learner-directed assessment in ESL. Mahwah, NJ:
Erlbaum.
Ellis, R. (1993). Quality Assurance for University Teaching. Buckingham: Open University
Press.
Elton, L. 1999. Facilitating change through self-assessment. Paper presented at the national
conference teaching and learning for the new millennium: Facilitating change through
Fox, J. D. (2009). Moderating top-down policy impact and supporting EAP curricular renewal:
Fulcher, G. (1997). An English language placement test: issues in reliability and validity.
Fulcher, G., & Davidson, F. (2007). Language testing and assessment. London, NY:
Routledge.
197
Garrison, D. R. (1997). Self-directed learning: Toward a comprehensive model. Adult
Gass, S. M., & Varonis, E. M. (1994). Input, interaction, and second language production.
Gitlin, A., & Smyth, J. (1989). Teacher evaluation: Educational alternatives. Lewes: Falmer
Press.
Green, A. Weir, J. (2004). Can placement tests inform instructional decisions? Language
Green, C. (2005). Integrating extensive reading in the task-based curriculum. ELT Journal,
59(4), 306-311.
Halliday, J. (1994). Quality in education: Meaning and prospects. Educational Philosophy and
51(1), 12-20.
Holland, P. W., & Dorans, N. J. (2006). Linking and equating. Educational measurement, 4,
187-220.
Holec, H. (1981). Autonomy and foreign language learning. Oxford: Pergamon Press.
Hughes, A., Weir C. and Porter, D. (1995). The Global Placement Test. Reading: Centre for
Hughes, A. (2003). Testing for language teachers, (2nd edition). Cambridge: Cambridge
198
University Press.
Huhta, A., Luoma, S., Oscarson, M., Sajavaara, K., & Teasdale, A. (2002). DIALANG, A
Publications/Open-Doors/Data/Intensive-English-Programs/Leading-Places-of-
Origin/2013-14
Chung, & J. Xu (Eds.), Towards Adaptive CALL: Natural Language Processing for
Diagnostic Language Assessment (pp. 117-131). Amer, IA: Iowa Scaner University.
Javid, C. Z., Al-thubaiti, T. S., & Uthman, A. (2013). Effects of English language proficiency
Johnson, D. (2015). Saudi students and IEP teachers: converging and diverging perspectives
Jones, N. (2002). Relating the ALTE Framework to the Common European Framework of
Reference, In: Council of Europe (Eds). Case Studies on the use of the Common
Press.
Kahn, A. B., Butler, F. A., Weigle, S. C., & Sato, E. Y. (1994). Adult ESL Placement Procedures
199
in California: A Summary of Survey Results. Adult ESL Assessment Project.
Kaplan, R. B. (2010). Whence applied linguistics: The twentieth century. Oxford handbook of
applied linguistics, 3-33. Kiely, R., & Rea-Dickens, P. (2005). Program evaluation in
Kirsch, I., Jamieson, J., Taylor, C., & Eignor, D. (1998). Computer familiarity among TOEFL
Kunnan, A. (2000). Fairness and justice for all. Fairness and validation in language
Associates.
Lantolf, J. P., & Poehner, M. E. (2008). Dynamic assessment. In Encyclopedia of language and
University Press.
Lee, Y. J., & Greene, J. (2007). The Predictive Validity of an ESL Placement Test A Mixed
200
Methods Approach. Journal of Mixed Methods Research, 1(4), 366-389.
Lewis, J. (1990) 'Self-assessment in the classroom: a case study'. In G. Brindley (ed.) The
Second Language Curriculum in Action. Sydney: National Centre for English Language
Little, D., Lazenby Simpson, B., & O’Connor, F. (2002). Meeting the English language needs
Little, D. (2006). The Common European Framework of Reference for Languages: Content,
Long, M. H. (1981). Input, interaction, and second‐language acquisition. Annals of the New
MacKay, R. (1998). Program evaluation and quality control. TESL Canada Journal, 5(2), 33-42.
McDonald, B., & Boud, D. (2003). The impact of self-assessment on achievement: the effects
McNamara, M., & Deane, D. (1995). Self-assessment activities toward autonomy in language
McNamara, F. & Roever, C. (2006). Language Testing: The Social Dimension. Oxford:
Blackwell.
Menges, R. J., & Weimer, M. (1996). Teaching on Solid Ground: Using Scholarship to Improve
201
Practice. San Francisco: Jossey-Bass Inc.
(Eds.) A handbook for language program administrators (2nd Edition), (pp. 117-136).
Messick, S. (1996). Validity and washback in language testing. Language Testing. 13, 241-
256.
Ministry of Economy and Planning (2015). The Saudi Ninth Development Plan. Retrieved
http://services.mep.gov.sa/themes/GoldenCarpet/index.jsp#1457986308681
Mitchell, C. B. & Vidal, K. E. (2001). Weighing the Ways of the Flow: Twentieth Century
MOHE. (2015). Saudi Higher Education Institutions. Retrieved March 8th, 2016, from
http://www.moe.gov.sa/ar/Pages/default.aspx
MOHE. (2014). The History of Study Abroad Scholarships in the Kingdom. Retrieved March
Public-relations/BooksList/book2eng.pdf
Mohr, L. B. (1995). Impact analysis for program evaluation. Thousand Oaks, CA: Sage
Publications, Inc.
202
Morphew, C. C., & Hartley, M. (2006). Mission statements: A thematic analysis of rhetoric
NCAAA. (2013). Standards for Quality Assurance and Accreditation of Higher Education
http://www.ncaaa.org.sa/en/Releases/Pages/Handbooks.aspx
NCAAA. (2014). Mission, Vision & Values. Retrieved March 08, 2016, from
http://www.ncaaa.org.sa/en/AboutUs/Pages/Vision.aspx
Norris, J.M., Davis, J.M., Sinicrope, C. & Watanabe, Y. (Eds.). (2009) Toward Useful Program
evaluation: Introduction to the special issue. Language Teaching Research, 13(1), 7-13.
Norris, J. M. (2016). Language Program Evaluation. The Modern Language Journal, 100(S1),
169-189.
North, B. (2007). The CEFR illustrative descriptor scales. The Modern Language Journal,
91(4), 656-659.
Nunan, D. (1989). Designing tasks for the communicative classroom. Cambridge, England:
203
Clapham and D. Corson (Eds), Encyclopedia of language and education, Volume 7:
Academic.
Posavac, E. (2015). Program evaluation: Methods and case studies. New York: Pearson Prentice
Hall.
Rea-Dickins, P. (1994). Evaluation and English language teaching. Language Teaching, 27(02),
71-91.
Reed, D. J., & Stansfield, C. W. (2004). Using the Modern Language Aptitude Test to Identify a
1(2-3), 161-176.
Reid, J. M. (1995). Learning Styles in the ESL/EFL Classroom. Florence, KY: Heinle & Heinle.
Roever, C. (2001). Web-based language testing. Language Learning & Technology, 5(2), 84-94.
Royse, D., Thyer, B., & Padgett, D. (2009). Program evaluation: An introduction. Belmont, CA:
Brooks-Cole.
Saba, M. S. (2014). Writing in a New Environment: Saudi ESL Students Learning Academic
https://vtechworks.lib.vt.edu/bitstream/handle/10919/54012/Saba_MS_D_2014.pdf?sequen
ce=1&isAllowed=y
SACM (2013). Overview on the ESL Department. Retrieved March 15th, 2016, from
204
http://www.sacm-usa.gov.sa/Departments/ESL/about.aspx
SACM (2013). Recommended ESL Schools in the U.S. Retrieved March 18th, 2016, from
http://esllist.sacm.org/
Research Association.
Sekely, A. (2014). Assessing the Validity of the Trauma Symptom Inventory on Military Patients
content/uploads/2014/09/Thesis05.04.2014.pdf
Shohamy, E., Gordon, C., & Kraemer, R. (1992). The effect of raters’ background and training
on the reliability of direct writing tests. The Modern Language Journal, 76 (1), 27-33.
Smith, L., & Abouammoh, A. (2013). Higher Education in Saudi Arabia. The Netherlands:
Springer.
Tannenbaum, R. J., & Wylie, E. C. (2008). Linking English-Language Test Scores Onto the
Taras, M. (2001). The use of tutor feedback and student self-assessment in summative
205
assessment tasks: Toward transparency for students and for tutors. Assessment and
Taras, M. (2008). Issues of power and equity in two models of self-assessment. Teaching in
Taylor, C. Albasri, W. (2014). The Impact of Saudi Arabia King Abdullah’s Scholarship
TESOL (2010). Position Statement on the Acquisition of Academic Proficiency in English at the
Van Damme, D. (2004). Standards and Indicators in Institutional and Program Accreditation in
Education, 127-159.
van Teijlingen, E., & Hundley, V. (2001). The importance of pilot studies. Social research
Wall, D., & Alderson, J. C. (1993). Examining washback: the Sri Lankan impact study.
Wall, D., Clapham, C., & Alderson, J. C. (1994). Evaluating a placement test. Language
206
11(2), 197-223.
23(2), 183-194.
Westerheijden, D. F., Stensaker, B., & Rosa, M. J. (2007). Quality assurance in higher
Wiggins, G. (1993). Assessment: Authenticity, context, and validity. Phi Delta Kappan,
75(3), 200-08.
2004&res_dat=xri:pqdiss&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&rft_dat=xri:pqdi
ss:3346271
207