Professional Documents
Culture Documents
Orbe DevelopingaCEFR Basedcan DoStatementsforListening - IMECFINALdoc
Orbe DevelopingaCEFR Basedcan DoStatementsforListening - IMECFINALdoc
Abstract
Listening comprehension skills are the most important skills in communication. However,
it is often neglected in English teaching and learning. To measure the English language
listening comprehension skills in samples of non-native-English speakers, a standard
listening sample test was administered to first-year university students in the Philippines.
Using item analysis, difficult items were analyzed. Results show that students’ difficulties
focused on the responses to tag questions, idioms, and making inferences and
generalizations. Informed by the findings of the study, a Can-Do Self-Evaluation for
Listening Skills, adapted from the Common European Framework of Reference for
Languages (CEFR) was developed and pilot-tested. This study recommends the inclusion
of activities, training, and assessment in listening across the curriculum to enhance the
listening comprehension skills of students.
Keywords: Item analysis, listening comprehension, receptive skills, CEFR, listening skills
inventory, self-evaluation, TOEIC
Introduction
Diagnosing listening comprehension skills is rarely done in the classrooms (Lynch, 2011;
Vandergrift, 2007). Generally, listening comprehension skills are tested along with
reading comprehension in language classes. For employment purposes and entry to
international schools, the admission requirements include definite levels of language
efficiency measured through the International English Language Testing System (IELTS),
1
International Maritime English Conference
Test of English to Speakers of a Foreign Language (TOEFL), or the Test of English for
International Communication (TOEIC). These tests are designed to assess the language
ability of non-native speakers of English who intend to study or work where English is
the language of communication.
In other countries, the CEFR levels of language proficiency are widely used. The CEFR
distinguishes 34 scales for listening, reading, speaking and writing, which cover both
general competences and communicative language competences for a range of domains to
define six levels of proficiency (Little, 2007). The CEFR classifies language proficiency
and divides proficiency into three categories with a total of six levels (A1, A2, B1, B2,
C1, and C2).
The CEFR has contributed to the innovation of language teaching in many important
ways. Focusing on providing language learners a tool for self-evaluation, it has promoted
a major change in the perception of language proficiency from a teacher-oriented product
to a learner-oriented process (Runnels, 2016). Asian countries like Japan, China,
Malaysia, Taiwan, Vietnam, among others have developed CEFR-based language
proficiency level. Thus, as a component of Association of Southeast Asian Nations
(ASEAN) integration, the Philippines may need to assess the language proficiency levels
of maritime students based on the CEFR. Using the CEFR has been noted to allow
students to define their own learning abilities and plan the direction of their learning
(COE, 2001; Glover, 2011; Little, 2006). Specific focus could be given to assessing the
listening skills of students given the dearth of studies focussing on this receptive skill.
Realizing the importance of developing listening among the students as preparation for
the workplace, this study is focused on diagnosing the listening comprehension skills of
university students, identifying the areas of improvement, and developing a list of Can-
2
International Maritime English Conference
The current paper reports on a project which aims to standardize the assessment of
language production in the academic context. This project resulted from the innovation of
the language teaching curriculum in agreement with the CEFR guidelines. The aim of the
project was to produce a set of listening learning targets that are representative for the
different components of the CEFR levels in the local and maritime contexts, which could
then be used for standardization of a listening assessment. Although this project was not
set up as an empirical evaluation of the CEFR, attempts to embed the CEFR into the day-
to-day teaching practice serve as a relevant verification of its feasibility. This study is
guided by the following research questions:
1. What is the current level of listening comprehension skills of the students?
2. What are the focus areas in listening that should be developed among the
students?
3. How accurate are the self-assessment evaluations of the students?
This study forms part of an on-going project of the College of Marine Transportation to
provide baseline data and explore innovations for enhancement of language proficiency
of its students.
2. Methods
A total of 481 first year students from a private tertiary university participated in this
study. Participants were unfamiliar with performing self-assessment using can-do
statements and the CEFR. All the participants have taken the Test of English for
International Communication (TOEIC) for the first time.
This study was conducted in two phases. First, a standard Test of English for International
Communication (TOEIC) listening test was given to the participants. The second phase
pertained to a self-assessment through the Can-Do Statements in Listening. Participants
indicated the extent of their agreement to 40 ordered can-do statements from the CEFR on
a five-point Likert scale. For the five-point scale, the categories were 0=I can’t do this at
all; 1=I can do this with a lot of help from my instructor; 2=I can do this a little; 3= I can
do this fairly well; 4= I can do this really well; and 5=I can do this almost perfectly. The
survey was administered through a questionnaire.
To determine the level of listening proficiency among the students, the listening test
scores of the students in the 100-item test were converted into a TOEIC Listening Score.
Descriptive statistics was used to describe the data. The mean scores of the students were
compared.
Item analysis was run to determine the difficulty indices of each item. Based on their
TOEIC listening scores, difficult items were identified and examined for possible causes
of difficulties. Items with difficulty indices were categorized in terms of linguistic focus
areas, themes or topics, and skills targeted.
3
International Maritime English Conference
Results and findings of this study are presented in the order of the research questions as
follows.
1. What is the current level of listening comprehension skills of the students?
The scores of the participants were well distributed (M=58.94, SD=13.75). Figures 1
shows the frequency distribution of scores. The median score is 60; the highest score is
92 and the lowest score is 21. About 2% of the participants comprise the top scorers while
1 percent comprise the low performers. The mean score of the 481 participants is 59.
About 19% of the participants got scores between 53-60. More than a quarter of the group
(23%) got a score within the range of 61-68.
4
International Maritime English Conference
These listening scores were converted into TOEIC Listening Scores as shown in Figure 2.
With these scores, the current level of listening comprehension skills of the students is A2
based on the conversion table provided by Waikato Institute of Education (2019).
Students at this level can “understand phrases and the highest frequency vocabulary
related to areas of most immediate personal relevance (e.g. very basic personal and
family information, shopping, local area, employment).” Moreover, students at this level
can “catch the main point in short, clear, simple messages and announcements” (COE,
2001).
This result seems to suggest that the level of listening skills of the participants is below
the general requirement and that there is a need to train students in order to enhance their
listening comprehension skills.
2. What are the focus areas in listening that should be developed among the students?
Table 1 presents the topics, communicative activity, and skills focus of the test. As used
in the TOEIC test, focus area pertains to the tasks involved in the listening test, such as
picture clues, questions and answers, dialogues or conversations between two speakers, or
short talks by a single speaker. Common topics tackled in the listening test pertain to
common personal activities like shopping, calling, listening to announcements and
instructions, or office matters. The types of communicative activities used in the analysis
were adapted from the listening competencies described in the CEFR can-do statements.
To determine the difficulty indices of the test items, an Item Analysis matrix was
generated. The listening test registered a reliability index of 88%. Out of the 100 items,
42 were considered average questions; 36 were easy questions, and 9 were very easy
items. About 12 items were found to be difficult and one item considered as very difficult
for the participants based on their listening performance (p=0.1476~0.3971). Items with
difficult and very difficult indices were selected and qualitatively analyzed. The results of
this examination may be useful for enhancement and enrichment activities for the
students.
5
International Maritime English Conference
In terms of focus areas, the short talks by a single speaker seemed to be the most
challenging task for the participants. The topics in these items included real estate, office
reports, specific historical buildings, announcements about restaurants, the weather, at the
airport, and instructions about phone services. These topics should be the focus of
listening activities in the classrooms. More specifically, these tasks involved abilities such
as understanding main ideas and important details, understanding key vocabulary, making
inferences about relationships among ideas, and integrating information with one’s own
knowledge.
It was observed that the participants had difficulty relating to the topics discussed such as
real estate, airports, and very specific historical buildings and places. The context of the
communicative tasks may have been unfamiliar to the participants which accounted for
the difficulty in listening comprehension. In an earlier study of Orbe (2018) the main
difficulties among maritime students in terms of listening comprehension focused on
unfamiliarity with some vocabulary words, phrasal verbs, and in making inferences. In
this present study, the lack of prior knowledge of the participants on specific contexts, for
example, the airport and the announcements about weather changes, could have
contributed to the difficulty in comprehension. North (2007) has explained that lack of
familiarity with tasks contribute to the perceived increase in task difficulty. These
contexts could be introduced to students as part of their language enhancement activities.
Informed by the results of the listening tests and the common difficulties in the test items,
the Can-Do List of Listening Skills was developed through a document analysis of
listening instruments developed in China, Japan, Canada and Britain, which were all
based on the CEFR.
Using meta-analysis, a local version of the CEFR-based Self-Evaluation of Listening
Skills appropriate for context was developed. The final listening self-evaluation
6
International Maritime English Conference
instrument or Can- Do List consisted of 40 statements which were adapted from the
EAQUALS Core for Inventory for General English (North, Ortega & Sheehan, 2010).
The Can-Do List for Listening Skills now called the Self-Evaluation for Listening Skills
has fewer statements than the Canadian Language Benchmarks but more statements than
the CEFR-J (Runnel, 2013) and the China Standards of English (CSE). It was also
subjected to a readability test using an online grammar and plagiarism software. Based on
the sentence structure and complexity of the statements, the document was found to be
readable. Figure 3 describes the main features of the listening self-evaluation survey.
The difficult listening behaviors for the students as shown in this study were consistent
with language descriptors at A2 level.
7
International Maritime English Conference
Based on the findings of this study the following conclusions are derived:
1. Listening level is A1 that is below the standard level for university level.
2. Common cause of difficulty is unfamiliarity with context, vocabulary words, and
idiomatic expressions.
3. The Self-Evaluation for Listening Skills (Can-Do List for Listening) is a learner-
centered, action-oriented approach that provides students the learning outcomes
for each desired level.
In the light of student difficulties in certain listening abilities, the inclusion of activities,
training, and assessment in listening across the curriculum to enhance the listening
comprehension skills of students is highly recommended. More focus should be given to
exposure to conversations between native speakers in varied communication contexts.
Further studies should also be made to determine if there were empirical differences in
the responses to the CEFR scales in listening and the actual performance in the listening
test. The levels of agreement must be compared across the entire population through an
analysis of variance (ANOVA) in a future research. Further ANOVAs may be performed
within each of the levels. If any differences were to be found, LSD post-hoc tests may be
used to determine which can-do statement resulted in significantly different mean
responses across the rating scales.
5. References
Aryadoust, V. (2013). Building a validity argument for a listening test of academy
proficiency. New Castle: Cambridge Scholars Publishing.
Buck, G. (2001). Assessing listening. Cambridge: Cambridge University Press.
Council of Europe. (2001). Common European Framework of Reference for Languages:
Learning, teaching, assessment (CEFR). Cambridge: Cambridge University Press.
Davidson, F. and Henning, G. (1985). A self-rating scale of English difficulty: Rasch
scalar analysis of items and rating categories. Language testing, 2, 164-179.
Field, J. (2008). Listening is the language classroom. Cambridge: Cambridge University
Press.
Gilman, R. A. and Moody, L.M. (1984). What practitioners say about listening: Research
implications for the classroom. Foreign Language Annals, 17 (4), 331-334.
8
International Maritime English Conference
Glover, P. (2011). Using CEFR level descriptors to raise university students’ awareness
of their speaking skills. Language Awareness, 20, 121-133.
Little, D. (2006). The Common European Framework for Languages: Content, purpose,
origin, reception and impact. Language Testing, 39, 167-190.
Lynch, T. (2011). Academic listening in the 21 st century: Reviewing a decade of research.
Journal of English for Academic Purposes, 10 (2), 79-88.
North, B. (2007, February). The CEFR Common Reference Levels: Validated reference
points and local strategies. Paper presented at the Common European Framework
of Reference for languages (CEFR) and the development of language policies:
Challenges and responsibilities, Strasbourg. Retrieved from
http//www.coe.int/T/DG4/North-Forum-paper.
9
International Maritime English Conference
North, B., Ortega, A. and Sheehan, S. (2010) The British Council – EAQUALS Core
Inventory for General English (CIGE). Retrieved from
http//www.teachingenglish.org.uk/publications/british-council-eaquals-coreo-
inventory-general-english.
Orbe, M.C. (2018). Listening comprehension skills of selected MAAP midshipmen.
MAAP Research Journal, 4.
Orbe, M.C. (2019). Developing a CEFR-based can-do statements for listening. Paper
presented at the 6th MAAP Research Forum, Mariveles, Bataan, Philippines.
Runnels, J. (2016). Self-assessment accuracy: Correlations between Japanese English
learners’ self-assessment on the CEFR-Japan’s can-do statements and scores on
the TOEIC. Taiwan Journal of TESOL, 13(1),105-137.
Vandergrift, L. (2007). Recent developments in second and foreign language listening
comprehension research. Language Teaching, 40, 191-210.
10