Running head: LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 1

Similar documents
Running head: METACOGNITIVE STRATEGIES FOR ACADEMIC LISTENING 1. The Relationship between Metacognitive Strategies Awareness

Linking the Common European Framework of Reference and the Michigan English Language Assessment Battery Technical Report

TAIWANESE STUDENT ATTITUDES TOWARDS AND BEHAVIORS DURING ONLINE GRAMMAR TESTING WITH MOODLE

English for Specific Purposes World ISSN Issue 34, Volume 12, 2012 TITLE:

Evidence-Centered Design: The TOEIC Speaking and Writing Tests

The Effect of Extensive Reading on Developing the Grammatical. Accuracy of the EFL Freshmen at Al Al-Bayt University

Psychometric Research Brief Office of Shared Accountability

ELS LanguagE CEntrES CurriCuLum OvErviEw & PEDagOgiCaL PhiLOSOPhy

Listening and Speaking Skills of English Language of Adolescents of Government and Private Schools

Developing a College-level Speed and Accuracy Test

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

ESL Curriculum and Assessment

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Corpus Linguistics (L615)

Further, Robert W. Lissitz, University of Maryland Huynh Huynh, University of South Carolina ADEQUATE YEARLY PROGRESS

Project Based Learning Debriefing Form Elementary School

OVERVIEW OF CURRICULUM-BASED MEASUREMENT AS A GENERAL OUTCOME MEASURE

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

Metacognitive Strategies that Enhance Reading Comprehension in the Foreign Language University Classroom

ROLE OF SELF-ESTEEM IN ENGLISH SPEAKING SKILLS IN ADOLESCENT LEARNERS

How to Judge the Quality of an Objective Classroom Test

Linking the Ohio State Assessments to NWEA MAP Growth Tests *

International Conference on Education and Educational Psychology (ICEEPSY 2012)

Empowering Students Learning Achievement Through Project-Based Learning As Perceived By Electrical Instructors And Students

Assessing speaking skills:. a workshop for teacher development. Ben Knight

Conceptual and Procedural Knowledge of a Mathematics Problem: Their Measurement and Their Causal Interrelations

Lower and Upper Secondary

The Effect of Syntactic Simplicity and Complexity on the Readability of the Text

DOES OUR EDUCATIONAL SYSTEM ENHANCE CREATIVITY AND INNOVATION AMONG GIFTED STUDENTS?

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

Concept mapping instrumental support for problem solving

Effect of Word Complexity on L2 Vocabulary Learning

Testing for the Homeschooled High Schooler: SAT, ACT, AP, CLEP, PSAT, SAT II

CONSULTATION ON THE ENGLISH LANGUAGE COMPETENCY STANDARD FOR LICENSED IMMIGRATION ADVISERS

Interpreting ACER Test Results

Understanding Language

COURSE SYNOPSIS COURSE OBJECTIVES. UNIVERSITI SAINS MALAYSIA School of Management

Designing a Rubric to Assess the Modelling Phase of Student Design Projects in Upper Year Engineering Courses

Learning and Retaining New Vocabularies: The Case of Monolingual and Bilingual Dictionaries

Intermediate Algebra

Standards-Based Bulletin Boards. Tuesday, January 17, 2012 Principals Meeting

Exploring the adaptability of the CEFR in the construction of a writing ability scale for test for English majors

The Task. A Guide for Tutors in the Rutgers Writing Centers Written and edited by Michael Goeller and Karen Kalteissen

How do we balance statistical evidence with expert judgement when aligning tests to the CEFR?

The Effect of Personality Factors on Learners' View about Translation

What is PDE? Research Report. Paul Nichols

Rules and Discretion in the Evaluation of Students and Schools: The Case of the New York Regents Examinations *

Welcome to the session on ACCUPLACER Policy Development. This session will touch upon common policy decisions an institution may encounter during the

Evidence for Reliability, Validity and Learning Effectiveness

VIEW: An Assessment of Problem Solving Style

Sheila M. Smith is Assistant Professor, Department of Business Information Technology, College of Business, Ball State University, Muncie, Indiana.

EFL teachers and students perspectives on the use of electronic dictionaries for learning English

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

TU-E2090 Research Assignment in Operations Management and Services

To test or not to test? The selection and analysis of an instrument to assess literacy skills of Indigenous children: a pilot study.

Planning a Dissertation/ Project

Language Acquisition Chart

Laporan Penelitian Unggulan Prodi

An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District

ASSESSMENT OF LEARNING STYLES FOR MEDICAL STUDENTS USING VARK QUESTIONNAIRE

Handbook for Graduate Students in TESL and Applied Linguistics Programs

JOURNAL OF LANGUAGE AND LINGUISTIC STUDIES ISSN: X Journal of Language and Linguistic Studies, 13(2), ; 2017

ALEKS. ALEKS Pie Report (Class Level)

Teachers development in educational systems

TEXT FAMILIARITY, READING TASKS, AND ESP TEST PERFORMANCE: A STUDY ON IRANIAN LEP AND NON-LEP UNIVERSITY STUDENTS

Student Morningness-Eveningness Type and Performance: Does Class Timing Matter?

Norms How were TerraNova 3 norms derived? Does the norm sample reflect my diverse school population?

and secondary sources, attending to such features as the date and origin of the information.

Writing a Basic Assessment Report. CUNY Office of Undergraduate Studies

A Guide to Adequate Yearly Progress Analyses in Nevada 2007 Nevada Department of Education

On-the-Fly Customization of Automated Essay Scoring

National Standards for Foreign Language Education

ScienceDirect. Noorminshah A Iahad a *, Marva Mirabolghasemi a, Noorfa Haszlinna Mustaffa a, Muhammad Shafie Abd. Latif a, Yahya Buntat b

MGMT 479 (Hybrid) Strategic Management

DOES RETELLING TECHNIQUE IMPROVE SPEAKING FLUENCY?

5. UPPER INTERMEDIATE

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Effects of Self-Regulated Strategy Development on EFL Learners Reading Comprehension and Metacognition

Investment in e- journals, use and research outcomes

Spanish Users and Their Participation in College: The Case of Indiana

STA 225: Introductory Statistics (CT)

STUDENT SATISFACTION IN PROFESSIONAL EDUCATION IN GWALIOR

English Language Arts Summative Assessment

The Effect of Close Reading on Reading Comprehension. Scores of Fifth Grade Students with Specific Learning Disabilities.

Monticello Community School District K 12th Grade. Spanish Standards and Benchmarks

Georgia Department of Education

Lexical Collocations (Verb + Noun) Across Written Academic Genres In English

Higher Education Six-Year Plans

AND TRANSLATION STUDIES (IJELR)

Second Language Acquisition in Adults: From Research to Practice

AC : DEVELOPMENT OF AN INTRODUCTION TO INFRAS- TRUCTURE COURSE

Audit Documentation. This redrafted SSA 230 supersedes the SSA of the same title in April 2008.

Interdisciplinary Journal of Problem-Based Learning

DATE ISSUED: 11/2/ of 12 UPDATE 103 EHBE(LEGAL)-P

Effects of connecting reading and writing and a checklist to guide the reading process on EFL learners learning about English writing

Taking into Account the Oral-Written Dichotomy of the Chinese language :

12-WEEK GRE STUDY PLAN

NAME OF ASSESSMENT: Reading Informational Texts and Argument Writing Performance Assessment

re An Interactive web based tool for sorting textbook images prior to adaptation to accessible format: Year 1 Final Report

ATW 202. Business Research Methods

Transcription:

Running head: LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 1 Assessing Students Listening Comprehension of Different University Spoken Registers Tingting Kang Applied Linguistics Program Northern Arizona University

LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 2 Abstract The awareness of varied registers has given rise to the general field of English for Academic Purposes (EAP), which focuses on teaching language that is used in university registers. This paper aimed to design an achievement test for formative purposes to measure advanced language learners (N = 25) mastery in different university spoken registers through listening. The intended consequences of the test are to provide stakeholders the opportunity to evaluate instruction and learning processes, examine the course objectives, reinforce administration goals, and in the long run, make the program administration more effective. The test results showed that most of the students passed the test. Following Bachman and Palmer s (2010c) four claims in Assessment Use Argument (AUA), various evidence has been provided to support the validity argument of this test: (a) consequences were beneficial; (b) specific decisions made on basis of scores; (c) results could be interpreted as indicator of construct; (d) records were consistent. Keywords: EAP, listening comprehension, university register

LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 3 Assessing Students Listening Comprehension of Different University Registers Background To ensure students academic success, it is not enough that they have vocabulary and grammatical knowledge; they should also be equipped with the knowledge of register differences (Biber & Conrad, 2009). Similarly, the awareness of varied registers has given rise to the general field of English for Academic Purposes (EAP). The purpose of this test was to measure advanced language learners listening comprehension of different university spoken registers and further help stakeholders evaluate instruction and learning processes, examine the course objectives, reinforce administration goals, and make the program administration more effective. Target language use (TLU) task was defined as a specific language use tasks which test takers are likely to perform in specific setting (p. 60), which fell inside the TLU domain (Bachman & Palmer, 2010). In terms of the spoken language in the university settings, it includes various university spoken registers, such as university office hours, lectures, class discussion, study group, student presentation, and service encounters (Biber, Conrad, Reppen, Byrd, Helt, Clark, Cortes, Csomay, & Urzua, 2004). The four TLU tasks that served as basis for my test purpose were university office hours, lectures, university encounter services, and university student conversations. According to Buck (2001), listening comprehension involves two types of knowledge: linguistic knowledge and non-linguistic knowledge. In this listening comprehension test, the three subconstructs of listening comprehension included two aspects of linguistic knowledge (i.e., vocabulary and cognitive understanding) and one aspect of non-linguistic knowledge (i.e., situational context) were examined.

LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 4 Hypotheses My first hypothesis was that main idea and detail questions would be easier than the situational context and vocabulary questions, and the hardest questions should be about vocabulary. Second, since the three subconstructs were hypothesized to measure the same construct, listening comprehension of different university registers, their internal consistency should be very high. Third, students scores on this test should match with their listening comprehension abilities in classroom performance. Methods The participants were PIE Level 5a (n = 14) and 6 (n = 11) students, enrolled in Fall 2013 Listening and Speaking courses. The listening comprehension test was administered by the test developer during participants Listening and Speaking class for about 30 minutes. The Table of Specification (Miller, Linn, & Gronlund, 2009c) for this test is in Appendix A. It is a two way chart that relates the subconstructs (i.e., situational context, cognitive understand, and vocabulary) and the test tasks (i.e., university office hours (listening 1), lectures (listening 2), service encounters (listening 3), and student conversations (listening 4)). Each task was worth 25% of the overall score. There were 10 questions per listening passage, which included (a) three situational context questions (i.e., participants, setting, and topic) (b) four vocabulary questions, and (c) three cognitive questions (i.e., one main idea question and two detail questions). The subconstructs accounted for 30% (situational context), 30% (cognitive understanding), and 40% (vocabulary) of the total score. The questions assigned for each subconstruct were participants (Question 3, 10, 17, and 24), setting (Question 4, 11, 18, and 25), topic (Question 6, 13, 20, and 27), main idea (Question 5, 12, 19, and 26), detail (Question 1, 2,

LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 5 8, 9, 15, 16, 22, and 23), and vocabulary (Question 7, 14, 21, and 28). Each question was worth one point. Results In terms of the descriptive statistics, the average scores were 33.56 (SD = 5.49, K = 40, total), 10.88 (SD = 1.30, K = 12, situational context), 10.28 (SD = 1.10, K = 12, cognitive understanding), and 12.40 (SD = 5.24, K = 16, vocabulary). Regarding reliability (i.e., consistency of measurement (Miller, Linn, & Gronlund, 2009b)), the results of Cronbach s Alpha for item internal consistency were.86 (total),.46 (situational context), -.06 (cognitive understanding), and.96 (vocabulary). Since this was a criterion-referenced test, the results of agreement coefficient, which measured the consistency of mastery and non-mastery classifications, were between.96 and.86 and exceed the minimum requirement,.75 (Subkoviak, 1988). The scores of SEM, the margin of errors in test scores (Test Service Bulletin No. 50, 1956), ranged from 2.02 to.93. The cutoff percentage for mastery in this test was 70%. Therefore, the cut score was 28 for the overall test, 8.4 for situational context and cognitive understanding, and 11.2 for vocabulary. Most of the students passed the test and each subconstruct. Also, students who failed the test and each subconstruct were mostly from Level 5a and very few from Level 6. Discussion After reviewing the item statistics, items with low P and D values were indicated as problematic items because they were hard items and failed to discriminate higher and lower proficiency students(e.g., Question 2, 4, 17, and 26). They had characteristics of being too hard, which led to guessing, or including highly possible distracters. Accordingly, revisions could be

LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 6 made to change some of the distracters. For example, in Question 26, the mostly selected distracter, c, could be They want to drink coffee together. Even though the interpretations and decisions that are made on the basis of assessment results can never be considered justified in any absolute sense (Bachman & Palmer, 2010c, p. 95), the evidence provided for the validity argument in the previous section were convincing. Some evidence such as theoretical cutoff percentages, teachers feedback, and SEM scores on TOEFL ibt made the validity argument more comprehensible. Overall, this test achieved its purpose, measuring advanced language learners listening comprehension of different university spoken registers. Also, the results confirmed some of the hypotheses. First, vocabulary questions were the hardest subconstructs, and it contained the highest number of non-masters (n = 8). Second, the overall internal consistency of this test was also very high (Cronbach s Alpha =.86). Lastly, according to teachers feedback on this test, students scores matched with their listening comprehension abilities in the classroom performance. However, owning to the low internal consistency for cognitive understanding, future teachers and test developers should consider deleting this subconstruct to measure students listening comprehension of university registers.

LISTENING COMPREHENSION OF UNIVERSITY REGISTERS 7 References Bachman, L., & Palmer, A. (2010). Language assessment in practice. New York, NY: Oxford University Press. Biber, D., & Conrad, S. (2009). Register, genre, and style. Cambridge, UK: Cambridge University Press. Biber, D., Conrad, S., Reppen, R., Byrd, P., Helt, M., Clark, V., Cortes, V., Csomay, E., & Urzua, A. (2004). Representing language use in the university: Analysis of the TOELF 2000 spoken and written academic language corpus. ETS TOEFL Monograph Series, MS-25. Princeton, NJ: Education Testing Service. Buck, G. (2001). Assessing listening. New York, NY: Cambridge University Press. Miller, M. D., Linn, R., & Gronlund, N. (2009). Measurement and evaluation in teaching (10th Edition). New York, NY: Pearson. Test Service Bulletin No. 50. (1956). How accurate is a test score? NY: The Psychological Corporation.

Appendix B Tables of Specifications Constructs Situational Context Cognitive Understanding University Spoken Registers Participants Setting Topic Main idea Detail Listening 1: Office hours Length: 3 25 Words per min: 181 Listening 2: Lectures Length: 3 01 Words per min: 137 Listening 3: Service encounters Length: 2 33 Words per min: 162 Listening 4: Student conversations Length: 2 57 Words per min: 193 # of questions 3 10 17 4 11 18 6 13 20 5 12 19 1, 2 [2] 8, 9 [2] 15, 16 [2] Vocabulary # of questions % 7-a, b, c, d [4] 10 25 14-a, b, c, d [4] 10 25 21-a, b, c, d [4] 10 25 24 25 27 26 22, 23 [2] 28-a, b, c, d [4] 10 25 Total questions: 4 4 4 4 8 16 40 Points per question 1 1 1 1 1 1 Points per sub construct 12 12 16 Total points: 40 % per subconstruct 30 30 40 100