What do teachers want to know about their student s elearning? A study of 70 evaluation plans

Similar documents
Web-based Learning Systems From HTML To MOODLE A Case Study

An application of student learner profiling: comparison of students in different degree programs

Concept mapping instrumental support for problem solving

Using Team-based learning for the Career Research Project. Francine White. LaGuardia Community College

Graduate Program in Education

Blended E-learning in the Architectural Design Studio

An Industrial Technologist s Core Knowledge: Web-based Strategy for Defining Our Discipline

Requirements-Gathering Collaborative Networks in Distributed Software Projects

Assessment System for M.S. in Health Professions Education (rev. 4/2011)

Person Centered Positive Behavior Support Plan (PC PBS) Report Scoring Criteria & Checklist (Rev ) P. 1 of 8

2013/Q&PQ THE SOUTH AFRICAN QUALIFICATIONS AUTHORITY

School Leadership Rubrics

A pilot study on the impact of an online writing tool used by first year science students

Focus on. Learning THE ACCREDITATION MANUAL 2013 WASC EDITION

Using Virtual Manipulatives to Support Teaching and Learning Mathematics

Course Specification Executive MBA via e-learning (MBUSP)

Running Head: Implementing Articulate Storyline using the ADDIE Model 1. Implementing Articulate Storyline using the ADDIE Model.

National Survey of Student Engagement (NSSE)

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Susan K. Woodruff. instructional coaching scale: measuring the impact of coaching interactions

Individualising Media Practice Education Using a Feedback Loop and Instructional Videos Within an elearning Environment.

Course Syllabus. Course Information Course Number/Section OB 6301-MBP

PIRLS. International Achievement in the Processes of Reading Comprehension Results from PIRLS 2001 in 35 Countries

BHA 4053, Financial Management in Health Care Organizations Course Syllabus. Course Description. Course Textbook. Course Learning Outcomes.

CONNECTICUT GUIDELINES FOR EDUCATOR EVALUATION. Connecticut State Department of Education

SEN SUPPORT ACTION PLAN Page 1 of 13 Read Schools to include all settings where appropriate.

The Incentives to Enhance Teachers Teaching Profession: An Empirical Study in Hong Kong Primary Schools

Higher education is becoming a major driver of economic competitiveness

New Ways of Connecting Reading and Writing

Statewide Strategic Plan for e-learning in California s Child Welfare Training System

Delaware Performance Appraisal System Building greater skills and knowledge for educators

Programme Specification. MSc in Palliative Care: Global Perspectives (Distance Learning) Valid from: September 2012 Faculty of Health & Life Sciences

Mathematics Program Assessment Plan

Internet Journal of Medical Update

Evaluation of Respondus LockDown Browser Online Training Program. Angela Wilson EDTECH August 4 th, 2013

Youth Sector 5-YEAR ACTION PLAN ᒫᒨ ᒣᔅᑲᓈᐦᒉᑖ ᐤ. Office of the Deputy Director General

EGE. Netspace/iinet. Google. Edmodoo. /enprovides. learning. page, provider? /intl/en/abou t. Coordinator. post in forums, on. message, Students to

BPS Information and Digital Literacy Goals

ABET Criteria for Accrediting Computer Science Programs

Introduction to Moodle

Kentucky s Standards for Teaching and Learning. Kentucky s Learning Goals and Academic Expectations

DISTRICT ASSESSMENT, EVALUATION & REPORTING GUIDELINES AND PROCEDURES

Blended Learning Module Design Template

ASSESSMENT OF STUDENT LEARNING OUTCOMES WITHIN ACADEMIC PROGRAMS AT WEST CHESTER UNIVERSITY

Programme Specification. BSc (Hons) RURAL LAND MANAGEMENT

Assessment and Evaluation

The Characteristics of Programs of Information

Final Teach For America Interim Certification Program

Motivation to e-learn within organizational settings: What is it and how could it be measured?

Davidson College Library Strategic Plan

Textbook Evalyation:

BSM 2801, Sport Marketing Course Syllabus. Course Description. Course Textbook. Course Learning Outcomes. Credits.

EXECUTIVE SUMMARY. TIMSS 1999 International Science Report

Programme Specification. MSc in International Real Estate

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Do Mobile Device Applications Affect Learning?

Strategic Planning for Retaining Women in Undergraduate Computing

National Survey of Student Engagement Executive Snapshot 2010

eportfolio Guide Missouri State University

1 Use complex features of a word processing application to a given brief. 2 Create a complex document. 3 Collaborate on a complex document.

University of Waterloo School of Accountancy. AFM 102: Introductory Management Accounting. Fall Term 2004: Section 4

Carolina Course Evaluation Item Bank Last Revised Fall 2009

Assessing and Providing Evidence of Generic Skills 4 May 2016

Systematic reviews in theory and practice for library and information studies

The Impact of Honors Programs on Undergraduate Academic Performance, Retention, and Graduation

Curriculum and Assessment Policy

Group Assignment: Software Evaluation Model. Team BinJack Adam Binet Aaron Jackson

Maintaining Resilience in Teaching: Navigating Common Core and More Online Participant Syllabus

TOPIC VN7 PAINTING AND DECORATING

ACS HONG KONG_INTERNATIONAL CHEMICAL SCIENCES CHAPTER 2011 ANNUAL REPORT

COUNSELLING PROCESS. Definition

STUDENT MOODLE ORIENTATION

Wide Open Access: Information Literacy within Resource Sharing

The Moodle and joule 2 Teacher Toolkit

REVIEW CYCLES: FACULTY AND LIBRARIANS** CANDIDATES HIRED ON OR AFTER JULY 14, 2014 SERVICE WHO REVIEWS WHEN CONTRACT

Policy for Hiring, Evaluation, and Promotion of Full-time, Ranked, Non-Regular Faculty Department of Philosophy

Automating Outcome Based Assessment

How to Develop and Evaluate an etourism MOOC: An Experience in Progress

2017 FALL PROFESSIONAL TRAINING CALENDAR

Module Title: Teaching a Specialist Subject

Week 4: Action Planning and Personal Growth

Practice Examination IREB

22264VIC Graduate Certificate in Bereavement Counselling and Intervention. Student Application & Agreement Form

The ELA/ELD Framework Companion: a guide to assist in navigating the Framework

P. Belsis, C. Sgouropoulou, K. Sfikas, G. Pantziou, C. Skourlas, J. Varnas

Submission of a Doctoral Thesis as a Series of Publications

Ministry of Education, Republic of Palau Executive Summary

STANDARDS AND RUBRICS FOR SCHOOL IMPROVEMENT 2005 REVISED EDITION

Promotion and Tenure standards for the Digital Art & Design Program 1 (DAAD) 2

Programme Specification

Formative Assessment in Mathematics. Part 3: The Learner s Role

MPA Internship Handbook AY

Virtual Seminar Courses: Issues from here to there

Vodcasting: A case study in adaptability to meet learners needs and preferences

Bold resourcefulness: redefining employability and entrepreneurial learning

Bluetooth mlearning Applications for the Classroom of the Future

The Future of Consortia among Indian Libraries - FORSA Consortium as Forerunner?

Nottingham Trent University Course Specification

BSc (Hons) Banking Practice and Management (Full-time programmes of study)

Transcription:

McNaught and Lam 433 What do teachers want to know about their student s elearning? A study of 70 evaluation plans Carmel McNaught and Paul Lam Centre for Learning Enhancement and Research The Chinese University of Hong Kong Abstract The e3learning (enrich, extend, evaluate learning; e3l) project in Hong Kong has conducted 70 evaluations of web-assisted courses with teachers in three Hong Kong universities. In this paper the evaluation intentions of teachers are examined. What do teachers wish to learn when they evaluate their teaching? In each e3l sub-project a formal evaluation plan is produced with a range of evaluation questions. Our 70 evaluation plans contained 457 separate evaluation questions. These questions were examined in order to portray the range of Hong Kong university teachers evaluation intentions and to see whether these changed during the 2½ years of the project. During the project there was a good balance of interest in looking across the learning environment, learning processes and learning outcomes, with increasing discrimination being shown in articulating what is meant by learning outcomes. In addition, teachers became more interested in conducting needs analyses and in having formative as well as summative evaluations. Keywords evaluation questions; evaluation focus; evaluation plan; evaluation trends Background: The e3learning model of evaluation The e3learning (enrich, extend, evaluate learning; e3l) project has been designed to assist teachers in webassisted teaching and learning. It offers a range of design, development and evaluation services. For each sub-project or case, a team is formed whereby teachers have access to educational and technical support. Full details of the design of this project are in James et al. (2003) and the project website. The e3l project operates across three universities, the Hong Kong Polytechnic University, the City University of Hong Kong and The Chinese University of Hong Kong. The nature of the websites developed varies a great deal. A snapshot of the design features in our 70 cases is shown in Table 1. These are reasonably standard ways to conduct elearning, including interactive design elements. Table 1: Design features of 70 e3learning websites Interaction Web-assisted function No. of instances Course background 10 Notes/ PowerPoints 17 Low interactivity Required readings 8 Past papers 1 Project/ assignment archives 1 Students interacting with materials and tasks Learning tools such as glossaries 1 Learning resources/ links 24 Multimedia materials 22 Self-assessment 16 Grade-giving assessments 4 Games 2 Teacher-student (one-on-one) communication 8 Interactions about specific information Assignment submission 4 Formal discussion forums with a range of designs 28 More public interaction space Exhibition of assignments 4 Creation and exhibition of multimedia projects 2 Inter-cultural discussions (across countries) 1 Total 153

434 ascilite 2005: Balance, Fidelity, Mobility: maintaining the momentum? The project has an evaluation officer (second author) who actively assists teachers with evaluating the websites developed under the auspices of the e3l project. The overall design of the evaluation is a reflection improvement model in which the findings of the evaluation contribute to further improvements in each of the web-assisted courses under investigation. There are evaluation resources already available, such as toolkits (eg. Oliver et al., 2002) or cookbooks (eg. Learning Technology Dissemination Initiative, 1998) but we chose to use a process mediated by an evaluation officer in order to fairly rapidly build up a set of cases of good evaluation practice for Hong Kong university teachers to refer to. Our system (like all others) is not value-free and tends towards a naturalistic model (Guba & Lincoln, 1981; Alexander & Hedberg, 1994). The process of evaluation consists of several stages, as shown in Figure 1. Although each of the e3learning websites go through all the stages in this overall evaluation process, the exact evaluation design and schedule varies (Lam & McNaught, 2004). A customized evaluation plan is produced for each e3l website. (See http://e3learning.edc.polyu.edu.hk/evaluate_scenarios2_plan.htm for a sample evaluation plan.) In this paper we are focussing just on the evaluation intentions of the teachers, as portrayed by the evaluation questions they decide to ask. Frame evaluation a questions DECISIONS & e ACTION! a Select strategies b Develop a plan (with timeline) Evaluation process c Collect data leads to Analyse d data Evaluation activities a Meeting with teachers. Getting a clear picture of how the teacher wishes to use the web. Deciding on the evaluation questions. Deciding the types of data to collect and the instruments to use. b Putting the decisions into an evaluation plan. c The evaluation officer designs the various evaluation instruments needed. Teachers help to finalize the instruments. Teachers help to administer questionnaires or tests to the class. The evaluation officer collects the feedback. d The evaluation officer analyzes the feedback. The evaluation officer writes reports. e Reports are handed over to the teachers. Meeting with teachers to look at the reports together and think about follow-up actions. Figure 1: The evaluation activities in the different stages of the e3l evaluation process Evaluation questions vary in accordance with the design purposes for the elements of the website. The evaluation questions also vary according to the focus the evaluation has on the various stages of the learning experience. Figure 2 illustrates various aspects of students learning experience (Bain, 1999). Evaluation can focus on one or more of these stages of the learning experience, For example, if we want to know whether students find a particular content-rich site valuable, there is a focus on the learning environment the students are experiencing. Evaluation questions such as the following might be appropriate. Is the navigation clear? Do the resources download easily? Are these resources appropriate for the desired learning outcomes? Is the number of resources about right? Etc. If we ask how often and when the students complete the online quizzes and exercises in an assessment-rich site, we are considering the learning processes. Lastly, if we ask whether students are able to demonstrate critical-thinking skills in some online debate activities held in the forum of a communication-rich site, or whether their performance on cognitively demanding tasks is satisfactory, the focus is on the learning outcomes that the students attain.

McNaught and Lam 435 work within Learning environments Students going through Learning outcomes in order to achieve Learning processes Figure 2: Stages of the learning experience Evaluation questions as a central component of evaluation plans In the two-and-a-half-year period of the e3learning project, 70 website evaluations were carried out. The evaluation plans were sequenced in time according to the date they were finalized (Table 2). There were, on average, six separate evaluation questions in each evaluation plan. Table 2: Scale of the 70 evaluation plans Number of evaluations Total no. of separate evaluation questions 2 nd semester of 2002-03 0203(2) 7 41 1 st semester of 2003-04 0304(1) 18 110 2 nd semester of 2003-04 0304(2) 16 111 1 st semester of 2004-05 0405(1) 18 122 2 nd semester of 2004-05 0405(2)* 11 73 Totals 70 457 some cases in the 0405(2) are still ongoing and evaluations are expected to continue in 0506(1). This paper is focused on comparing the evaluation questions contained in the evaluation plans of these 70 cases of the e3learning project. While the evaluation questions were negotiated between teachers and e3l staff, e3l staff have taken care to ensure that the ownership of the design and evaluation processes is clearly with the teachers. To a large extent we therefore feel confident that the nature and range of the final evaluation questions in each plan reflect the evaluation intentions of the teacher(s) in each of the sub-projects. The evaluation questions were extracted out of each evaluation plan. The questions were categorized by a process of first separating them into the four broad areas of: pre-development investigations; doing a needs analysis the learning environment; this could involve examination of actual materials and also broad cultural aspects of learning teaching and learning processes learning outcomes. Then we further grouped the evaluation questions by a process of iterative sorting in order to preserve key elements in diversity while achieving a manageable way to look for trends across the 457 evaluation questions. Our final decisions resulted in 28 evaluation themes, which are shown in Table 3. The learning general theme has a number of sub-categories. Our analysis then examined the percentage of sub-projects or cases where particular categories of evaluation questions had been asked.

436 ascilite 2005: Balance, Fidelity, Mobility: maintaining the momentum? Table 3: Classification of the 457 evaluation questions Main evaluation themes Description Pre-development investigations/ needs analysis 1 Evaluation of previous work Checking previous versions to collect ideas for the new development. 2 Expectation Collecting ideas about what students need and wish to see on the website. 3 IT skills/ habits of students Finding out the IT competence and habits of the students before teachers plan to introduce elearning in their classes. Learning environment 4 Mid-development formative evaluation Checking the web materials at the mid-stage of development. 5 Strength/ weaknesses Finding out detail of the good and weak points of the web materials. 6 Improvement ideas Being open to ways to improve the sites. 7 Usability Collecting opinions on site design, layout and navigation. 8 Opinions on content (difficulty/ coverage/ potential benefits) 9 Opinion on web teaching in general Collecting opinions on the quality and quantity of the teaching and learning content on the site. Commenting on the elearning experience as a whole, often contrasting this new way of learning with the traditional face-to-face only mode. 10 Appreciation of effort Finding out whether students valued having elearning experiences. Teaching and learning processes 11 Class management Checking whether the site supported students managing their studies. 12 Technical processes Looking at what hardware and software students prefer to use. 13 Engagement Checking on the visits paid to the site and how students spent time there. 14 Patterns of use Concerning matters such as when students assess the site, what they do on the site, and how the elearning experience affect other learning activities. 15 Meaningful/ effective communication Concerning how students interact with the teachers or their peers online. 16 Workload Finding out the amount of work involved in the elearning activities. 17 Enjoyment Finding out about students enjoyment while completing the elearning activities. Learning outcome 18 Motivation and affect Finding out whether the elearning experience has resulted in higher motivation to learn the subject. 19 Confidence Finding out whether the students feel more confidence in their studies. 20 Approaches to learning/ learning styles Finding out about a range of factors that can broadly be called a student-centred approach to learning. 21 Generic learning skills Checking on whether generic learning skills (eg. communication and problem-solving) have been developed through the elearning experience. 22 Subject-specific skills Checking on whether subject-specific skills (eg. clinical or language skills) have been developed through the elearning experience. 23 Thinking skills Checking on thinking skills such as the ability to evaluate or criticize. 24 Reflective learning Finding out about students ability to reflect on their own learning. 25 Learning general Finding out whether the sites have helped learning of the subject in a broad sense remember/ understand Learning mainly in remembering or understanding of subject content. apply/ analyze Learning to apply theories and analyze various situations. evaluate/ create Learning how to evaluate a situation, build on concepts and construct artifacts. 26 Retention of learning Checking the retention period of the knowledge acquired. 27 Relationship/ sense of community Finding out whether the elearning experience has resulted in better relationships. 28 Intercultural awareness Checking whether the online activities have promoted intercultural contacts.

McNaught and Lam 437 Trends and patterns in teachers evaluation intentions A balanced view of learning It is satisfying that there is consistent attention paid to the three aspects of learning (environment, processes and outcomes), as shown in Figure 3. Of course we present this as a model of learning but we do try not to foist our views on teachers. For us this affirms our beliefs that educational design input does assist teachers in developing more sophisticated views on learning and what they need to know about their students. For each case or sub-project in the e3l project there is an initial meeting between the teacher(s) teaching the course under consideration, the e3l project manager and the evaluation officer. Normally, one of the three e3l academic supervisors is also present. This is an excellent opportunity for teachers to articulate their needs and for the e3l staff to suggest ways in which these ideas might be enacted in a website. The conversation ranges across both educational and technical design ideas. This brainstorming session is seen as a vital component of the e3l model. As the e3l project has progressed, we have had more examples of finished websites to show teachers in new sub-projects. These local examples really enrich these design meetings and have assisted a great deal in enabling teachers to take a wider perspective on learning. Understanding the need to have data for planning Figure 3 also shows that teachers are now turning to pre-development investigations and are more seriously conducting needs analyses before embarking on design and development. Figure 4 shows the nature of these needs analyses. Mostly teachers are deciding to gauge students expectations, though a few doing an analysis of students skills. Only one teacher embarked on a focused evaluation of an existing site. Percentages of cases in the main categories 10 8 6 4 Pre-development 11.1% 62.5% 16.7% 27.3% Environment 85.7% 83.3% 93.8% 88.9% 72.7% T & L process 10 88.9% 93.8% 88.9% 54.5% Learning outcome 10 72.2% 68.8% 83.3% 63.6% Pre-development Environment T & L process Learning outcome Figure 3: Balance of evaluations across the four major categories

438 ascilite 2005: Balance, Fidelity, Mobility: maintaining the momentum? 6 5 4 3 1 IT skills/ habits of students Expectation Evaluation of previous development IT skills/ habits of students 5.6% 12.5% Expectation 5.6% 56.3% 16.7% 27.3% Evaluation of previous development 5.6% Figure 4: The nature of needs analyses undertaken From the general to the particular in evaluating the learning environment and processes Figure 5 shows that, at the beginning of the e3l project, teachers asked very general questions about the learning environment. Since then the evaluation questions have honed in on specific usability issues. Further, as our pool of cases has increased, teachers can see the need to carry out formative evaluations where students responses can have a direct impact on the nature of the website, so as to maximize the benefits the teaching teams can achieve from their websites (Figure 6). This trend has increased the cost-effectiveness of the e3l project. 9 8 7 6 5 4 3 1 Formative/ mid-development 6.3% 5.6% 54.5% Strength/ weaknesses 14.3% 11.1% 18.8% Improvement ideas 71.4% 77.8% 75.0% 66.7% 45.5% Usability 14.3% 66.7% 18.8% 44.4% 81.8% Opinions on content (diff/ coverage/ L 71.4% 83.3% 62.5% 66.7% 63.6% potentials) Opinion on web teaching in general 42.9% 38.9% 68.8% 16.7% 9.1% Appreciation of effort 14.3% Formative/ middevelopment Strength/ weaknesses Improvement ideas Usability Opinions on content (diff/ coverage/ L potentials) Opinion on web teaching in general Appreciation of effort Figure 5: Increasing diversity in evaluations of the learning environments

McNaught and Lam 439 6 5 4 3 1 formative/middevelopment 6.3% 5.6% 54.5% Figure 6: A growing interest in formative evaluation The same trend is seen in terms of evaluations of the learning process. We have moved from a fairly simplistic notion of engagement, usually measured by a satisfaction questionnaire, to interest in patterns of use and the nature of communication interactions online (Figure 7). While this is anecdotal evidence, the second semester of the 0203 academic year was the time of SARS in Hong Kong and, while tragic as a crisis, it changed Hong Kong universities from face-to-face universities to distance euniversities essentially overnight. This has had a lasting impact on teachers understanding of the web as a communication vehicle (McNaught, 2004). That legacy has remained. 10 9 8 7 6 5 4 3 1 Class management 14.3% 11.1% Technicality 5.6% 6.3% Engagement 85.7% 77.8% 93.8% 77.8% 45.5% Patterns of use 14.3% 27.8% 6.3% 22.2% 54.5% Meaningful/effective communication 28.6% 5.6% 5 11.1% 63.6% Workload 5.6% 9.1% Enjoyment 18.8% 16.7% 9.1% Class management Technicality Engagement Patterns of use Meaningful/effective communication Workload Enjoyment Figure 7: Increasing diversity in evaluations of learning processes

440 ascilite 2005: Balance, Fidelity, Mobility: maintaining the momentum? 10 9 8 7 6 5 4 3 1 Learning (remember/ understand) 27.8% 12.5% 72.2% 54.5% Learning (apply/ analyze) 28.6% 16.7% 18.8% 38.9% 45.5% Learning (evaluate/ create) 14.3% 18.8% 11.1% 9.1% Learning in general 10 61.1% 75.0% 83.3% 81.8% Learning (remember/ understand) Learning (apply/ analyze) Learning (evaluate/ create) Learning in general Figure 8: Increasing awareness of evaluating more detail in aspects of learning outcomes Greater detail in evaluating learning outcomes When considering evaluation of learning outcomes, teachers have become much more interested in finding out more detail (Figure 8). There are fewer questions of the type whether the site has helped students learn and more of the type whether the site has helped students put theories into practice, or support students to be more able to play a professional role in the field, etc. In addition to looking at more detail in the cognitive domain, our teachers do seek evidence of learning outcomes apart from the knowledge domain. Figure 9 shows that there is consistent attention paid to seeking evidence of the development of aspects of learning including a range of skills, motivational changes, learning style changes and reflective capacity. 35.0% 3 Motivation and affect 25.0% 15.0% 1 5.0% Confidence Approaches to learning/learning styles Generic learning skills Subject-specific skills Thinking skills Motivation and affect 14.3% 16.7% 12.5% 27.8% 9.1% Confidence 16.7% 6.3% 5.6% 9.1% Approaches to learning/learning styles 14.3% 16.7% 6.3% 11.1% 9.1% Generic learning skills 14.3% 16.7% 18.8% 33.3% Subject-specific skills 11.1% 6.3% 11.1% 27.3% Thinking skills 14.3% 18.8% 11.1% Reflective learning 14.3% 5.6% 6.3% 5.6% 9.1% Relationship/sense of community 5.6% Intercultural awareness 5.6% 5.6% Retention of learning 9.1% Reflective learning Relationship/sense of community Intercultural awareness Retention of learning Figure 9: Evaluation of learning as being more than cognitive change

McNaught and Lam 441 In the first year of e3l there were no interactional studies. However, since then there have been three studies where teachers wanted to investigate whether time spent on the website was related to learning outcomes, measured in a variety of ways. These are simple studies but show an interest on teachers part in trying to understand the dynamics of elearning. In the most recent semester we have one study in progress relating students learning styles to higher order cognitive learning outcomes. Matching evaluation strategies to teachers evaluation intentions The data above is based on a consideration of the evaluation questions teachers in 70 e3l sub-projects asked. To what extent did the actual evaluations that were conducted (or, in a few cases, will be conducted next semester) enable teacher to get answers to those questions? Table 4 shows the range of evaluation instruments used over the years and Table 5 shows this as data types. Teachers in Hong Kong are accustomed to the use of student questionnaires for evaluation. This we realized in the first semester of the project and recognized that we needed to invest more time in describing the variety of possible evaluation strategies to teachers. Since then, we have managed to obtain a range of data beyond that obtained by self-reporting. Our website lists a number of evaluation strategies (see http://e3learning.edc.polyu.edu.hk/strategies.htm) and these are discussed in detail. There is no doubt that strategies such as focus groups, interviews, and detailed analysis of forum postings (eg. see Lam, Cheng & McNaught, 2005) and students assessment tasks are expensive in time and effort. However, our experience suggests that these evaluation strategies are essential if we are begin to understand how students operate in and learn from web environments. Table 4: Evaluation instruments and strategies used in the e3l project Student surveys Course-start Task-end Mid-term Course-end Surveys (teachers) Focus groups Interviews Web logs Forum postings analysis Assignments/ exams Others (expert-review/ user try-out sessions) 0203(2) Feb 03 June 03 1 6 0 8 8 4 3 2 4 1 0 0304(1) Sept 03 Jan 04 2 0 1 28 20 5 3 15 3 2 1 0304(2) Feb 04 June 04 1 0 0 13 13 8 0 14 6 1 0 0405(1) Sept 04 Jan 05 4 1 1 16 13 8 0 17 1 4 1 0405(2) Feb 05-June 05* 0 0 0 3 2 1 0 11 1 2 4 Total 8 7 2 68 56 26 6 59 15 10 6 some 0405(2) evaluations are still ongoing Table 5: Data types used in the e3l project Semester Opinion data Web log action data Learning outcome data Other 0203(2) 30 2 5 0 0304(1) 59 15 5 1 0304(2) 35 14 7 0 0405(1) 43 17 5 1 0405(2)* 6 11 3 4 some 0405(2) evaluations are still ongoing

442 ascilite 2005: Balance, Fidelity, Mobility: maintaining the momentum? Conclusions This paper has charted the evaluation intentions of teachers in 70 web-supported courses in Hong Kong. Its intention is not to describe the specific evaluations or to describe the overall outcomes of these evaluations. This is being reported elsewhere (eg. McNaught & Lam, in press; Lam & McNaught, in press). We designed the e3l project on the premise that teachers need both educational and technical support if they are to achieve useful results in web-assisted courses. If there is no support many teachers will just give up or remain very basic in their use of the web. Our evidence is that the teachers in these 70 cases have become more sophisticated in their educational thinking during the e3l project. Some of the sub-projects are repeat clients and their planning shows this development. This affirms the beliefs of the e3l staff that, while retaining control and ownership of their educational websites, teachers can quite rapidly learn from colleagues experience and from expert elearning professionals, and begin to engage with educational evaluation on a deeper level. Increased interest in investigating the dynamics and outcomes of students elearning is highly likely to lead to improved elearning designs, and we are confident that many of our e3l teachers have developed a critical reflective stance in relation to their own eteaching and their students elearning. References Alexander, S., & Hedberg, J. (1994). Evaluating technology-based learning: Which model? In K. Beattie, C. McNaught, & S. Wills (Eds.), Multimedia in higher education: Designing for change in teaching and learning (pp. 233 244). Amsterdam: Elsevier. Bain, J. D. (1999). Introduction. Special issue: Learning-centred evaluation of innovation in higher education. Higher Education Research & Development, 18(2), 165 172. e3l project website. (2005). Retrieved June 19, 2005, from http://e3learning.edc.polyu.edu.hk/main.htm Guba, E. G., & Lincoln, Y. S. (1981). Effective evaluation. San Francisco: Jossey Bass. James, J., McNaught, C., Csete, J., Hodgson, P., & Vogel, D. (2003). From MegaWeb to e 3 Learning: A model of support for university academics to effectively use the web for teaching and learning. In D. Lassner & C. McNaught (Eds.), ED-MEDIA 2003: Proceedings of the 15th annual world conference on Educational Multimedia, Hypermedia & Telecommunications (pp. 3303 3310). Norfolk, VA: Association for the Advancement of Computers in Education. Lam, P., Cheng, K. F., & McNaught, C. (2005). Asynchronous online discussion: Empirical evidence on quantity and quality. In G. Richards & P. Kommers (Eds.), ED-MEDIA 2005: Proceedings of the 17th annual world conference on Educational Multimedia, Hypermedia & Telecommunications (pp. 3209 3215). Norfolk, VA: Association for the Advancement of Computers in Education. Lam, P., & McNaught, C. (in press). Evaluating designs for web assisted peer and group assessment. In T. Roberts (Ed.), Self, peer, and group assessment in e-learning. Hershey, PA: Idea Group. Lam, P., & McNaught, C. (2004). Evaluating educational web sites: A system for multiple web sites at multiple universities. In L. Cantoni & C. McLoughlin (Eds.), ED-MEDIA 2004: Proceedings of the 16 th annual world conference on Educational Multimedia, Hypermedia & Telecommunications (pp. 1066 1073). Norfolk, VA: Association for the Advancement of Computers in Education. Learning Technology Dissemination Initiative. (1998). Evaluation cookbook. Retrieved June 19, 2005, from http://www.icbl.hw.ac.uk/ltdi/cookbook/contents.html McNaught, C. (2004). Using narrative to understand the convergence of distance and campus-based learning during the time of SARS in Hong Kong. Educational Media International, 41(3), 183 193. McNaught, C., & Lam, P. (in press). Building an evaluation culture and evidence base for e-learning in three Hong Kong universities. British Journal of Educational Technology. Oliver, M., McBean, J., Conole, G., & Harvey, J. (2002). Using a toolkit to support the evaluation of learning. Journal of Computer Assisted Learning, 18, 199 208. Acknowledgments Funding support from the University Grants Committee in Hong Kong is gratefully acknowledged, as is the collaborative support of many colleagues in our three universities. All teachers generously agreed to their data being used in this paper.

McNaught and Lam 443 Author contact details Carmel McNaught Professor of Learning Enhancement Centre for Learning Enhancement And Research The Chinese University of Hong Kong carmel.mcnaught@cuhk.edu.hk http://www.cuhk.edu.hk/clear/staff/staff7.htm Paul Lam Post-doctoral Fellow Centre for Learning Enhancement And Research The Chinese University of Hong Kong paul.lam@cuhk.edu.hk http://www.cuhk.edu.hk/clear/staff/staff11.htm Copyright 2005 Carmel McNaught and Paul Lam The author(s) assign to ascilite and educational non-profit institutions a non-exclusive licence to use this document for personal use and in courses of instruction provided that the article is used in full and this copyright statement is reproduced. The author(s) also grant a non-exclusive licence to ascilite to publish this document on the ascilite web site (including any mirror or archival sites that may be developed) and in printed form within the ascilite 2005 conference proceedings. Any other usage is prohibited without the express permission of the author(s).