Measuring student knowledge and skills. The PISA 2000 Experience. By Andreas Schleicher and Claudia Tamassia. Poland Greece Portugal.

Similar documents
Department of Education and Skills. Memorandum

National Academies STEM Workforce Summit

The Survey of Adult Skills (PIAAC) provides a picture of adults proficiency in three key information-processing skills:

TIMSS Highlights from the Primary Grades

PIRLS. International Achievement in the Processes of Reading Comprehension Results from PIRLS 2001 in 35 Countries

Summary and policy recommendations

Introduction Research Teaching Cooperation Faculties. University of Oulu

Students with Disabilities, Learning Difficulties and Disadvantages STATISTICS AND INDICATORS

SOCRATES PROGRAMME GUIDELINES FOR APPLICANTS

PROGRESS TOWARDS THE LISBON OBJECTIVES IN EDUCATION AND TRAINING

Twenty years of TIMSS in England. NFER Education Briefings. What is TIMSS?

Impact of Educational Reforms to International Cooperation CASE: Finland

The European Higher Education Area in 2012:

15-year-olds enrolled full-time in educational institutions;

Overall student visa trends June 2017

The recognition, evaluation and accreditation of European Postgraduate Programmes.

EXECUTIVE SUMMARY. TIMSS 1999 International Mathematics Report

EXECUTIVE SUMMARY. TIMSS 1999 International Science Report

Science and Technology Indicators. R&D statistics

May To print or download your own copies of this document visit Name Date Eurovision Numeracy Assignment

The International Coach Federation (ICF) Global Consumer Awareness Study

INSTRUCTION MANUAL. Survey of Formal Education

International House VANCOUVER / WHISTLER WORK EXPERIENCE

international PROJECTS MOSCOW

Assessment of Generic Skills. Discussion Paper

DEVELOPMENT AID AT A GLANCE

Teaching Practices and Social Capital

School Size and the Quality of Teaching and Learning

UPPER SECONDARY CURRICULUM OPTIONS AND LABOR MARKET PERFORMANCE: EVIDENCE FROM A GRADUATES SURVEY IN GREECE

HIGHLIGHTS OF FINDINGS FROM MAJOR INTERNATIONAL STUDY ON PEDAGOGY AND ICT USE IN SCHOOLS

Higher Education Review of University of Hertfordshire

REFLECTIONS ON THE PERFORMANCE OF THE MEXICAN EDUCATION SYSTEM

Welcome to. ECML/PKDD 2004 Community meeting

Tailoring i EW-MFA (Economy-Wide Material Flow Accounting/Analysis) information and indicators

ROA Technical Report. Jaap Dronkers ROA-TR-2014/1. Research Centre for Education and the Labour Market ROA

How to Search for BSU Study Abroad Programs

OECD THEMATIC REVIEW OF TERTIARY EDUCATION GUIDELINES FOR COUNTRY PARTICIPATION IN THE REVIEW

Information needed to facilitate the clarity, transparency and understanding of mitigation contributions

CHAPTER 4: REIMBURSEMENT STRATEGIES 24

CONNECTICUT GUIDELINES FOR EDUCATOR EVALUATION. Connecticut State Department of Education

School Inspection in Hesse/Germany

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

DISCUSSION PAPER. In 2006 the population of Iceland was 308 thousand people and 62% live in the capital area.

PIRLS 2006 ASSESSMENT FRAMEWORK AND SPECIFICATIONS TIMSS & PIRLS. 2nd Edition. Progress in International Reading Literacy Study.

Mathematics subject curriculum

Research Update. Educational Migration and Non-return in Northern Ireland May 2008

CONSULTATION ON THE ENGLISH LANGUAGE COMPETENCY STANDARD FOR LICENSED IMMIGRATION ADVISERS

GCSE English Language 2012 An investigation into the outcomes for candidates in Wales

Report on organizing the ROSE survey in France

MODERNISATION OF HIGHER EDUCATION PROGRAMMES IN THE FRAMEWORK OF BOLOGNA: ECTS AND THE TUNING APPROACH

Document number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering

AUTHORITATIVE SOURCES ADULT AND COMMUNITY LEARNING LEARNING PROGRAMMES

EXECUTIVE SUMMARY. Online courses for credit recovery in high schools: Effectiveness and promising practices. April 2017

Measuring up: Canadian Results of the OECD PISA Study

FINNISH KNOWLEDGE IN MATHEMATICS AND SCIENCES IN 2002

IAB INTERNATIONAL AUTHORISATION BOARD Doc. IAB-WGA

The development of national qualifications frameworks in Europe

BENCHMARK TREND COMPARISON REPORT:

Pupil Premium Impact Assessment

Australia s tertiary education sector

Longitudinal Analysis of the Effectiveness of DCPS Teachers

The Achievement Gap in California: Context, Status, and Approaches for Improvement

Rethinking Library and Information Studies in Spain: Crossing the boundaries

Student Assessment and Evaluation: The Alberta Teaching Profession s View

Alignment of Australian Curriculum Year Levels to the Scope and Sequence of Math-U-See Program

Introduction. Background. Social Work in Europe. Volume 5 Number 3

Developing an Assessment Plan to Learn About Student Learning

ehealth Governance Initiative: Joint Action JA-EHGov & Thematic Network SEHGovIA DELIVERABLE Version: 2.4 Date:

Universities as Laboratories for Societal Multilingualism: Insights from Implementation

Challenges for Higher Education in Europe: Socio-economic and Political Transformations

Politics and Society Curriculum Specification

INSTITUTIONAL FACT SHEET

CALL FOR PARTICIPANTS

Psychometric Research Brief Office of Shared Accountability

Peer Influence on Academic Achievement: Mean, Variance, and Network Effects under School Choice

Post-intervention multi-informant survey on knowledge, attitudes and practices (KAP) on disability and inclusive education

The Talent Development High School Model Context, Components, and Initial Impacts on Ninth-Grade Students Engagement and Performance

The Effectiveness of Realistic Mathematics Education Approach on Ability of Students Mathematical Concept Understanding

Integrating Common Core Standards and CASAS Content Standards: Improving Instruction and Adult Learner Outcomes

ACADEMIC AFFAIRS GUIDELINES

A pilot study on the impact of an online writing tool used by first year science students

Financiación de las instituciones europeas de educación superior. Funding of European higher education institutions. Resumen

Academic profession in Europe

GREAT Britain: Film Brief

Audit Of Teaching Assignments. An Integrated Analysis of Teacher Educational Background and Courses Taught October 2007

Sample Reports. for Progress Test in Maths.

EUROPEAN UNIVERSITIES LOOKING FORWARD WITH CONFIDENCE PRAGUE DECLARATION 2009

Principal vacancies and appointments

Organising ROSE (The Relevance of Science Education) survey in Finland

Alternative education: Filling the gap in emergency and post-conflict situations

Iowa School District Profiles. Le Mars

Department of Sociology and Social Research

Abu Dhabi Grammar School - Canada

MGT/MGP/MGB 261: Investment Analysis

Centre for Evaluation & Monitoring SOSCA. Feedback Information

RELATIONS. I. Facts and Trends INTERNATIONAL. II. Profile of Graduates. Placement Report. IV. Recruiting Companies

CONSTRUCTION OF AN ACHIEVEMENT TEST Introduction One of the important duties of a teacher is to observe the student in the classroom, laboratory and

Navitas UK Holdings Ltd Embedded College Review for Educational Oversight by the Quality Assurance Agency for Higher Education

EQE Candidate Support Project (CSP) Frequently Asked Questions - National Offices

UNIVERSITY AUTONOMY IN EUROPE II

Transcription:

STATISTICS BRIEF Measuring student knowledge and skills The PISA 00 Experience By Andreas Schleicher and Claudia Tamassia June 02 No. 4 In this issue 1 What is PISA? 3 Target population and samples 3 Content coverage and implementation How did PISA ensure a fair assessment across OECD countries, measure student proficiency, 6 decompose variation in student performance? 7 Glossary 8 Further information How well do school systems perform in providing young people with a solid foundation of knowledge and skills, and in preparing them for life and learning beyond school? International comparisons of the outcomes of education systems have in the past been elusive. While it has been possible to compare basic structural characteristics of educational programmes and qualifications across countries, such as their entry requirements, their labour-market destination or typical patterns of student participation (see OECD, 1999b), there are no agreed standards that would allow to compare the level of content and quality of the underlying educational activities and services. It is thus difficult to make inferences regarding the knowledge and skills individuals have actually attained from such comparisons. Moreover, knowledge and skills are acquired not just through formal education but also, and increasingly, through formal and informal learning outside regular educational programmes. However, since 1997 governments of the OECD have been working on establishing a comparative framework to assess how well their education systems meet core objectives. The result has been the OECD Programme for International Student Assessment (PISA), the most comprehensive and rigorous international effort to date to assess learning outcomes and to identify the policy levers that may help improving the performance of education systems. What is PISA? The first PISA survey was conducted in 00 in 32 countries (including 28 OECD Member countries), with more than a quarter of a million sampled 1-year-olds taking internationally standardised tests in schools under independently supervised testing conditions. PISA 00 surveyed reading, mathematical, and scientific literacy, with a primary focus on reading. The survey will be repeated every three years, with successive cycles psychometrically Performance 80 Chart 1. Student performance and spending per student Relationship between average performance across the combined reading, mathematical and scientific literacy scales and cumulative expenditure on educational institutions up to age 1 60 40 00 480 460 Finland Japan United Kingdom Australia Czech Republic Ireland Belgium Austria Norway Switzerland Hungary France Spain Germany Denmark Poland Greece Portugal 440 4 400 380 0 Mexico 10 000 000 30 000 40 000 0 000 60 000 70 000 80 000 Cumulative expenditure per student (US dollars converted using purchasing power parities) Source: Knowledge and Skills for Life: First Results from PISA 00, OECD, 01. Organisation for Economic Co-operation and Development

linked so as to allow for comparisons over time, and with the primary focus shifting to mathematics in 03, science in 06 and back to reading in 09. First results from PISA were published in 01, showing how well 1-year-olds across the OECD can apply knowledge and skills in key subject areas and what they are like as learners. For some countries, these results were disappointing, showing that their students performance lags considerably behind that of their counterparts, sometimes by the equivalent of several school years, and sometimes despite high investments in education, both in terms of government spending and student learning time (see Chart 1). Overall, however, the results provided encouraging insights: the performance of countries such as Finland, Japan and suggests that excellence in education is an attainable goal, and at reasonable cost. Equally important, the results show that high performance standards do not have to come at the price of large disparities in student performance: six out of the eight countries with the smallest disparities in mathematics knowledge and skills the most school-bound subject assessed by PISA all performed well overall. Finally, the examples of Canada, Finland, Japan, and, which combined high performance levels with an exceptionally moderate impact of social background on student performance, underline that poor performance in school does not automatically follow from a disadvantaged socioeconomic background of students. The high performance standards achieved by some countries set ambitious goals for others. The question is what they can learn from them to help students to learn better, teachers to teach better, and schools to be more effective. While a cross-national survey such as PISA cannot show which policies or practices cause success, it does allow one to observe some common characteristics of students, schools and education systems, which do well. To this end, PISA 00 also collected data that could be used to construct indicators pointing to social, cultural, economic and educational factors that are associated with student performance. These allow one to address differences: between countries in the relationships between student level factors (such as gender, motivation, and family context) and achievement; in the relationships between school level factors and achievement across countries; in the proportion of variation in achievement between (rather than within) schools, and differences in this value across countries; between countries in the extent to which schools moderate or increase the effects of individual-level student factors and student achievement; in education systems and national context that are related to differences in student achievement across countries; and, PISA 00 an internationally standardised assessment of 1-year-olds Sample size 20,000 1-year-olds in the 32 participating countries were sampled. 11 additional countries in 02. Coverage three domains: reading, mathematical and scientific literacy. ability to use knowledge and skills to meet real-life challenges. mastery of processes, understanding of concepts, and ability to function in various situations within each domain. information on students attitudes to learning. Methods pencil-and-paper assessments, lasting two hours for each student. multiple-choice items and questions requiring students to construct their own answers. total of seven hours of assessment items included, with different students taking different combinations of the assessment items. questionnaires on student background as well as learning and study practices. school principals completed a questionnaire about their school. Outcomes a profile of knowledge and skills among 1-year-olds. contextual indicators relating results to student and school characteristics. a knowledge base for policy analysis and research. trend indicators showing how results change over time, once data become available from subsequent cycles of PISA. Future assessments three-year cycles, next in 03. 2

in the future, changes in any or all of these relationships over time. Through the collection of such information at the student and school level on a cross-nationally comparable basis, PISA adds significantly to the knowledge base that was previously available from aggregate national statistics. Such new insights do, however, come at a cost. PISA is both a resource intensive and a methodologically complex undertaking. This Statistics Brief describes some of those methodologies, along with the features that have ensured that PISA provides high quality data that can support policy formation and review. 1 Target population and samples PISA 00 was designed to provide an assessment of the cumulative yield of education and learning at a point at which most young adults are still enrolled in initial education. A major challenge for an international survey of this kind is to operationalise such a concept in ways that ensure the international comparability of national target populations. Differences between countries in the nature and extent of pre-primary education and care, the age of entry to formal schooling, and the institutional structure of educational systems do not allow the definition of internationally comparable grade levels of schooling. Some previous international assessments have defined their target population on the basis of the grade level that provide maximum coverage of a particular age cohort. A disadvantage of this approach is that slight variations in the age distribution of students across grade levels often lead to the selection of different target grades in different countries, raising serious questions about the comparability of results across countries. In addition, because not all students of the desired age are usually represented in grade-based samples, there may be a more serious potential bias in the results if the unrepresented students are typically enrolled in the next higher grade in some countries and the next lower grade in others. This would exclude students with potentially higher levels of performance in the former countries and students with potentially lower levels of performance in the latter. In order to resolve this problem, PISA 00 uses an agebased definition for its target population, i.e. a definition that is not tied to the institutional structures of national education systems: PISA assesses students who are 1 years of age at the time of testing and are enrolled in an educational institution, regardless of the grade levels 1. For further information, see the OECD, 02b. or type of institution in which they were enrolled, and of whether they were in full-time or part-time education. PISA 00 thus makes statements about the knowledge and skills of a group of individuals who were born within a comparable reference period, but who may have undergone different educational experiences both within and outside schools. The precise composition of the population of 1-year-old students was dependent on the time of testing, which each country selected so as to yield comparable data across countries in terms of student ages. All countries attempted to maximise the coverage of 1- year-olds enrolled in education in their national samples, including students enrolled in special educational institutions. The sampling standards used in PISA did, however, permit countries to exclude, under certain defined circumstances, up to a total of per cent of the relevant population either by excluding schools or by excluding students within schools. The ceiling for population exclusions of per cent ensures that potential bias resulting from exclusions is likely to remain within about one standard error of sampling. Quality standards, procedures, instruments and verification mechanisms were developed for PISA that ensured that the national probability samples yielded comparable data and that the results can be compared with confidence. Data quality standards in PISA required minimum participation rates for schools as well as for students. In the case of countries meeting these standards, it can be assumed that any bias resulting from non-response will be negligible, i.e. typically smaller than the sampling error. Content coverage and implementation The most challenging stage in the development of PISA was to build consensus among OECD countries on the knowledge and skills that should be used as benchmarks for an assessment of the outcomes of education systems, and then to define and operationalise these in ways that are cross-nationally valid. Following much discussion, Member countries agreed to focus the first PISA assessment on reading, mathematical and scientific literacy. A panel of renowned international experts led, in close consultation with OECD countries, the identification of the range of skills and competencies that were, in the three literacy areas, considered to be crucial for an individual s capacity to fully participate in and contribute to a successful modern society. Each assessment area was then defined to reflect: firstly, knowledge of a set of fundamental skills and understandings that are specific to the assessment area and secondly, a capacity to utilise those skills to address real-life issues and problems. 3

Finland Canada New Zealand Australia Ireland United Kingdom Japan Austria Belgium Iceland Norway France Denmark Switzerland Spain Czech Republic Germany Liechtenstein Hungary Poland Greece Portugal Russian Federation Latvia Luxembourg Mexico Brazil Performance of countries on the reading literacy scale The figure below summarises the performance of countries on the reading literacy scale. It also indicates whether countries perform significantly higher or lower than the comparison countries as well as the estimated rank order position of each country. For example, Finland, with all triangles pointed up performed significantly better than all other countries while Canada performed significantly lower than Finland shown by a triangle pointed down, similarly to New Zealand, Australia, Ireland and Japan shown by a circle and significantly better than all other countries shown by a triangle pointed up. Multiple comparisons of mean performance on the combined reading literacy scale Above the country average Not statistically significant from the country average Below the country average Average performance significantly higher than comparison country No statistical difference from comparison country Average performance significantly lower than comparison country Countries Finland Canada New Zealand Australia Ireland United Kingdom Japan Austria Belgium Iceland Norway France Denmark Switzerland Spain Czech Republic Germany Liechtenstein Hungary Poland Greece Portugal Russian Federation Latvia Luxembourg Mexico Brazil Possible ranks* Mean 46 34 29 28 27 2 23 22 16 07 07 07 0 0 04 497 494 493 492 487 484 483 480 479 474 470 462 48 441 422 396 SE (2.6) (1.6) (2.8) (3.) (3.2) (2.4) (2.6) (.2) (2.2) (2.4) (3.6) (1.) (2.8) (2.7) (7.0) (2.4) (4.2) (2.7) (2.4) (2.9) (2.) (4.1) (4.0) (4.) (.0) (4.) (4.2) (.3) (1.6) (3.3) (3.1) 46 (2.6) 34 (1.6) 29 (2.8) 28 (3.) 27 (3.2) 2 (2.4) 23 (2.6) 22 (.2) 16 (2.2) 07 (2.4) 07 (3.6) 07 (1.) 0 (2.8) 0 (2.7) 04 (7.0) 497 (2.4) 494 (4.2) 493 (2.7) 492 (2.4) 487 (2.9) 484 (2.) 483 (4.1) 480 (4.0) 479 (4.) 474 (.0) 470 (4.) 462 (4.2) 48 (.3) 441 (1.6) 422 (3.3) 396 (3.1) Upper 1 2 2 2 3 4 3 9 11 11 11 11 11 10 16 16 17 17 19 21 21 21 23 24 27 27 30 31 32 Lower 1 4 8 9 9 9 9 10 11 16 16 1 16 16 19 21 21 21 24 2 26 26 27 28 28 29 29 30 31 32 Source: Knowledge and Skills for Life: First Results from PISA 00, OECD, 01. * Because data are based on samples, it is not possible to report exact rank order positions for countries. However, it is possible to report the range of rank order positions within which the country mean lies with 9 per cent likelihood. A detailed operationalisation of the assessment areas referred to as the assessment framework was then developed and agreed at both scientific and policy levels. On the basis of the framework, assessment tasks were then developed and piloted in a Field Trial in all participating countries before a final set of tasks was selected for PISA 00. Each PISA task consisted of a stimulus (or multiple stimuli) and a set of one or more questions. The tasks included multiple-choice questions as well as open constructed-response questions that allowed for open individual responses and differing viewpoints. The latter usually asked students to relate information or ideas in the stimulus text to their own experience or opinions, with the acceptability depending less on the position taken by the student than on the ability to use 4

what they had read when justifying or explaining that position. Partial credit was often given for partially correct or less sophisticated answers, and all of these tasks were marked by hand. 2 How did PISA ensure a fair assessment across OECD countries? PISA devoted significant attention and resources to reflecting the national, cultural and linguistic variety among OECD countries. As part of this effort, the assessment included material contributed by participating countries. In addition, each participating country rated the cultural relevance and appropriateness of each of the assessment tasks in the national context and the results of this exercise were, together with the empirical results from the Field Trial, taken into consideration in the establishment of the PISA 00 instruments. Nevertheless, the use of an international assessment based on tasks drawn from a wide range of countries and cultural contexts raises the question for individual countries as to what extent its relative standing in the international comparisons might have been different if a task set had been used that reflected more closely the specific national educational, cultural and linguistic context. Since participating countries had rated the appropriateness of each of the PISA 00 assessment tasks, it was possible to explore how countries would have performed if the PISA assessment would have been based only on those tasks which each country considered most appropriate in the national context. Such an analysis was undertaken and showed a high degree of consistency between the results based on the international task set and those based only on the national preferred tasks. All but two countries ( and Norway) show either no differences at all or would have scored only one or two rank order positions higher if their national preferred task set had been used as the basis of comparison. Another requirement was to administer PISA 00 in 16 different languages. The challenge of making tests linguistically equivalent in each of these languages is a considerable one. PISA addressed this in several ways: first, PISA prepared and distributed all of its assessment materials in two source languages; second, participating countries developed national versions based upon independent translations from the two source versions, followed by a reconciliation of the resulting two national versions; and third, all national versions were verified against the English and French source versions by specially trained trilingual verifiers. The development of national versions derived from 2. For a sample of the tasks used see OECD, 02a. two different source languages is shown to have many advantages over other methods. Finally, detailed guidelines contributed to a marking process of student responses that was accurate and consistent across countries. In each country, a sub-sample of test booklets was marked independently by four markers on the basis of precise guidelines and examined by the PISA Consortium (see glossary). In order to examine the consistency of this marking process in more detail within each country and to estimate the magnitude of the variance components associated with the use of markers, PISA conducted an inter-marker reliability study on a sub-sample of assessment booklets. Homogeneity analysis was applied to the national sets of multiple marking and compared with the results of the Field Trial. At the between-country level, an inter-country reliability study was carried out on a sub-set of tasks. The aim was to check whether the marking given by national markers was of equal severity in each country, both overall and for particular tasks. The survey was implemented through standardised procedures. The PISA Consortium provided comprehensive manuals that explained the implementation of the survey, including precise instructions for the work of School Coordinators and scripts for the Test Administrators for use during the testing sessions. To minimise the burden on the participating schools, to establish the credibility of PISA as valid and as unbiased and to encourage uniformity in the administration of the testing sessions, independent Test Administrators conducted the assessment based on detailed and internationally agreed protocols. National Quality Monitors appointed by the international PISA Consortium visited all national centres to review data-collection procedures. Finally, School Quality Monitors from the PISA Consortium visited a sample of 2 per cent of the schools during the assessment. How did PISA measure student proficiency? PISA 00 required each sampled student to respond to a 1-minute test. Because a single 1-minute test is not sufficient to allow a representative coverage of the content to be assessed in reading, mathematics and science, a test rotation system was employed. In this system a total of 8 different reading, mathematics and science tasks were distributed around 9 two-hour booklets. By randomly assigning one of these 9 booklets to each student and employing item response scaling methods, it was possible to cover a wide-range of material within a two-hour testing period for each individual student. In these scaling methods, the tasks are described

by a fixed set of unknown task difficulty parameters. The task difficulty parameters are then estimated through a model based on the combined international data set with countries given equal weight. On the basis of the international item difficulty parameters, estimates of student performance are then derived. For each of the PISA 00 areas, the scale was standardised such that the student mean score across OECD Member-countries was 00 and the standard deviation was 100. To give substantive meaning to the numerical scale scores, levels of proficiency were constructed that were derived based on the skill requirements of the underlying tasks. Finally, a content-based interpretation for scores in each level was developed. Chart 2 presents the percentage of students performing at each of the proficiency levels on the PISA 00 reading literacy scale. How did PISA decompose variation in student performance? The fact that PISA measures proficiency not at the aggregate country level, but at the level of individual students, makes it possible to also examine variation in student performance within countries. Such variation may result from the socio-economic backgrounds of students and schools, from the human and financial resources available to schools, from curricular differences, from selection policies and practices and from the way in which teaching is organised and delivered. Some countries have non-selective school systems that seek to provide all students with the same opportunities for learning and that allow each school to cater for the full range of student performance. Other countries respond to diversity explicitly by forming groups of students of similar performance levels through selection either within or between schools, with the aim of serving students according to their specific needs. In many countries, combinations of the two approaches occur. Even in comprehensive school systems, there may be significant variation between schools due to the socio-economic and cultural characteristics of the communities that the schools serve or to geographical differences (such as differences between regions, provinces or states in federal systems, or differences between rural and urban areas). Chart 2. Percentage of students performing at each of the proficiency levels on the combined reading literacy scale Percentage of students: Below level 1 At level 1 At level 2 At level 3 At level 4 At level 100 100 80 60 40 0 18 32 29 6 31 39 17 10 14 29 28 27 28 33 30 19 18 16 11 12 9 9 11 8 26 2 24 26 26 2 24 24 24 2 26 27 30 26 30 31 28 31 12 8 9 4 7 9 6 21 4 2 21 22 21 19 19 19 18 17 17 11 1 6 29 33 27 28 31 31 27 28 29 28 27 2 19 19 30 3 4 13 14 27 2 1 3 13 80 60 40 0 40 60 80 14 2 19 18 18 18 17 19 7 7 8 9 9 9 1 9 2 3 3 3 4 3 17 11 8 22 22 10 11 11 4 4 6 22 21 23 21 11 12 12 13 4 6 6 7 26 2 26 12 11 14 4 6 22 13 10 24 2 26 2 1 16 16 17 9 7 9 10 27 21 14 30 28 16 23 1 8 29 18 9 26 18 13 28 33 23 40 60 80 100 100 Finland Canada Japan Ireland New Zealand Australia United Kingdom Belgium Austria Iceland Norway France Denmark Switzerland Spain Czech Republic Germany Poland Hungary Greece Portugal Luxembourg Mexico Liechtenstein Russian Federation Latvia Brazil Source: Knowledge and Skills for Life: First Results from PISA 00, OECD, 01. 6

How do the policies and historical patterns that shape each country s school system affect and relate to the overall variation in student performance? The decomposition of the variation in student performance shown in Chart 3 provides some insights. Chart 3 shows the extent of variation attributable to different factors in each country. The length of the bars indicate the total observed variation in student performance on the PISA reading literacy scale. The bar for each country is aligned in the figure such that variation between schools is represented by the length of the bar to the left of the vertical line down the centre of the figure (e.g., 76.0 per cent for Belgium), and variation within schools is represented by the length of the bar to the right of that vertical line (e.g., 0.9 per cent for Belgium). If the sum of the between and within-school variance components is larger than 100, this indicates that variation in student performance is greater in the corresponding country than in a typical OECD country. Similarly, a value smaller than 100 indicates belowaverage variation in student performance. As shown, in most countries a considerable portion of the variation in student performance lies between schools. Further analysis of the data suggests that, in school systems with differentiated school types, the clustering of students with particular socio-economic characteristics in certain schools is greater than in systems where the curriculum does not vary significantly between schools. PISA results suggest that the identification of factors explaining why some schools or some countries have better results than others is complex. Further research Chart 3. Variation in student performance between and within schools on the combined reading literacy scale Expressed as a percentage of the average variation in student performance in OECD countries Belgium Germany Japan United Kingdom Australia New Zealand Canada Finland Variation explained by geographical, systemic and institutional factors Between school variation 76.0 74.8 0.9 36. 3.1 22.4.9.1 19.7 17.1 10.7 8.9 Within school variation 33.0 0.9 0.2 43.4 43.9 83.6 82.3 103.9 80.1 76. 90.6 83.0 1 80 40 0 40 80 1 Source: Knowledge and Skills for Life: First Results from PISA 00, OECD, 01. and analysis will be necessary to identify how these factors operate, interact with home background, and influence student performance. Glossary Reading literacy: understanding, using, and reflecting on written texts, in order to achieve one s goals, to develop one s knowledge and potential, and to participate in society. Mathematical literacy: capacity to identify and understand the role that mathematics plays in the world, to make well-founded judgements and to engage in mathematics, in ways that meet the needs of that individual s life as a constructive, concerned, and reflective citizen. Scientific literacy: capacity to use scientific knowledge, to identify questions and to draw evidence-based conclusions in order to understand and help make decisions about the natural world and the changes made to it through human capacity. PISA Consortium: The implementation of PISA was the responsibility of an international consortium led by the Australian Council for Educational Research. The other partners in this consortium were the National Institute for Educational Measurement in the Netherlands, Westat and the Educational Testing Service in the, and the National Institute for Educational Policy Research in Japan. Proficiency level: The PISA proficiency scales are divided into five levels of knowledge and skills, describing what students can do. Students at a particular level not only demonstrate the knowledge and skills of that level but also the ones required at lower levels. 7

Further information Access to the complete micro-level database for PISA as well as to relevant technical documentation and the PISA reports is provided through the website www.pisa.oecd.org. Measuring Student Knowledge and Skills: A New Framework for Assessment, OECD, 1999a. Conceptual framework on which the PISA 00 assessment is based, defining the domains of reading, mathematical and scientific literacy in terms of the content that students need to acquire, the processes that need to be performed, and the contexts in which knowledge and skills are applied. Classifying Educational Programmes. Manual for ISCED-97 Implementation in OECD Countries, OECD, 1999b. Presents the educational structures of the OECD countries into a comparable international framework. Knowledge and Skills for Life: First Results from PISA 00, OECD, 01. Presents evidence on student performance in reading, mathematical and scientific literacy, revealing factors that influence the development of these skills at home and at school, and examining what the implications are for policy development. Sample Tasks from the PISA 00 Assessment: Reading, Mathematical and Scientific Literacy, OECD, 02a. Description of the instruments underlying the PISA 00 assessment. It introduces the PISA approach to assessing reading, mathematical and scientific literacy with its three dimensions of process, content and context. It also presents a sample of PISA 00 tasks, and explains how these tasks were scored and how they relate to the conceptual framework underlying PISA. PISA 00 Technical Report, OECD, 02b, forthcoming. Report of all technical aspects related to the implementation of the PISA assessment including: test design and development; questionnaire design and development; sampling design; quality standards and quality monitoring procedures; weighting, scaling and error variance estimation procedures; scale construction; data processing and database preparation; and, questionnaire analysis. The Statistics Brief is published by the Statistics Directorate of the OECD Editor in chief: Enrico Giovannini Editor: Andreas Lindner Editorial team: Benoit Arnaud, Eileen Capponi, Lydia Deloumeaux, Brian Finn, Joscelyn Magdeleine For further information contact: The Editor STATISTICS BRIEF Statistics Directorate OECD 2, rue André-Pascal 777 Paris France email: STD.STATBRIEF@oecd.org Do you want to subscribe to Statistics Brief? Register now at www.oecd.org/oecddirect and choose Statistics Brief or e-mail/write to the Editor. OECD Worldwide OECD NORTH AMERICA OECD Washington Center 01 L Street N.W., Suite 60 Washington, DC 036-4922, USA Toll free: +1 (800) 46-6323 Fax: +1 (2) 78-030 General information: +1 (2) 78-6323 E-mail: washington.contact@oecd.org Internet: www.oecdwash.org OECD JAPAN OECD Tokyo Centre Nippon Press Center Bldg., 3rd floor 2-2-1 Uchisaiwaicho, Chiyoda-ku Tokyo 100-0011, Japan Tel.: +81 (3) 32 0021 Fax: +81 (3) 32 003 E-mail: center@oecdtokyo.org Internet: www.oecdtokyo.org OECD GERMANY OECD Berlin Centre Albrechtstrasse 9, 3. OG 10117 Berlin-Mitte Tel: +49 30 288 833 Fax: +49 30 288 834 E-mail: berlin.contact@oecd.org Internet: www.oecd.org/deutschland OECD Bonn Centre August Bebel Allee 6 317 Bonn Tel: +49 228 99 12 1 Fax: +49 228 99 12 18 E-mail: bonn.contact@oecd.org Internet: www.oecd.org/deutschland OECD MEXICO OECD Mexico Centre Av. Presidente Mazaryk 26 Colonia: Polanco C.P. 1160 Mexico D.F., Mexico Tel.: +2 2 81 38 10 Fax: +2 2 80 04 80 E-mail: mexico.contact@oecd.org Internet: rtn.net.mx/ocde/ OECD PARIS OECD Paris Centre 2, rue André-Pascal 777 Paris Cedex 16, France Tel: +33 1 4 24 81 67 Fax: +33 1 4 24 19 0 E-mail: sales@oecd.org Internet: www.oecd.org/bookshop 8 81783