ReaderBench: A Multi-lingual Framework for Analyzing Text Complexity

Similar documents
Designing Autonomous Robot Systems - Evaluation of the R3-COP Decision Support System Approach

Teachers response to unexplained answers

Specification of a multilevel model for an individualized didactic planning: case of learning to read

Smart Grids Simulation with MECSYCO

Towards a MWE-driven A* parsing with LTAGs [WG2,WG3]

A Novel Approach for the Recognition of a wide Arabic Handwritten Word Lexicon

User Profile Modelling for Digital Resource Management Systems

Specification of the Verity Learning Companion and Self-Assessment Tool

Students concept images of inverse functions

English Language and Applied Linguistics. Module Descriptions 2017/18

Automating the E-learning Personalization

Process Assessment Issues in a Bachelor Capstone Project

AQUA: An Ontology-Driven Question Answering System

Word Segmentation of Off-line Handwritten Documents

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

A Study of Synthetic Oversampling for Twitter Imbalanced Sentiment Analysis

Multi-Lingual Text Leveling

Learning Methods for Fuzzy Systems

Language specific preferences in anaphor resolution: Exposure or gricean maxims?

Raising awareness on Archaeology: A Multiplayer Game-Based Approach with Mixed Reality

AUTHORING E-LEARNING CONTENT TRENDS AND SOLUTIONS

Linking Task: Identifying authors and book titles in verbose queries

The 9 th International Scientific Conference elearning and software for Education Bucharest, April 25-26, / X

Probabilistic Latent Semantic Analysis

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

The Effect of Syntactic Simplicity and Complexity on the Readability of the Text

Reducing Features to Improve Bug Prediction

Florida Reading Endorsement Alignment Matrix Competency 1

Postprint.

THE WEB 2.0 AS A PLATFORM FOR THE ACQUISITION OF SKILLS, IMPROVE ACADEMIC PERFORMANCE AND DESIGNER CAREER PROMOTION IN THE UNIVERSITY

USER ADAPTATION IN E-LEARNING ENVIRONMENTS

DYNAMIC ADAPTIVE HYPERMEDIA SYSTEMS FOR E-LEARNING

CROSS-LANGUAGE INFORMATION RETRIEVAL USING PARAFAC2

CaMLA Working Papers

Language Independent Passage Retrieval for Question Answering

Robot Learning Simultaneously a Task and How to Interpret Human Instructions

Laboratorio di Intelligenza Artificiale e Robotica

MASTER OF SCIENCE (M.S.) MAJOR IN COMPUTER SCIENCE

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Axiom 2013 Team Description Paper

Intermediate Academic Writing

Arts, Literature and Communication International Baccalaureate (500.Z0)

Technology-mediated realistic mathematics education and the bridge21 model: A teaching experiment

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

Parsing of part-of-speech tagged Assamese Texts

Program Matrix - Reading English 6-12 (DOE Code 398) University of Florida. Reading

A Case Study: News Classification Based on Term Frequency

Ontologies vs. classification systems

Top US Tech Talent for the Top China Tech Company

Content-free collaborative learning modeling using data mining

OilSim. Talent Management and Retention in the Oil and Gas Industry. Global network of training centers and technical facilities

LEXICAL COHESION ANALYSIS OF THE ARTICLE WHAT IS A GOOD RESEARCH PROJECT? BY BRIAN PALTRIDGE A JOURNAL ARTICLE

Operational Knowledge Management: a way to manage competence

Multilingual Sentiment and Subjectivity Analysis

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Evidence for Reliability, Validity and Learning Effectiveness

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

CWIS 23,3. Nikolaos Avouris Human Computer Interaction Group, University of Patras, Patras, Greece

The Enterprise Knowledge Portal: The Concept

Detecting Wikipedia Vandalism using Machine Learning Notebook for PAN at CLEF 2011

E-learning Strategies to Support Databases Courses: a Case Study

Maeha a Nui: A Multilingual Primary School Project in French Polynesia

Linguistics. Undergraduate. Departmental Honors. Graduate. Faculty. Linguistics 1

A Comparison of Two Text Representations for Sentiment Analysis

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Agent-Based Software Engineering

Conference Presentation

The Impact of Neuroscience on Foreign Languages in School

Online Marking of Essay-type Assignments

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

(English translation)

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

the contribution of the European Centre for Modern Languages Frank Heyworth

Highlighting and Annotation Tips Foundation Lesson

International Conference on Education and Educational Psychology (ICEEPSY 2012)

InTraServ. Dissemination Plan INFORMATION SOCIETY TECHNOLOGIES (IST) PROGRAMME. Intelligent Training Service for Management Training in SMEs

arxiv: v1 [cs.cl] 2 Apr 2017

Applications of memory-based natural language processing

Speech Recognition at ICSI: Broadcast News and beyond

The Use of Statistical, Computational and Modelling Tools in Higher Learning Institutions: A Case Study of the University of Dodoma

The MEANING Multilingual Central Repository

GALICIAN TEACHERS PERCEPTIONS ON THE USABILITY AND USEFULNESS OF THE ODS PORTAL

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

A student diagnosing and evaluation system for laboratory-based academic exercises

Modern Languages. Introduction. Degrees Offered

Timeline. Recommendations

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

Prentice Hall Literature: Timeless Voices, Timeless Themes, Platinum 2000 Correlated to Nebraska Reading/Writing Standards (Grade 10)

Evolutive Neural Net Fuzzy Filtering: Basic Description

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

Nancy Guilloteau The University of Texas at Austin Department of French and Italian 201 W. 21st St. STOP B7600 Austin, Texas 78712

Chamilo 2.0: A Second Generation Open Source E-learning and Collaboration Platform

University of Exeter College of Humanities. Assessment Procedures 2010/11

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Assessing Entailer with a Corpus of Natural Language From an Intelligent Tutoring System

INPE São José dos Campos

Pearson Longman Keystone Book F 2013

A Coding System for Dynamic Topic Analysis: A Computer-Mediated Discourse Analysis Technique

Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability

Transcription:

ReaderBench: A Multi-lingual Framework for Analyzing Text Complexity Mihai Dascalu, Gabriel Gutu, Stefan Ruseti, Ionut Cristian Paraschiv, Philippe Dessus, Danielle Mcnamara, Scott Crossley, Stefan Trausan-Matu To cite this version: Mihai Dascalu, Gabriel Gutu, Stefan Ruseti, Ionut Cristian Paraschiv, Philippe Dessus, et al.. ReaderBench: A Multi-lingual Framework for Analyzing Text Complexity. É. Lavoué; H. Drachsler; K. Verbert; J. Broisin; M. Pérez-Sanagustín. Data Driven Approaches in Digital Education, Proc 12th European Conference on Technology Enhanced Learning, EC-TEL 2017, 2017, Tallinn, Estonia. Springer, pp.606-609, 2017, Data Driven Approaches in Digital Education 12th European Conference on Technology Enhanced Learning, EC-TEL 2017, Tallinn, Estonia, September 12 15, 2017, Proceedings... HAL Id: hal-01584870 https://hal.archives-ouvertes.fr/hal-01584870 Submitted on 10 Sep 2017

HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

ReaderBench: A Multi-Lingual Framework for Analyzing Text Complexity Mihai Dascalu 1,2, Gabriel Gutu 1, Stefan Ruseti 1, Ionut Cristian Paraschiv 1, Philippe Dessus 3, Danielle S. McNamara 4, Scott A. Crossley 5, Stefan Trausan-Matu 1,2 1 University Politehnica of Bucharest, Splaiul Independenței 313, 60042, Romania {mihai.dascalu, gabriel.gutu, stefan.ruseti, ionut.paraschiv, stefan.trausan}@cs.pub.ro 2 Academy of Romanian Scientists, Splaiul Independenţei 54, 050094, Bucharest, Romania 3 Laboratoire des Sciences de l Éducation, Univ. Grenoble Alpes, F38000 Grenoble, France philippe.dessus@univ-grenoble-alpes.fr 4 Institute for the Science of Teaching & Learning, Arizona State University, Tempe, USA dsmcnama@asu.edu 5 Georgia State University, Department of Applied Linguistics/ESL, Atlanta, 30303, USA scrossley@gsu.edu Abstract. Assessing textual complexity is a difficult, but important endeavor, especially for adapting learning materials to students and readers levels of understanding. With the continuous growth of information technologies spanning through various research fields, automated assessment tools have become reliable solutions to automatically assessing textual complexity. ReaderBench is a text processing framework relying on advanced Natural Language Processing techniques that encompass a wide range of text analysis modules available in a variety of languages, including English, French, Romanian, and Dutch. To our knowledge, ReaderBench is the only open-source multilingual textual analysis solution that provides unified access to more than 200 textual complexity indices including: surface, syntactic, morphological, semantic, and discourse specific factors, alongside cohesion metrics derived from specific lexicalized ontologies and semantic models. Keywords: Multi-Lingual Text Analysis, Textual Complexity, Comprehension Prediction, Natural Language Processing, Textual Cohesion, Writing Style 1 Introduction Two important and cumbersome tasks, which often face many teachers, are selecting reading materials suitable for their students levels of understanding, and assessing their written productions (e.g., essays, summaries). In order to support both tasks, ReaderBench [1], a multilingual, open-source framework centered on discourse analysis, was developed. From an architectural perspective, as shown in Figure 1, our framework comprises three layers: a) linguistic resources that provide solid language background knowledge and can be used to train the semantic models and compute various measures; b) linguistic services used to process and append semantic metainformation to text resources, and c) linguistic applications that rely on machine

learning and data mining techniques, and are designed for various educational experiments and visualizations. ReaderBench implements various metrics and categories of textual complexity indices that can be used to leverage the automated classification of datasets in multiple languages, such as English [2], French [3], Romanian [4] and Dutch [5]. Fig. 1. ReaderBench processing architecture. 2 Description of Textual Complexity Indices More than 200 textual complexity indices computed by the ReaderBench platform have been used in a number of experiments. ReaderBench integrates a multitude of indices, discussed briefly below, ranging from classic readability formulas, surface indices, morphology and syntax, as well as semantics and discourse structure. Surface indices. These are the simplest measures that consider only the form of the text. This category includes indices such as sentence length, word length, the number of unique words used, and word entropy. All these indices rely on the assumption that more complex texts contain more information and, inherently, more diverse concepts. Word complexity indices. This category of indices focuses on the complexity of words, but goes way beyond their form. Thus, the complexity of a word is estimated by the number of syllables and how different the flectional form is from its lemma or stem, considering that adding suffixes and prefixes increases the difficulty of using a given word. Moreover, a word s complexity is measured by considering the number of potential meanings derived from the word s senses available in WordNet, as well as a word s specificity reflected in its depth within the lexicalized ontology. Syntactic and morphologic indices. These indices are computed at the sentence level. The words corresponding parts of speech and the types of dependencies that appear in each sentence can be used as relevant measures, reflective of a text s complexity. In addition, named entity-based features are tightly correlated with the amount of cognitive resources required to understand the given text. Semantic cohesion indices. Cohesion plays an important role in text comprehension and our framework makes extensive usage of Cohesion Network

Analysis. ReaderBench estimates both local and global cohesion by considering lexical chains, different semantic models (semantic distances in different multilingual WordNets, LSA Latent Semantic Analysis, LDA Latent Dirichlet Allocation, and Word2Vec), as well as co-reference chains. Discourse structure indices. Specific discourse connectives and metrics derived from polyphonic model of discourse [1], which considers the evolution of expressed points of view, provide additional valuable insights in terms of the text s degree of elaboration. Word features and vectors from the integrated linguistic resources are also used to reflect specific discourse traits. 3 Validation Experiments Multiple experiments have been performed to validate ReaderBench as a multilingual text analysis software framework. This section focuses on the latest and most representative experiments conducted in English, French, Romanian, and Dutch languages. The first experiment [2] was performed on a set of 108 argumentative essays written in English and timed to 25 minutes. For the analysis, only essays that contained three or more paragraphs were considered in order to use global cohesion measures reflective of inter-paragraph relations. Individual difference measures such as vocabulary knowledge and reading comprehension scores were assessed. The results showed that writers with stronger vocabulary knowledge used longer words with multiple senses and higher entropy, but also created more cohesive essays. Also, students with higher reading comprehension scores created more cohesive and more lexically sophisticated essays, using longer words, and with higher entropy. The second experiment [3] relied on a set of 200 documents collected from primary school French manuals. The documents were pre-classified into five complexity classes mapped onto the first five primary grade levels of the French national education system. A Support Vector Machine (SVM) was used to classify the documents. The pre-trained model was used to determine the complexity for an additional set of 16 documents that were manually classified into three primary grades. Students belonging to the three classes had to read the texts and answer a posttest. Correlations between the textual complexity factors scores and the students average scores were computed. This allowed the computation of the impact for each factor in calculating the reliability of prediction of the textual complexity score for a given document. The third experiment [4] was conducted on a set of 137 documents written in Romanian language. The documents were collected from two time periods, 1941-1991 and 1992-present, and two regions, Bessarabia and Romania. The first period altered the Romanian language spoken in the country because of the implementation of the Russian language into the education system of Bessarabia. The aim of the experiment was to determine whether differences between the two regions and the two time periods could be observed in relation to the complexity of written texts. The analysis showed that more elaborated texts were created in the second period for both Bessarabia and Romania, while more unique words have been used in the second period for Bessarabia, but remained the same for Romania. The semantic cohesion of

the texts increased over time, but no significant differences were observed between the two regions. The fourth experiment [5] was run on a set of 173 technical reports written in Dutch language belonging to high or low performance students. Due to the length of the documents, a multi-level hierarchical structure was automatically generated based on the section headings. The experiment showed that students who received higher scores had longer reports, but also greater word entropy. They used more pronouns, discourse connectors and unique words, but also had lower inner cohesion scores per paragraph which is indicative of more sophisticated paragraphs. 4 Conclusion Many pedagogical scenarios can fully integrate the use of ReaderBench, thanks to its versatility. The wide range of textual assessment features can support both teachers assessment and learners writing self-regulation. Moreover, multiple learning contexts take advantage from ReaderBench s support: either individual textual production and reflection, or collaborative knowledge building. The presented experiments support the ReaderBench framework for determining the textual complexity of texts written in English, French, Romanian, and Dutch languages. Other languages, such as Spanish, Italian, and Latin are also partially supported. To our knowledge, ReaderBench is a unique multilingual system that provides access to a wide range of textual complexity indices and to various textual cohesion analyses. Acknowledgments. This research was partially supported by the FP7 2008-212578 LTfLL project, by the 644187 EC H2020 RAGE project, by the ANR-10-blan-1907-01 DEVCOMP project, as well as by University Politehnica of Bucharest through the Excellence Research Grants Program UPB GEX 12/26.09.2016. References 1. Dascalu, M.: Analyzing discourse and text complexity for learning and collaborating, Studies in Computational Intelligence, Vol. 534. Springer, Cham, Switzerland (2014) 2. Allen, L.K., Dascalu, M., McNamara, D.S., Crossley, S., Trausan-Matu, S.: Modeling Individual Differences among Writers Using ReaderBench. In: EduLearn16, pp. 5269 5279. IATED, Barcelona, Spain (2016) 3. Dascalu, M., Stavarache, L.L., Trausan-Matu, S., Dessus, P., Bianco, M.: Reflecting Comprehension through French Textual Complexity Factors. In: 26th Int. Conf. on Tools with Artificial Intelligence (ICTAI 2014), pp. 615 619. IEEE, Limassol, Cyprus (2014) 4. Gifu, D., Dascalu, M., Trausan-Matu, S., Allen, L.K.: Time Evolution of Writing Styles in Romanian Language. In: ICTAI 2016, pp. 1048 1054. IEEE, San Jose, CA (2016) 5. Dascalu, M., Westera, W., Ruseti, S., Trausan-Matu, S., Kurvers, H.: ReaderBench Learns Dutch: Building a Comprehensive Automated Essay Scoring System for Dutch. In: AIED 2017. Springer, Wuhan, China (in press)