AUTOMATIC LEARNING OBJECT CATEGORIZATION FOR INSTRUCTION USING AN ENHANCED LINEAR TEXT CLASSIFIER

Similar documents
SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

AQUA: An Ontology-Driven Question Answering System

CS Machine Learning

Rule Learning With Negation: Issues Regarding Effectiveness

Cross Language Information Retrieval

Word Segmentation of Off-line Handwritten Documents

Linking Task: Identifying authors and book titles in verbose queries

Beyond the Blend: Optimizing the Use of your Learning Technologies. Bryan Chapman, Chapman Alliance

A Case Study: News Classification Based on Term Frequency

Analyzing sentiments in tweets for Tesla Model 3 using SAS Enterprise Miner and SAS Sentiment Analysis Studio

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Lecture 1: Machine Learning Basics

Radius STEM Readiness TM

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Disambiguation of Thai Personal Name from Online News Articles

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Probability and Statistics Curriculum Pacing Guide

CREATING SHARABLE LEARNING OBJECTS FROM EXISTING DIGITAL COURSE CONTENT

Reducing Features to Improve Bug Prediction

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

On-Line Data Analytics

Combining a Chinese Thesaurus with a Chinese Dictionary

Learning Methods in Multilingual Speech Recognition

TIMSS ADVANCED 2015 USER GUIDE FOR THE INTERNATIONAL DATABASE. Pierre Foy

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Detecting Wikipedia Vandalism using Machine Learning Notebook for PAN at CLEF 2011

Specification of the Verity Learning Companion and Self-Assessment Tool

Probabilistic Latent Semantic Analysis

Rule Learning with Negation: Issues Regarding Effectiveness

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

A Bayesian Learning Approach to Concept-Based Document Classification

Controlled vocabulary

Switchboard Language Model Improvement with Conversational Data from Gigaword

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Multi-Lingual Text Leveling

Automatic document classification of biological literature

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Problems of the Arabic OCR: New Attitudes

Automating the E-learning Personalization

Developing True/False Test Sheet Generating System with Diagnosing Basic Cognitive Ability

PowerTeacher Gradebook User Guide PowerSchool Student Information System

Grade 6: Correlated to AGS Basic Math Skills

Iterative Cross-Training: An Algorithm for Learning from Unlabeled Web Pages

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

CS 446: Machine Learning

Interaction Design Considerations for an Aircraft Carrier Deck Agent-based Simulation

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Human Emotion Recognition From Speech

Speech Recognition at ICSI: Broadcast News and beyond

THE VERB ARGUMENT BROWSER

Exposé for a Master s Thesis

GACE Computer Science Assessment Test at a Glance

Using dialogue context to improve parsing performance in dialogue systems

Generative models and adversarial training

Hawai i Pacific University Sees Stellar Response Rates for Course Evaluations

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

The stages of event extraction

On the Combined Behavior of Autonomous Resource Management Agents

A Comparison of Two Text Representations for Sentiment Analysis

Python Machine Learning

South Carolina English Language Arts

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

On document relevance and lexical cohesion between query terms

Developing an Assessment Plan to Learn About Student Learning

Cross-Lingual Text Categorization

Analyzing Linguistically Appropriate IEP Goals in Dual Language Programs

Using Blackboard.com Software to Reach Beyond the Classroom: Intermediate

Bridging Lexical Gaps between Queries and Questions on Large Online Q&A Collections with Compact Translation Models

Trends in College Pricing

Term Weighting based on Document Revision History

Lecture 1: Basic Concepts of Machine Learning

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

Constructing Parallel Corpus from Movie Subtitles

National Literacy and Numeracy Framework for years 3/4

Modeling function word errors in DNN-HMM based LVCSR systems

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Designing e-learning materials with learning objects

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Learning Methods for Fuzzy Systems

Speech Emotion Recognition Using Support Vector Machine

Learning Disability Functional Capacity Evaluation. Dear Doctor,

A Topic Maps-based ontology IR system versus Clustering-based IR System: A Comparative Study in Security Domain

Australian Journal of Basic and Applied Sciences

Ensemble Technique Utilization for Indonesian Dependency Parser

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

Modeling function word errors in DNN-HMM based LVCSR systems

Loughton School s curriculum evening. 28 th February 2017

The development of a new learner s dictionary for Modern Standard Arabic: the linguistic corpus approach

Purdue Data Summit Communication of Big Data Analytics. New SAT Predictive Validity Case Study

Stefan Engelberg (IDS Mannheim), Workshop Corpora in Lexical Research, Bucharest, Nov [Folie 1] 6.1 Type-token ratio

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

CLASSIFICATION OF TEXT DOCUMENTS USING INTEGER REPRESENTATION AND REGRESSION: AN INTEGRATED APPROACH

Finding Translations in Scanned Book Collections

Context Free Grammars. Many slides from Michael Collins

Automating Outcome Based Assessment

As a high-quality international conference in the field

Word Sense Disambiguation

Transcription:

AUTOMATIC LEARNING OBJECT CATEGORIZATION FOR INSTRUCTION USING AN ENHANCED LINEAR TEXT CLASSIFIER THOMAS GEORGE KANNAMPALLIL School of Information Sciences and Technology, Pennsylvania State University, University Park, Pa 16802, USA thomasg@psu.edu ROBERT G. FARRELL Next Generation Web Dept, IBM T.J. Watson Research Center, Hawthorne, NY 10532, USA robfarr@us.ibm.com This paper explores the use of a machine learning algorithm to automate the task of classifying learning materials into categories useful for instruction. A collection of documents was segmented manually into independent learning objects. A regularized linear text classifier was trained to recognize four topic categories and eleven instructional use categories using manual category labels as training data. The classifier was able to categorize text-based learning objects into topic categories with high accuracy, but initial performance for instructional use classification was poor. An enhanced classifier was able to distinguish between conceptual and procedural categories of instructional use with high accuracy. 1 Introduction The World Wide Web (WWW) is increasingly used as a learning tool because of the wealth of information available on it. While search engines have been developed for the Web (Maudlin, 1998) and produce good results, there is an increasing number of specialized search engines for learning materials. These search engines improve browsing and retrieve through the use of learning object metadata The Institute for Electronic and Electrical Engineers (IEEE) has defined an international standard for learning object metadata (IEEE LOM, 2002) that is gaining acceptance in both corporate and university training and education communities. Increasingly, authoring tools allow content developers to enter metadata that fits this standard. However, manually entering metadata is a time-consuming process. Automated techniques are required for wider adoption of the standard. The IEEE learning object metadata standard defines several attributes that may be assigned to each learning object. We investigated whether two of these attributes could be derived automatically: classification and learning resource type. In many applications, the classification is by topic, so we concentrated on how to assign learning objects to one of several distinct topics (for e.g. topics like java or web services ). The learning resource type attribute describes the kind of learning object, such as an exercise or a simulation. We used a specialized vocabulary of eleven learning resource types focused on the instructional use of the learning object in web-based courseware (Farrell et. al., 2004). These categories include introduction, concepts, and procedures. 1

2 The topic and instructional use for a learning object are central to the operation of learning tools and applications, such as the second author s larger effort directed at dynamic assembly of learning objects (Farrell et. al, 2004). Automating the process of assigning topics and instructional use categories for new text-based learning materials could aid in building a large repository of learning objects for retrieval and assembly. Unfortunately, standard metadata editing tools do not automatically supply a value for these critical attributes. 2 Procedure In this work, we attempted to automatically infer topic and instructional use categories using machine learning techniques, based upon existing manual assignments made by experts. We chose a high performance text classification system developed at IBM (Zhang & Oles, 2001). We defined a small set of information technology topic categories based on interviews with a domain expert and two instructional experts on our team defined a set of instructional use categories. A team of independent subject-matter experts then assigned learning objects to both sets of categories. We then used these category assignments to train a linear text classifier. We chose three instructional use categories (procedures, concepts, and introduction) and four topic categories (j2ee, web services, websphere, and security) for our experiments, based upon the availability of sufficient training data assigned to these categories. The basic experimental set up is shown in figure 1. Technical reference documents were collected and stored in the DocBook XML format (Docbook). PowerPoint TM presentation documents were collected and saved as HTML. Figure 1: Experimental Setup for Automatic Learning Object Text Classification First, text-based learning objects were manually extracted from reference books and informational presentations and tagged with topic and instructional use categories (Farrell et. al. 2003). Next, the resulting category labels were saved in a learning object metadata (LOM) XML file. A random subset of the learning objects with associated text content

and category labels were then passed to a regularized linear text classifier for training. The trained classifier was then used to automatically categorize the remaining learning objects. 3 Results from Topic and Instructional Use Classification Sixty-eight learning object documents in each of the four categories ( websphere, j2ee, web services and security ) were used for this experiment. The linear classifier assigned 83 percent of the documents to one of these four categories correctly. When the percentage of training documents was increased to 80 percent, the classification accuracy was 94 percent, as would be expected. This level of accuracy is similar to that reported in the literature for subject-oriented text classification of news articles and other types of documents. For instructional use classification, the entire collection was utilized. When training with 20 percent of the documents in each category, 52.5 percent of the remaining documents were classified correctly The experiment was repeated using 80 percent of the documents with a resulting classification precision of 57 percent. The resulting improvement in precision is to be expected, given the additional training data. However, the low classification accuracy, even with 80 percent of the data used for training, was not expected. We attributed this result to an insufficient number of training documents in several of the instructional use categories. A second experiment was conducted for just procedures and concepts categories using 60 documents from each group. Training with 20 percent of the learning objects yielded 68 percent classification accuracy and this improved to 74 percent when using 80 percent of the documents for training. The result from this experiment is shown below in figure 3(a). While this is a large improvement, it is still a lower classification precision than desired. 4 Enhancing Instructional Use Classification We next investigated if we could boost the precision of the text classifier for instructional use categories by improving the weights of certain features. In particular, we artificially boosted the frequency of important words in the document text. We computed important words by three different techniques: highest frequency within a document corpus, words appearing in document titles, and the words appearing as synonyms of words in the category name from a predefined list appearing in WordNet, a widely available research thesaurus (Miller, 1985; WordNet). The experimental set up for the enhanced classifier is shown in figure 2. The basic experimental set up is similar to figure 1. We identified two mechanisms for enhancing the instructional use classification of the different categories: 1. To improve the accuracy of classifying learning objects as procedures vs. concepts, we boosted salient words from the text corpus within each of these two categories. 3

4 2. To improve the accuracy of classifying learning objects into the introduction: category, we repeated the salient words from the learning object s title and added words from the WordNet thesaurus. Figure 2: Enhanced Instructional Use Classification 4.1 Investigating procedures vs. concepts Instructional Use Classification For the procedures vs. concepts classification, a training file was first created using hand-coded data. Then, the top salient words were extracted for each category. First, all the words of the documents assigned to the category were extracted after removing stop words and performing stemming. Next, the number of occurrences of each resulting term was computed per category as a percentage of the total number of terms in that category. For the same term, the term frequency was calculated based on the entire corpus of terms in all categories. Thus for each term in a category, a salience value can be computed as follows: S(t) = (TF c (t) / n) (TF d (t)/ N) In this equation, S(t) = salience value of a term t; TF c (t) Term Frequency of a term t in the whole corpus; TF d (t) Term Frequency of the term t in a category ; N Number of terms in the entire corpus; n Number of terms in the category The salience value, S(t), was computed for the top hundred words in each category and these terms were then arranged in a decreasing order of salience. A positive salience value indicates that the particular word has a value greater than the normalized frequency in the corpus. The top 100 words of the category procedure were extracted. Words from

this list appearing in the training data were boosted by repeating words multiple times (up to n = 10 times). Example set of salient words from the procedures category included - Task, execute, manage, develop, enter, click, next, resource, process, select. Example set of salient words from the concepts category included - Application, Service, Theme, Role, Provide, System, Contain, Specifications. 4.2 Investigating using title words for introduction category For the introduction category, salient words were extracted from the text based on the frequency of the words appearing just in titles. The top 10 words were included in the testing file multiple times (up to a maximum of n = 10 times). A sample set of words in this category included words like start, overview, and begin. Along with this, the WordNet thesaurus was used to retrieve words similar to the word introduction.for example, WordNet words for introduction include start, begin, launch, debut. A combination of the title words and WordNet words were included in the training document to boost these words. Each word that appeared in the WordNet list or the title was repeated five times. WordNet gave different senses of each word, and these were each repeated (n = 5 times) in the testing file created. 4.3 Results and Evaluation Based on the enhanced feature set and feature boosting, classification of procedures vs concepts for 66 documents using 20 percent training yielded 74 percent classification accuracy. Increasing the training percentage to 80 percent resulted in 84 percent classification accuracy. By enhancing the feature set by repeating the salient words in a document (up to n = 10 times), the precision (classification accuracy) was also increased. The results are plotted in Figure 3 (b). (a) (b) Instructional Use Classification Enhanced Instructional Use 5 Precision (percentage) 100 90 80 70 60 50 40 30 20 10 0 68 74 Precision (percentage) 100 90 80 70 60 50 40 30 20 10 0 74 84 20 80 20 80 Percentage data used for training Percentage data used for training Figure 3: Instructional Use Classification (66 documents each of Procedures and Concepts) (a) before enhancements (b) after enhancements The enhancement using the title words and WordNet words did not result in any change in the performance efficiency for procedures vs. concepts and also did not improve classification for the introduction category. Upon further investigation, we determined that the WordNet thesaurus terms were too infrequent in our corpus. In addition, orthographic and grammatical variations may have created an artificially low term overlap. We also found that repeating terms in the section or slide titles also did not

6 boost classification accuracy. One explanation is that words like overview and introduction appeared in titles, but they also appeared with some frequency in the rest of the document text. 5 Conclusions This paper described the development of an enhanced linear classifier for performing topic and instructional use classification of learning objects. Though instructional use is important for learners, it is an extremely difficult classification problem. We were able to improve the precision of instructional use classification by weighting salient categoryspecific corpus terms. Our experiments indicate that the repeated use of stylized verbs in procedural texts can be used to accurately differentiate conceptual vs. procedural learning materials. Our technique is an initial step toward fully automatic and accurate tagging of learning materials with standardized learning object metadata. References Docbook, Available at: http://www.oasis-open.org/docbook/, Accessed on March 14,2005. Farrell, R. et.al. Learner-driven Assembly of Web-based Courseware. Proceedings of E- Learn 2003 (Phoenix AZ, Nov 2003). Farrell, R. G., Liburd, S. D., and Thomas J. C. (2004). Dynamic assembly of learning objects. WWW 2004 (Alternate Track Papers & Posters) 2004, 162-169. IEEE LOM 2002, IEEE Standard for Learning Object Metadata, IEEE Standards Department, Institute of Electrical and Electronic Engineers, Inc. (2002). Maudlin, M. (1998). A history of search engines. Available at: http://www.wiley.com/compbooks/sonnenreich/history.html, last accessed on March 14, 2005. Miller, G. A. (1985).WordNet: a dictionary browser, In Proceedings of the First International Conference on Information in Data, University of Waterloo, Waterloo. WordNet, Available at: http://wordnet.princeton.edu/cgi-bin/webwn, last accessed on March 14, 2005 Zhang, T., Oles, F.J., (2001). Text categorization based on regularized linear classification methods, Information Retrieval, 4, 5-31.