A NOTE ON UNDETECTED TYPING ERRORS

Similar documents
Senior Stenographer / Senior Typist Series (including equivalent Secretary titles)

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Disambiguation of Thai Personal Name from Online News Articles

An Interactive Intelligent Language Tutor Over The Internet

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Welcome to the Purdue OWL. Where do I begin? General Strategies. Personalizing Proofreading

A Case Study: News Classification Based on Term Frequency

University of Waterloo School of Accountancy. AFM 102: Introductory Management Accounting. Fall Term 2004: Section 4

Changing User Attitudes to Reduce Spreadsheet Risk

5. UPPER INTERMEDIATE

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

New Jersey Department of Education

Holy Family Catholic Primary School SPELLING POLICY

Evidence for Reliability, Validity and Learning Effectiveness

South Carolina English Language Arts

Calculators in a Middle School Mathematics Classroom: Helpful or Harmful?

Cross Language Information Retrieval

What s in a Step? Toward General, Abstract Representations of Tutoring System Log Data

TCH_LRN 531 Frameworks for Research in Mathematics and Science Education (3 Credits)

Interpreting ACER Test Results

Evaluating the Effectiveness of the Strategy Draw a Diagram as a Cognitive Tool for Problem Solving

Effective Instruction for Struggling Readers

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

K 1 2 K 1 2. Iron Mountain Public Schools Standards (modified METS) Checklist by Grade Level Page 1 of 11

Writing Research Articles

Predicting Students Performance with SimStudent: Learning Cognitive Skills from Observation

Identifying Novice Difficulties in Object Oriented Design

Xinyu Tang. Education. Research Interests. Honors and Awards. Professional Experience

Norms How were TerraNova 3 norms derived? Does the norm sample reflect my diverse school population?

Division Strategies: Partial Quotients. Fold-Up & Practice Resource for. Students, Parents. and Teachers

Providing student writers with pre-text feedback

Reading Horizons. Updating the Dolch Basic Sight Vocabulary. Jerry L. Johns JANUARY Volume 16, Issue Article 7

A 3D SIMULATION GAME TO PRESENT CURTAIN WALL SYSTEMS IN ARCHITECTURAL EDUCATION

An Evaluation of the Interactive-Activation Model Using Masked Partial-Word Priming. Jason R. Perry. University of Western Ontario. Stephen J.

CEFR Overall Illustrative English Proficiency Scales

Software Maintenance

Learning Microsoft Publisher , (Weixel et al)

On the Combined Behavior of Autonomous Resource Management Agents

10: The use of computers in the assessment of student learning

Data Structures and Algorithms

The Impact of Honors Programs on Undergraduate Academic Performance, Retention, and Graduation

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Language Acquisition Chart

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

The Good Judgment Project: A large scale test of different methods of combining expert predictions

Loughton School s curriculum evening. 28 th February 2017

INTERMEDIATE ALGEBRA PRODUCT GUIDE

6 Financial Aid Information

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Achievement Level Descriptors for American Literature and Composition

Longest Common Subsequence: A Method for Automatic Evaluation of Handwritten Essays

Diagnostic Test. Middle School Mathematics

Getting Started with Deliberate Practice

MADERA SCIENCE FAIR 2013 Grades 4 th 6 th Project due date: Tuesday, April 9, 8:15 am Parent Night: Tuesday, April 16, 6:00 8:00 pm

PART 1. A. Safer Keyboarding Introduction. B. Fifteen Principles of Safer Keyboarding Instruction

AQUA: An Ontology-Driven Question Answering System

Controlled vocabulary

Journalism 336/Media Law Texas A&M University-Commerce Spring, 2015/9:30-10:45 a.m., TR Journalism Building, Room 104

Session Six: Software Evaluation Rubric Collaborators: Susan Ferdon and Steve Poast

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

Individual Component Checklist L I S T E N I N G. for use with ONE task ENGLISH VERSION

READ 180 Next Generation Software Manual

Adaptation Criteria for Preparing Learning Material for Adaptive Usage: Structured Content Analysis of Existing Systems. 1

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

OPAC and User Perception in Law University Libraries in the Karnataka: A Study

EECS 571 PRINCIPLES OF REAL-TIME COMPUTING Fall 10. Instructor: Kang G. Shin, 4605 CSE, ;

Arizona s English Language Arts Standards th Grade ARIZONA DEPARTMENT OF EDUCATION HIGH ACADEMIC STANDARDS FOR STUDENTS

Progress Monitoring for Behavior: Data Collection Methods & Procedures

Many instructors use a weighted total to calculate their grades. This lesson explains how to set up a weighted total using categories.

Web as Corpus. Corpus Linguistics. Web as Corpus 1 / 1. Corpus Linguistics. Web as Corpus. web.pl 3 / 1. Sketch Engine. Corpus Linguistics

Linking Task: Identifying authors and book titles in verbose queries

21st Century Community Learning Center

Principal vacancies and appointments

Preprint.

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Northeastern University Online Course Syllabus

Physics 270: Experimental Physics

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

A Game-based Assessment of Children s Choices to Seek Feedback and to Revise

Pragmatic Use Case Writing

ADHD Classroom Accommodations for Specific Behaviour

No Child Left Behind Bill Signing Address. delivered 8 January 2002, Hamilton, Ohio

PowerTeacher Gradebook User Guide PowerSchool Student Information System

Problems of the Arabic OCR: New Attitudes

STARTS THE 2nd Week of APRIL!

Effect of Word Complexity on L2 Vocabulary Learning

Mandarin Lexical Tone Recognition: The Gating Paradigm

1 Use complex features of a word processing application to a given brief. 2 Create a complex document. 3 Collaborate on a complex document.

Improving Conceptual Understanding of Physics with Technology

Outline. Web as Corpus. Using Web Data for Linguistic Purposes. Ines Rehbein. NCLT, Dublin City University. nclt

Alberta Police Cognitive Ability Test (APCAT) General Information

Systematic reviews in theory and practice for library and information studies

Introduction. 1. Evidence-informed teaching Prelude

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Purpose of internal assessment. Guidance and authenticity. Internal assessment. Assessment

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Towards a Collaboration Framework for Selection of ICT Tools

Transcription:

SPkClAl SECT/ON A NOTE ON UNDETECTED TYPING ERRORS Although human proofreading is still necessary, small, topic-specific word lists in spelling programs will minimize the occurrence of undetected typing errors. JAMES L. PETERSON Computer programs to detect and correct spelling and typographic errors are fairly well understood and becoming quite common [8]. The interesting problems now deal mainly with the details of the design: the data structures and user interface, and not with the basic algorithms. Work is already proceeding on the construction of more advanced techniques for detecting syntactic and semantic errors [3, 61. Despite the obvious value of spelling checkers, there can be problems. Perhaps the most important problems are mistakes made by the spelling checker. As with all pattern-recognition algorithms, mistakes can be of two kinds: (1) failure to accept a correctly spelled word, and (2) failure to reject an incorrectly spelled word. How bad are these problems? Let us assume that the spelling checker is based on a lookup algorithm: The spelling checker maintains a word list of correctly spelled words. To check a word, the speller simply searches this list. If the word is found, it is assumed to be correctly spelled, and hence accepted; if it is not found, it is assumed to be incorrectly spelled, and hence rejected. Clearly, with this algorithm, the speller will fail to accept correctly spelled words only if they are not on its list of approved words. To reduce the probability of this mistake, we can simply add more words to the approved word list, increasing its size. Ignoring the problems of storage and search time for large This work was supported. in part. by the Department of Computer Sciences word lists, this approach may increase the probabilof the University of Texas at Austin, and the Computer Science Department of Carnegie-Mellon University. ity of the second kind of mistake: failure to reject an incorrectly spelled word. 1986 ACM OOOl-782/86/7-633 759 How can an incorrectly spelled word be mistak- luly 1986 Volume 29 Number 7 Communications of the ACM 633

enly recognized as correct by a spelling checker? Assume that the author/user meant to use the word x, but mistyped the word as y. (For example, house may be mistyped as horse.) This error will be undetected only if y is a correctly spelled word. It is admittedly not the desired word, but it is still a word on the approved word list. This kind of error can be detected only by more complex algorithms using syntactic or semantic information. The probability of this kind of error-misspelling or mistyping a word as another word-should increase as the size of the word list increases. As the size of the word list increases, more obscure and unusual words will be added to the list. These words will no longer be detected as errors, but will be thought correct by the spelling checker. This article reports on research to determine the probability that a word can be mistyped as another word, as a function of the size of the word list. TYPES OF ERRORS First, we must redefine the problem. In our previous research into spelling checkers [9], we were,unable to come up with any good information about how people misspell, so we have concentrated instead on how words are mistyped. The classical definition of typing errors by Damerau [4] indicates that 8 percent of typing errors are caused by 1. transposition of two adjacent letters, 2. one extra letter, 3. one missing letter, or 4. one wrong letter. Our own studies have shown similar results. We found 36 errors in a computer copy of Webster s Seventh New Collegiate Dictionary [ll] that had escaped detection since the file was keyboarded 15 years ago (including 1 errors in the original printed version). We also found 155 errors made by college students when they retyped the word division list of the U.S. Government Printing Office [5]. These errors were distributed as follows: GPO Web 7 Transposition 4 (2.6%) 47 (13.1%) One extra letter 29 (18.7%) 73 (2.3%) One missing letter 49 (31.6%) 124 (34.4%) One wrong letter 62 (4.%) 97 (26.9%) Total 144 (92.9%) 341 (94.7%) The next most common errors were two extra letters, two missing letters, or two letters transposed around a third (such as prodecure instead of procedure). Although these results are similar, they are not sufficiently accurate or stable to provide good estimates of the probabilities of different kinds of errors or of key-level errors. There is almost certainly an error distribution for each key, due in large part to keyboard layout; a t is more likely to be mistyped as an r than as a 4, while a 4 is more likely to be mistyped as an a than an r. We have assumed simply that these four types of errors are all equally probable. WORD LISTS Over the last few years, we have collected some 16 word lists from various sources, including l Webster s Seventh New Collegiate Dictiona y, derived from the work of Olney at SDC and Alberga at IM Yorktown [lo] (81,627 words): l Webster s Second international Dictiona y, obtained over the ARPAnet, apparently keyboarded by students (234,933 words); l Longman s Dictiona y of Contempora y English, a ritish dictionary; the computer files were derived from the printer s phototypesetting tapes (2,56 words); l Random House Dictionary, used in the Random House Proofreader, a spelling program from Aspen Software Company (5,3 words); l the Proof spelling checker developed by JM (7,399 words); l the spell program for Unix@ from ell Labs (24,473 words); l the word division book of the U.S. Government Printing Office (18,47 words). All 16 word lists were merged to produce one master list of 369,546 words in a file of 3,799,425 bytes (an average of 9.3 characters plus a separator byte per word). THE EXPERIMENT A program was written to read the master word list and generate each of the possible mistyped versions of the word, according to the four main typing errors. Only one occurrence of one error per mistyping was considered, and each mistyped word was then checked to see if it was some other word. It took five days on a dedicated VAX 11/78-some 158,583. seconds of CPU time (about 44 hours)-to execute this program. The result was a file of 15,69,965 bytes with 988,192 entries. Each entry consisted of two words and a code describing how one could be mistyped to Unix is a trademark of AT&T ell Laboratories 634 Commutlications of the ACM luly 1986 Volume 29 Number 7

be the other. Of the 369,546 unique words in our master list, 153,664 of them cannot be mistyped as some other word; 215,882 can be mistyped as some other word. In the extreme, for example, the word sat can be mistyped as 54 other words. Of the 988,192 possible words that can be mistyped as another word, 616,21 were due to one wrong letter, 18,559 were due to one extra letter, 18,559 were due to one missing letter, and 1,864 were due to transposing two letters. The numbers for one extra and one missing letter are the same because, if word x can be mistyped as word y by adding one extra letter, then y can be mistyped as x by deleting that extra letter. Similarly, the number of errors for wrong letters and transpositions should be even. How many typing errors are possible? Assuming an alphabet of 28 characters (the 26 letters plus the hyphen and the apostrophe) and a word with n letters, the four typing errors could cause the following numbers of errors: One wrong letter 27n One extra letter 28(n + 1) One missing letter n Transposition n-l Thus, 57n + 27 typing errors are possible for a word of length n. This allows, for our master list of 369,546 words, 25,48,845 possible mistyped words. Since only 988,192 of these are words, the probability of mistyping a word as another word is 988,192 + 25,48,845 or about half a percent, for our full list. VARYING THE SIZE OF THE WORD LIST Of course, no one uses our entire word list in a spelling checker; we need to repeat this study for smaller word lists. There are, however, 236g,546 sublists for a word list of 369,546 words. It would be impossible (and pretty useless) to do this for every such sublist. We assumed instead that a word list of length m would be the m most-common words. Then any word on a list of size m would also be on each list of size m > m. This allows us to determine the words in a word list as its size varies, if we know the frequency of the words. There have been at least three major word frequency studies: the rown Corpus [7] (5,46 unique words from a sample of 1,14,232), the American Heritage Intermediate Corpus [2] (86,741 unique words from a sample of 5,88,721), and a study of word frequency in the New York Times wire service stories [l] (7,144 unique words from a sample of 833,155). Although these studies provided some frequency information, they were not broad enough to provide complete information; only 22,439 words were on all three studies. To produce complete frequency information, we averaged the frequencies of the three major studies. Words that were in none of the studies were assigned a frequency corresponding to the number of times they occurred in our set of word lists. This produced a frequency for each word ranging from 1 to 323,32,72. Summing these frequencies and dividing by the sum give probabilities (ranging from.52 to.16). Sorting by frequency allows us to assign ranks to the words. The most frequent word is the, with of, and, to, a, in, is, that, it, and he completing the top 1. Words that have the same frequency (to the accuracy available from the existing studies) are all assigned the same rank. Thus, the 27,943 words that occurred in only one word list were all assigned the same rank: 369,546. Now assume that a word x of rank p can be mistyped as a word y of rank 4. If p < q (i.e., x is more frequent than y), then the mistake will be detected for all lists with only words of rank less than 4. If the word list expands to 4 or more words, the mistake will go undetected. In general, the mistake will be undetected if the length of the word list is greater than or equal to max(p, 4). For example, the word house is rank 139; horse is rank 577. Thus, if house is mistyped as horse, it will be undetected if the length of the word list is greater than or equal to 577. If horse is mistyped as house, it will also be an undetected error for a word list of more than 577 words. (In fact, this error would be undetected for word lists from 1% to 577 words also, but in this case horse would not be considered a valid word as it is not on the word list. We have not considered the problem of mistyping a nonword as a word.) RESULTS Figure 1 (p. 636) shows the number of undetected errors as the size of the word list varies, from 1 to 369,546 words. The number of undetected errors appears to increase linearly with the size of the word list. Clearly, however, the number of errors that are undetected will almost always increase as the size of the word list increases; undetected errors remain undetected as more words are added to the word list. We should look, however, not only at the total number of undetected errors, but at the fraction of all errors that are undetected as well. As the word list grows, the probability that a given mistyping luly 1986 Volume 29 Number 7 Communications of the ACM 635

E 7, 2 ; 6, E x g 5, 5 $ 4, 5 3, 3oo~oool I n I I I I 2, -_ RI 1, / 1, 2, 3, / 4( 3,O 21,611 words of 1 characters (out of a potential 296,196,766,695,424) are valid. Figure 3 reinforces the conclusion that most undetected typing errors come from the short frequent words that are in every word list by measuring the probability of an undetected typing error in running text. Since small words are both more frequent and more likely to produce undetected typing errors, the probability of undetected errors increases rapidly for the first 1, words and then grows much more slowly. In actual usage, the probability of an undetected typing error varies directly with the size of the word list, ranging from 2 percent to almost 16 percent of all typing errors. It is important to note, however, that these numbers are only rough indicators of the actual problem. Although we used the best information available to us, the following questions are not well answered: l What is the frequency of words in English? Our averaging of the three major word frequency studies indicates that some words may have a frequency as low as one occurrence out of 6,,, words. Accordingly, an accurate FIGURE 1. The Number of Undetected Typing Errors according to Word List Size.16.14 would go undetected increases, but of course, with more words on the word list, there are more mistypings possible. Figure 2 shows the fraction of undetected typing errors as the size of the word list varies. For a word list of size m, this is the number of undetected typing errors for m (from Figure 1) divided by the sum, over all m words in the list, of the number of possible typing errors (57~ + 27 for each word of length n). Although the fraction of undetected errors does increase as the word list grows to 5, or 6, words, it then levels off and is fairly stable, over the long run, at about.5 percent. It would appear, then, that only 1 word out of every 2 mistyped words would accidentally become another word, escaping detection by a spelling checker. This result, however, ignores two important observations. First, frequent words tend to be shorter than less frequent words. And, second, short words, if mistyped, are more likely to be undetected than longer words, since proportionally more short words are on a word list. Of the 784 (28 X 28) possible twoletter words, 431 (55 percent) are valid (in the complete 369,546 word list). On the other hand, only I?.12 (? $ F ;.1 j g.8 c 2.k.6!i t.4.2 FIGURE 2. 1, 2, 3, 4, The Fraction of Possible Undetected Typing Errors 636 Communications of the ACM luly 1986 Volume 29 Number 7

Special.1.9.8 52 P &I.7 F ii *.6 73 d 8 g.5 c 3 s.4 C t.3 -I programs should be kept small; a large word list is not necessarily a better word list. In particular, word lists used for spelling should be tailored for the particular author and topic for which it is to be used. Word lists used for checking computer science papers should generally not include medical, legal, and geographic words, for example. We also see a need for more intelligent checking programs, such as syntax and semantics checkers. A simple spelling checker cannot detect a word that is mistyped as another word. More complex approaches (using part of speech information, for example) may be able to detect such errors. Finally, although programs can catch many spelling and typing problems, human proofreading is still necessary to detect and correct many errors..2.1 1, 2, 3, 4, FIGURE 3. The Fraction of Possible Undetected Typing Errors Weighted by Frequency study may require billions of words to be analyzed. l What is the frequency of the four main typing errors? l What is the distribution of errors caused by keyboard layout? l How are words misspelled (rather than simply mistyped)? In addition, we would expect all of these numbers to vary with the author/typist. Despite these unanswered questions and their effect on the actual probability of an undetected spelling error, it would seem reasonable to conclude that a significant number of spelling and typing errors may be undetectable by a spelling program, particularly with large word lists. CONCLUSIONS There is a real danger that longer word lists will result in a substantial number of undetected typing errors. Almost one typing error out of six may be undetected. Therefore, word lists used in spelling REFERENCES 1. Amsler. R. Private communication. Stanford Research Institute. Calif.. 1982. 2. Carroll, J... Davies, P., and Richman.. Word Frequerry ook. Houghton Mifflin, oston, Mass., 1971. 3. Cherry, L.L. Writing tools-the STYLE and DICTION programs. Tech. Rep. 9. Computer Science, ell Laboratories, Murray Hill, N.J., 198. 4. Damerau. F.J. A technique for computer detection and correction of spelling errors. Commurt. ACM 7. 3 (Mar. 1964), 171-176. 5. Government Printing Office. Word Divisim, A Suppleme~tf to Government Pritlfirzg Office Style Manual. 7th ed. Government Printing Office, Washington, D.C.. 1968. 6. Heidorn. GE.. Jensen, K., Miller, L.A.. yrd. R.J.. and Chodorow. MS The EPISTLE text-critiquing system. IM Syst. \. 21, 3 (1982). 35-326. 7. Kucera. H.. and Francis, W.N. Computational Analysis of Present-Day Americarf English. rown University Press, Providence. R.I.. 1967. 8. Peterson, J.L. Computer programs for detecting and correcting spelling errors. Commun. ACM 23, 12 (Dec. 198). 676-687. 9. Peterson, J.L. Computer Program for Spelling Correction. Vol. 96, LPCfure Notes in Computer Science. Springer-Verlag. New York, 198. 1. Peterson. J.L. Webster s Seventh New Collegiate Dictionary: A computer-readable file format. Tech. Rep. TR-196, Dept. of Computer Sciences, Univ. of Texas. Austin. May 1982. 11. Peterson, J.L. Use of Webster s Seventh New Collegiate Dictionary to construct a master hyphenation list. In Proceedings 1 the National Conrputer Co$erence (Houston. Tex., June 7-1). AFIPS Press, Arlington, Va.. 1982. pp. 667-67. CR Categories and Subject Descriptors: H.1.2 [Models and Principles]: User/Machine Systems-human factors; H.4.1 [Information Sys- tems Applications]: Office Automation--word processing: 1.7.1 [Text Processing]: Text Editing-spelling General Terms: Documentation. Human Factors, Measurement Additional Key Words and Phrases: typing errors, word lists Author s Present Address: James L. Peterson, Microelectronics and Computer Technology Corporation (MCC). 943 Research oulevard. Austin. TX 78759. Permission to copy without fee all or part of this material is granted provided that the copies are not made or distributed for direct commercial advantage, the ACM copyright notice and the title of the publication and its date appear. and notice is given that copying is by permission of the Association for Computing Machinery. To copy otherwise. or to republish. requires a fee and/or specific permission. ]uly 1986 Volume 29 Number 7 Communications of the ACM 637