A Domain Ontology Development Environment Using a MRD and Text Corpus

Similar documents
AQUA: An Ontology-Driven Question Answering System

Linking Task: Identifying authors and book titles in verbose queries

Prediction of Maximal Projection for Semantic Role Labeling

METHODS FOR EXTRACTING AND CLASSIFYING PAIRS OF COGNATES AND FALSE FRIENDS

The MEANING Multilingual Central Repository

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

Using dialogue context to improve parsing performance in dialogue systems

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

A DISTRIBUTIONAL STRUCTURED SEMANTIC SPACE FOR QUERYING RDF GRAPH DATA

Cross Language Information Retrieval

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

Probabilistic Latent Semantic Analysis

A Comparison of Two Text Representations for Sentiment Analysis

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

Compositional Semantics

A Bayesian Learning Approach to Concept-Based Document Classification

Vocabulary Usage and Intelligibility in Learner Language

Computerized Adaptive Psychological Testing A Personalisation Perspective

Multilingual Sentiment and Subjectivity Analysis

Concept Acquisition Without Representation William Dylan Sabo

Learning Methods in Multilingual Speech Recognition

A Case Study: News Classification Based on Term Frequency

SEMAFOR: Frame Argument Resolution with Log-Linear Models

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Detecting English-French Cognates Using Orthographic Edit Distance

The Internet as a Normative Corpus: Grammar Checking with a Search Engine

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Evaluation of Usage Patterns for Web-based Educational Systems using Web Mining

Constraining X-Bar: Theta Theory

Some Principles of Automated Natural Language Information Extraction

Combining a Chinese Thesaurus with a Chinese Dictionary

Modeling Attachment Decisions with a Probabilistic Parser: The Case of Head Final Structures

Semantic Inference at the Lexical-Syntactic Level for Textual Entailment Recognition

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

A Semantic Similarity Measure Based on Lexico-Syntactic Patterns

Leveraging Sentiment to Compute Word Similarity

An Introduction to the Minimalist Program

arxiv: v1 [cs.cl] 2 Apr 2017

Matching Similarity for Keyword-Based Clustering

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Abstractions and the Brain

Organizational Knowledge Distribution: An Experimental Evaluation

2.1 The Theory of Semantic Fields

Radius STEM Readiness TM

Data Integration through Clustering and Finding Statistical Relations - Validation of Approach

Parsing of part-of-speech tagged Assamese Texts

The Role of String Similarity Metrics in Ontology Alignment

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

A Graph Based Authorship Identification Approach

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Ontologies vs. classification systems

A heuristic framework for pivot-based bilingual dictionary induction

Informatics 2A: Language Complexity and the. Inf2A: Chomsky Hierarchy

The Smart/Empire TIPSTER IR System

Constructing Parallel Corpus from Movie Subtitles

THE VERB ARGUMENT BROWSER

Word Sense Disambiguation

Exploiting Phrasal Lexica and Additional Morpho-syntactic Language Resources for Statistical Machine Translation with Scarce Training Data

EdIt: A Broad-Coverage Grammar Checker Using Pattern Grammar

Learning Methods for Fuzzy Systems

Reinforcement Learning by Comparing Immediate Reward

ARKANSAS TECH UNIVERSITY

THE ROLE OF DECISION TREES IN NATURAL LANGUAGE PROCESSING

A Version Space Approach to Learning Context-free Grammars

Axiom 2013 Team Description Paper

Knowledge Elicitation Tool Classification. Janet E. Burge. Artificial Intelligence Research Group. Worcester Polytechnic Institute

Data-driven Type Checking in Open Domain Question Answering

2/15/13. POS Tagging Problem. Part-of-Speech Tagging. Example English Part-of-Speech Tagsets. More Details of the Problem. Typical Problem Cases

Automating the E-learning Personalization

Product Feature-based Ratings foropinionsummarization of E-Commerce Feedback Comments

The stages of event extraction

Mining Association Rules in Student s Assessment Data

Rule Learning With Negation: Issues Regarding Effectiveness

Assignment 1: Predicting Amazon Review Ratings

Teachers: Use this checklist periodically to keep track of the progress indicators that your learners have displayed.

Can Human Verb Associations help identify Salient Features for Semantic Verb Classification?

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Loughton School s curriculum evening. 28 th February 2017

Friday, October 3, 2014 by 10: a.m. EST

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Using Semantic Relations to Refine Coreference Decisions

Conceptual Framework: Presentation

Trend Survey on Japanese Natural Language Processing Studies over the Last Decade

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

*Net Perceptions, Inc West 78th Street Suite 300 Minneapolis, MN

Assessing System Agreement and Instance Difficulty in the Lexical Sample Tasks of SENSEVAL-2

Part III: Semantics. Notes on Natural Language Processing. Chia-Ping Chen

Guidelines for Writing an Internship Report

On document relevance and lexical cohesion between query terms

BYLINE [Heng Ji, Computer Science Department, New York University,

Learning Optimal Dialogue Strategies: A Case Study of a Spoken Dialogue Agent for

Rule discovery in Web-based educational systems using Grammar-Based Genetic Programming

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Transcription:

A Domain Ontology Development Environment Using a MRD and Text Corpus Naomi Nakaya 1 and Masaki Kurematsu 2 and Takahira Yamaguchi 1 1 Faculty of Information, Shizuoka University 3-5-1 Johoku Hamamatsu Shizuoka 432-8011 Japan {nakaya, yamaguti}@ks.cs.inf.shizuoka.ac.jp 2 Faculty of Software and Information Science, Iwate Prefectural University 152-52 Takizawasugo Takizawa Iwate 020-0193 Japan kure@soft.iwate-pu.ac.jp Abstract. In this paper, we describe how to exploit a machine-readable dictionary (MRD) and domain-specific text corpus in supporting the construction of domain ontologies that specify taxonomic and non-taxonomic relationships among given domain concepts. A) In building taxonomic relationships (hierarchically structure) of domain concepts, some hierarchically structure can be extracted from a MRD with marked sub-trees that may be modified by a domain expert, using matching result analysis and trimmed result analysis. B) In building non-taxonomic relationships (specification templates) of domain concepts, we construct concept specification templates that come from pairs of concepts extracted from text corpus, using WordSpace and an association rule algorithm. A domain expert modifies taxonomic and non-taxonomic relationships later. Through the case study with CISG, we make sure that our system can work to support the process of constructing domain ontologies with a MRD and text corpus. 1 Introduction Although ontologies have been very popular in many application areas, we still face the problem of high cost associated with building up them manually. In particular, since domain ontologies have the meaning specific to application domains, human experts have to make huge efforts for constructing them entirely by hand. In order to reduce the costs, automatic or semi-automatic methods have been proposed using knowledge engineering techniques and natural language processing ones (cf. Ontosaurus [1]). The authors have also developed a domain ontology rapid development environment called DODDLE [2], using a machine-readable dictionary. However, these environments facilitate the construction of only a hierarchically structured set of domain concepts, in other words, taxonomic conceptual relationships. As domain ontologies have been applied to widespread areas, such as knowledge sharing, knowledge reuse and software agents, we need software environments that support a human expert in constructing domain ontologies with not only taxonomic conceptual relationships

but also non-taxonomic ones. In order to develop environments, it seems better to put together two or more techniques such as knowledge engineering, natural language processing, machine learning and data engineering(e.g. [3]). Here in this paper, we extend DODDLE into DODDLE II that constructs both taxonomic and non-taxonomic conceptual relationships, exploiting WordNet [4] and domain-specific text corpus (text corpus) with the automatic analysis of lexical co-occurrence statistics and an association rule algorithm[5]. Furthermore, we evaluate how DODDLE II works in the field of law, the Contracts for the International Sale of Goods(CISG). 2 DODDLE II: A Domain Ontology Rapid Development Environment Figure 1 shows an overview of DODDLE II, a Domain Ontology rapid Development Environment that has the following two components: Taxonomic relationship acquisition module using WordNet and Non-taxonomic relationship learning module using text corpus. A domain expert who is a user of DODDLE II gives a set of domain terms to the system. A) The taxonomic relationship acquisition module (TRA module) does spell match between input domain terms and WordNet. The spell match links these terms to WordNet. Thus the initial model from the spell match results is a hierarchically structured set of all the nodes on the path from these terms to the root of WordNet. However the initial model has unnecessary internal terms (nodes). They do not contribute to keeping topological relationships among matched nodes, such as parent-child relationship and sibling relationship. So we get a trimmed model by trimming the unnecessary internal nodes from the initial model. In order to refine the trimmed model, we have the following two strategies that we will describe later: Matched result analysis and Trimmed result analysis. B) The non-taxonomic relationship learning module (NTRL module) extracts the pairs of terms that should be related by some relationship from text corpus, analyzing lexical cooccurrence statistics, based on WordSpace (WS) [6] and an associate rule algorithm(ar). Thus the pairs of terms extracted from text corpus are the candidates for non-taxonomic relationships. NTRL module also extracts candidates for taxonomic relationships from these pairs, analyzing the distance between terms in a document. We can built concept specification templates by putting together taxonomic and nontaxonomic relationships for the input domain terms. The relationships should be identified in the interaction with a human expert. 3 Taxonomic Relationship Acquisition After getting the trimmed model, TRA module refines it by interaction with a domain expert, using Matched result analysis and Trimmed result analysis. First, TRA module divides the trimmed model into a PAB (a PAth including only Best spell-matched nodes) and a STM (a Sub-Tree that includes best spell-matched nodes and other nodes and so can be Moved) based on the distribution of best-matched nodes. A PAB is a path that includes only best-matched nodes that have the senses good for given domain specificity. Because all nodes have already been adjusted to the domain in PABs, PABs can stay in the trimmed model. A STM is such a sub-tree that an internal node is a root and the subordinates are only best-matched nodes. Because internal nodes have not been confirmed to have the senses good for a given domain, a STM can be moved in the trimmed model.

MRD (WordNet) A) Taxnomic Relationship Acquisition Module a Set of Domain Terms User (A Domain Expert) Domain Specific Text Corpus B) Non-Taxnomic Relationship Learning Module matched result analysis trimmed result analysis extraction of 4-grams construction of WordSpace finding associations modification using the syntactic strategies extraction of similar concept pairs extraction of concept pairs modification using a candidate for domain-specific hierarchy structure extraction of candidates for taxonomic relationships Taxonomic Relationship Non-Taxonomic Relationship Concept Specification Template Extension & Modification A Domain Ontology Figure 1: DODDLE II overview In order to refine the trimmed model, DODDLE II can use trimmed result analysis. Taking some sibling nodes with the same parent node, there may be many differences about the number of trimmed nodes between them and the parent node. When such a big difference comes up on a sub-tree in the trimmed model, it is be better to change the structure of it. DODDLE II asks a human expert if the sub-tree should be reconstructed or not. Based on the empirical analysis, the sub-trees with two or more differences may be reconstructed. Finally DODDLE II completes taxonomic relationships of the input domain terms manually from the user. 4 Non-Taxonomic Relationship Learning NTRL module almost comes from WS. WS derives lexical co-occurrence information from a large text corpus and is a multi-dimension vector space (a set of vectors). The inner product between two word vectors works as the measure of their semantic relatedness. When two words inner product is beyond some upper bound, there are possibilities to have some nontaxonomic relationships between them. NTRL module also use an AR algorithm to find associations between terms in text corpus. When an AR between terms exceeds user-defined thresholds, there are possibilities to have some non-taxonomic relationships between them. 4.1 Construction of WordSpace WS is constructed as shown in Figure 2. 1. extraction of high-frequency 4-grams Since letter-by-letter co-occurrence information becomes too much and so often irrelevant, we take term-by-term co-occurrence information in four words (4-gram) as the primitive to make up co-occurrence matrix useful to represent

7H[WV 1JUDP DUUD\ «J J J J J 7H[WV 1JUDP DUUD\ «J J J J J Z Z J J J J J J J J J J J «ºÆÅ˼ÏËË ÊºÆǼ J J J J Z J J J J J J Z Z Z Z «:RUG1HW V\QVHW Figure 2: Construction Flow of WS context of a text based on experimented results. We take high frequency 4-grams in order to make up WS. 2. construction of context vectors A context vector represents context of a word or phrase in a text. Element a i,j in a context vector w i is the number of 4-gram g j which comes up a around appearance place of a word or phrase w i (called context scope). The concept vector counts how many other 4-grams come up around a word or phrase. 3. construction of word vectors A word vector W i is a sum of context vectors w i at all appearance places of a word or phrase w i within texts. A set of word vectors is WS. 4. construction of vector representations of all concepts The best matched synset of each input terms in WordNet is already specified, and a sum of the word vector contained in these synsets is concept vector set to the vector representation of a concept corresponding to a input term. The concept label is the input term. A concept vector C can be expressed with the following formula. Here, A(w) is all appearance places of a word or phrase w in a text, and w (i) is a context vector at a appearance place i of a word or phrase w. C = ( w synset(c ) i A(w) w (i)) 5. construction of a set of similar concept pairs Vector representations of all concepts are obtained by constructing WS. Similarity between concepts is obtained from inner products in all the combination of these vectors. Then we define certain threshold for this similarity. A concept pair with similarity beyond the threshold is extracted as a similar concept pair. 4.2 Finding Association Rules between Input Terms The basic AR algorithm is provided with a set of transactions, T := {t i i =1..n},where each transaction t i consists of a set of items, t i = {a i,j j =1..m i,a i,j C} and each item a i,j is form a set of concepts C. The algorithm finds ARs X k Y k :(X k,y k C, X k Y k = {}) such that measures for support and confidence exceed user-defined thresholds. Thereby, support of a rule X k Y k is the percentage of transactions that contain X k Y k as a subset and confidence for the rule is defined as the percentage of transactions that Y k is seen when X k appears in a transaction. support(x k Y k )= {t i X k Y k t i } n

confidence(x k Y k )= {t i X k Y k t i } {t i X k t i } As we regard input terms as items and sentences in text corpus as transactions, DODDLE II finds associations between terms in text corpus. Based on experimented results, we define the threshold of support as 0.4% and the threshold of confidence as 80%. When an association rules between terms exceeds thresholds, the pair of terms are extracted as candidates for nontaxonomic relationships. 4.3 Constructing and Modifying Concept Specification Templates A set of similar concept pairs from WS and term pairs from the AR algorithm becomes concept specification templates. Both of the concept pairs, whose meaning is similar (with taxonomic relation), and has something relevant to each other (with non-taxonomic relation), are extracted as concept pairs with above-mentioned methods. However, by using taxonomic information from TRA module with co-occurrence information, DODDLE II distinguishes the concept pairs which are hierarchically close to each other from the other pairs as TAX- ONOMY. A user constructs a domain ontology by considering the relation with each concept pair in the concept specification templates, and deleting unnecessary concept pairs. 4.4 Extracting Taxonomic Relationships from Text Corpus NTRL module tries to extract pairs of terms which form part of a candidate for domainspecific hierarchy structure. Because we suppose that there are taxonomic relationships in text corpus. In order to do that, we pay attention to the distance between two terms in a document. In this paper, the distance between two terms means the number of words between them. If the distance between two terms is small and the similarity between them is close, we suppose that one term explains the other. If the distance is large and the similarity is close, we suppose that they have taxonomic relationships. According to above-mentioned idea, we calculate the proximally rate between two terms within a certain scope. It is the number of times both terms occur within the scope divided by the number of times only one term occurs within it. We define certain threshold for this proximally rate. Pairs of terms whose proximally rate is within this threshold and the similarity between them is beyond the threshold are extracted as part of a candidate for non-taxonomic relationships. DODDLE II asks the domain expert if the hierarchy structure from TRA module should be changed into unified one or not. 5 Case Studies for Taxonomic Relationship Acquisition In order to evaluate how DODDLE is doing in practical fields, case studies have been done in a particular field of law called Contracts for the International Sale of Goods (CISG)[7]. Two lawyers joined the case studies. They were users of DODDLE II in case studies. In the first case study, input terms are 46 legal terms from CISG Part-II. the second case study, they are 103 terms including general terms in an example case and legal terms from CISG articles related with the case. One lawyer did the first case study and the other lawyer did the second.

Table 1: The Case Studies Results The first The second The number of X case study case study Input terms 46 103 Small DT(Component terms) 2(6) 6(25) Nodes matched with WordNet(Unmatched) 42(0) 71(4) Salient Internal Nodes(Trimmed nodes) 13(58) 27(83) Small DT integrated into a trimmed model(unintegrated) 2(0) 5(1) Modification by the user(addition) 17(5) 44(7) Evaluation of strategy1 4/16(25.0%) 9/29(31.0%) Evaluation of strategy2 3/10(30.0%) 4/12(33.3%) Nodes matched with WordNet is the number of input terms which have be selected proper senses in WordNet and Unmatched is not the case. The number of suggestions accepted by a user/the number of suggestions generated by DODDLE Table 2: significant 46 concepts in CISG part II acceptance delivery offer reply act discrepancy offeree residence addition dispatch offerer revocation address effect party silence assent envelope payment speech act circumstance goods person telephone communication system holiday place of business telex conduct indication price time contract intention proposal transmission counteroffer invitation quality withdrawal day letter quantity delay modification rejection Table 1 shows the result of the case studies. Generally speaking, in constructing legal ontologies, 70 % or more support comes from DODDLE. About half portion of the final legal ontology results in the information extracted form WordNet. Because the two strategies just imply the part where concept drift may come up, the part generated by them has low component rates and about 30 % hit rates. So one out of three indications based on the two strategies work well in order to manage concept drift. Because the two strategies use such syntactical feature as matched and trimmed results, the hit rates are not so bad. In order to manage concept drift smartly, we may need to use more semantic information that is not easy to come up in advance in the strategies. 6 Case Studies for Non-Taxonomic Relationship Learning DODDLE II is being implemented on Perl/Tk now. Figure 3 shows the ontology editor. Subsequently, as a case study for non-taxonomic relationship acquisition, we constructed the concept definition for significant 46 concepts of having used on the first case study (Table 2) with editing the concept specification template using DODDLE II, and verified usefulness. The concept hierarchy, which the lawyer actually constructed using DODDLE in the first case study was used here (Figure 4). 6.1 Construction of WordSpace High-frequency 4-grams were extracted from CISG (about 10,000 words) and 543 kinds of 4- grams were obtained. The extraction frequency of 4-grams must be adjusted according to the

Figure 3: The Ontology Editor scale of text corpus. As CISG is the comparatively small-scale text, the extraction frequency was set as 7 times this case. In order to construct a context vector, a sum of 4-gram around appearance place circumference of each of 46 concepts was calculated. One article of CISG consists of about 140 4-grams. In order to construct a context scope from some 4-grams, it consists of putting together 60 4-grams before the 4-gram and 10 4-grams after the 4-gram. For each of 46 concepts, the sum of context vectors in all the appearance places of the concept in CISG was calculated, and the vector representations of the concepts were obtained. The set of these vectors is used as WS to extract concept pairs with context similarity. Having calculated the similarity from the inner product for the 1035 concept pairs which is all the combination of 46 concepts, and having used threshold as 0.87, 77 concept pairs were extracted. 6.2 Finding Associations between Input Terms In this case, DODDLE II extracted 55 pairs of terms from text corpus using the abovementioned AR algorithm. There are 15 pairs out of them in a set of similar concept pairs extracted using WS.

quantity time delay day holiday attribute silence quality price discrepancy abstraction communication content indication of intention statement offer withdrawal reply counteroffer CONCEPT act rejection letter relation regal relation contract intention conduct addition change modification revocation departure dispatch deed transmission payment delivery assent acceptance speech act invitation proposal person party offeree offeror entity location state inanimate object address place of business circumstance effect instrumentality goods residence communication system Figure 4: domain concept hierarchy of CISG part II telex envelope telephone goods non-taxonomy : quality non-taxonomy : payment non-taxonomy : quantity Figure 5: The concept specification templates for goods goods ATTRIBUTE : quality ATTRIBUTE : quantity MATERIAL : offer MATERIAL : contract Figure 6: The concept definition for goods with editing the templates 6.3 Constructing and Modifying Concept Specification Templates Concept specification templates were constructed from two sets of concept pairs extracted by WS and AR. Table 3 is the list of the extracted similar concepts corresponding to each concept. In Table 3, a concept in bold letters is either an ancestor, descendant or a sibling to the left concept in the concept hierarchy constructed using DODDLE in the first case study. In concept specification templates, such a concept is distinguished as TAXONOMY relation. As taxonomic and non-taxonomic relationships may be mixed in the list based on only context similarity, the concept pairs which may be concerned with non-taxonomic relationships are obtained by removing the concept pairs with taxonomic relationships. Figure 5 shows concept specification templates extracted about the concept goods. The final concept definition is constructed from consideration of concept pairs in the templates. Figure 6 shows the definition of the concept goods constructed from the templates. 6.4 Extracting Taxonomic Relationships from Text Corpus In this case, we defined the threshold for the proximally rate as 0.78 and the certain scope as the same sentence. DODDLE II extracted 128 pairs of concepts regarded as having taxonomic relationships from text corpus. 8 pairs out of them have occurred in the concept hierarchy constructed by the user and have not occurred in the trimmed model. That is, they are the same as modifications by the user. It shows that DODDLE II can extract taxonomic relationships,

Table 3: the concept pairs extracted according to context similarity (threshold 0.87) CONCEPT acceptance act assent communication conduct contract delay delivery dispatch effect goods indication intention offer offeree offeror party payment person place price proposal quality quantity telex time withdrawal CONCEPT LIST IN SIMILAR CONTEXT communication, offer, indication, telex offeror, assent, effect, payment, person, quantity, time, goods, delivery, dispatch, price, contract, delay, withdrawal, offeree, place, quality offeror, act, effect, offer, person, offeree, withdrawal, time, proposal acceptance, offer, telex, conduct, indication party, telex, communication effect, act, person, delivery, payment, quantity delivery, offer, act, payment payment, quantity, goods, place, act, delay, time, contract, person, effect, quality goods, price, act, person, quantity, offeror person, assent, act, offeror, contract, proposal, payment, time, withdrawal, party, delivery dispatch, quantity, delivery, payment, act, person, price, quality intention, acceptance, communication indication acceptance, assent, communication, delay withdrawal, offeror, assent, act, price act, assent, withdrawal, offeree, person, effect, time, price, dispatch conduct, effect, place, person quantity, delivery, place, act, goods, quality, delay, effect, person, contract, time effect, offeror, act, proposal, goods, assent, withdrawal, contract, dispatch, payment, delivery, party, place, price payment, delivery, time, quantity, party, act, person dispatch, act, offeror, goods, withdrawal, offeree, person person, effect, withdrawal, assent quantity, payment, goods, act, delivery payment, delivery, goods, act, quality, dispatch, place, contract, time conduct, communication, acceptance act, offeror, delivery, place, effect, payment, quantity, assent offeree, offeror, person, price, act, assent, effect, proposal which are not included in a MRD, from text corpus. But the rate of accepted taxonomic relationships is about 6%(8/128) and is not good. So, we have to improve do that. 6.5 Results and Evaluation The user evaluated the following two sets of concept pairs: ones extracted by WS and ones extracted by AR. Figure 7 shows three different sets of concept pairs from user, WS and AR. Table 4 shows the details of evaluation by the user, computing precision and recall with the numbers of concept pairs extracted by WS and AR, accepted by the user and rejected by the user. Looking at the field of Precision in Table 4, there in almost no differences among three kinds of results from WS, AR and the join of WS and AR. However, looking at the field of Recall in Table 4, the recall from the join WS and AR is higher than that from each WS and AR, and then goes over 0.5. Generating non-taxonomic relationships of concepts is harder than modifying and deleting them. Therefore, taking the join of WS and AR with high recall, it supports the user in constructing non-taxonomic relationships. 7 Related Work In the research using verb-oriented method, the relation of a verb and nouns modified with it is described, and the concept definition is constructed from these information (e.g. [9]). In [8],

Table 4: Evaluation by the user with legal knowledge # Extracted # Accepted # Rejected Precision Recall concept pairs concept pairs concept pairs WS 77 18 59 0.23(18/77) 0.38(18/48) AR 55 13 42 0.24(13/55) 0.27(13/48) the join of 117 27 90 0.23(27/117) 0.56(27/48) WS and AR Figure 7: Three different sets of concept pairs from user, WS and AR taxonomic relationships and Subcategorization Frame of verbs (SF) are extracted from technical texts using a machine learning method. The nouns in two or more kinds of different SF with a same frame-name and slot-name is gathered as one concept, base class. And ontology with only taxonomic relationships is built by carrying out clustering of the base class further. Moreover, in parallel, Restriction of Selection (RS) which is slot-value in SF is also replaced with the concept with which it is satisfied instantiated SF. However, proper evaluation is not yet done. Since SF represents the syntactic relationships between verb and noun, the step for the conversion to non-taxonomic relationships is necessary. On the other hand, in ontology learning using data-mining method, discovering nontaxonomic relationships using a AR algorithm is proposed by [3]. They extract concept pairs based on the modification information between terms selected with parsing, and made the concept pairs a transaction. By using heuristics with shallow text processing, the generation of a transaction more reflects the syntax of texts. Moreover, RLA, which is their original learning accuracy of non-taxonomic relationships using the existing taxonomic relations, is proposed. The concept pair extraction method in our paper does not need parsing, and it can also run off context similarity between the terms appeared apart each other in texts or not mediated by the same verb. 8 Conclusions In this paper, we discussed how to construct a domain ontology using existing MRD and text corpus. In order to acquire taxonomic relationships, two strategies have been proposed: matched result analysis and trimmed result analysis. Furthermore, in order to learn nontaxonomic relationships, concept pairs have been extracted from text corpus with WS and AR. Taking the join of WS and AR, the recall goes over 0.5 and so it works to support the user in constructing non-taxonomic relationships of concepts.

The future work comes as follows: the system integration of taxonomic relationship acquisition module and non-taxonomic relationship learning module and the application to large size domains. Acknowledgments We would like to express our thanks to Mr. Takamasa Iwade (a graduate student of Shizuoka university) and Mr.Takuya Miyabe (a student of Shizuoka university) and the members in the Yamaguchi-Lab. References [1] Bill Swartout, Ramesh Patil, Kevin Knight and Tom Russ: Toward Distributed Use of Large-Scale Ontologies, Proc. of the 10th Knowledge Acquisition Workshop (KAW 96), (1996) [2] Rieko Sekiuchi, Chizuru Aoki, Masaki Kurematsu and Takahira Yamaguchi: DODDLE : A Domain Ontology Rapid Development Environment, PRICAI98, (1998) [3] Alexander Maedche, Steffen Staab: Discovering Conceptual Relations from Text, ECAI2000, 321 325 (2000) [4] C.Fellbaum ed: Wordnet, The MIT Press, 1998. see also URL: http://www.cogsci.princeton.edu/ wn/ [5] Rakesh Agrawal, Ramakrishnan Srikant : Fast algorithms for mining association rules,, Proc. of VLDB Conference, 487 499 (1994) [6] Marti A. Hearst, Hinrich Schutze: Customizing a Lexicon to Better Suit a Computational Task, in Corpus Processing for Lexical Acquisition edited by Branimir Boguraev & James Pustejovsky, 77 96 [7] Kazuaki Sono, Masasi Yamate: United Nations convention on Contracts for the International Sale of Goods, Seirin-Shoin(1993) [8] David Faure, Claire Nédellec, Knowledge Acquisition of Predicate Argument Structures from Technical Texts Using Machine Learning: The System ASIUM, EKAW 99 [9] Udo Hahn, Klemens Schnattinger Toward Text Knowledge Engineering, AAAI98, IAAAI-98 proceedings, 524-531 (1998)