Dependency-based Sentence Synthesis Component for Czech

Similar documents
Adding syntactic structure to bilingual terminology for improved domain adaptation

The College Board Redesigned SAT Grade 12

A Minimalist Approach to Code-Switching. In the field of linguistics, the topic of bilingualism is a broad one. There are many

Word Stress and Intonation: Introduction

Semi-supervised Training for the Averaged Perceptron POS Tagger

Proof Theory for Syntacticians

CS 598 Natural Language Processing

A Framework for Customizable Generation of Hypertext Presentations

Loughton School s curriculum evening. 28 th February 2017

Developing a TT-MCTAG for German with an RCG-based Parser

Specification and Evaluation of Machine Translation Toy Systems - Criteria for laboratory assignments

Enhancing Unlexicalized Parsing Performance using a Wide Coverage Lexicon, Fuzzy Tag-set Mapping, and EM-HMM-based Lexical Probabilities

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Developing Grammar in Context

Intensive English Program Southwest College

Introduction to HPSG. Introduction. Historical Overview. The HPSG architecture. Signature. Linguistic Objects. Descriptions.

LING 329 : MORPHOLOGY

Writing a composition

What the National Curriculum requires in reading at Y5 and Y6

Reading Grammar Section and Lesson Writing Chapter and Lesson Identify a purpose for reading W1-LO; W2- LO; W3- LO; W4- LO; W5-

Modeling full form lexica for Arabic

Parsing of part-of-speech tagged Assamese Texts

PAGE(S) WHERE TAUGHT If sub mission ins not a book, cite appropriate location(s))

Character Stream Parsing of Mixed-lingual Text

Oakland Unified School District English/ Language Arts Course Syllabus

ELD CELDT 5 EDGE Level C Curriculum Guide LANGUAGE DEVELOPMENT VOCABULARY COMMON WRITING PROJECT. ToolKit

1. Introduction. 2. The OMBI database editor

Language Acquisition by Identical vs. Fraternal SLI Twins * Karin Stromswold & Jay I. Rifkin

Underlying and Surface Grammatical Relations in Greek consider

Target Language Preposition Selection an Experiment with Transformation-Based Learning and Aligned Bilingual Data

5 Star Writing Persuasive Essay

Some Principles of Automated Natural Language Information Extraction

Grammars & Parsing, Part 1:

Linguistic Variation across Sports Category of Press Reportage from British Newspapers: a Diachronic Multidimensional Analysis

CEFR Overall Illustrative English Proficiency Scales

Constraining X-Bar: Theta Theory

Mercer County Schools

FOREWORD.. 5 THE PROPER RUSSIAN PRONUNCIATION. 8. УРОК (Unit) УРОК (Unit) УРОК (Unit) УРОК (Unit) 4 80.

ENGBG1 ENGBL1 Campus Linguistics. Meeting 2. Chapter 7 (Morphology) and chapter 9 (Syntax) Pia Sundqvist

A High-Quality Web Corpus of Czech

Senior Stenographer / Senior Typist Series (including equivalent Secretary titles)

Opportunities for Writing Title Key Stage 1 Key Stage 2 Narrative

Approaches to control phenomena handout Obligatory control and morphological case: Icelandic and Basque

Syntactic Dependencies for Multilingual and Multilevel Corpus Annotation

CAAP. Content Analysis Report. Sample College. Institution Code: 9011 Institution Type: 4-Year Subgroup: none Test Date: Spring 2011

National Literacy and Numeracy Framework for years 3/4

Subject: Opening the American West. What are you teaching? Explorations of Lewis and Clark

Beyond the Pipeline: Discrete Optimization in NLP

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Syntax Parsing 1. Grammars and parsing 2. Top-down and bottom-up parsing 3. Chart parsers 4. Bottom-up chart parsing 5. The Earley Algorithm

Improved Effects of Word-Retrieval Treatments Subsequent to Addition of the Orthographic Form

GERM 3040 GERMAN GRAMMAR AND COMPOSITION SPRING 2017

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

ReinForest: Multi-Domain Dialogue Management Using Hierarchical Policies and Knowledge Ontology

SINGLE DOCUMENT AUTOMATIC TEXT SUMMARIZATION USING TERM FREQUENCY-INVERSE DOCUMENT FREQUENCY (TF-IDF)

TABE 9&10. Revised 8/2013- with reference to College and Career Readiness Standards

On-Line Data Analytics

First Grade Curriculum Highlights: In alignment with the Common Core Standards

AQUA: An Ontology-Driven Question Answering System

Candidates must achieve a grade of at least C2 level in each examination in order to achieve the overall qualification at C2 Level.

Common Core State Standards for English Language Arts

Emmaus Lutheran School English Language Arts Curriculum

Case government vs Case agreement: modelling Modern Greek case attraction phenomena in LFG

Procedia - Social and Behavioral Sciences 154 ( 2014 )

Sample Goals and Benchmarks

FinnTreeBank: Creating a research resource and service for language researchers with Constraint Grammar

Applications of memory-based natural language processing

Derivational and Inflectional Morphemes in Pak-Pak Language

Advanced Grammar in Use

Linking Task: Identifying authors and book titles in verbose queries

Tutoring First-Year Writing Students at UNM

English Language and Applied Linguistics. Module Descriptions 2017/18

Chunk Parsing for Base Noun Phrases using Regular Expressions. Let s first let the variable s0 be the sentence tree of the first sentence.

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Chapter 9 Banked gap-filling

THE VERB ARGUMENT BROWSER

11/29/2010. Statistical Parsing. Statistical Parsing. Simple PCFG for ATIS English. Syntactic Disambiguation

Abstractions and the Brain

Specifying a shallow grammatical for parsing purposes

The Strong Minimalist Thesis and Bounded Optimality

UNIVERSITY OF OSLO Department of Informatics. Dialog Act Recognition using Dependency Features. Master s thesis. Sindre Wetjen

Pseudo-Passives as Adjectival Passives

Adjectives tell you more about a noun (for example: the red dress ).

Parallel Syntactic Annotation of Multiple Languages

ECE-492 SENIOR ADVANCED DESIGN PROJECT

Natural Language Processing. George Konidaris

Arizona s English Language Arts Standards th Grade ARIZONA DEPARTMENT OF EDUCATION HIGH ACADEMIC STANDARDS FOR STUDENTS

An Interactive Intelligent Language Tutor Over The Internet

Universal Grammar 2. Universal Grammar 1. Forms and functions 1. Universal Grammar 3. Conceptual and surface structure of complex clauses

Accurate Unlexicalized Parsing for Modern Hebrew

University of Groningen. Systemen, planning, netwerken Bosman, Aart

Epping Elementary School Plan for Writing Instruction Fourth Grade

A Computational Evaluation of Case-Assignment Algorithms

Context Free Grammars. Many slides from Michael Collins

5 th Grade Language Arts Curriculum Map

Project in the framework of the AIM-WEST project Annotation of MWEs for translation

Ontologies vs. classification systems

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Ch VI- SENTENCE PATTERNS.

A Grammar for Battle Management Language

Transcription:

MTT 2007, Klagenfurt, May 21 24, 2007 Wiener Slawistischer Almanach, Sonderband 69, 2007 Dependency-based Sentence Synthesis Component for Czech Jan Ptáček, Zdeněk Žabokrtský Institute of Formal and Applied Linguistics Malostranské náměstí 25, 11800 Prague, Czech Republic {ptacek,zabokrtsky}@ufal.mff.cuni.cz Abstract We propose a complex rule-based system for generating Czech sentences out of tectogrammatical trees, as introduced in Functional Generative Description (FGD) and implemented in the Prague Dependency Treebank 2.0 (PDT 2.0). Linguistically relevant phenomena including valency, diathesis, condensation, agreement, word order, punctuation and vocalization have been studied and implemented in Perl using software tools shipped with PDT 2.0. Parallels between generation from the tectogrammatical layer in FGD and deep syntactic representation in Meaning-Text Theory are also briefly sketched. Keywords Natural Language Generation, Prague Dependency Treebank. 1 Introduction Natural Language Generation (NLG) is a sub-domain of Computational Linguistics; its aim is studying and simulating the production of written (or spoken) discourse. Usually the discourse is generated from a more abstract, semantically oriented data structure. The most prominent application of NLG is probably transfer-based machine translation, but NLG is relevant also for dialog systems, systems for text summarizing, systems for generating technical documentation, etc. In this paper, the NLG task is formulated as follows: given a Czech tectogrammatical tree as introduced in Functional Generative Description (Sgall, 1967), (Sgall et al., 1986) and recently elaborated in more detail within the PDT 2.0 1 project generate a Czech sentence the meaning of which corresponds to the content of the input tree. Note that in the context of PDT 2.0, synthesis of written sentences can be viewed as a process inverse to treebank 1 http://ufal.mff.cuni.cz/pdt2.0/

Jan Ptáček & Zdeněk Žabokrtský annotation. Not surprisingly, the presented research is motivated by the idea of transfer-based machine translation with the usage of tectogrammatics as the highest abstract representation. The NLG task is thoroughly explored. We mention here the text-generation systems based on the Meaning-Text Theory (Mel čuk, 1988). As first comparison criterion, we examine the data structure expected on input to the generation procedure. No standard input is defined for a general NLG problem. The generation can be based on non-linguistic data stored in a database or obtained interactively (in systems like LSF or AlethGen (Iordanskaja et al., 1992), (Coch, 1996)). Unlike our generator, such systems are focused on a particular domain and deal with text and sentence planning. But the generation process can also start from a deep syntactic representation such as system RealPro (Lavoie & Rambow, 1997). We differ though in the definition of the deep syntactic representation. The second criterion is the mechanism of grammar rules application. A graph rewriting approach suggested by Mel čuk (Mel čuk, 1988) dominates here. Such approach treats grammar as a separable resource and needs a nontrivial framework (such as MATE (Bohnet & Wanner, 2001)) for its processing. Our grammar of Czech is hardwired ; written in the Perl programming language. It is modularized and uses pluggable resources as seen in Figure 2. Procedural design results in quick prototyping and also natural order of operations is highlighted. 2 PDT 2.0 in a Nutshell In the Prague Dependency Treebank 2.0 annotation scenario, based on the theoretical framework of Prague Functional Generative Description, three layers of annotation are added to Czech sentences (Jan Hajič et al., 2006): Morphological layer (m-layer), on which each token in each sentence of the source texts is lemmatized and tagged with a positional POS-tag. 2 Analytical layer (a-layer), on which a sentence is represented as a rooted ordered tree with labeled nodes and edges, corresponding to the surface-syntactic relations; each a-layer node corresponds to exactly one m-layer token. Tectogrammatical layer (t-layer), on which the sentence is represented as a deep-syntactic dependency tree structure (t-tree) built of nodes and edges. T-layer nodes represent autosemantic words (including pronouns and numerals) while functional words such as prepositions, subordinating conjunctions and auxiliary verbs have no nodes of their own in the tree. Each tectogrammatical node is a complex data structure it can be viewed as a set of attribute-value pairs, or even as a typed feature structure. Word forms occurring in the original surface expression are substituted with their t-lemmas. Only semantically 2 Technically, there is also one more layer called w-layer (word layer) below the mlayer; on this lowest layer the original raw text is only segmented into documents, paragraphs and tokens, and all these units are enriched with identifiers.

Dependency-based Sentence Synthesis Component for Czech indispensable morphological categories (called grammatemes) are stored in the nodes (such as number for nouns, or degree of comparison for adjectives), but not the categories imposed by government (i.e. case for nouns) or agreement (congruent categories such as person for verbs or gender for adjectives). Each edge in the t-tree is labeled with a functor representing the deep-syntactic dependency relation. 3 Coreference and topic-focus articulations are annotated in t-trees as well. See (Mikulová et al., 2005) for a detailed description of the t-layer. Figure 1: Illustration of the individual steps of the generating procedure when applied on a t- tree fragment corresponding to the expression se starším bratrem (lit. with older brother). 3 Synthesis Procedure Unlike stochastic end-to-end solutions, a rule-based approach, which we adhere to in this paper, requires a careful decomposition of the task. (Due to the very complex nature of the task, a monolithic implementation could hardly be maintainable). The decomposition was not trivial to find, because many linguistic phenomena are to be considered and some of them may interfere with others; the presented solution results from several months of experiments and a few re-implementations. In our system, the input tectogrammatical tree is gradually changing in each step, new node attributes and/or new nodes are added. Step by step, the structure becomes more and more similar to an a-layer tree. After the last step, the resulting sentence is obtained simply by concatenating word forms which are already filled in the individual nodes, the ordering of which is also already specified. 3 Edge labels are in fact treated and visualized as attributes of dependent nodes.

Jan Ptáček & Zdeněk Žabokrtský Figure 2: Data-flow diagram representing the process of sentence synthesis. A simplified data-flow diagram corresponding to the generating procedure is displayed in Figure 2. All the main phases of the generating procedure will be outlined in the following subsections, some of them are illustrated on an artificial t-tree fragment in Figure 1, or on an authentic sentence from PDT 2.0 in Figures 3 and 4. The procedure has been implemented in Perl within the tred/btred 4 tree processing environment developed by Petr Pajas. 3.1 Formeme Selection, Diatheses, Derivations In this phase, the input tree is traversed in the depth-first fashion, and so called formeme is specified for each node. Under this term we understand a set of constraints on how the given node can be expressed on the surface (i.e., what morphosyntactic form is used). Possible values are for instance simple case gen (genitive), prepositional case pod+7 (preposition pod and instrumental), v-inf (non-finite verb), že+v-fin (subordinating clause introduced with subordinating conjunction že), adj (syntactic adjective), etc. Several types of information are used when deriving the value of the new formeme attribute. At first, the valency lexicon 5 is consulted: if the governing node of the current node has a nonempty valency frame, and the valency frame specifies constraints on the surface form for the functor of the current node, then these constraints imply the set of possible formemes. In case of verbs, it is also necessary to specify which diathesis should be used (active, passive, reflexive passive, etc.; depending on the type of diathesis, the valency frame from the lexicon undergoes certain transformations). If the governing node does not have a valency frame, then the formeme default for the functor of the current node (and subfunctor, which specifies the type of the dependency relations in more detail) is used. For instance, the default formeme for the functor ACMP (accompaniment) and subfunctor basic is s+7 (with), whereas for ACMP.wout it is bez+2 (without). 4 5 http://ufal.mff.cuni.cz/ pajas/tred/ There is the valency lexicon PDT-VALLEX ((Hajič et al., 2003)) associated with PDT 2.0. On the t-layer of the annotated data, all semantic verbs and some semantic nouns and adjectives are equipped with a reference to a nonempty valency frame in PDT-VALLEX, which was used in the given sentence.

Dependency-based Sentence Synthesis Component for Czech It should be noted that the formeme constraints depend also on the possible word-forming derivations applicable on the current node. For instance, the functor APP (appurtenance) can be typically expressed by formemes gen (genitive) and adj (possessive adjective), but in some cases only the former one is possible (some Czech nouns do not form derived possessive adjectives). 3.2 Propagating Values of Congruent Categories In Czech, which is a highly inflectional language, several types of dependencies are manifested by agreement of morphological categories (agreement in gender, number, and case between a noun and its adjectival attribute, agreement in number, gender, and person between a finite verb and its subject, agreement in number and gender between relative pronoun in a relative clause and the governor of the relative clause, etc.). As already mentioned, the original tectogrammatical tree contains only those morphological categories which are semantically indispensable. After the formeme selection phase, value of case should be also known for all nouns. In this phase, oriented agreement arcs (corresponding to the individual types of agreement) are conceived between nodes within the tree, and the values of morphological categories are iteratively spread along these arcs until the unification process is completed. 3.3 Expanding Complex Verb Forms Only now, when person, number, and gender of finite verbs are known, it is possible to expand complex verb forms where necessary. New nodes corresponding to reflexive particles (e.g., in the case of reflexiva tantum), to auxiliary verbs (e.g., in the case of complex future tense), or to modal verbs (if deontic modality of the verb is specified) are attached below the original autosemantic verb. 3.4 Adding Prepositions and Subordinating Conjunctions In this phase, new nodes corresponding to prepositions and subordinating conjunctions are added into the tree. Their lemmas are already implied by the value of node formemes. 3.5 Determining Inflected Word Forms After the agreement step, all information necessary for choosing the appropriate inflected form of the lemma of the given node should be available in the node. To perform the inflection, we employ morphological tools (generator and analyzer) developed by Hajič (Hajič, 2004). The generator tool expects a lemma and a positional tag (as specified in (Hana et al., 2002)) on the input, and returns the inflected word form. Thus the task of this phase is effectively reduced to composing the positional morphological tag; the inflection itself is performed by the morphological generator.

Jan Ptáček & Zdeněk Žabokrtský Figure 3: (Simplified) PDT 2.0 t-tree corresponding to the sentence Přesto uvedením lhůty ve smlouvě by se bylo předešlo četným nedorozuměním, která se nyní objevila a která nás mrzí. (But still, stating the period in the contract would have prevented frequent misunderstandings which have now arisen and which we are sorry about.) 3.6 Special Treatment of Definite Numerals Definite numerals in Czech (and thus also in PDT 2.0 t-trees) show many irregularities (compared to the rest of the language system), that is why it seems advantageous to generate their forms separately. Generation of definite numerals is discussed in (Ptáček, 2005). 3.7 Reconstructing Word Order Ordering of nodes in the annotated t-tree is used to express information structure of the sentences, and does not directly mirror the ordering in the surface shape of the sentence. The word order of the output sentence is reconstructed using simple syntactic rules (e.g., adjectival attribute precedes the governing noun) and topic-focus articulation. Special treatment is required for clitics: they should be located in the second position in the clause (Wackernagel position); if there are more clitics in the same clause, simple rules for specifying their relative ordering are used (for instance, the clitic by always precede short reflexive pronouns). 3.8 Adding Punctuation Marks In this phase, missing punctuation marks are added to the tree, especially (i) the terminal punctuation (derived from the sentmod grammateme), (ii) punctuations delimiting boundaries of clauses, of parenthetical constructions, and of direct speeches, (iii) and punctuations in multiple coordinations (commas in expressions of the form A, B, C and D).

Dependency-based Sentence Synthesis Component for Czech Figure 4: One of the intermediate phases during the processing of the t-tree from Figure 3. Almost all processing steps are performed (see the added nodes with functional words and punctuation marks, the inflected word forms, properly placed clitics etc). After performing the last step concatenation of the word forms into one string the following synthesized sentence is obtained: Přesto uvedením lhůty ve smlouvě by se bylo předešlo četným nedorozuměním, která se nyní objevila a která nás mrzí. Besides adding punctuation marks, the first letter of the first token in the sentence is also capitalized in this phase. 3.9 Vocalizing Prepositions Vocalization is a phonological phenomenon: the vowel -e or -u is attached to a preposition if the pronunciation of the prepositional group would be difficult without the vowel (e.g., ve výklenku instead of *v výklenku). We have adopted vocalization rules precisely formulated in (Petkevič, 1995) (technically, we converted them into the form of an XML file, which is loaded by the vocalization module). 3.10 Linearization At this moment, the resulting structure has roughly the shape of surface-syntactic tree (one inflected word form or punctuation mark per node, see Figure 4). The last thing to do is to merge the tokens into the final sentence string, which is a trivial task complicated only by the question of placement of spaces around quotation marks and other special symbols. 4 Final Remarks In this paper we have presented our approach to generating Czech sentences from tectogrammatical trees. More information about the system (including some implementation details and evaluation of the generator performance by measuring BLEU-score distance between the original sentences in the PDT 2.0 and the generated sentences) is given in (Ptáček, 2005).

Jan Ptáček & Zdeněk Žabokrtský Finally, we would like to note that the task of generating sentences from t-trees is in our opinion very similar to generating sentences from DSyntR (Deep-Syntactic Representation) as defined in Meaning-Text Theory. Most of the consequences of the common features could have been seen in the previous section. However, in the following paragraphs, we try to make them explicit using the list of resemblances between t-trees and DSyntR enumerated in (Žabokrtský, 2005). (1) The skeleton of both representations is formed by dependency tree (unordered in MTT, ordered according to information structure in FGD). In other words, lexicalization and hierarchization of a message (and each sentence in particular) is more or less specified already in the input of the generating procedure (unlike the case of generation e.g., from SemR). (2) Only semantically full lexemes (autosemantic words) do have nodes of their own (semantically empty lexemes/synsemantic words, such as prepositions, subordinating conjunctions, auxiliary verbs, etc. are introduced only in the surface-syntactic structure). This implies two things: both in the generation from t-layer and DSyntR, full lexemes must undergo inflection and functional words have to be added. (3) Each lexeme is associated with appropriate semantically full grammemes (grammatemes in FGD terminology); grammemes imposed only by government and agreement are excluded. During generation, values of grammemes have to be distributed along the agreement links also to the places, where they are not semantically indispensable, but are manifested by inflection. (4) Each dependency tree is accompanied with (non-tree) grammatical coreferential relations, together forming dag (directed acyclic graph). To generate a grammatical sentence, coreferential links cannot be ignored: they are important e.g., for detection of reflexive pronoun or agreement of relative pronouns. We also believe that the two approaches could mutually enrich each other: for example, it would be very useful to adopt the notion of lexical functions for FGD, especially when a similar notion is de facto used in PDT for relating e.g., deadjectival adverbs with their primary adjectives, or possessive adjectives with their primary nouns. Acknowledgements The research has been funded by projects 1ET101120503, 1ET201120505 and MSM0021620838. Bibliography Bohnet, B. & L. Wanner. 2001. On Using a Parallel Graph Rewriting Grammar Formalism in Generation. In Proceedings of the 8th European Natural Language Generation Workshop at the Annual Meeting of the Association for Computational Linguistics, Toulouse.

Dependency-based Sentence Synthesis Component for Czech Coch, J. 1996. Overview of AlethGen. In Proceedings of the Eighth International Workshop on Natural Language Generation, Demonstrations Volume. Brighton : Information Technology Research Institute, University of Brighton, 25 28. Hajič, J., J. Panevová, Z. Urešová, A. Bémová, V. Kolářová-Řezníčková & P. Pajas. 2003. PDT-VALLEX: Creating a Large-coverage Valency Lexicon for Treebank Annotation. In Proceedings of The Second Workshop on Treebanks and Linguistic Theories, Vaxjo: Vaxjo University Press, 57 68. Hajič, J. 2004. Disambiguation of Rich Inflection Computational Morphology of Czech. Prague: The Karolinum Press. Hana, J., H. Hanová, J. Hajič, B. Vidová-Hladká & E. Jeřábek. 2002. Manual for Morphological Annotation. Technical Report TR-2002-14, Prague: ÚFAL MFF, Charles University. Hajič, J. et al. 2006. Prague Dependency Treebank 2.0. Linguistic Data Consortium, CAT LDC2006T01, ISBN 1-58563-370-4. Iordanskaja, L., M. Kim, R. Kittredge, B. Lavoie & A. Polguere. 1992. Generation of extended bilingual statistical reports. In Proceedings of the 14th International Conference on Computational Linguistics, Nantes, 1019 1023. Lavoie, B. & O. Rambow. 1997. A fast and portable realizer for text generation systems. In Proceedings of the 5th Conference on Applied Natural Language Processing, Washington, 265 268. Mel čuk, I. 1988. Dependency Syntax: Theory and Practice, Albany: State University of New York Press. Mikulová, M., A. Bémová, J. Hajič, E. Hajičová, J. Havelka, V. Kolářová, M. Lopatková, P. Pajas, J. Panevová, M. Razímová, P. Sgall, J. Štěpánek, Z. Urešová, K. Veselá, Z. Žabokrtský & L. Kučová. 2005. Anotace na tektogramatické rovině Pražského závislostního korpusu, Anotátorská příručka. Technical Report TR-2005-28, Prague: ÚFAL MFF, Charles University. Petkevič, V. 1995. Vocalization of Prepositions. In Petkevič, V. (ed). Linguistic Problems of Czech, Final Research Report for the JRP PECO, Prague: Charles University, 147 157. Ptáček, J. 2005. Generování vět z tektogramatických stromů Pražského závislostního korpusu. Master s thesis, Prague: MFF, Charles University. Sgall, P. 1967. Generativní popis jazyka a česká deklinace. Prague: Academia. Sgall, P., E. Hajičová & J. Panevová. 1986. The Meaning of the Sentence in Its Semantic and Pragmatic Aspects. Dordrecht: D. Reidel Publishing Company. Žabokrtský, Z. 2005. Resemblances between Meaning-Text Theory and Functional Generative Description. In J.D. Apresjan, L.L. Iomdin, (ed). Proceedings of the 2nd International Conference of Meaning-Text Theory, Moscow, 549 557.