| { |
| "paper_id": "W96-0405", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T04:58:46.953107Z" |
| }, |
| "title": "Input Specification in the WAG Sentence Generation System", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "O'donnell", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "University of Edinburgh", |
| "location": { |
| "addrLine": "80 South Bridge", |
| "postCode": "EH1 1HN", |
| "settlement": "Edinburgh", |
| "country": "UK" |
| } |
| }, |
| "email": "micko@aisb.ed.ac.uk" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper describes the input specification language of the WAG Sentence Generation system. The input is described in terms of Halliday's (1978) three meaning components, ideational meaning (the propositional content to be expressed), interactional meaning (what the speaker intends the listener to do in making the utterance), and textual meaning (how the content is structured as a message, in terms of theme, reference, etc.).", |
| "pdf_parse": { |
| "paper_id": "W96-0405", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper describes the input specification language of the WAG Sentence Generation system. The input is described in terms of Halliday's (1978) three meaning components, ideational meaning (the propositional content to be expressed), interactional meaning (what the speaker intends the listener to do in making the utterance), and textual meaning (how the content is structured as a message, in terms of theme, reference, etc.).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "This paper describes the input specification language of the WAG Sentence Generation system. The input is described in terms of Halliday's (1978) three meaning components, ideational meaning (the propositional content to be expressed), interactional meaning (what the speaker intends the listener to do in making the utterance), and textual meaning (how the ideational content is structured as a message, in terms of theme, reference, etc.).", |
| "cite_spans": [ |
| { |
| "start": 128, |
| "end": 145, |
| "text": "Halliday's (1978)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "One motivation for this paper is the lack of descriptions of input-specifications for sentence generators. I have been asked at various times to fill this gap. Perhaps a better motivation is the need to argue for a more abstract level of input. Many of the available sentence generators require specification of syntactic information within the input specification. This means that any text-planner which uses this system as its realisation module needs to concern itself with these fiddling details. One of the aims in the WAG system has been to lift the abstractness of sentence specification to a semantic level. This paper discusses this representation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The WAG Sentence Generation System is one component of the Workbench for Analysis and Generation (WAG), a system which offers various tools for developing Systemic resources (grammars, semantics, lexicons, etc.) , maintaining these resources (lexical acquisition tools, network graphers, hypertext browsers, etc.), and processing (sentence analysis -O' Donnell 1993 Donnell , 1994 ; sentence generation O'Donnell 1995b; knowledge representation -O'Donnell 1994; corpus tagging and exploration-O'Donnell 1995a).", |
| "cite_spans": [ |
| { |
| "start": 174, |
| "end": 211, |
| "text": "(grammars, semantics, lexicons, etc.)", |
| "ref_id": null |
| }, |
| { |
| "start": 353, |
| "end": 365, |
| "text": "Donnell 1993", |
| "ref_id": null |
| }, |
| { |
| "start": 366, |
| "end": 380, |
| "text": "Donnell , 1994", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The Sentence Generation component of this system generates single sentences from a semantic input. This semantic input could be supplied by a human user. Alternatively, the semantic input can be generated as the output of a multi-sentential text generation system, allowing such a system to use the WAG system as its realisation component. The sentence generator can thus be treated a blackbox unit. Taking this approach, the designer of the multi-sentential generation system can focus on multi-sentential concerns without worrying about sentential issues. WAG improves on earlier sentence generators in various ways. Firstly, it provides a more abstract level of input than many other systems (Mumble: McDonald 1980; Meteer et al. 1987; FUF: Elhadad 1991) , as will be demonstrated throughout this paper. The abstractness improves even over the nearest comparable system, Penman (Mann 1983; Mann 8z Matthiessen 1985) , in its treatment of textual information (see below). Other sentence generators, while working from abstract semantic specifications, do not represent a generalised realiser, but are somewhat domain specific in implementation, e.g., Proteus (Davey 1974 (Davey /1978 ; Slang (Patten 1988) . Other systems do not allow generation independent from user interaction, for instance, Genesys (Faw-cett & Tucker 1990) requires the user to make decisions throughout the generation process.", |
| "cite_spans": [ |
| { |
| "start": 695, |
| "end": 718, |
| "text": "(Mumble: McDonald 1980;", |
| "ref_id": null |
| }, |
| { |
| "start": 719, |
| "end": 738, |
| "text": "Meteer et al. 1987;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 739, |
| "end": 757, |
| "text": "FUF: Elhadad 1991)", |
| "ref_id": null |
| }, |
| { |
| "start": 881, |
| "end": 892, |
| "text": "(Mann 1983;", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 893, |
| "end": 918, |
| "text": "Mann 8z Matthiessen 1985)", |
| "ref_id": null |
| }, |
| { |
| "start": 1161, |
| "end": 1172, |
| "text": "(Davey 1974", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 1173, |
| "end": 1185, |
| "text": "(Davey /1978", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 1194, |
| "end": 1207, |
| "text": "(Patten 1988)", |
| "ref_id": "BIBREF23" |
| }, |
| { |
| "start": 1305, |
| "end": 1329, |
| "text": "(Faw-cett & Tucker 1990)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Against WAG, it does not yet have the grammatical and semantic coverage of Penman, FUF or Mumble, although its coverage is reasonable, and growing quickly.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The input to the WAG Sentence generation system is a specification of an utterance on the semantic stratum. We thus need to explore further the nature of Systemic semantic representation. Halliday (1978) divides semantic resources into three areas, called metafunctions:", |
| "cite_spans": [ |
| { |
| "start": 188, |
| "end": 203, |
| "text": "Halliday (1978)", |
| "ref_id": "BIBREF8" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic Metafunctions", |
| "sec_num": "1.1" |
| }, |
| { |
| "text": "1. Interactional Metafunction: viewing language as interaction, i.e., an activity involving speakers and listeners, speechacts, etc. Interactional meaning includes the attitudes, social roles, illocutionary goals, etc of interactants.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic Metafunctions", |
| "sec_num": "1.1" |
| }, |
| { |
| "text": "2. Ideational Metafunction: concerned with the propositional content of the text, structured in terms of processes (mental, verbal, material, etc.) , the participants in the process (Actor, Actee, etc.), and the circumstances surrounding the process (Location, Manner, Cause, etc.).", |
| "cite_spans": [ |
| { |
| "start": 115, |
| "end": 147, |
| "text": "(mental, verbal, material, etc.)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Semantic Metafunctions", |
| "sec_num": "1.1" |
| }, |
| { |
| "text": "is constructed as a message conveying information. This concerns, for instance, the thematic structuring of the ideation presented in the text, its presentation as recoverable or not, the semantic relevance of information, etc.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Textual Metafunction: how the text", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Although these metafunctions apply to both the semantics of sentence-size and multisentential texts, this paper will focus on sentential semantics, since we are dealing with the input to a sentence generator. Below we explore the nature of this semantic specification in more detail.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Textual Metafunction: how the text", |
| "sec_num": "3." |
| }, |
| { |
| "text": "Interactional representation views the text as part of the interaction between participants. Sentences themselves serve an important part in interaction, they form the basic units -the movesof which interactions are composed. Moves are also called speech-acts. Note that WAG serves in monologic as well as dialogic interactions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Interactional Specification", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The input to the WAG generator is basically a speech-act specification, although this specification includes ideational and textual specification. Figure 1 shows a sample speech-act specification, from which the generator would produce: I'd like information on some body repairers. The distinct contributions of the three meta-functions are separated by the grey boxes. Say is the name of the lisp function which evaluates the speech-act specification, calling the generator, dialog-5 is the name of this particular speech-act (each speech-act is given a unique identifier, its unit-id).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 147, |
| "end": 155, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Interactional Specification", |
| "sec_num": "2" |
| }, |
| { |
| "text": "In specifying the speech-act, there are several important things which need to be specified:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Interactional Specification", |
| "sec_num": "2" |
| }, |
| { |
| "text": "\u2022 Speech-Function: what does the speaker requires the hearer to do in regard to the encoded proposition? 1 This is called in Systemics the speech-function. Is the hearer supposed to accept the content as a fact? Or are they supposed to complete the proposition in some way? Or perform some action in response to the utterance? Participants: who is uttering the speechact (the Speaker), and who is it addressed to (the Hearer).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Interactional Specification", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The participant roles do not need to be included in every sentence-specification, but may be in some, for the following reasons:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 Pronominalisation: If the filler of the Speaker or Hearer role happens to play some role in the ideational specification, then an appropriate pronoun will be used in the generated string (e.g., T, 'you').", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "\u2022 Voice Selection: If the spoken output mode is used, WAG will select a voice of the same gender as the speaker entity.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "1For ease of writing, I use the terms 'speaker' and 'hearer' to apply to the participants in both spoken and written language. Paris 1993): lexico-grammatical choices can be constrained by reference to attributes specified in the Speaker and Hearer roles. 2 This has not, however, been done at present: while the implementation is set up to handle this tailoring, the resources have not yet been appropriately constrained.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "WAG's semantic input improves over that of Penman in regards to the relationship between the speech-act and the proposition. In Penman, the ideational specification is central: a semantic specification is basically an ideational specification, with the speech-act added as an additional (and optional) field. This approach is taken because Penman was designed with monologic text in mind, so the need for varied speech-acts is not well integrated. 3 2Since the fillers of the Speaker and Hearer roles are ideational units, they can be extensively specified for user-modelling purposes, including the place of origin, social class, social roles, etc of the participant. Relations between the participants can also be specified, for instance, parent/child, or doctor/patient relations. Lexico-grammatical decisions can be made by reference to this information: tailoring the language to the speaker's and hearer's descriptions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "a~VAG also allows the representation of non-verbal moves (e.g., the representation of system or user physical actions), which allows WAG to model interaction in a wider sense.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "WAG however takes the speech-act as central, the semantic specification is a specification of a speech-act. The ideational specification is provided as a role of the speech-act (the :proposition role). WAG thus integrates with more ease into a system intended for dialogic interaction, such as a tutoring system. In particular, it simplifies the representation of speech-acts with no ideational content, such as greetings, thank-yous, etc. Figure 2 shows the systems of the speech-act network used in WAG (based on O 'Donnell 1990 , drawing upon Berry 1981 Martin 1992) . The main systems in this network are as follows:", |
| "cite_spans": [ |
| { |
| "start": 517, |
| "end": 530, |
| "text": "'Donnell 1990", |
| "ref_id": null |
| }, |
| { |
| "start": 531, |
| "end": 556, |
| "text": ", drawing upon Berry 1981", |
| "ref_id": null |
| }, |
| { |
| "start": 557, |
| "end": 569, |
| "text": "Martin 1992)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 440, |
| "end": 448, |
| "text": "Figure 2", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Content: what proposition is being negotiated between the speaker and hearer?", |
| "sec_num": null |
| }, |
| { |
| "text": "I n i t i a t i o n : The grammatical form used to realise a particular utterance depends on whether the speaker/writer is initiating a new exchange, or responding to an existing exchange (e.g., an answer to a question). Responding moves reflect a far higher degree of ellipsis than initiating moves. In particular, a move responding to a whquestion usually only needs to provide the whelement in their reply. An elicit move indicates that the speaker requires some contentfull response, while a propose move may require changes of state of belief in the hearer, support moves indicate the speaker's acceptance of the prior speaker's proposition. Other speech-functions cater to various alternative responses in dialogue, for instance: deny-knowledge -the speaker indicates that they are unable to answer a question due to lack of knowledge; contradict: the speaker indicates that they disagree with the prior speaker's proposition; requestrepeat: the speaker indicates that they did not fully hear the prior speaker's move.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "T y p e s o f S p e e c h -A c t s", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "\u2022 Object of Negotiation: Speech-acts can negotiate information (questions, statements, etc.), or action (commands, permission, etc.). A move with features (:and elicit negotiate-action) would be realised as a request for action (e.g., Will you go now?), while a move with features (:and propose negotiate-action) would be realised as a command (e.g., Go now. O.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "T y p e s o f S p e e c h -A c t s", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "In writing a speech-act specification, the :is field is used to specify the the speech-act type (the same key is used to specify ideational types in propositional units). The speech-act of figure 1 is specified to be (:and initiate propose). Feature-specifications can be arbitrarily complex, consisting of either a single feature, or a logical combination of features (using any combination of :and, :or or :not). One does not need to specify features which are systemically implied, e.g., specifying propose is equivalent to specifying (:and move speech-act negotiatory propose). Hovy (1993) points out that as the input specification language gets more powerful, the amount of information required in the specification gets larger, and more complex. WAG thus allows elements of the semantic specification to take a default value if not specified. For instance, the example in figure 1 does not specify a choice between negotiate-information or negotiate-action (the first is the default). Other aspects are also defaulted, for instance, the relation between the speaking time and the time the event takes place (realised as tense).", |
| "cite_spans": [ |
| { |
| "start": 582, |
| "end": 593, |
| "text": "Hovy (1993)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "T y p e s o f S p e e c h -A c t s", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Once we have specified what the speech-act is doing, and who the participants are, we need to specify the ideational content of the speechact.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Specification", |
| "sec_num": "3" |
| }, |
| { |
| "text": "When talking about ideational specification, we need to distinguish ideational potentialthe specification of what possible ideational structures we can have; and ideational instanrials -actual ideational structures. The first is sometimes termed terminological knowledge knowledge about terms and their relations, the second, assertional knowledge -knowledge about actual entities and their relations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Ideational Potential: Ideational potential is represented in terms of an ontology of semantic types, a version of Penman's Upper Model (UM) (Bateman 1990; Bateman et al. 1990) . 4 The root of this ontology is shown in figure 3 . Many of the types in this ontology will have associated role constraints, for instance, a mental-process requires a Sensor role, which must be filled by a conscious entity. The UM thus constrains the possible ideational structures which can be produced.", |
| "cite_spans": [ |
| { |
| "start": 140, |
| "end": 154, |
| "text": "(Bateman 1990;", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 155, |
| "end": 175, |
| "text": "Bateman et al. 1990)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 218, |
| "end": 226, |
| "text": "figure 3", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The UM provides a generalised classification system of conceptual entities. For representing concepts which are domain-specific (e.g., bodyrepairer), users provide domain-models, where domain-specific concepts are subsumed to concepts in the UM.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Ideational Structure: An ideational specification is a structure of entities (processes, t-hings and qualities), and the relations between these entities. Such a structure is specified by providing two sets of information for each entity (as in the propositional slot of figure 1):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "\u2022 Type Information: a specification of the semantic types of the entity, derived from the UM, or associated domainmodel.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "4WAG's Upper Model has been re-represented in terms of system networks, rather than the more loosely defined type-lattice language used in Penman. WAG thus uses the same formalism for representing ideational, inteLctional and lexico-grammatical information.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "ideational-unit -I \"c\u00b0nsci\u00b0us q-human... Typically, a text-planner has a knowledgebase (KB) to express, and produces a set of sentence-specifications to express it. The form of the sentence-specifications differs depending on the degree of integration between the textplanner and the sentence-realiser. In most systems, the sentence-realiser has no access to the KB of the text-planner. This is desirable so that the sentence-realiser is independent of the text-planner -it can act as an independent module, making no assumptions as to the internal representations of the text-planner. The sentence-realiser can thus be used in connection with many different textplanners.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The sole communication between the two systems is through a sentence-specification -the text-planner produces a sentencespecification, which the sentence-realiser takes as input.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "The text-planner thus needs to re-express the contents of its KB into the ideational notation used by the sentencerealiser. This approach has been followed with systems such as Penman, FUF, and Mumble. Each of these has been the platform supporting various text-planners (often experimental).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "WAG also has been designed to support this planner-realiser separation, if need be. WAG can thus act as a stand-alone sentence realiser. The sentence specifihation of figure 1 reflects this mode of generation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "However, WAG supports a second mode of generation, allowing a higher degree of integration between the text-planner and the sentence realiser. In this approach, both processes have access to the KB. Ideational material thus does not need to be included within the input specification. Rather, the input specification provides only a pointer into the attached KB. Since the information to be expressed is already present in the KB, why does it need to be re-expressed in the semantic specification? Taking this approach, the role of the semantic specification is to describe how the information in the KB is to be expressed, including both interactional and textual shaping.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "This integration allows economies of generation not possible where content used for textplanning and content used for sentence generation are represented distinctly. One benefit involves economy of code -many of the processes which need to be coded to deal with ideation for a text as a whole can also be used to deal with ideation for single sentences. Another involves the possibility of integrating the two processes -since the sentence realiser has access to the same knowledge as the multi-sentential planner, it can make decisions without requiring explicit informing from the planner.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Another economy arises because translation between representations is avoided. In the stand-alone approach, the sentence-planner needs knowledge of how ideational specifications are formulated in the sentence specification language. It needs to map from the language of its KB to the language of the sentence specification. This is not necessary in an integrated approach.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "To demonstrate this integrated approach to sentence generation, we show below the generation of some sentences in two stagesfirstly, assertion of knowledge into the KB, and secondly, the evaluation of a series of speechacts, which selectively express components of this knowledge.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "; Participants (tell John :is male :name \"John\") (tell Mary :is female :name \"Mary\") (tell Party :is spatial) ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Ideational Representation", |
| "sec_num": "3.1" |
| }, |
| { |
| "text": "Now we are ready to express this knowledge. The following sentence-specification indicates that the speaker is proposing in/ormation, and that the leaving process is to be the semantic head of the expression. It also indicates which of the roles of each entity are relevant for expression (and are thus expressed if possible), -and which entities are identifiable in context (and can thus be referred to by name). The generation process, using this specification, produces the sentence shown after the form. As stated above, this approach does not require the sentence-specification to include any ideational-specification, except for a pointer into the KB. The realiser operates directly on the KB, using the information within the sentence-specification to tailor the expression.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Selective Expression of KB", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "Alternative sentence-specificati0ns result in different expressions of the same information, for instance, including more or less detail, changing the speech-act, or changing the textual status of various entities. The expression can also be altered by selecting a ~different entity as the head of the utterance. For instance, the following sentence-specification is identical to the previous, except the cause relation is now taken as the head, producing a substantially different sentence: We will now turn to the textual component of the input specification, which iS responsible for tailoring the expression of the ideational content.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Selective Expression of KB", |
| "sec_num": "3.2.2" |
| }, |
| { |
| "text": "Textual semantics concerns the role of the text and its components as a message, While creating a text (whether a single utterance or a whole book), we have a certain amount of content we wish to encode. But there are various ways to encode this information, to present our message. The textual semantics represents the various strategies for structuring the message.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Textual Specification", |
| "sec_num": "4" |
| }, |
| { |
| "text": "One of the main steps in the text generation process involves content selection -the selection of information from the speaker's knowledge base for presentation. Such a process must decide what information is relevant at each point of the unfolding discourse. In some systems, content selection is driven through the construction of the rhetorical structure of the text (e.g., Moore & Paris 199 ). As we build a rhetorical structure tree, the ideation which is necessary for each rhetorical relation is selected. For instance, if we add an evidence relation to an existing RST tree, the ideation which functions as evidence is selected for expression. The rhetorical structure thus organises the ideational content to be expressed, selecting out those parts of the ideation-base which are relevant to the achievement of the discourse goals at each point of the text. I use the term rhetorical relevance to refer to this sort of relevance. 5", |
| "cite_spans": [ |
| { |
| "start": 377, |
| "end": 394, |
| "text": "Moore & Paris 199", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant-Roles", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Rhetorical relevance is dynamic -it changes as the text progresses. It represents a shifting focus on the ideation base (Halliday ~ Matthiessen, 1995, pp373-380) . What is relevant changes as the text unfolds, as the rhetorical structure is realised. Relevance forms what Grosz (1977/86 ) calls a focus space. 6 Halliday & Matthiessen (1995) extend Grosz's notion of focus space to include other types of textual spaces: thematic spaces, identifiability spaces, new spaces, etc. (p376). Each of these spaces can be though of as a pattern stated over the ideation base.", |
| "cite_spans": [ |
| { |
| "start": 120, |
| "end": 161, |
| "text": "(Halliday ~ Matthiessen, 1995, pp373-380)", |
| "ref_id": null |
| }, |
| { |
| "start": 272, |
| "end": 286, |
| "text": "Grosz (1977/86", |
| "ref_id": null |
| }, |
| { |
| "start": 310, |
| "end": 341, |
| "text": "6 Halliday & Matthiessen (1995)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant-Roles", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "According to Grosz, focus is \"that part of the knowledge base relevant at a given point of a dialog.\" (p353). However, Grosz's notion of relevance is based on the needs of a text understanding system -which objects in the knowledge-base can be used to interpret the utterance. My sense of relevance is derived from relevance in generation -what information has been selected as relevant to the speaker's unfolding discourse goals. She is dealing with a set of objects which may potentially appear in the text at this point, while I am dealing with the set of objects which most probably do appear in the text.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant-Roles", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "To represent the relevance space in a sentence specification, I initially provided a :relevant-entities field, which listed those ideational entities which were relevant for expression. However, problems soon arose with this approach. Take for instance a situation where Mark owns both a dog and a house, and the dog destroyed the house. Now, we might wish to express a sentence to the effect that A dog destroyed Mark's house, which ignores Mark's ownership of the dog. In a system where relevance is represented as a list of entities, we could not produce this sentence.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant-Roles", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "What we need is a representation of the relevant relations in the KB. To this end, WAG's input specification allows a field :relevant-roles, which records the roles of each entity which are currently relevant for expression, e.g., as was used in the examples of section 3.2.2. 7", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant-Roles", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "While constructing a sentence, the sentence generator refers to this list at various points, to see if a particular semantic role is relevant, and on the basis of this, chooses one syntactic structure over another. At present, the ordering of roles in the list is not significant, but it could be made so, to constrain grammatical salience, etc.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Relevant-Roles", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "The :theme field of the speech-act specifies the unit-id of the ideational entity which is thematic in the sentence. If a participant in a process, it will typically be made Subject of the sentence. If the Theme plays a circumstantial role in the proposition, it is usually realised as a sentence-initial adjunct. WAG's treatment of Theme needs to be extended to handle the full range of thematic phenomena. Theme specification in WAG is identical to that used in Penman.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Theme", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "The participants in an interaction each possess a certain amount of information, some of which is shared, and some unshared. I use the term information status to refer to the status of information as either shared or unshared.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Information Status", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "The information status of ideational entities affects the way in which those items can be referred to. Below we discuss two dimensions of information status:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Information Status", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "TIf the explicit ideational specification is included in the say form (as in figure 1) , then the relevance space need not be stated, it is assumed that all the entities included within the specification axe relevant, and no others.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 77, |
| "end": 86, |
| "text": "figure 1)", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Information Status", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "entities which the speaker believes are known to the hearer can be referred to using identifiable reference, e.g., definite deixis, e.g., the President; and naming, e.g., Ronald Reagan. Entities which are not believed to be shared require some form of indefinite deixis, e.g., a boy called John; Eggs; Some eggs, etc.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": ". Shared Entities:", |
| "sec_num": null |
| }, |
| { |
| "text": "A speaker uses indefinite deixis to indicate that he believes the entity unknown to the hearer. It is thus a strategy used to introduce unshared entities into the discourse. Once the entity is introduced, some form of definite reference is appropriate.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": ". Shared Entities:", |
| "sec_num": null |
| }, |
| { |
| "text": "are part of the immediate discourse context can be referred to using pronominalisation (e.g., she, them, it, this, etc.) ; substitution (e.g., I saw one;); or ellipsis (the non-mention of an entity, e.g., Going to the shop?). The immediate discourse context includes entities introduced earlier in the discourse; and also entities within the immediate physical context of the discourse, e.g., the discourse participants (speaker, hearer, or speaker+hearer) and those entities which the participants can point at, for instance, a nearby table, or some person.", |
| "cite_spans": [ |
| { |
| "start": 94, |
| "end": 120, |
| "text": "she, them, it, this, etc.)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Recoverable Entities: Entities which", |
| "sec_num": "2." |
| }, |
| { |
| "text": "Two fields in the semantic specification allow the user to specify the information status of ideational entities -and thus how they can be referred to in discourse s (these lists will typically be maintained by the text-planner as part of its model of discourse context):", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Recoverable Entities: Entities which", |
| "sec_num": "2." |
| }, |
| { |
| "text": "\u2022 The Shared-Entities Field: a list of the ideational entities which the speaker wishes to indicate as known by the hearer, e.g., by using definite reference.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Recoverable Entities: Entities which", |
| "sec_num": "2." |
| }, |
| { |
| "text": "\u2022 The Recoverable-entities Field: a list of t the ideational entities which are recoverable from context, whether from the prior text, or from the immediate interactional context. SInformation status only partially constrains the choice of referential form -the choice between the remaining possibilities can be made by the sentence planner, by specifying directly grammatical preferences.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Recoverable Entities: Entities which", |
| "sec_num": "2." |
| }, |
| { |
| "text": "The input specification for the WAG sentence generator is a' speech-act, which includes an indication of which relations in the KB are relevant for expression at this point. Other information in the input specification helps tailor the expression of the content, such as an indicator of which KB element to use as the head of the generated form, which is theme, which elements are recoverable and identifiable.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In taking this approach, WAG attempts to extend the degree to which surface forms can be constrained by semantic specification. In many sentence generation systems, direct specifications of grammatical choices or forms is often needed, or, in the case Of Penman, the user needs to include arcane inquiry preselections -interventions in the interstratal mapping component, perhaps more arcane than grammar-level intervention.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "By providing a more abstract form of representation, text-planners using WAG need less knowledge of grammatical forms, and can spend more of their efforts dealing with issues of text-planning. I say 'less' here because, although WAG has extended the level at which surface forms can be specified semantically, there are still gaps. To allow for this, WAG allows input specifications to directly constrain the surface generation, either by directly specifying the grammatical feature(s) a given unit must have, or alternatively, specifying grammatical defaults: grammatical features which will be preferred if there is a choice.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The advantages of WAG's input specification language are summarised below:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "1. Interactional Specification: By placing the proposition as a role of the speech-act, rather than visa-versa, WAG allows cleaner integration into systems intended for dialogic interaction. WAG's input specification also allows a wider range of specification of the speech-act type than used in Penman and other sentencegeneration systems.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "2. Ideational Specification: WAG allows two modes of expressing the KB -in one mode, each sentence specification is a self-contained specification, containing all the ideational information needed (the 'black-box' mode). In the other, a sentence specification contains only a pointer into the KB, allowing finer integration between text-planner and sentence realiser. The availability of both alternatives means that WAG can fit a wider range of generation environments.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": ". Textual Specification: WAG introduces a high level means of representing the textual status of information to be expressed. Following Grosz (1977/86), and Halliday & Matthiessen (1995) , I use the notion of textual spaces, partitionings of the ideation base, each of which shifts dynamically as the discourse unfolds. I have outlined:", |
| "cite_spans": [ |
| { |
| "start": 136, |
| "end": 156, |
| "text": "Grosz (1977/86), and", |
| "ref_id": null |
| }, |
| { |
| "start": 157, |
| "end": 167, |
| "text": "Halliday &", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 168, |
| "end": 186, |
| "text": "Matthiessen (1995)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 a relevance space: the information which is rhetorically relevant at the present point of the discourse;", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 a shared-entity space: the information which is part of the shared knowledge of the speaker and hearer.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "\u2022 a recoverability space: the information which has entered the discourse context, including the entities which have been mentioned up to this point in the discourse. Information in the recoverability space can be presumed, or pronominalised.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "While the WAG generator has only been under development for a few years, and by a single author, in many aspects it meets, and in some ways surpasses, the functionality and power of the Penman system, as discussed above. It is also easier to use, having been designed to be part of a Linguist's Workbencha tool aimed at linguists without programming skills.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "The main advantage of the Penman system over the WAG system is the extensive linguistic resources available. Penman comes with a large grammar and semantics of English (and other languages). WAG comes with a mediumsized grammar of English. 9 Penman also supports a wider range of multi-lingual processing. 9While the WAG system can work with the grammar and lexicons of the Nigel resources, the resources which map grammar and semantics in Nigel are in a form incompatible with WAG).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "5" |
| }, |
| { |
| "text": "5SeePattabhiraman & Cercone (1990) for a good computational treatment of relevance, and its relation to salience.6Various earlier linguists and computational linguists have also used the notion of 'spaces' to represent textual status, see for instance,Reichman (1978);Grimes (1982).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Upper Modeling: organizing knowledge for natural language processing", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Bateman", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proceedings of the Fifth International Workshop on Natural Language Generation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bateman, John .1990 \"Upper Modeling: organizing knowledge for natural language processing\", Pro- ceedings of the Fifth International Workshop on Natural Language Generation, June 1990, Pitts- burgh.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "A General Organisation of Knowledge for Natural Language Processing: the Penman Upper Model", |
| "authors": [ |
| { |
| "first": "John", |
| "middle": [], |
| "last": "Bateman", |
| "suffix": "" |
| }, |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "Kasper", |
| "suffix": "" |
| }, |
| { |
| "first": "Johanna Moore & Richard", |
| "middle": [], |
| "last": "Whitney", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "USC/Information Sciences Institute", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bateman, John, Robert Kasper, Johanna Moore & Richard Whitney 1990 \"A General Organisation of Knowledge for Natural Language Processing: the Penman Upper Model\", USC/Information Sciences Institute Technical Report.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Systemic linguistics and discourse analysis: a multi-layered approach to exchange structure", |
| "authors": [ |
| { |
| "first": "Margaret", |
| "middle": [], |
| "last": "Berry", |
| "suffix": "" |
| } |
| ], |
| "year": 1981, |
| "venue": "Studies in Discourse Analysis", |
| "volume": "", |
| "issue": "", |
| "pages": "120--145", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Berry, Margaret 1981 \"Systemic linguistics and dis- course analysis: a multi-layered approach to ex- change structure\" in Coulthard M. & Montgomery M. (eds.) Studies in Discourse Analysis, London: Boston-Henly: Routledge & Kegan Paul, 120-145.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Discourse Production: a computer model of some aspects o] a speaker", |
| "authors": [ |
| { |
| "first": "Anthony", |
| "middle": [], |
| "last": "Davey", |
| "suffix": "" |
| } |
| ], |
| "year": 1974, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Davey, Anthony 1974/1978 Discourse Production: a computer model of some aspects o] a speaker, Ed- inburgh University Press, Edinburgh, 1978. Pub- lished version of Ph.D. dissertation, University of Edinburgh, 1974.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "FUF: The Universal Unifier User Manual Version 5.0", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "Elhadad", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Elhadad, Michael 1991 \"FUF: The Universal Uni- fier User Manual Version 5.0\", Technical Report CUCS-038-91, Columbia University, New York, 1991.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Demonstration of GENESYS: a very large semantically based Systemic Functional Grammar", |
| "authors": [ |
| { |
| "first": "Robin", |
| "middle": [ |
| "P" |
| ], |
| "last": "Fawcett", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "-Gordon", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Tucker", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proceedings of the 13th Int. Con]. on Computational Linguistics (COLING '90)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Fawcett, Robin P. -Gordon H. Tucker (1990) \"Demonstration of GENESYS: a very large se- mantically based Systemic Functional Grammar\". In Proceedings of the 13th Int. Con]. on Computa- tional Linguistics (COLING '90).", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "Reference Spaces in Text", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [ |
| "E" |
| ], |
| "last": "Grimes", |
| "suffix": "" |
| } |
| ], |
| "year": 1982, |
| "venue": "Proceedings of the 51st Nobel Symposium", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Grimes, J. E. 1982 \"Reference Spaces in Text\", in Proceedings of the 51st Nobel Symposium, Stock- holm.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "The Representation and Use of Focus in Dialog Understanding", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Grosz", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "Readings in Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Grosz, B. 1977/86 \"The Representation and Use of Focus in Dialog Understanding\", Technical Re- port 151, Artificial Intelligence Centre, SRI Inter- national, California. Reprinted in B.J. Grosz, K. Sparck-Jones, & B.L. Webber (eds.), Readings in Natural Language Processing, Morgan Kaufmann Publishers, Los Altos, CA, 1986.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Language as social semiotic. The social interpretation of language and meaning", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "A K" |
| ], |
| "last": "Halliday", |
| "suffix": "" |
| } |
| ], |
| "year": 1978, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Halliday, M.A.K. 1978 Language as social semiotic. The social interpretation of language and meaning. London: Edward Arnold.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Construing experience through meaning: a languagebased approach to cognition", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [ |
| "A K C I M" |
| ], |
| "last": "Halliday", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Matthiessen", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Halliday, M.A.K. & C.I.M. Matthiessen 1995 Con- struing experience through meaning: a language- based approach to cognition. Pinter: London.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "On the Generator Input of the Future", |
| "authors": [ |
| { |
| "first": "Eduard", |
| "middle": [], |
| "last": "Hovy", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "New Concepts in Natural Language Generation: Planning, Realisation and Systems", |
| "volume": "", |
| "issue": "", |
| "pages": "283--287", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hovy, Eduard 1993 \"On the Generator Input of the Future\", in Helmut Horacek & Michael Zock (eds.), New Concepts in Natural Language Genera- tion: Planning, Realisation and Systems, London: Pinter, p283-287.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "An Overview of the Penman Text Generation System", |
| "authors": [ |
| { |
| "first": "William", |
| "middle": [ |
| "C" |
| ], |
| "last": "Mann", |
| "suffix": "" |
| } |
| ], |
| "year": 1983, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mann, William C. 1983 \"An Overview of the Pen- man Text Generation System \", USC/ISI Technical Report RR-84-127.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Demonstration of the Nigel Text Generation Computer Program", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [ |
| "C C I M" |
| ], |
| "last": "Mann", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "Systemic Perspectives on Discourse", |
| "volume": "1", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Mann, W. C. & C. I. M. M:~tthiessen 1985 \"Demon- stration of the Nigel Text Generation Computer Program\", In Benson and Greaves (eds.), Systemic Perspectives on Discourse, Volume 1. Norwood: Ablex.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Text: system and structure. Amsterdam: Benjamins", |
| "authors": [ |
| { |
| "first": "James", |
| "middle": [ |
| "R" |
| ], |
| "last": "Martin", |
| "suffix": "" |
| } |
| ], |
| "year": 1992, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Martin, James R. 1992English Text: system and structure. Amsterdam: Benjamins.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Language Production as a Process of Decision-making under Constraints, MIT Ph.D. Dissertation", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Mcdonald", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "McDonald, D. 1980 Language Production as a Pro- cess of Decision-making under Constraints, MIT Ph.D. Dissertation, 1980. MIT Report.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Mumble-86: Design and Implementation", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Meteer", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Mcdonald", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Anderson", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Forster", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Gay", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Huettner", |
| "suffix": "" |
| }, |
| { |
| "first": "&", |
| "middle": [ |
| "P" |
| ], |
| "last": "Sibun", |
| "suffix": "" |
| } |
| ], |
| "year": 1987, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Meteer, M., D. McDonald, S. Anderson, D. Forster, L. Gay, A. Huettner, & P. Sibun. 1987 \"Mumble- 86: Design and Implementation\", COINS Technical Report 87-87, University of Massachusetts at Am- herst, Computer and Information Science.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Planning Text for Advisory Dialogues: Capturing Intentional and Rhetorical Information", |
| "authors": [ |
| { |
| "first": "Johanna & Cecile", |
| "middle": [], |
| "last": "Moore", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Paris", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Computational Linguistics", |
| "volume": "19", |
| "issue": "4", |
| "pages": "651--694", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Moore, Johanna & CEcile Paris 1993 \"Planning Text for Advisory Dialogues: Capturing Inten- tional and Rhetorical Information.\" Computational Linguistics Vol 19, No 4, pp651-694, 1993.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "A Dynamic Model of Exchange", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "O'donnell", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "in Word", |
| "volume": "41", |
| "issue": "3", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "O'Donnell, Michael 1990 \"A Dynamic Model of Ex- change\" in Word, vol. 41, no. 3 Dec. 1990", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "O'Donnell, Michael 1994 Sentence Analysis and Generation -A Systemic Perspective", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "O'donnell", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "Proceedings of the Third International Workshop on Parsing Technologies", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "O'Donnell, Michael 1993 \"Reducing Complexity in a Systemic Parser \", in Proceedings of the Third International Workshop on Parsing Technologies, Tilburg, the Netherlands, August 10-13, 1993. O'Donnell, Michael 1994 Sentence Analysis and Generation -A Systemic Perspective. Ph.D., De- partment of Linguistics, University of Sydney.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "From Corpus to Codings: Semi-Automating the Acquisition of Linguistic Features", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "O'donnell", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings of the AAAI Spring Symposium on Empirical Methods in Discourse Interpretation and Generation", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "O'Donnell, Michael 1995a \"From Corpus to Cod- ings: Semi-Automating the Acquisition of Lin- guistic Features\", in Proceedings of the AAAI Spring Symposium on Empirical Methods in Dis- course Interpretation and Generation, Stanford University, California, March 27 -29.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Sentence Generation Using the Systemic WorkBench", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [], |
| "last": "O'donnell", |
| "suffix": "" |
| } |
| ], |
| "year": 1995, |
| "venue": "Proceedings of the Fifth European Workshop on Natural Language Generation", |
| "volume": "", |
| "issue": "", |
| "pages": "235--238", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "O'Donnell, Michael 1995b \"Sentence Generation Using the Systemic WorkBench\", in Proceedings of the Fifth European Workshop on Natural Language Generation, 20-22 May, Leiden, The Netherlands, pp 235-238.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "User Modelling in Text Generation", |
| "authors": [ |
| { |
| "first": "C~cile", |
| "middle": [], |
| "last": "Paris", |
| "suffix": "" |
| } |
| ], |
| "year": 1993, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Paris,C~cile 1993 User Modelling in Text Genera- tion, London & New York: Pinter.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Selection: Salience, Relevance and the Coupling between Domain-Level Tasks and Text Planning", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Pattabhiraman", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Nick Cercone", |
| "suffix": "" |
| } |
| ], |
| "year": 1990, |
| "venue": "Proceedings of the 5th International Workshop on Natural Language Generation", |
| "volume": "", |
| "issue": "", |
| "pages": "3--6", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Pattabhiraman, T. & Nick Cercone 1990 \"Se- lection: Salience, Relevance and the Coupling between Domain-Level Tasks and Text Planning\", in Proceedings of the 5th International Workshop on Natural Language Generation, 3-6 June, 1990, Dawson, Pennsylvania.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Systemic text generation as problem solving", |
| "authors": [ |
| { |
| "first": "Terry", |
| "middle": [], |
| "last": "Patten", |
| "suffix": "" |
| } |
| ], |
| "year": 1988, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Patten, Terry 1988 Systemic text generation as problem solving, Cambridge: Cambridge University Press.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "Conversational Coherency", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Reichman", |
| "suffix": "" |
| } |
| ], |
| "year": 1978, |
| "venue": "Cognitive Science", |
| "volume": "2", |
| "issue": "", |
| "pages": "283--327", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Reichman, R. 1978 \"Conversational Coherency\", Cognitive Science 2, pp283-327.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Speech-Act Representation \u2022 U s e r M o d e l l i n g : In theory, the Speaker and Hearer fields are available for usermodelling purposes (cf.", |
| "num": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "N e g o t i a t o r y vs. S a l u t o r y : negotiatory speech-acts contribute towards the construction of an ideational proposition. while salutory moves do not, rather speechThe Speech-Act Network serving a phatic function, for instance, greetings, farewells, and thank-yous. \u2022 Speech Function: The speech-function is the speaker's indication of what they want the hearer to do with the utterance.", |
| "num": null |
| }, |
| "FIGREF3": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "The Upper Model\u2022 Role Information: a specification of the roles of the entity, and of the entities which fill these roles.", |
| "num": null |
| }, |
| "FIGREF4": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "Building a Knowledge-Base 3.2.1 Assertion of Knowledge into KBFigure 4shows the forms which assert some knowledge about John and Mary into the KB. The information basically tells that Mary left a party because John arrived at the party, tell is a lisp macro form used to assert knowledge into the KB.", |
| "num": null |
| }, |
| "FIGREF5": { |
| "type_str": "figure", |
| "uris": null, |
| "text": ") ) : identifiable-entities (John Mary)) => Mary left because John arrived.", |
| "num": null |
| }, |
| "FIGREF6": { |
| "type_str": "figure", |
| "uris": null, |
| "text": "entities (John lMary)) => John's arrival caused Mary ~to leave.", |
| "num": null |
| } |
| } |
| } |
| } |