| { |
| "paper_id": "C80-1021", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T13:05:48.943970Z" |
| }, |
| "title": "", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Wahlster", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "Germanisches Seminar Universitaet Hamburg Von-Melle", |
| "location": { |
| "addrLine": "Park 6", |
| "postCode": "D-2000", |
| "settlement": "Hamburg 13", |
| "country": "Federal Republic of Germany" |
| } |
| }, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper discusses aspects of a computational model for the semantics of why-questions which are relevant to the implementation of an explanation component in a natural language dialogue system. After a brief survey of all of the explanation components which have been implemented to date, some of the distinguishing features of the explanation component designed and implemented by the author are listed. In the first part of the paper the major types of signals which, like the word whV, can be used to set the explanation component into action are listed, and some ways of recognizing them automatically are considered. In addition to these linguistic signals, communicative and cognitive conditions which can have the same effect are discussed. In the second part the various schemata.for argumentative dialogue sequences which can be handled by the explanation component in question are examined, Particular attention is paid to problems arising in connection with the iteration of why-questions and the verbalization of multiple justifications. Finally schemata for metacommunicative why-questions and for why-questions asked by the user are investigated.", |
| "pdf_parse": { |
| "paper_id": "C80-1021", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper discusses aspects of a computational model for the semantics of why-questions which are relevant to the implementation of an explanation component in a natural language dialogue system. After a brief survey of all of the explanation components which have been implemented to date, some of the distinguishing features of the explanation component designed and implemented by the author are listed. In the first part of the paper the major types of signals which, like the word whV, can be used to set the explanation component into action are listed, and some ways of recognizing them automatically are considered. In addition to these linguistic signals, communicative and cognitive conditions which can have the same effect are discussed. In the second part the various schemata.for argumentative dialogue sequences which can be handled by the explanation component in question are examined, Particular attention is paid to problems arising in connection with the iteration of why-questions and the verbalization of multiple justifications. Finally schemata for metacommunicative why-questions and for why-questions asked by the user are investigated.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "The explanation component of a natural language AI system is that component whose job it is to generate, in response to a why-question an explanation which is both understandable to the user and appropriate to the current state of the dialogue.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "Although there has been relatively little research into the semantics and pragmatics of why-questions1,5,9, 17 and the cognitive processes underlying the answering of them, several AI systems do exist which are capable of handling certain types of why-questions. The practical value of the incorporation of an explanation component lies essentially in the fact that, as Stallman and Sussman have put it, '~such programs are more convincing when right and easier to debug when wrong\".~5 Figure I provides an overview and comparison of the explanation components which have been implemented to date: BLAH 22, DIGITALIS ADVISOR 16, EL Is, EXPOUND ~, HAM-RPMI~, 21, LUIG113, MYCIN12, ~, NOAH 11, PROSPECTOR 7, SHRDLU ~, TKP2 I\u00b0 (The symbol \"-\" signifies that the attribute in question is not applicable to the given system).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 486, |
| "end": 494, |
| "text": "Figure I", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "This paper presents some results of my experience in designing and implementing an explanation com-ponentS1; together, they represent a step toward a computational model for the semantics of whyquestions. The explanation component was designed as a module which could in principle be incorporated into any natural language AI system. It has been tested within the natural language dialogue system HAM-RPM 6, which converses with a human partner in colloquial German about limited but interchangeable scenes.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "In implementing HAM-RPM we have taken into account the human ability to deduce useful information even in the case of fuzzy knowledge by approximate reasoning. The model of fuzzy reasoning used in HAM-RPM can be characterized by the following four properties2\u00b0: (a) A fuzzy inference rule represents a weak implication; a particular 'implication strength' must thus be associated with each such rule. (b) The premises of a fuzzy inference rule are often fulfilled only to a certain degree. (c) The applicability of a fuzzy inference rule in the derivation of a particular conclusion is likewise a matter of degree. (d) Several mutually independent fuzzy inference rules can corroborate each other in the derivation of a particular conclusion.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "The explanation component which I have developed differs from BLAH 22, one of the most advanced explanation components which have similar goals, in that on the one hand fuzzy inference rules and facts can be modified by appropriate hedges (in accordance with (a) through (c) above), and on the other hand the system is able in the course of a dialogue to generate multiple justifications for an explanandum (in accordance with (d) above). A further important difference between this explanation component and the other systems included in Figure I is that the system is equipped with a fairly sophisticated natural language generator, which is ATN-based and includes algorithms for generating pronouns and definite descriptions 19.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 539, |
| "end": 547, |
| "text": "Figure I", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "Only two aspects of this explanation component will be discussed in this paper: The signals on the basis of which the explanation component generates an argumentative answer to a question asked by the user and the speech act schemata for the argumentative dialogue sequences which can be The purpose of the present section is to list the major types of signals which are capable of setting an explanation component into action. The resulting classification of linguistic expressions does not, of course, imply that all of the expressions in a given category are completely synonymous.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": null |
| }, |
| { |
| "text": "From the point of view of algorithmic recognition, the simplest case is that in which the user elicits an argumentative answer from the system by asking a direct question. The word why can often be interpreted as a signal for an argumentative answer. On the other hand, its exact meaning depends on the dialogue context and it can be used within speech acts which have nothing to do with explanation, such as making a suggestion or a comment 5. In spite of its ambiguity, the word why represents the only means of eliciting an argumentative answer in most AI systems which have an explanation component.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Signals for Argumentative Answers in the User's Utterances", |
| "sec_num": null |
| }, |
| { |
| "text": "Special idiomatic expressions such as those listed in (LI) can have the same function as the word why. Indirect questions such as those in (LI) require that the system be able to divide the utterance into matrix sentence and embedded sentence syntactically; only then can it process the latter using the same means as in the case of direct questions containing why or the questions in (L1).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Signals for Argumentative Answers in the User's Utterances", |
| "sec_num": null |
| }, |
| { |
| "text": "Further types of signals include direct (see LJ) and indirect (see L4) requests. The problem of", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(L2) Please tell me why A, I'd like to know why A", |
| "sec_num": null |
| }, |
| { |
| "text": "how indirect speech acts such as the requests in (L4) can be recognized automatically is one which has recently been attracting much attention from natural language AI researchersJ, 8", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "The word why and the expressions in (LI) needn't accompany the proposition to be explained within a single utterance, as they do in the example (El); they can also be used alone after the system has answered a question to elicit an explanation of the answer (cf. E2). The expressions in (LJ) and (L4) can also be used to achieve just the opposite: An argumentative answer is requested in advance, before the corresponding question has been asked of the system. the truth of the answer.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "A kind of signal which suggests an argumentative answer in a still more obvious manner is the category of utterances by the user which indicate an opinion contrary to that expressed by the system (cf. L6). The idiomatic expressions in (L5) and (L6) which always express doubt or a contrary opinion no matter what the current dialogue context may be, can be handled adequately if information concerning their implications is stored in the system's 'idiom lexicon '6.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "A further way in which the user can indirectly ask a why-question is by himself suggesting an explanation of what the system has just asserted, while at the same time indicating a desire to have the explanation confirmed by the system. For example, after the system has given the answer (E2. is to be realized in a dialogue system, the program must be able to recognize (E2.6) as a proposed explanation. Algorithms which recognize explanations in certain contexts have been developed, e.g., for the ICAI system ACE TM and the text-understanding system PAM 23.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "Leading and rhetorical questions which suggest an affirmative answer may be seen as containing an implicit request to justify the answer if it is negative. is not something like (E3.2), but rather simply Yes, in fact I am, the system isn't exhibiting the sort of cooperative behavior which we would like to have in a natural language dialogue system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "These last two types of speech acts cannot at present be handled adequately by AI systems. The same is true of explanations within the schema reproach-justification (cf. E4.1 and E4.2). Communicative and Cognitive Conditions as Signals for Arj'umenEatliv ~ Answers Two further kinds of signals which suggest argumentative answers deserve mention in this section. In contrast to the preceding types they can be incorporated without difficulty into existing AI systems, e.g. HAM-RPM 21.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "Both kinds of signal lead to the question s being oucr-~we2..\u00a3d in that they suggest an argumentative answer in the absence of any explicit or implicit request for such an answer in the user's question.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "On the one hand, the system may offer an unsolicited explanation for reasons of p(z.,utneA tae2./\u00a3~ if it has already noticed that the user seems to have a tendency to ask for explanations of answers 6.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "On the other hand, over-answering may even be reasonably expected of the system in the case where the answer is based on uncertain beliefs and approximate or hypothetical reasoning. This kind of behavior can be modelled to a limited extent if the system is programmed so as to attempt to generate an explanation as soon as its confidence in its own answer sinks below a certain threshold, e.g., because the implication strength (see (a) above) of one of the inference rules it has used is low (cf. E5.1, E5. generation of an argumentative answer in such a context falls outside the usual scope of linguistic analysis; it is a good example of an application of the AI paradigm in that the condition which gives rise to the generation of an argumentative answer is a certain property of a cognitive process, namely the inference process by which the answer is derived. of an u./_-t/mcc.tC cxpZ~noJCio~. Though it is debatable whether ultimate explanations in a philosophical sense are in fact possible, it is clear that participants in everyday dialogues frequently offer explanations which they are not in a position to explain further. Some typical formulations which are used in such cases are listed in (L7).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(LJ) Please explain why A, prove that A (L4) I'd be interested in hearing why you think that A, Are you prepared to justify your conclusion that A?", |
| "sec_num": null |
| }, |
| { |
| "text": "The Ambiguity of Iterated Why-questions A further problem in connection with iterated why-questions is the ambiguity which they regularly involve. Each of the why-questions after the first one can refer either to (a) the assertion which constituted the explanans, or (b) the inferential relationship between the explanans and the explanandum.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(L7) It's obvious, That's the way it is, Can't you see it?", |
| "sec_num": null |
| }, |
| { |
| "text": "($3.1) U: Why Q? ($3.2) S: Because P.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "(L7) It's obvious, That's the way it is, Can't you see it?", |
| "sec_num": null |
| }, |
| { |
| "text": "Why P? It is of course possible to eliminate this ambiguity with a more precise formulation of the whyquestion, as when, for example, ($2.1) is replaced with ($2.1').", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "J %", |
| "sec_num": null |
| }, |
| { |
| "text": "(S2.1') U: I know that. But why does that make you think that Q7", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "J %", |
| "sec_num": null |
| }, |
| { |
| "text": "Although interpretation (a) is far more common than (b) in nontechnical dialogues, the occurrence of questions such as ($2.1') shows that it is nonetheless worthwhile to provide an AI system with the ability to answer in accordance with either of the possible interpretations. For interpretation (b), this means that the system must be able,'like HAM-RPM al, to verbalize the inference rules it uses. Jf the system is requested, via a further why-question, to explain an inference rule that it has verbalized in this way, the existence of a third type of argument in addition to the presentation of factual evidence and the verbalisation of inference rules becomes evident: The system may supply a bacl./n9 Is for its inference rule. A backing usually refers to a convention, a theory, or observations.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "J %", |
| "sec_num": null |
| }, |
| { |
| "text": "An explanation component which uses backings must have access to the corresponding metaknowledge about its inference rules.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "J %", |
| "sec_num": null |
| }, |
| { |
| "text": "A further variant of schema 2 can be used to exhibit the step-by-step elicitation of a multiple justification. Instead of simply asking another why-question, the user specifically requests further corroborating evidence for the explanandum. Some typical expressions are listed in (L8). .Djalo@ue Schemata with Metacommunicative Why-quations in all of the dialogue schemata we have examined so far, a why-question asked by the user followed an answer by the system to a previous question.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Elicitation of a Multiple dustific@tion", |
| "sec_num": null |
| }, |
| { |
| "text": "In this section we shall discuss dialogue sequences in which why-questions refer to questions or requests. In fact, of course, any kind of speech act, e.g. a threat or an insult, can give rise to a metacommunicative why-question; the two types to be discussed here are those most relevant to foreseeable applications of natural language AI systems. Schema 5 will serve as a starting point. In clarification dialogues schema 6,a variant of schema 5, can be instantiated. ", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Elicitation of a Multiple dustific@tion", |
| "sec_num": null |
| }, |
| { |
| "text": "Although all of the why-questions considered so far have been asked by the user, the system can also ask why the user has made a particular input\u2022 This situation is described by schema 5 except that the roles of USER (U) and SYSTEM (S) are reversed\u2022 Providing an application-oriented AI system with the ability to ask such why-questions is worthwhile because there are many situations in which the system requires further information about the user's intention to guide its search for an answer or to help to formulate its answer in a communicatively adequate manner\u2022 Of course, the system can only make use of the user's answer to such a why-question if it is equipped with the ability to analyse argumentative answers. The example (E8) might occur in one of HAM-RPM's dialogue situations, in which the system simulates a hotel manager who is anxious to rent a particular room to a caller who is inquiring about it. It illustrates the way information about the dialogue partner's intentions can influence the way a particular state of affairs is described. The schemata we have investigated in this and the previous sections can also be embedded in one another, as can be seen from schema 7. In this schema, (S7.4),but not ($7.3), is a metacommunicative why-question. In mixed-Z~.).2J.o.~..i_u\u00a2 systems, in which either of the partners can initiate a dialogue sequence, the system must be able both to ask and to answer why-questions, including those of a metacommunicative nature.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Why-questions Asked by the System", |
| "sec_num": null |
| }, |
| { |
| "text": "Summary and Integration of All Argumentative Dialogue Schemata Relevant to AI Systems Figure 3 summarizes and integrates the schemata for argumentative dialogue sequences discussed in the preceding sections. The arrows joining the rectangles indicate that one speech act follows another in time. If arrows join two rectangles in both directions, loops such as those discussed in connection with iterated why-questions are possible. Double vertical lines on the left-or right-hand side of a rectangle indicate that the speech act in question can be the first or the last speech act in a sequence, respectively. The system's criteria for recognizing at each point which of the possible speech acts the user has performed and for selecting its own speech acts are not included in the diagram. If one extends Figure 3 by permitting the reversal of the roles of system and user, all schemata for argumentative dialogue sequences 21 are included which are relevant for foreseeable applications in dialogue systems with mixed-initiative.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 86, |
| "end": 94, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| }, |
| { |
| "start": 805, |
| "end": 813, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Why-questions Asked by the System", |
| "sec_num": null |
| }, |
| { |
| "text": "A non-compiled version of HAM-RPM is running on the DECsystem 1070 (PDP-10) of the Fachbereich fuer Informatik of the University of Hamburg under the TOPSI0 operating system. Comprising approximately 600 LISP/FUZZY procedures, the current version occupies 150K of 36-bit words and requires from one to Fifteen seconds for a response.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Technical Data", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "I wish to thank Anthony Jameson for careful reading of an earlier draft of this paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Acknowledgements", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Why-questions", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Bromberger", |
| "suffix": "" |
| } |
| ], |
| "year": 1966, |
| "venue": "Mind and Cosmos", |
| "volume": "", |
| "issue": "", |
| "pages": "86--111", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Bromberger, S. (1966): Why-questions. In: Colodny, R. (ed.): Mind and Cosmos. Pittsburgh: Univ. Press, p. 86-111", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "The translation of formal proofs into English", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Chester", |
| "suffix": "" |
| } |
| ], |
| "year": 1976, |
| "venue": "Artificial Intelligence", |
| "volume": "7", |
| "issue": "3", |
| "pages": "261--278", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Chester, D. (1976): The translation of for- mal proofs into English. In: Artifi- cial Intelligence, 7, 3, p. 261-278", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "On knowing what to say: Planning speech acts. Univ. of Toronto", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [ |
| "R" |
| ], |
| "last": "Cohen", |
| "suffix": "" |
| } |
| ], |
| "year": 1978, |
| "venue": "Dept. of Computer Science", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Cohen, P.R. (1978): On knowing what to say: Planning speech acts. Univ. of Toron- to, Dept. of Computer Science, Tech- nical Report No. 118", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Applications of meta level knowledge to the construction, maintenance and use of large knowledge bases", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Davis", |
| "suffix": "" |
| } |
| ], |
| "year": 1976, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Davis, R. (1976): Applications of meta level knowledge to the construction, main- tenance and use of large knowledge bases. Stanford Univ., Technical Re- port STAN-CS-76-562", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "A pragmatic analysis of tenseless why-questions", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Freeman", |
| "suffix": "" |
| } |
| ], |
| "year": 1976, |
| "venue": "Papers of the twelfth regional meeting of the Chicago Linguistic Society", |
| "volume": "", |
| "issue": "", |
| "pages": "208--219", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Freeman, C. (1976): A pragmatic analysis of tenseless why-questions. In: Mufwene, S.S., Walker, C.A., Steeuer, S.B. (eds.): Papers of the twelfth region- al meeting of the Chicago Linguistic Society. Chicago: Chicago Linguistic Society, p. 208-219", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "The anatomy of the natural language dialogue system HAM-RPM", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Hahn", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Hoeppner", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Jameson", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Wahlster", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "v. Hahn, W., Hoeppner, W., Jameson, A., Wahlster, W. (1980): The anatomy of the natural language dialogue system HAM-RPM. In: Bolc, L. (ed.): Natural language based computer systems. Munich: Hanser/Macmillan", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "PROSPECTOR -A computer-based consultation system for mineral exploration", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [ |
| "E" |
| ], |
| "last": "Hart", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [ |
| "O" |
| ], |
| "last": "Duda", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "Technical Note", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hart, P.E., Duda, R.O. (1977): PROSPECTOR - A computer-based consultation system for mineral exploration. Stanford Research International, AI Center, Technical Note 155", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "An anatomy of graceful interaction in spoken and written man-machine communication", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Hayes", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Reddy", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Hayes, P., Reddy, R. (1979): An anatomy of graceful interaction in spoken and written man-machine communication. Carnegie-Mellon-Univ., Dept. of Com- puter Science, Technical Report CMU- CS-79-144", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Praktische Semantik", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [ |
| "J" |
| ], |
| "last": "Heringer", |
| "suffix": "" |
| } |
| ], |
| "year": 1974, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Heringer, H.J. (1974): Praktische Semantik. Stuttgart: Klett", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "An automatic theorem prover generating a proof in natural language", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Nakamishi", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Nagata", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Ueda", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "IJCAI-79", |
| "volume": "", |
| "issue": "", |
| "pages": "636--638", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Nakamishi, M., Nagata, M., Ueda, K. (1979): An automatic theorem prover genera- ting a proof in natural language. In: IJCAI-79, Tokyo, p. 636-638", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "A structure for plans and behavior", |
| "authors": [ |
| { |
| "first": "E", |
| "middle": [ |
| "D" |
| ], |
| "last": "Sacerdoti", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sacerdoti, E.D. (1977): A structure for plans and behavior. N.Y.: Elsevier", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "Explanation capabilities of production-based consultation systems", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "A" |
| ], |
| "last": "Scott", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Clancey", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Davis", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [ |
| "H" |
| ], |
| "last": "Shortliffe", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "American Journal of Computational Linguistics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Scott, C.A., Clancey, A., Davis, R., Shortliffe, E.H. (1977): Explanation capabilities of production-based con- sultation systems. In: American Jour- nal of Computational Linguistics, Microfiche 62", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "Answering questions about processes", |
| "authors": [ |
| { |
| "first": "G", |
| "middle": [ |
| "W" |
| ], |
| "last": "Scragg", |
| "suffix": "" |
| } |
| ], |
| "year": 1974, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Scragg, G.W. (1974): Answering questions about processes. Univ. of California, San Diego, Ph.D. Thesis", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "ACE: A system which analyses complex explanations", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [ |
| "H" |
| ], |
| "last": "Sleeman", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [ |
| "J" |
| ], |
| "last": "Hendley", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "International Journal of Man-Machine Studies", |
| "volume": "11", |
| "issue": "", |
| "pages": "125--144", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Sleeman, D.H., Hendley, R.J. (1979): ACE: A system which analyses complex ex- planations. In: International Journal of Man-Machine Studies, 11, p. 125-144", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Forward reasoning and dependency-directed backtracking in a system for computer-aided circuit analysis", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [ |
| "M" |
| ], |
| "last": "Stallman", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "J" |
| ], |
| "last": "Sussman", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "Artificial Intelligence", |
| "volume": "9", |
| "issue": "", |
| "pages": "135--196", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Stallman, R.M., Sussman, G.J. (1977): For- ward reasoning and dependency-direc- ted backtracking in a system for com- puter-aided circuit analysis. In: Artificial Intelligence 9, P. 135-196", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "A Digitalis therapy advisor with explanations. MIT Lab. for Computer Science", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [ |
| "R" |
| ], |
| "last": "Swartout", |
| "suffix": "" |
| } |
| ], |
| "year": 1977, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Swartout, W.R. (1977): A Digitalis therapy advisor with explanations. MIT Lab. for Computer Science, Technical Re- port TR-176", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Semantics of the question in a problem-solving situation", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Tondl", |
| "suffix": "" |
| } |
| ], |
| "year": 1969, |
| "venue": "Czech Academy of Science", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tondl, L. (1969): Semantics of the question in a problem-solving situation. Czech Academy of Science, Prague", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "The uses of argument", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Tou]min", |
| "suffix": "" |
| } |
| ], |
| "year": 1969, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Tou]min, S. (1969): The uses of argument, Cambridge: Univ. Press (Ist ed. 1958)", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "Glancing, referring and explaning in the dialogue system HAM-RPM", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Wahlster", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Jameson", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Hoeppner", |
| "suffix": "" |
| } |
| ], |
| "year": 1978, |
| "venue": "", |
| "volume": "77", |
| "issue": "", |
| "pages": "53--67", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wahlster, W., Jameson, A., Hoeppner, W. (1978): Glancing, referring and ex- planing in the dialogue system HAM-RPM. In: American Journal of Computational Linguistics, Microfiche 77, P. 53-67", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "Implementing fuzziness in dialogue systems", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Wahlster", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "Empirical Semantics", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wahlster, W. (1980): Implementing fuzziness in dialogue systems. In: Rieger, B. (ed.): Empirical Semantics. Bochum: Brockmeyer", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "Automatic generation of natural language explanations for conclusions based on fuzzy inferences", |
| "authors": [ |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Wahlster", |
| "suffix": "" |
| } |
| ], |
| "year": 1980, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wahlster, W. (1980): Automatic generation of natural language explanations for con- clusions based on fuzzy inferences. (in preparation)", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "The structure of natural explanation: Theory and application. System Development Corporation", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [ |
| "L" |
| ], |
| "last": "Weiner", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Weiner, J.L. (1979): The structure of natural explanation: Theory and application. System Development Corporation, Santa Monica, Technical Report SP-4035", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Why John maried Mary: Understanding stories involving recurring goals", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Wilensky", |
| "suffix": "" |
| } |
| ], |
| "year": 1978, |
| "venue": "Cognitive Science", |
| "volume": "2", |
| "issue": "", |
| "pages": "235--266", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Wilensky, R. (1978): Why John maried Mary: Understanding stories involving re- curring goals. In: Cognitive Science, 2, p. 235-266", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "Understanding natural language", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Winograd", |
| "suffix": "" |
| } |
| ], |
| "year": 1972, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Winograd, T. (1972): Understanding natural language. N.Y.: Academic", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF1": { |
| "uris": null, |
| "num": null, |
| "text": "Comparison of all explanation components implemented to date --145--systemls simulated actions theorems realized in the system. A Formal Description of the Signals Suggesting an Argumentative Answer", |
| "type_str": "figure" |
| }, |
| "FIGREF2": { |
| "uris": null, |
| "num": null, |
| "text": "In the system HAM-RPM expressions like (LI) How come, what ... for, how do you know these are recognized through pattern matching during lexical analysis 6.", |
| "type_str": "figure" |
| }, |
| "FIGREF3": { |
| "uris": null, |
| "num": null, |
| "text": "(El) USER (U): Why is Glenbrook Drive closed? (E2.1) USER (U) : Is Olenbrook Drive closed? (E2.2) SYSTEM (S): Yes. (E2.3) USER (U): Hew do you explain that?", |
| "type_str": "figure" |
| }, |
| "FIGREF4": { |
| "uris": null, |
| "num": null, |
| "text": "2), the user should be able, by asking the question (E2.6), to elicit an explanation like (E2.7) from the system. If this kind of behavior (E2.6) U: Because of an accident? (\u00a32.7) S: No, because they're repaying it.", |
| "type_str": "figure" |
| }, |
| "FIGREF5": { |
| "uris": null, |
| "num": null, |
| "text": "If the system's answer to (EJ.I) (E3.1) U: You aren't going to restrict me to 40k of core today again, are you? (\u00a33.2) S: Yes, in fact I am. I've got 47 jobs logged-in in the moment.", |
| "type_str": "figure" |
| }, |
| "FIGREF6": { |
| "uris": null, |
| "num": null, |
| "text": "E4.1) U: You erased my file COLING. TMP# (E4.2) S: Yeah, your log-out quota was exceeded.", |
| "type_str": "figure" |
| }, |
| "FIGREF7": { |
| "uris": null, |
| "num": null, |
| "text": "2)\u2022 The (E5\u2022I) U: I wonder if the Mercedes is cheap. (E5.2) S: I imagine so --.it's pretty old and rusty.", |
| "type_str": "figure" |
| }, |
| "FIGREF8": { |
| "uris": null, |
| "num": null, |
| "text": "summarizes the various signals for argumentative answers which have been discussed in this section (types of signals which have been implemented in HAM-RPM's explanation component are indicated by a *).", |
| "type_str": "figure" |
| }, |
| "FIGREF9": { |
| "uris": null, |
| "num": null, |
| "text": "doubt in user * \u2022 evidence of a contrary opinion in user * \u2022 inadequate explanation suggested by user \u2022 unexpected answer to a leading or rhetorical question \u2022 evidence of reproach in user \u2022 /partner-tactics * \"\u00b0ver-answer~ng\"-~--uncertainty about own answer * Signals which can elicit an argumentative answer", |
| "type_str": "figure" |
| }, |
| "FIGREF10": { |
| "uris": null, |
| "num": null, |
| "text": "Iterated Why-questions and Ultimate ExplanationsA sequence such as (E6.1) through (E6.4) may be continued by one or more repititions of schema 2, in which the user requests that the system's argumentative answer itself be explained.($2.1) U: Why? ($2.2) S: <argumentative answer> Schema 2: Iteration of a why-question The dialogue sequence (E6.5) through (E6.8) is a continuation of (E6) in which two further whyquestions occur. The answer (E6.8) is an example (E6.5) U: Why? (E6.6) S: It's in need of repairs because its rear axle is bent. (\u00a36.7) U: How come? (E6.8) S: That's just the way it is.", |
| "type_str": "figure" |
| }, |
| "FIGREF11": { |
| "uris": null, |
| "num": null, |
| "text": "The ambiguity of an iterated why-question If the second sort of interpretation is applied to the question (E6.7), an answer such as (E6.9) becomes appropriate.(E6.9) S: A machine is in need of repairs when one of its parts is in need of repairs.", |
| "type_str": "figure" |
| }, |
| "FIGREF12": { |
| "uris": null, |
| "num": null, |
| "text": ".1) S: <question>,<request> (55.2) U: Why? (55.3) S: <argumentative answer> ($5.4) U: <response to S5.1> interpretation of $5.2 by S: What was the intention underlying the speech act in $5.1? Schema 5: A dialogue sequence with a metacommunicative why-question ($6.1) U: <question> ($6.2) S: <clarification question concerning S6.1>,<request for a paraphrase of $6.1> ($6.3) U: Why? ($6.4) S: <argumentative answer> (S6.5) U: <response to S6.2> (S6.6) S: <response to $6.1> Schema 6: A metacommunicative why-question within a clarification dialogue Here is a dialogue sequence containing a metacommunicative why-question asked by the user: (E7.1) U: Please list all articles since 1978 on the subject of 'presposition'. (E7.2) S: Do you really mean 'presposition'? (E7.3) U: Why do you ask? (E7.4) S: I don't know this word. (E7.5) U: I meant 'presupposition' (E7.6) S: I have the following entries: ...", |
| "type_str": "figure" |
| }, |
| "FIGREF13": { |
| "uris": null, |
| "num": null, |
| "text": "E8.1) U: Has the room got a big desk? (E8.2) S: Why do you ask? (E8.3) U: Because I've got a lot of work to do. (E8.4) S: Yes, the desk is fairly large. (E8.3') U: I hate big desks. (E8.4') S: It isn't particularly big.", |
| "type_str": "figure" |
| }, |
| "FIGREF14": { |
| "uris": null, |
| "num": null, |
| "text": ".1) U: <yes-no-question> ($7.2) S: <affirmative answer> (with restricting hedge) (S7.3) U: Why? ($7.4) S: Why do you ask? ($7.5) U: <argumentative answer to $7.4> ($7.6) S: <argumentative answer to $7.3> Schema 7: Successive why-questions of different types", |
| "type_str": "figure" |
| } |
| } |
| } |
| } |