| { |
| "paper_id": "J81-1008", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T02:43:02.882948Z" |
| }, |
| "title": "The Transformational Question Answering (TQA) System: Description, Operating Experience, and Implications", |
| "authors": [ |
| { |
| "first": "Robert", |
| "middle": [], |
| "last": "Wilensky", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Roger", |
| "middle": [ |
| "C" |
| ], |
| "last": "Schank", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Pat", |
| "middle": [], |
| "last": "Hayes", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Mrs", |
| "middle": [ |
| "A C Toni" |
| ], |
| "last": "Shelter", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Fred", |
| "middle": [ |
| "J" |
| ], |
| "last": "Damerau", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Janet", |
| "middle": [ |
| "L" |
| ], |
| "last": "Kolodner", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Michael", |
| "middle": [ |
| "G" |
| ], |
| "last": "Dyer", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| }, |
| { |
| "first": "Wendy", |
| "middle": [ |
| "G" |
| ], |
| "last": "Lehnert", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper sketches the structure of the TQA system with some examples, discusses some of the results of running the system in a user site during the year 1978, and outlines a few of the implications of a natural language query system for larger and more diverse data bases. It is apparent that a number of problems remain to be solved, but it also seems likely that a useful system can be constructed for some domains.", |
| "pdf_parse": { |
| "paper_id": "J81-1008", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper sketches the structure of the TQA system with some examples, discusses some of the results of running the system in a user site during the year 1978, and outlines a few of the implications of a natural language query system for larger and more diverse data bases. It is apparent that a number of problems remain to be solved, but it also seems likely that a useful system can be constructed for some domains.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "ry retrieval. Emphasis is placed on human question answering abilities, and the heuristics needed to simulate these phenomena. An example narrative processed by BORIS is discussed in detail and used to illustrate design decisions. Cognitive Science 4, 3 (July-Sept. 1980), 209-241. In its earliest drafts, this paper was a structured argument, presenting a comprehensive view of cognitive science, criticizing prevailing approaches to the study of language and thought and advocating a new way of looking at things. Although I strongly believed in the approach it outlined, somehow it didn't have the convincingness on paper that it had in my own reflection. After some discouraging attempts at reorganization and rewriting, I realized that there was a mismatch between the nature of what I wanted to say and the form in which I was trying to communicate.", |
| "cite_spans": [ |
| { |
| "start": 241, |
| "end": 281, |
| "text": "Science 4, 3 (July-Sept. 1980), 209-241.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Computer", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "What Does it Mean to Understand Language? Terry Winograd", |
| "sec_num": null |
| }, |
| { |
| "text": "The understanding on which it was based does not have the form of a carefully structured framework into which all of cognitive science can be placed. It is more an orientation --a way of approaching the phenomena --that has grown out of many different experiences and influences and that bears the marks of its history. I found myself wanting to describe a path rather than justify its destination, finding that in the flow, the ideas came across more clearly. Since this collection was envisioned as a panorama of contrasting individual views, I have taken the liberty of making this chapter explicitly personal and describing the evolution of my own understanding. My interests have centered around natural language. I have been engaged in the design of computer programs that in some sense could be said to \"understand language,\" and this has led to looking at many aspects of the problems, including theories of meaning, representation formalisms, and the design and construction of complex computer systems. There has been a continuous evolution in my understanding of just what it means to say that a person or computer \"understands,\" and this story can be read as recounting that evolution. It is long, because it is still too early to look back and say \"what I was really getting at for all those years was the one basic idea that ...\" I am too close and too involved in its continuation to see beyond the twists and turns. The last sections of the paper describe a viewpoint that differs in significant ways from most current approaches, and that offers new possibilities for a deeper understanding of language and a grasp on some previously intractable or unrecognized problems. I hope that it will give some sense of where the path is headed.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "What Does it Mean to Understand Language? Terry Winograd", |
| "sec_num": null |
| }, |
| { |
| "text": "Yale University New Haven, Connecticut 06520", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Department of Computer Science", |
| "sec_num": null |
| }, |
| { |
| "text": "Cognitive Science 4, 3 (July-Sept. 1980), 243-284. This paper outlines some of the issues\" and basic philosophy that have guided my work and that of my students in the last ten years. It describes the progression of conceptual representational theories developed during that time, as well as some of the research models built to implement those theories. The paper concludes with a discussion of my most recent work in the area of modelling memory. It presents a theory of MOPs (Memory Organization Packets), which serve as both processors and organizers of information in memory. This enables effective categorization of experiences in episodic memory, which in turn enables better predictive understanding of new experiences. Recently there has been some very interesting AI work done on the representation of knowledge of large scale maps and their use in the construction of routes through some space (Kuipers, 1978; McDermott, 1980) . The stress in this work has been on how spatial information is best represented and how reasoning is performed with the assumed representation.", |
| "cite_spans": [ |
| { |
| "start": 10, |
| "end": 50, |
| "text": "Science 4, 3 (July-Sept. 1980), 243-284.", |
| "ref_id": null |
| }, |
| { |
| "start": 905, |
| "end": 920, |
| "text": "(Kuipers, 1978;", |
| "ref_id": null |
| }, |
| { |
| "start": 921, |
| "end": 937, |
| "text": "McDermott, 1980)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Department of Computer Science", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper I would like to address the natural language processing of texts giving directions and make a claim that during a casual first reading, there actually does not need to be much spatial reasoning going on at all. When I read a written set of directions, I am primarily interested whether the directions seem clear and sensible -not in constructing a map or program specifying all the turns, distances, and locates that will be involved. Certainly there are times when I have to make some kind of route or map structure, such as when I am the subject in a spatial reasoning experiment, or when I am lost and ask someone on the street for directions. But I do not need to work so hard when I am given a piece of paper with a set of directions, and I know that I will have this piece of paper with me when I actually make the trip. I know that I can always read the directions again to figure out what to do, when I actually set out on the trip. A number of criticisms of a recent paper by Black and Wilensky (1979) are made. 1In attempting to assess the observational adequacy of story grammars, they state that a context-free grammar cannot handle discontinuous elements; however, they do not show that such elements occur in the domain to which the grammars apply. Further, they do not present adequate evidence for their claim that there are acceptable stories not accounted for by existing grammars and that the grammars will accept non-stories such as procedures. (2) They state that it has been proven that under natural conditions children cannot learn transformational grammars, which is a misrepresentation of the learnability proofs which have been offered. 3Most important, they take an unduly narrow approach to story understanding by claiming that people only understand story content and do not have knowledge of story structure which is useful in comprehension or memory. Counter-evidence from the literature is cited which indicates that such knowledge is both useful and used, and a number of methods for assessing the psychological adequacy of structural models are discussed.", |
| "cite_spans": [ |
| { |
| "start": 999, |
| "end": 1024, |
| "text": "Black and Wilensky (1979)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Department of Computer Science", |
| "sec_num": null |
| }, |
| { |
| "text": "Cognitive Science 4, 3 (July-Sept. 1980), 313-316. In their recent article entitled \"An Evaluation of Story Grammars,\" Black and Wilensky (1979) offer a critique of the recent work on this topic. They argue that story grammars (or story schemata as I prefer to call them) are not a productive approach to the study of story understanding, and they offer three main lines of argumentation. First, they argue that story grammars are not formally adequate in as much as most of them are represented as a set of context free rewrite rules which are known to be inadequate even for sentence grammars. Second, they argue that story grammars are not empirically adequate in as much as there are stories which do not seem to follow story grammars and there are nonstories which do. Finally, they argue that story grammars could not form an adequate basis for a comprehension model since in order to apply the grammar you need to have interpreted the story. These arguments are in my opinion, indicative of a misunderstanding of the enterprise that I and others working on these issues have been engaged in. I believe that they are all based on a misunderstanding about what grammars might be good for and about how comprehension might occur. In this response, I wish to clarify the nature of story schemata as I understand them, clarify the nature of Black and Wilensky's misunderstandings and show how each of their arguments fails to address the important issues about story grammars and story schemata. This is an excerpt from the Handbook of Artificial Intelligence, a compendium of hundreds of articles about AI ideas, techniques, and programs being prepared at Stanford University by AI researchers and students from across the country. This article, which is from the chapter on Natural Language Understanding, presents a brief sketch of the history of natural language processing research and gives an idea of the current state of the art. The other articles in the NL chapter of the Handbook include a historical sketch of machine translation, technical articles on grammars and parsing techniques, and an article on text generation. Finally, there are several articles describing the NL programs themselves. The Handbook also includes chapters on speech understanding and knowledge representation. Two premises, reflected in the title, underlie the perspective from which I will consider research in natural language processing in this paper. First, progress on building computer systems that process natural languages in any meaningful sense (i.e., systems that interact reasonably with people in natural language) requires considering language as part of a larger communicative situation. In this larger situation, the participants in a conversation and their states of mind are as important to the interpretation of an utterance as the linguistic expressions from which it is formed. A central concern when language is considered as communication is its function in building and using shared models of the world. Second, as the phrase \"utterance and objective\" suggests, regarding language as communication requires consideration of what is said literally, what is intended, and the relationship between the two. Recently, the emphasis in research in natural language processing has begun to shift from an analysis of utterances as isolated linguistic phenomena to a consideration of how people use utterances to achieve certain objectives. But, in considering objectives, it is important not to ignore the utterances themselves. A consideration of a speaker's underlying goals and motivations is critical, but so is an analysis of the particular way in which that speaker expresses his thoughts.", |
| "cite_spans": [ |
| { |
| "start": 10, |
| "end": 50, |
| "text": "Science 4, 3 (July-Sept. 1980), 313-316.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "On Evaluating Story Grammars David E. Rumelhart University of California, San Diego La Jolla, California 92093", |
| "sec_num": null |
| }, |
| { |
| "text": "This paper examines three consequences of these claims for the development of language processing theories and the construction of language processing programs: (1) language processing requires a combination of language-specific mechanisms and general common-sense reasoning mechanisms, (2) language systems must be able to represent the beliefs and knowledge of multiple individual agents, and (3) utterances must be viewed as having effects along multiple dimensions.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "On Evaluating Story Grammars David E. Rumelhart University of California, San Diego La Jolla, California 92093", |
| "sec_num": null |
| }, |
| { |
| "text": "Technical Note 203, Dec. 1979. In this paper, planning models developed in artificial intelligence are applied to the kind of planning that must be carried out by participants in a conversation. A planning mechanism is defined, and a short fragment of a free-flowing videotaped conversation is described. The bulk of the paper is then devoted to an attempt to understand the conversation in terms of the planning mechanism. This microanalysis suggests ways in which the planning mechanisms must be augmented, and reveals several important conversational phenomena that deserve further investigation.", |
| "cite_spans": [ |
| { |
| "start": 20, |
| "end": 30, |
| "text": "Dec. 1979.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conversation as Planned Behavior Jerry R. Hobbs and David A. Evans Artificial Intelligence Center SRI International 333 Ravenswood Avenue Menlo Park, California 94025", |
| "sec_num": null |
| }, |
| { |
| "text": "Technical Note 204, Dec. 1979. The importance of spatial and other metaphors is demonstrated. An approach to handling metaphor in a computational framework is described, based on the idea of selective inferencing. Three examples of metaphors are examined in detail in this light m a simple metaphor, a spatial metaphor schema, and a novel metaphor. Finally, there is a discussion, from this perspective, of the analogical processes that underlie metaphor in this approach and what the approach says about several classical questions about metaphor. Task Technical Report CMU-CS-80-123, April 1980. Command interfaces to current interactive systems often appear inflexible and unfriendly to casual and expert users alike. We are constructing an interface that will behave more cooperatively (by correcting spelling and grammatical errors, asking the user to resolve ambiguities in subparts of commands, etc.). Given that present-day interfaces often absorb a major portion of implementation effort, such a gracefully interacting interface can only be practical if it is independent of the specific tool or functional subsystem with which it is used.", |
| "cite_spans": [ |
| { |
| "start": 20, |
| "end": 30, |
| "text": "Dec. 1979.", |
| "ref_id": null |
| }, |
| { |
| "start": 571, |
| "end": 597, |
| "text": "CMU-CS-80-123, April 1980.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 549, |
| "end": 553, |
| "text": "Task", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Metaphor, Metaphor Schemata and Selective Inferencing Jerry R. Hobbs Artificial Intelligence Center SRI International 333 Ravenswood Avenue Menlo Park, California 94025", |
| "sec_num": null |
| }, |
| { |
| "text": "Our interface is tool-independent in the sense that all its information about a particular tool is expressed in a declarative tool description. This tool description contains schemas for each operation that the tool can perform, and for each kind of object known to the system. The operation schemas describe the relevant parameters, their types and defaults, and the object schemas give corresponding structural descriptions in terms of defining and derived subcomponents. The schemas also include input syntax, display formats, and explanatory text. We discuss how these schemas can be used by the tool-independent interface to provide a graceful interface to the tool they describe. Technical Report CMU-CS-80-122, May 1980. When people use natural language in natural settings, they often use it ungrammatically, missing out or repeating words, breaking-off and restarting, speaking in fragments, etc. Their human listeners are usually able to cope with these deviations with little difficulty. If a computer system wishes to accept natural language input from its users on a routine basis, it must display a similar indifference. In this paper, we outline a set of parsing flexibilities that such a system should provide.", |
| "cite_spans": [ |
| { |
| "start": 703, |
| "end": 727, |
| "text": "CMU-CS-80-122, May 1980.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representation of", |
| "sec_num": null |
| }, |
| { |
| "text": "We go on to describe FlexP, a bottom-up pattern-matching parser that we have designed and implemented to provide these flexibilities for restricted natural language input to a limited-domain computer system. Frazier and Fodor, 1978) . One of the major results which Frazier and Fodor bring forward in support of their proposal concerns a parsing strategy which, following Kimball (1973) , they call Right Association. The center-piece of their argument concerns an interaction between this parsing strategy and another one, which they call Minimal Attachment. Frazier and Fodor (henceforth FF) provide interesting evidence that the language user makes tacit use of both strategies to resolve temporary syntactic ambiguities that arise during parsing. FF then proceed to argue that the existence of these strategies, as well as the apparent interaction between them, can be fully explained if we assume that the language user's parsing system is configured along the lines of the Sausage Machine. In FF's view, the Augmented Transition Network (ATN) runs a very poor second to the Sausage Machine, for according to FF's argument, it is impossible even to describe the two parsing strategies within the ATN framework. In effect then, FF are claiming that the Sausage Machine achieves explanation adequacy in this case while the ATN fails to reach the level of descriptive adequacy.", |
| "cite_spans": [ |
| { |
| "start": 208, |
| "end": 232, |
| "text": "Frazier and Fodor, 1978)", |
| "ref_id": null |
| }, |
| { |
| "start": 372, |
| "end": 386, |
| "text": "Kimball (1973)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representation of", |
| "sec_num": null |
| }, |
| { |
| "text": "These are strong and potentially important claims. If correct, they obviously provide grounds for pursuing parsing models built along the lines of the Sausage Machine rather than the ATN. However, when FF's arguments are examined at close range, the comparison between parsing systems comes out rather differently than they claim. In particular, it appears that the Sausage Machine explanation of Right Association and its interaction with Minimal Attachment is empirically incorrect. The inadequacy of this explanation completely cancels the Sausage Machine's ability to describe the interaction between strategies that FF have observed. This follows because FF aspire to an explanation that renders independent description of the parsing strategies unnecessary. The Sausage Machine contains no apparatus for describing strategies. Hence, the failure to achieve explanatory adequacy automatically entails descriptive failure as well. In contrast, and in contradiction of FF's negative claim, the ATN can provide a perfectly general description for each strategy in terms of scheduling principles that constrain the order in which arcs in an ATN grammar are attempted. Moreover, when these scheduling principles are coupled with an ATN version of the grammar FF tacitly employed to generate their pivotal cases, FF's observations about the interactions between strategies are completely accounted for. Thus, although the ATN framework does not provide an explanation for either parsing strategy, it appears to achieve descriptive adequacy. Moreover, the descriptive framework of the ATN makes it possible to discern just what phenomena require explanation and to speculate in a reasonable way about the explanatory principles that underlie the parsing strategies FF have discovered. ACM Trans. Prog. Lang. Syst. 2, 3 (July 1980), 415-462. A new algorithm for recognizing and parsing arbitrary context-free languages is presented, and several new results are given on the computational complexity of these problems. The new algorithm is of both practical and theoretical interest. It is conceptually simple and allows a variety of efficient implementations, which are worked out in detail. Two versions are given which run in faster than cubic time. Surprisingly close connections between the Cocke-Kasami-Younger and Earley algorithms are established which reveal that the two algorithms are \"almost\" identical. One significant use of general context-free methods is as part of a system of processing natural languages such", |
| "cite_spans": [ |
| { |
| "start": 1794, |
| "end": 1838, |
| "text": "Prog. Lang. Syst. 2, 3 (July 1980), 415-462.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Representation of", |
| "sec_num": null |
| }, |
| { |
| "text": "American Journal of Computational Linguistics, Volume 7, Number 1, January-March 1981", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "as English, viewing the grammar/parser as a convenient control structure for directing the analysis. Computing Surveys 12, 2 (June 1980), 213-253. The Hearsay-II system, developed at Carnegie-Mellon University during the D ARPA-sponsored fiveyear speech-understanding research program, represents both a specific solution to the speechunderstanding problem and a general framework for coordinating independent processes to achieve cooperative problem-solving behavior. Speech-understanding, as a computational problem, reflects a large number of intrinsically interesting issues. Spoken sounds are achieved by a long chain of successive transformations, from intentions, through semantic and syntactic structuring, eventually resulting in audible acoustic waves. As a consequence, interpreting speech means effectively inverting these transformations to recover the speaker's intention from the sound. At each step in the interpretive process, ambiguity and uncertainty arise.", |
| "cite_spans": [ |
| { |
| "start": 111, |
| "end": 146, |
| "text": "Surveys 12, 2 (June 1980), 213-253.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "annex", |
| "sec_num": null |
| }, |
| { |
| "text": "UnderstandingThe Hearsay-II problem-solving framework reconstructs an intention from hypothetical interpretations formulated at various levels of abstraction. In addition, it allocates limited processing resources first to the most promising incremental actions. The final configuration of the Hearsay-II system comprises problem-solving components to generate and evaluate speech hypotheses, and a focus-of-control mechanism to identify potential actions of greatest value. Many of these specific procedures reveal novel approaches to speech problems. Most importantly, the system successfully integrates and coordinates all of these independent activities to resolve uncertainty and control combinatorics. Several adaptations of the Hearsay-II framework have already been undertaken in other problem domains and we anticipate this trend will continue; many future systems necessarily will integrate diverse sources of knowledge to solve complex problems cooperatively. This paper discusses the characteristics of the speech problem in particular, the special kinds of problem-solving uncertainty in that domain, the structure of the Hearsay-II system developed to cope with that uncertainty, and the relationship between Hearsay-II's structure and those of other speechunderstanding systems. The paper is intended for the general computer science audience and presupposes no speech or artificial intelligence background. Meth. Inform. Med. 19, 2 (April 1980), 99-105. This paper describes an automated procedure for morphosemantic analysis and semantic interpretation of medical compound word forms ending in -ITIS. The requirements for morphosemantic analysis of -ITIS forms include: a) semantic classification of morphosemantic constituents forming -ITIS words forms, b) establishment of morphosemantic distribution patterns occurring in -ITIS form, c) preparation of paraphrasing rules. Newsletter 13, 3 (Sept. 1980), 6-12. We present a computerized method for reducing inflected German words to their stem. The German language has many possibilities of inflecting words (nouns have case endings, endings of adjectives depend on case and gender of the noun they belong to, etc.) Therefore it frequently happens that an inflected word can be reduced to more than one stem. In this case all possible dictionary entries to which the word can be reduced are stored and the final selection can only be done by semantic means. The program is written in PASCAL and runs on a 16 bit machine.", |
| "cite_spans": [ |
| { |
| "start": 1429, |
| "end": 1469, |
| "text": "Inform. Med. 19, 2 (April 1980), 99-105.", |
| "ref_id": null |
| }, |
| { |
| "start": 1892, |
| "end": 1928, |
| "text": "Newsletter 13, 3 (Sept. 1980), 6-12.", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "The Hearsay-II Speech-", |
| "sec_num": null |
| } |
| ], |
| "bib_entries": {}, |
| "ref_entries": { |
| "TABREF0": { |
| "content": "<table/>", |
| "num": null, |
| "type_str": "table", |
| "text": "", |
| "html": null |
| }, |
| "TABREF1": { |
| "content": "<table/>", |
| "num": null, |
| "type_str": "table", |
| "text": "", |
| "html": null |
| }, |
| "TABREF3": { |
| "content": "<table/>", |
| "num": null, |
| "type_str": "table", |
| "text": "", |
| "html": null |
| } |
| } |
| } |
| } |