| { |
| "paper_id": "E09-1012", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T10:49:34.696269Z" |
| }, |
| "title": "Incremental Parsing Models for Dialog Task Structure", |
| "authors": [ |
| { |
| "first": "Srinivas", |
| "middle": [], |
| "last": "Bangalore", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "AT&T Labs -Research, Inc", |
| "location": { |
| "addrLine": "180 Park Avenue, Florham Park", |
| "postCode": "07932", |
| "region": "NJ", |
| "country": "USA" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Amanda", |
| "middle": [ |
| "J" |
| ], |
| "last": "Stent", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "AT&T Labs -Research, Inc", |
| "location": { |
| "addrLine": "180 Park Avenue, Florham Park", |
| "postCode": "07932", |
| "region": "NJ", |
| "country": "USA" |
| } |
| }, |
| "email": "stent@research.att.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "In this paper, we present an integrated model of the two central tasks of dialog management: interpreting user actions and generating system actions. We model the interpretation task as a classi\uf0decation problem and the generation task as a prediction problem. These two tasks are interleaved in an incremental parsing-based dialog model. We compare three alternative parsing methods for this dialog model using a corpus of human-human spoken dialog from a catalog ordering domain that has been annotated for dialog acts and task/subtask information. We contrast the amount of context provided by each method and its impact on performance.", |
| "pdf_parse": { |
| "paper_id": "E09-1012", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "In this paper, we present an integrated model of the two central tasks of dialog management: interpreting user actions and generating system actions. We model the interpretation task as a classi\uf0decation problem and the generation task as a prediction problem. These two tasks are interleaved in an incremental parsing-based dialog model. We compare three alternative parsing methods for this dialog model using a corpus of human-human spoken dialog from a catalog ordering domain that has been annotated for dialog acts and task/subtask information. We contrast the amount of context provided by each method and its impact on performance.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Corpora of spoken dialog are now widely available, and frequently come with annotations for tasks/games, dialog acts, named entities and elements of syntactic structure. These types of information provide rich clues for building dialog models (Grosz and Sidner, 1986) . Dialog models can be built of\uf0dfine (for dialog mining and summarization), or online (for dialog management).", |
| "cite_spans": [ |
| { |
| "start": 243, |
| "end": 267, |
| "text": "(Grosz and Sidner, 1986)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "A dialog manager is the component of a dialog system that is responsible for interpreting user actions in the dialog context, and for generating system actions. Needless to say, a dialog manager operates incrementally as the dialog progresses. In typical commercial dialog systems, the interpretation and generation processes operate independently of each other, with only a small amount of shared context. By contrast, in this paper we describe a dialog model that (1) tightly integrates interpretation and generation, (2) makes explicit the type and amount of shared context, (3) includes the task structure of the dialog in the context, (4) can be trained from dialog data, and (5) runs incrementally, parsing the dialog as it occurs and interleaving generation and interpretation.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "At the core of our model is a parser that incrementally builds the dialog task structure as the dialog progresses. In this paper, we experiment with three different incremental tree-based parsing methods. We compare these methods using a corpus of human-human spoken dialogs in a catalog ordering domain that has been annotated for dialog acts and task/subtask information. We show that all these methods outperform a baseline method for recovering the dialog structure.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The rest of this paper is structured as follows: In Section 2, we review related work. In Section 3, we present our view of the structure of taskoriented human-human dialogs. In Section 4, we present the parsing approaches included in our experiments. In Section 5, we describe our data and experiments. Finally, in Section 6, we present conclusions and describe our current and future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "There are two threads of research that are relevant to our work: work on parsing (written and spoken) discourse, and work on plan-based dialog models.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Discourse Parsing Discourse parsing is the process of building a hierarchical model of a discourse from its basic elements (sentences or clauses), as one would build a parse of a sentence from its words. There has now been considerable work on discourse parsing using statistical bottom-up parsing (Soricut and Marcu, 2003) , hierarchical agglomerative clustering (Sporleder and Lascarides, 2004) , parsing from lexicalized tree-adjoining grammars (Cristea, 2000) , and rulebased approaches that use rhetorical relations and discourse cues (Forbes et al., 2003; Polanyi et al., 2004; LeThanh et al., 2004) . With the exception of Cristea (2000) , most of this research has been limited to non-incremental parsing of textual monologues where, in contrast to incremental dialog parsing, predicting a system action is not relevant.", |
| "cite_spans": [ |
| { |
| "start": 298, |
| "end": 323, |
| "text": "(Soricut and Marcu, 2003)", |
| "ref_id": "BIBREF28" |
| }, |
| { |
| "start": 364, |
| "end": 396, |
| "text": "(Sporleder and Lascarides, 2004)", |
| "ref_id": "BIBREF29" |
| }, |
| { |
| "start": 448, |
| "end": 463, |
| "text": "(Cristea, 2000)", |
| "ref_id": "BIBREF11" |
| }, |
| { |
| "start": 540, |
| "end": 561, |
| "text": "(Forbes et al., 2003;", |
| "ref_id": "BIBREF12" |
| }, |
| { |
| "start": 562, |
| "end": 583, |
| "text": "Polanyi et al., 2004;", |
| "ref_id": "BIBREF21" |
| }, |
| { |
| "start": 584, |
| "end": 605, |
| "text": "LeThanh et al., 2004)", |
| "ref_id": "BIBREF17" |
| }, |
| { |
| "start": 630, |
| "end": 644, |
| "text": "Cristea (2000)", |
| "ref_id": "BIBREF11" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "The work on discourse parsing that is most similar to ours is that of Baldridge and Lascarides (2005) . They used a probabilistic headdriven parsing method (described in (Collins, 2003) ) to construct rhetorical structure trees for a spoken dialog corpus. However, their parser was not incremental; it used global features such as the number of turn changes. Also, it focused strictly in interpretation of input utterances; it could not predict actions by either dialog partner. In contrast to other work on discourse parsing, we wish to use the parsing process directly for dialog management (rather than for information extraction or summarization). This in\uf0dfuences our approach to dialog modeling in two ways. First, the subtask tree we build represents the functional task structure of the dialog (rather than the rhetorical structure of the dialog). Second, our dialog parser must be entirely incremental.", |
| "cite_spans": [ |
| { |
| "start": 70, |
| "end": 101, |
| "text": "Baldridge and Lascarides (2005)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 170, |
| "end": 185, |
| "text": "(Collins, 2003)", |
| "ref_id": "BIBREF9" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Related Work", |
| "sec_num": "2" |
| }, |
| { |
| "text": "Plan-based approaches to dialog modeling, like ours, operate directly on the dialog's task structure. The process of task-oriented dialog is treated as a special case of AI-style plan recognition (Sidner, 1985; Litman and Allen, 1987; Rich and Sidner, 1997; Carberry, 2001; Bohus and Rudnicky, 2003; Lochbaum, 1998) . Plan-based dialog models are used for both interpretation of user utterances and prediction of agent actions. In addition to the hand-crafted models listed above, researchers have built stochastic plan recognition models for interaction, including ones based on Hidden Markov Models (Bui, 2003; Blaylock and Allen, 2006 ) and on probabilistic context-free grammars (Alexandersson and Reithinger, 1997; Pynadath and Wellman, 2000) .", |
| "cite_spans": [ |
| { |
| "start": 196, |
| "end": 210, |
| "text": "(Sidner, 1985;", |
| "ref_id": "BIBREF27" |
| }, |
| { |
| "start": 211, |
| "end": 234, |
| "text": "Litman and Allen, 1987;", |
| "ref_id": "BIBREF18" |
| }, |
| { |
| "start": 235, |
| "end": 257, |
| "text": "Rich and Sidner, 1997;", |
| "ref_id": "BIBREF24" |
| }, |
| { |
| "start": 258, |
| "end": 273, |
| "text": "Carberry, 2001;", |
| "ref_id": "BIBREF8" |
| }, |
| { |
| "start": 274, |
| "end": 299, |
| "text": "Bohus and Rudnicky, 2003;", |
| "ref_id": "BIBREF6" |
| }, |
| { |
| "start": 300, |
| "end": 315, |
| "text": "Lochbaum, 1998)", |
| "ref_id": "BIBREF19" |
| }, |
| { |
| "start": 601, |
| "end": 612, |
| "text": "(Bui, 2003;", |
| "ref_id": "BIBREF7" |
| }, |
| { |
| "start": 613, |
| "end": 637, |
| "text": "Blaylock and Allen, 2006", |
| "ref_id": "BIBREF5" |
| }, |
| { |
| "start": 683, |
| "end": 719, |
| "text": "(Alexandersson and Reithinger, 1997;", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 720, |
| "end": 747, |
| "text": "Pynadath and Wellman, 2000)", |
| "ref_id": "BIBREF22" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Plan-Based Dialog Models", |
| "sec_num": null |
| }, |
| { |
| "text": "In this area, the work most closely related to ours is that of Barrett and Weld (Barrett and Weld, 1994) , who build an incremental bottom-up parser Figure 2 : Sample output (subtask tree) from a parse-based model for the catalog ordering domain.", |
| "cite_spans": [ |
| { |
| "start": 63, |
| "end": 104, |
| "text": "Barrett and Weld (Barrett and Weld, 1994)", |
| "ref_id": "BIBREF3" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 149, |
| "end": 157, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Plan-Based Dialog Models", |
| "sec_num": null |
| }, |
| { |
| "text": "to parse plans. Their parser, however, was not probabilistic or targeted at dialog processing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Plan-Based Dialog Models", |
| "sec_num": null |
| }, |
| { |
| "text": "We consider a task-oriented dialog to be the result of incremental creation of a shared plan by the participants (Lochbaum, 1998) . The shared plan is represented as a single tree T that incorporates the task/subtask structure, dialog acts, syntactic structure and lexical content of the dialog, as shown in Figure 1 . A task is a sequence of subtasks ST \u2208 S. A subtask is a sequence of dialog acts DA \u2208 D. Each dialog act corresponds to one clause spoken by one speaker, customer (c u ) or agent (c a ) (for which we may have acoustic, lexical, syntactic and semantic representations). Figure 2 shows the subtask tree for a sample dialog in our domain (catalog ordering). An order placement task is typically composed of the sequence of subtasks opening, contact-information, order-item, related-offers, summary. Subtasks can be nested; the nesting can be as deep as \uf0deve levels in our data. Most often the nesting is at the leftmost or rightmost frontier of the subtask tree.", |
| "cite_spans": [ |
| { |
| "start": 113, |
| "end": 129, |
| "text": "(Lochbaum, 1998)", |
| "ref_id": "BIBREF19" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 308, |
| "end": 316, |
| "text": "Figure 1", |
| "ref_id": "FIGREF0" |
| }, |
| { |
| "start": 587, |
| "end": 595, |
| "text": "Figure 2", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "As the dialog proceeds, an utterance from a participant is accommodated into the subtask tree in an incremental manner, much like an incremental syntactic parser accommodates the next word into a partial parse tree (Alexandersson and Reithinger, 1997 ). An illustration of the incremental evolution of dialog structure is shown in Figure 4 . However, while a syntactic parser processes input from a single source, our dialog parser parses user-system exchanges: user utterances are interpreted, while system utterances are generated. So the steps taken by our dialog parser to incorporate an utterance into the subtask tree depend on whether the utterance was produced by the agent or the user (as shown in Figure 3 ). User utterances Each user turn is split into clauses (utterances). Each clause is supertagged Interpretation of a user's utterance:", |
| "cite_spans": [ |
| { |
| "start": 215, |
| "end": 250, |
| "text": "(Alexandersson and Reithinger, 1997", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 331, |
| "end": 339, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 707, |
| "end": 715, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "DAC : da u i = argmax d u \u2208D P (d u |c u i , ST i\u22121 i\u2212k , DA i\u22121 i\u2212k , c i\u22121 i\u2212k ) (1) ST C : st u i = argmax s u \u2208S P (s u |da u i , c u i , ST i\u22121 i\u2212k , DA i\u22121 i\u2212k , c i\u22121 i\u2212k )", |
| "eq_num": "(2)" |
| } |
| ], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Generation of an agent's utterance:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "ST P : st a i = argmax s a \u2208S P (s a |ST i\u22121 i\u2212k , DA i\u22121 i\u2212k , c i\u22121 i\u2212k )", |
| "eq_num": "(3)" |
| } |
| ], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "DAP : da a i = argmax : Equations used for modeling dialog act and subtask labeling of agent and user utterances. c u i /c a i = the words, syntactic information and named entities associated with the i th utterance of the dialog, spoken by user/agent u/a. da u i /da a i = the dialog act of the i th utterance, spoken by user/agent u/a. st u i /st a i = the subtask label of the i th utterance, spoken by user/agent u/a. DA i\u22121 i\u2212k represents the dialog act tags for utterances i \u2212 1 to i \u2212 k.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "EQUATION", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [ |
| { |
| "start": 0, |
| "end": 8, |
| "text": "EQUATION", |
| "ref_id": "EQREF", |
| "raw_str": "d a \u2208D P (d a |st a i , ST i\u22121 i\u2212k , DA i\u22121 i\u2212k , c i\u22121 i\u2212k )", |
| "eq_num": "(4)" |
| } |
| ], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "and labeled with named entities 1 . Interpretation of the clause (c u i ) involves assigning a dialog act label (da u i ) and a subtask label (st u i ). We use", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "ST i\u22121 i\u2212k , DA i\u22121 i\u2212k", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": ", and c i\u22121 i\u2212k to represent the sequence of preceeding k subtask labels, dialog act labels and clauses respectively. The dialog act label da u i is determined from information about the clause and (a k th order approximation of) the subtask tree so", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "far (T i\u22121 = (ST i\u22121 i\u2212k , DA i\u22121 i\u2212k , c i\u22121 i\u2212k ))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": ", as shown in Equation 1 (Table 1) . The subtask label st u i is determined from information about the clause, its dialog act and the subtask tree so far, as shown in Equation 2. Then, the clause is incorporated into the subtask tree. Agent utterances In contrast, a dialog system starts planning an agent utterance by identifying the subtask to contribute to next, st a i , based on the subtask tree so far (", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 25, |
| "end": 34, |
| "text": "(Table 1)", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "T i\u22121 = (ST i\u22121 i\u2212k , DA i\u22121 i\u2212k , c i\u22121 i\u2212k ))", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": ", as shown in Equation 3 (Table 1) . Then, it chooses the dialog act of the utterance, da a i , based on the subtask tree so far and the chosen subtask for the utterance, as shown in Equation 4. Finally, it generates an utterance, c a i , to realize its communicative intent (represented as a subtask and dialog act pair, with associated named entities) 2 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 25, |
| "end": 34, |
| "text": "(Table 1)", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "Note that the current clause c u i is used in the conditioning context of the interpretation model (for user utterances), but the corresponding clause for the agent utterance c a i is to be predicted and hence is not part of conditioning context in the generation model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Structure", |
| "sec_num": "3" |
| }, |
| { |
| "text": "A dialog parser can produce a \"shallow\" or \"deep\" tree structure. A shallow parse is one in which utterances are grouped together into subtasks, but the dominance relations among subtasks are not tracked. We call this model a chunk-based dialog model (Bangalore et al., 2006) . The chunkbased model has limitations. For example, dominance relations among subtasks are important for dialog processes such as anaphora resolution (Grosz and Sidner, 1986) . Also, the chunkbased model is representationally inadequate for center-embedded nestings of subtasks, which do occur in our domain, although less frequently than the more prevalent \"tail-recursive\" structures.", |
| "cite_spans": [ |
| { |
| "start": 251, |
| "end": 275, |
| "text": "(Bangalore et al., 2006)", |
| "ref_id": "BIBREF2" |
| }, |
| { |
| "start": 427, |
| "end": 451, |
| "text": "(Grosz and Sidner, 1986)", |
| "ref_id": "BIBREF13" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Dialog Parsing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "We use the term parse-based dialog model to refer to deep parsing models for dialog which not only segment the dialog into chunks but also predict dominance relations among chunks. For this paper, we experimented with three alternative methods for building parse-based models: shiftreduce, start-complete and connection path. Each of these operates on the subtask tree for the dialog incrementally, from left-to-right, with access only to the preceding dialog context, as shown in Figure 4 . They differ in the parsing actions and the data structures used by the parser; this has implications for robustness to errors. The instructions to reconstruct the parse are either entirely encoded in the stack (in the shift-reduce method), or entirely in the parsing actions (in the start-complete and connection path methods). For each of the four types of parsing action required to build the parse tree (see Figure 4 : An illustration of incremental evolution of dialog structure a feature vector containing contextual information for the parsing action (see Section 5.1). These feature vectors and the associated parser actions are used to train maximum entropy models (Berger et al., 1996) . These models are then used to incrementally incorporate the utterances for a new dialog into that dialog's subtask tree as the dialog progresses, as shown in Figure 3 .", |
| "cite_spans": [ |
| { |
| "start": 1165, |
| "end": 1186, |
| "text": "(Berger et al., 1996)", |
| "ref_id": "BIBREF4" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 481, |
| "end": 489, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 903, |
| "end": 911, |
| "text": "Figure 4", |
| "ref_id": null |
| }, |
| { |
| "start": 1347, |
| "end": 1355, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Dialog Parsing", |
| "sec_num": "4" |
| }, |
| { |
| "text": "In this method, the subtask tree is recovered through a right-branching shift-reduce parsing process (Hall et al., 2006; Sagae and Lavie, 2006) . The parser shifts each utterance on to the stack. It then inspects the stack and decides whether to do one or more reduce actions that result in the creation of subtrees in the subtask tree. The parser maintains two data structures -a stack and a tree. The actions of the parser change the contents of the stack and create nodes in the dialog tree structure. The actions for the parser include unaryreduce-X, binary-reduce-X and shift, where X is each of the non-terminals (subtask labels) in the tree. Shift pushes a token representing the utterance onto the stack; binary-reduce-X pops two tokens off the stack and pushes the non-terminal X; and unary-reduce-X pops one token off the stack and pushes the non-terminal X. Each type of reduce action creates a constituent X in the dialog tree and the tree(s) associated with the reduced elements as subtree(s) of X. At the end of the dialog, the output is a binary branching subtask tree.", |
| "cite_spans": [ |
| { |
| "start": 101, |
| "end": 120, |
| "text": "(Hall et al., 2006;", |
| "ref_id": "BIBREF15" |
| }, |
| { |
| "start": 121, |
| "end": 143, |
| "text": "Sagae and Lavie, 2006)", |
| "ref_id": "BIBREF25" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Shift-Reduce Method", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "Consider the example subdialog A: would you like a free magazine? U: no. The process-ing of this dialog using our shift-reduce dialog parser would proceed as follows: the STP model predicts shift for st a ; the DAP model predicts YNP(Promotions) for da a ; the generator outputs would you like a free magazine?; and the parser shifts a token representing this utterance onto the stack. Then, the customer says no. The DAC model classi\uf0dees da u as No; the STC model clas-si\uf0dees st u as shift and binary-reduce-special-offer; and the parser shifts a token representing the utterance onto the stack, before popping the top two elements off the stack and adding the subtree for special-order into the dialog's subtask tree.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Shift-Reduce Method", |
| "sec_num": "4.1" |
| }, |
| { |
| "text": "In the shift-reduce method, the dialog tree is constructed as a side effect of the actions performed on the stack: each reduce action on the stack introduces a non-terminal in the tree. By contrast, in the start-complete method the instructions to build the tree are directly encoded in the parser actions. A stack is used to maintain the global parse state. The actions the parser can take are similar to those described in (Ratnaparkhi, 1997) . The parser must decide whether to join each new terminal onto the existing left-hand edge of the tree, or start a new subtree. The actions for the parser include start-X, n-start-X, complete-X, u-complete-X and b-complete-X, where X is each of the nonterminals (subtask labels) in the tree. Start-X pushes a token representing the current utterance onto the stack; n-start-X pushes non-terminal X onto the stack; complete-X pushes a token representing the current utterance onto the stack, then pops the top two tokens off the stack and pushes the non-terminal X; u-complete-X pops the top token off the stack and pushes the non-terminal X; and b-complete-X pops the top two tokens off the stack and pushes the non-terminal X. This method produces a dialog subtask tree directly, rather than producing an equivalent binary-branching tree.", |
| "cite_spans": [ |
| { |
| "start": 425, |
| "end": 444, |
| "text": "(Ratnaparkhi, 1997)", |
| "ref_id": "BIBREF23" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Start-Complete Method", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "Consider the same subdialog as before, A: would you like a free magazine? U: no. The processing of this dialog using our start-complete dialog parser would proceed as follows: the STP model predicts start-special-offer for st a ; the DAP model predicts YNP(Promotions) for da a ; the generator outputs would you like a free magazine?; and the parser shifts a token representing this utterance onto the stack. Then, the customer says no. The DAC model classi\uf0dees da u as No; the STC model classi\uf0dees st u as complete-special-offer; and the parser shifts a token representing the utterance onto the stack, before popping the top two elements off the stack and adding the subtree for special-order into the dialog's subtask tree.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Start-Complete Method", |
| "sec_num": "4.2" |
| }, |
| { |
| "text": "In contrast to the shift-reduce and the startcomplete methods described above, the connection path method does not use a stack to track the global state of the parse. Instead, the parser directly predicts the connection path (path from the root to the terminal) for each utterance. The collection of connection paths for all the utterances in a dialog de\uf0denes the parse tree. This encoding was previously used for incremental sentence parsing by (Costa et al., 2001) . With this method, there are many more choices of decision for the parser (195 decisions for our data) compared to the shiftreduce (32) and start-complete (82) methods.", |
| "cite_spans": [ |
| { |
| "start": 445, |
| "end": 465, |
| "text": "(Costa et al., 2001)", |
| "ref_id": "BIBREF10" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Connection Path Method", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "Consider the same subdialog as before, A: would you like a free magazine? U: no. The processing of this dialog using our connection path dialog parser would proceed as follows. First, the STP model predicts S-special-offer for st a ; the DAP model predicts YNP(Promotions) for da a ; the generator outputs would you like a free magazine?; and the parser adds a subtree rooted at special-offer, with one terminal for the current utterance, into the top of the subtask tree. Then, the customer says no. The DAC model classi-\uf0dees da u as No and the STC model classi\uf0dees st u as S-special-offer. Since the right frontier of the subtask tree has a subtree matching this path, the parser simply incorporates the current utterance as a terminal of the special-offer subtree.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Connection Path Method", |
| "sec_num": "4.3" |
| }, |
| { |
| "text": "To evaluate our parse-based dialog model, we used 817 two-party dialogs from the CHILD corpus of telephone-based dialogs in a catalog-purchasing domain. Each dialog was transcribed by hand; all numbers (telephone, credit card, etc.) were removed for privacy reasons. The average dialog in this data set had 60 turns. The dialogs were automatically segmented into utterances and automatically annotated with part-ofspeech tag and supertag information and named entities. They were annotated by hand for dialog acts and tasks/subtasks. The dialog act and task/subtask labels are given in Tables 2 and 3 .", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 586, |
| "end": 600, |
| "text": "Tables 2 and 3", |
| "ref_id": "TABREF4" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data and Experiments", |
| "sec_num": "5" |
| }, |
| { |
| "text": "In our experiments we used the following features for each utterance: (a) the speaker ID; (b) unigrams, bigrams and trigrams of the words; (c) unigrams, bigrams and trigrams of the part of speech tags; (d) unigrams, bigrams and trigrams of the supertags; (e) binary features indicating the presence or absence of particular types of named entity; (f) the dialog act (determined by the parser); (g) the task/subtask label (determined by the parser); and (h) the parser stack at the current utterance (deter-mined by the parser). Each input feature vector for agent subtask prediction has these features for up to three utterances of left-hand context (see Equation 3). Each input feature vector for dialog act prediction has the same features as for agent subtask prediction, plus the actual or predicted subtask label (see Equation 4). Each input feature vector for dialog act interpretation has features ah for up to three utterances of left-hand context, plus the current utterance (see Equation 1). Each input feature vector for user subtask classi\uf0decation has the same features as for user dialog act interpretation, plus the actual or classi\uf0deed dialog act (see Equation 2). The label for each input feature vector is the parsing action (for subtask classi\uf0decation and prediction) or the dialog act label (for dialog act clas-si\uf0decation and prediction). If more than one parsing action takes place on a particular utterance (e.g. a shift and then a reduce), the feature vector is repeated twice with different stack contents.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Features", |
| "sec_num": "5.1" |
| }, |
| { |
| "text": "We randomly selected roughly 90% of the dialogs for training, and used the remainder for testing.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Training Method", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "We separately trained models for: user dialog act classi\uf0decation (DAC, Equation 1); user task/subtask classi\uf0decation (STC, Equation 2); agent task/subtask prediction (STP, Equation 3); and agent dialog act prediction (DAP, Equation 4). In order to estimate the conditional distributions shown in Table 1 , we use the general technique of choosing the MaxEnt distribution that properly estimates the average of each feature over the training data (Berger et al., 1996) . We use the machine learning toolkit LLAMA (Haffner, 2006) , which encodes multiclass classi\uf0decation problems using binary MaxEnt classi\uf0deers to increase the speed of training and to scale the method to large data sets.", |
| "cite_spans": [ |
| { |
| "start": 444, |
| "end": 465, |
| "text": "(Berger et al., 1996)", |
| "ref_id": "BIBREF4" |
| }, |
| { |
| "start": 510, |
| "end": 525, |
| "text": "(Haffner, 2006)", |
| "ref_id": "BIBREF14" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 294, |
| "end": 301, |
| "text": "Table 1", |
| "ref_id": "TABREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Training Method", |
| "sec_num": "5.2" |
| }, |
| { |
| "text": "The decoding process for the three parsing methods is illustrated in Figure 3 and has four stages: STP, DAP, DAC, and STC. As already explained, each of these steps in the decoding process is modeled as either a prediction task or a classi\uf0decation task. The decoder constructs an input feature vector depending on the amount of context being used. This feature vector is used to query the appropriate classi\uf0deer model to obtain a vector of labels with weights. The parser action labels (STP and STC) are used to extend the subtask tree. For example, in the shift-reduce method, shift results in a push action on the stack, while reduce-X results in popping the top two elements off the stack and pushing X on to the stack. The dialog act labels (DAP and DAC) are used to label the leaves of the subtask tree (the utterances).", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 69, |
| "end": 77, |
| "text": "Figure 3", |
| "ref_id": "FIGREF1" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Decoding Method", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "The decoder can use n-best results from the classi\uf0deer to enlarge the search space. In order to manage the search space effectively, the decoder uses a beam pruning strategy. The decoding process proceeds until the end of the dialog is reached. In this paper, we assume that the end of the dialog is given to the decoder 3 .", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Decoding Method", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "Given that the classi\uf0deers are error-prone in their assignment of labels, the parsing step of the decoder needs to be robust to these errors. We exploit the state of the stack in the different methods to rule out incompatible parser actions (e.g. a reduce-X action when the stack has one element, a shift action on an already shifted utterance). We also use n-best results to alleviate the impact of classi\uf0decation errors. Finally, at the end of the dialog, if there are unattached constituents on the stack, the decoder attaches them as sibling constituents to produce a rooted tree structure. These constraints contribute to robustness, but cannot be used with the connection path method, since any connection path (parsing action) suggested by the classi\uf0deer can be incorporated into the incremental parse tree. Consequently, in the connection path method there are fewer opportunities to correct the errors made by the classi\uf0deers.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Decoding Method", |
| "sec_num": "5.3" |
| }, |
| { |
| "text": "We evaluate dialog act classi\uf0decation and prediction by comparing the automatically assigned dialog act tags to the reference dialog act tags. For these tasks we report accuracy. We evaluate subtask classi\uf0decation and prediction by comparing the subtask trees output by the different parsing methods to the reference subtask tree. We use the labeled crossing bracket metric (typically used in the syntactic parsing literature (Harrison et al., 1991) ), which computes recall, precision and crossing brackets for the constituents (subtrees) in a hypothesized parse tree given the reference parse tree. We report F-measure, which is a combination of recall and precision.", |
| "cite_spans": [ |
| { |
| "start": 424, |
| "end": 447, |
| "text": "(Harrison et al., 1991)", |
| "ref_id": "BIBREF16" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "For each task, performance is reported for 1, 3, 5, and 10-best dynamic decoding as well as oracle (Or) and for 0, 1 and 3 utterances of context. Figure 5 shows the performance of the different methods for determining the subtask tree of the dialog. Wider beam widths do not lead to improved performance for any method. One utterance of context is best for shift-reduce and start-join; three is best for the connection path method. The shiftreduce method performs the best. With 1 utterance of context, its 1-best f-score is 47.86, as compared with 34.91 for start-complete, 25.13 for the connection path method, and 21.32 for the chunkbased baseline. These performance differences are statistically signi\uf0decant at p < .001. However, the best performance for the shift-reduce method is still signi\uf0decantly worse than oracle.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 146, |
| "end": 154, |
| "text": "Figure 5", |
| "ref_id": "FIGREF2" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Evaluation Metrics", |
| "sec_num": "5.4" |
| }, |
| { |
| "text": "All of the methods are subject to some 'stickiness', a certain preference to stay within the current subtask rather than starting a new one. Also, all of the methods tended to perform poorly on parsing subtasks that occur rarely (e.g. callforward, order-change) or that occur at many different locations in the dialog (e.g. out-of-domain, order-problem, check-availability) . For example, the shift-reduce method did not make many shift errors but did frequently b-reduce on an incorrect non-terminal (indicating trouble identifying subtask boundaries). Some non-terminals most likely to be labeled incorrectly by this method (for both agent and user) are: call-forward, orderchange, summary, order-problem, opening and out-of-domain.", |
| "cite_spans": [ |
| { |
| "start": 318, |
| "end": 373, |
| "text": "(e.g. out-of-domain, order-problem, check-availability)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5.5" |
| }, |
| { |
| "text": "Similarly, the start-complete method frequently mislabeled a non-terminal in a complete action, e.g. misc-other, check-availability, summary or contact-info. It also quite frequently mislabeled nonterminals in n-start actions, e.g. order-item, contact-info or summary. Both of these errors indicate trouble identifying subtask boundaries.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5.5" |
| }, |
| { |
| "text": "It is harder to analyze the output from the connection path method. This method is more likely to mislabel tree-internal nodes than those immediately above the leaves. However, the same non-terminals show up as error-prone for this method as for the others: out-of-domain, checkavailability, order-problem and summary. Figure 6 shows accuracy for classi\uf0decation of user dialog acts. Wider beam widths do not lead to sign\uf0decantly improved performance for any method. Zero utterances of context gives the highest accuracy for all methods. All methods perform fairly well, but no method signi\uf0decantly outperforms any other: with 0 utterances of context, 1-best accuracy is .681 for the connection path method, .698 for the start-complete method and .698 for the shift-reduce method. We note that these results are competitive with those reported in the literature (e.g. (Poesio and Mikheev, 1998; Sera\uf0den and Eugenio, 2004) ), although the dialog corpus and the label sets are different.", |
| "cite_spans": [ |
| { |
| "start": 864, |
| "end": 890, |
| "text": "(Poesio and Mikheev, 1998;", |
| "ref_id": "BIBREF20" |
| }, |
| { |
| "start": 891, |
| "end": 916, |
| "text": "Sera\uf0den and Eugenio, 2004)", |
| "ref_id": "BIBREF26" |
| } |
| ], |
| "ref_spans": [ |
| { |
| "start": 319, |
| "end": 327, |
| "text": "Figure 6", |
| "ref_id": "FIGREF3" |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5.5" |
| }, |
| { |
| "text": "The most common errors in dialog act classi\uf0decation occur with dialog acts that occur 40 times or fewer in the testing data (out of 3610 testing utterances), and with Not(Information). Figure 7 shows accuracy for prediction of agent dialog acts. Performance for this task is lower than", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 184, |
| "end": 192, |
| "text": "Figure 7", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Results", |
| "sec_num": "5.5" |
| }, |
| { |
| "text": "This results in a syntactic parse of the clause and could be done incrementally as well.2 We do not address utterance realization in this paper.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "This is an unrealistic assumption if the decoder is to serve as a dialog model. We expect to address this limitation in future work.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [ |
| { |
| "text": "that for dialog act classi\uf0decation because this is a prediction task. Wider beam widths do not generally lead to improved performance for any method. Three utterances of context generally gives the best performance. The shift-reduce method performs signi\uf0decantly better than the connection path method with a beam width of 1 (p < .01), but not at larger beam widths; there are no other signi\uf0decant performance differences between methods at 3 utterances of context. With 3 utterances of context, 1-best accuracies are .286 for the connection path method, .329 for the start-complete method and .356 for the shift-reduce method.The most common errors in dialog act prediction occur with rare dialog acts, Not(Information), and the prediction of Acknowledge at the start of a turn (we did not remove grounding acts from the data). With the shift-reduce method, some YNQ acts are commonly mislabeled. With all methods, dialog acts pertaining to Order-Info and Product-Info acts are commonly mislabeled, which could potentially indicate that these labels require a subtle distinction between information pertaining to an order and information pertaining to a product. Table 4 shows the parsing actions performed by each of our methods on the dialog snippet presented in Figure 4 . For this example, the connection path method's output is correct in all cases.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 1161, |
| "end": 1168, |
| "text": "Table 4", |
| "ref_id": null |
| }, |
| { |
| "start": 1263, |
| "end": 1271, |
| "text": "Figure 4", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "annex", |
| "sec_num": null |
| }, |
| { |
| "text": "In this paper, we present a parsing-based model of task-oriented dialog that tightly integrates interpretation and generation using a subtask tree representation, can be trained from data, and runs incrementally for use in dialog management. At the core of this model is a parser that incrementally builds the dialog task structure as it interprets user actions and generates system actions. We experiment with three different incremental parsing methods for our dialog model. Our proposed shiftreduce method is the best-performing so far, and performance of this method for dialog act classi\uf0decation and task/subtask modeling is good enough to be usable. However, performance of all the methods for dialog act prediction is too low to be useful at the moment. In future work, we will explore improved models for this task that make use of global information about the task (e.g. whether each possible subtask has yet been completed; whether required and optional task-related concepts such as shipping address have been \uf0delled). We will also separate grounding and task-related behaviors in our model.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions and Future Work", |
| "sec_num": "6" |
| } |
| ], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Learning dialogue structures from a corpus", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Alexandersson", |
| "suffix": "" |
| }, |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Reithinger", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Proceedings of Eurospeech", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Alexandersson and N. Reithinger. 1997. Learning dialogue structures from a corpus. In Proceedings of Eurospeech.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Probabilistic head-driven parsing for discourse", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Baldridge", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Lascarides", |
| "suffix": "" |
| } |
| ], |
| "year": 2005, |
| "venue": "Proceedings of CoNLL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Baldridge and A. Lascarides. 2005. Probabilistic head-driven parsing for discourse. In Proceedings of CoNLL.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Learning the structure of task-driven human-human dialogs", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Bangalore", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "Di" |
| ], |
| "last": "Fabbrizio", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Stent", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of COLING/ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Bangalore, G. Di Fabbrizio, and A. Stent. 2006. Learning the structure of task-driven human-human dialogs. In Proceedings of COLING/ACL.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Task-decomposition via plan parsing", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Barrett", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Weld", |
| "suffix": "" |
| } |
| ], |
| "year": 1994, |
| "venue": "Proceedings of AAAI", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Barrett and D. Weld. 1994. Task-decomposition via plan parsing. In Proceedings of AAAI.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "A Maximum Entropy Approach to Natural Language Processing", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Berger", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [ |
| "D" |
| ], |
| "last": "Pietra", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [ |
| "D" |
| ], |
| "last": "Pietra", |
| "suffix": "" |
| } |
| ], |
| "year": 1996, |
| "venue": "Computational Linguistics", |
| "volume": "22", |
| "issue": "1", |
| "pages": "39--71", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Berger, S.D. Pietra, and V.D. Pietra. 1996. A Max- imum Entropy Approach to Natural Language Pro- cessing. Computational Linguistics, 22(1):39-71.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Hierarchical instantiated goal recognition", |
| "authors": [ |
| { |
| "first": "N", |
| "middle": [], |
| "last": "Blaylock", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [ |
| "F" |
| ], |
| "last": "Allen", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of the AAAI Workshop on Modeling Others from Observations", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "N. Blaylock and J. F. Allen. 2006. Hierarchical instan- tiated goal recognition. In Proceedings of the AAAI Workshop on Modeling Others from Observations.", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "RavenClaw: Dialog management using hierarchical task decomposition and an expectation agenda", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Bohus", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Rudnicky", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of Eurospeech", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Bohus and A. Rudnicky. 2003. RavenClaw: Dialog management using hierarchical task decomposition and an expectation agenda. In Proceedings of Eu- rospeech.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "A general model for online probabalistic plan recognition", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [ |
| "H" |
| ], |
| "last": "Bui", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of IJCAI", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H.H. Bui. 2003. A general model for online probabal- istic plan recognition. In Proceedings of IJCAI.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Techniques for plan recognition", |
| "authors": [ |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Carberry", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "User Modeling and User-Adapted Interaction", |
| "volume": "11", |
| "issue": "1-2", |
| "pages": "31--48", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S. Carberry. 2001. Techniques for plan recogni- tion. User Modeling and User-Adapted Interaction, 11(1-2):31-48.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Head-driven statistical models for natural language parsing", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Collins", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Computational Linguistics", |
| "volume": "29", |
| "issue": "4", |
| "pages": "589--638", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Collins. 2003. Head-driven statistical models for natural language parsing. Computational Linguis- tics, 29(4):589-638.", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Wide coverage incremental parsing by learning attachment preferences", |
| "authors": [ |
| { |
| "first": "F", |
| "middle": [], |
| "last": "Costa", |
| "suffix": "" |
| }, |
| { |
| "first": "V", |
| "middle": [], |
| "last": "Lombardo", |
| "suffix": "" |
| }, |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Frasconi", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Soda", |
| "suffix": "" |
| } |
| ], |
| "year": 2001, |
| "venue": "Proceedings of the Conference of the Italian Association for Artificial Intelligence (AIIA)", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "F. Costa, V. Lombardo, P. Frasconi, and G. Soda. 2001. Wide coverage incremental parsing by learning at- tachment preferences. In Proceedings of the Con- ference of the Italian Association for Artificial Intel- ligence (AIIA).", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "An incremental discourse parser architecture", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Cristea", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of the 2nd International Conference on Natural Language Processing", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Cristea. 2000. An incremental discourse parser ar- chitecture. In Proceedings of the 2nd International Conference on Natural Language Processing.", |
| "links": null |
| }, |
| "BIBREF12": { |
| "ref_id": "b12", |
| "title": "D-LTAG system: Discourse parsing with a lexicalized tree-adjoining grammar", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Forbes", |
| "suffix": "" |
| }, |
| { |
| "first": "E", |
| "middle": [], |
| "last": "Miltsakaki", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Prasad", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Sarkar", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Joshi", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Webber", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Journal of Logic, Language and Information", |
| "volume": "12", |
| "issue": "3", |
| "pages": "261--279", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Forbes, E. Miltsakaki, R. Prasad, A. Sarkar, A. Joshi, and B. Webber. 2003. D-LTAG system: Discourse parsing with a lexicalized tree-adjoining grammar. Journal of Logic, Language and Informa- tion, 12(3):261-279.", |
| "links": null |
| }, |
| "BIBREF13": { |
| "ref_id": "b13", |
| "title": "Attention, intentions and the structure of discourse", |
| "authors": [ |
| { |
| "first": "B", |
| "middle": [ |
| "J" |
| ], |
| "last": "Grosz", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "L" |
| ], |
| "last": "Sidner", |
| "suffix": "" |
| } |
| ], |
| "year": 1986, |
| "venue": "Computational Linguistics", |
| "volume": "12", |
| "issue": "3", |
| "pages": "175--204", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "B.J. Grosz and C.L. Sidner. 1986. Attention, inten- tions and the structure of discourse. Computational Linguistics, 12(3):175-204.", |
| "links": null |
| }, |
| "BIBREF14": { |
| "ref_id": "b14", |
| "title": "Scaling large margin classi\uf0deers for spoken language understanding", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Haffner", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Speech Communication", |
| "volume": "48", |
| "issue": "3-4", |
| "pages": "239--261", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. Haffner. 2006. Scaling large margin classi\uf0deers for spoken language understanding. Speech Communi- cation, 48(3-4):239-261.", |
| "links": null |
| }, |
| "BIBREF15": { |
| "ref_id": "b15", |
| "title": "Discriminative classi\uf0deers for deterministic dependency parsing", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Hall", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nivre", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Nilsson", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of COLING/ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Hall, J. Nivre, and J. Nilsson. 2006. Discriminative classi\uf0deers for deterministic dependency parsing. In Proceedings of COLING/ACL.", |
| "links": null |
| }, |
| "BIBREF16": { |
| "ref_id": "b16", |
| "title": "Evaluating syntax performance of parser/grammars of English", |
| "authors": [ |
| { |
| "first": "P", |
| "middle": [], |
| "last": "Harrison", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Abney", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Fleckenger", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Gdaniec", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Grishman", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Hindle", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Ingria", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Marcus", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [], |
| "last": "Santorini", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Strzalkowski", |
| "suffix": "" |
| } |
| ], |
| "year": 1991, |
| "venue": "Proceedings of the Workshop on Evaluating Natural Language Processing Systems, ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "P. Harrison, S. Abney, D. Fleckenger, C. Gdaniec, R. Grishman, D. Hindle, B. Ingria, M. Marcus, B. Santorini, and T. Strzalkowski. 1991. Evaluating syntax performance of parser/grammars of English. In Proceedings of the Workshop on Evaluating Nat- ural Language Processing Systems, ACL.", |
| "links": null |
| }, |
| "BIBREF17": { |
| "ref_id": "b17", |
| "title": "Generating discourse structures for written texts", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Lethanh", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [], |
| "last": "Abeysinghe", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Huyck", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H. LeThanh, G. Abeysinghe, and C. Huyck. 2004. Generating discourse structures for written texts. In Proceedings of COLING.", |
| "links": null |
| }, |
| "BIBREF18": { |
| "ref_id": "b18", |
| "title": "A plan recognition model for subdialogs in conversations", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Litman", |
| "suffix": "" |
| }, |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Allen", |
| "suffix": "" |
| } |
| ], |
| "year": 1987, |
| "venue": "Cognitive Science", |
| "volume": "11", |
| "issue": "2", |
| "pages": "163--200", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D. Litman and J. Allen. 1987. A plan recognition model for subdialogs in conversations. Cognitive Science, 11(2):163-200.", |
| "links": null |
| }, |
| "BIBREF19": { |
| "ref_id": "b19", |
| "title": "A collaborative planning model of intentional structure", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Lochbaum", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Computational Linguistics", |
| "volume": "24", |
| "issue": "4", |
| "pages": "525--572", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Lochbaum. 1998. A collaborative planning model of intentional structure. Computational Linguistics, 24(4):525-572.", |
| "links": null |
| }, |
| "BIBREF20": { |
| "ref_id": "b20", |
| "title": "The predictive power of game structure in dialogue act recognition: experimental results using maximum entropy estimation", |
| "authors": [ |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Poesio", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mikheev", |
| "suffix": "" |
| } |
| ], |
| "year": 1998, |
| "venue": "Proceedings of ICSLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "M. Poesio and A. Mikheev. 1998. The predictive power of game structure in dialogue act recognition: experimental results using maximum entropy esti- mation. In Proceedings of ICSLP.", |
| "links": null |
| }, |
| "BIBREF21": { |
| "ref_id": "b21", |
| "title": "A rule based approach to discourse parsing", |
| "authors": [ |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Polanyi", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Culy", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Van Den", |
| "suffix": "" |
| }, |
| { |
| "first": "G", |
| "middle": [ |
| "L" |
| ], |
| "last": "Berg", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Thione", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Ahn", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of SIGdial", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "L. Polanyi, C. Culy, M. van den Berg, G. L. Thione, and D. Ahn. 2004. A rule based approach to discourse parsing. In Proceedings of SIGdial.", |
| "links": null |
| }, |
| "BIBREF22": { |
| "ref_id": "b22", |
| "title": "Probabilistic state-dependent grammars for plan recognition", |
| "authors": [ |
| { |
| "first": "D", |
| "middle": [ |
| "V" |
| ], |
| "last": "Pynadath", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [ |
| "P" |
| ], |
| "last": "Wellman", |
| "suffix": "" |
| } |
| ], |
| "year": 2000, |
| "venue": "Proceedings of UAI", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "D.V. Pynadath and M.P. Wellman. 2000. Probabilistic state-dependent grammars for plan recognition. In Proceedings of UAI.", |
| "links": null |
| }, |
| "BIBREF23": { |
| "ref_id": "b23", |
| "title": "A linear observed time statistical parser based on maximum entropy models", |
| "authors": [ |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Ratnaparkhi", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Proceedings of EMNLP", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "A. Ratnaparkhi. 1997. A linear observed time statis- tical parser based on maximum entropy models. In Proceedings of EMNLP.", |
| "links": null |
| }, |
| "BIBREF24": { |
| "ref_id": "b24", |
| "title": "COLLAGEN: When agents collaborate with people", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Rich", |
| "suffix": "" |
| }, |
| { |
| "first": "C", |
| "middle": [ |
| "L" |
| ], |
| "last": "Sidner", |
| "suffix": "" |
| } |
| ], |
| "year": 1997, |
| "venue": "Proceedings of the First International Conference on Autonomous Agents", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. Rich and C.L. Sidner. 1997. COLLAGEN: When agents collaborate with people. In Proceedings of the First International Conference on Autonomous Agents.", |
| "links": null |
| }, |
| "BIBREF25": { |
| "ref_id": "b25", |
| "title": "A best-\uf0derst probabilistic shift-reduce parser", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Sagae", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Lavie", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "Proceedings of COL-ING/ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Sagae and A. Lavie. 2006. A best-\uf0derst proba- bilistic shift-reduce parser. In Proceedings of COL- ING/ACL.", |
| "links": null |
| }, |
| "BIBREF26": { |
| "ref_id": "b26", |
| "title": "FLSA: Extending latent semantic analysis with features for dialogue act classi\uf0decation", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Sera\uf0den", |
| "suffix": "" |
| }, |
| { |
| "first": "B", |
| "middle": [ |
| "Di" |
| ], |
| "last": "Eugenio", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of ACL", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Sera\uf0den and B. Di Eugenio. 2004. FLSA: Extending latent semantic analysis with features for dialogue act classi\uf0decation. In Proceedings of ACL.", |
| "links": null |
| }, |
| "BIBREF27": { |
| "ref_id": "b27", |
| "title": "Plan parsing for intended response recognition in discourse", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [ |
| "L" |
| ], |
| "last": "Sidner", |
| "suffix": "" |
| } |
| ], |
| "year": 1985, |
| "venue": "Computational Intelligence", |
| "volume": "1", |
| "issue": "1", |
| "pages": "1--10", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C.L. Sidner. 1985. Plan parsing for intended re- sponse recognition in discourse. Computational In- telligence, 1(1):1-10.", |
| "links": null |
| }, |
| "BIBREF28": { |
| "ref_id": "b28", |
| "title": "Sentence level discourse parsing using syntactic and lexical information", |
| "authors": [ |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Soricut", |
| "suffix": "" |
| }, |
| { |
| "first": "D", |
| "middle": [], |
| "last": "Marcu", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "Proceedings of NAACL/HLT", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "R. Soricut and D. Marcu. 2003. Sentence level dis- course parsing using syntactic and lexical informa- tion. In Proceedings of NAACL/HLT.", |
| "links": null |
| }, |
| "BIBREF29": { |
| "ref_id": "b29", |
| "title": "Combining hierarchical clustering and machine learning to predict high-level discourse structure", |
| "authors": [ |
| { |
| "first": "C", |
| "middle": [], |
| "last": "Sporleder", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Lascarides", |
| "suffix": "" |
| } |
| ], |
| "year": 2004, |
| "venue": "Proceedings of COLING", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C. Sporleder and A. Lascarides. 2004. Combining hi- erarchical clustering and machine learning to pre- dict high-level discourse structure. In Proceedings of COLING.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "type_str": "figure", |
| "num": null, |
| "text": "A schema of a shared plan tree for a dialog.", |
| "uris": null |
| }, |
| "FIGREF1": { |
| "type_str": "figure", |
| "num": null, |
| "text": "Dialog management process", |
| "uris": null |
| }, |
| "FIGREF2": { |
| "type_str": "figure", |
| "num": null, |
| "text": "Performance of parse-based methods for subtask tree building", |
| "uris": null |
| }, |
| "FIGREF3": { |
| "type_str": "figure", |
| "num": null, |
| "text": "Performance of dialog act assignment to user's utterances.", |
| "uris": null |
| }, |
| "TABREF1": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "", |
| "html": null |
| }, |
| "TABREF2": { |
| "content": "<table><tr><td/><td/><td colspan=\"5\">Order Item Task</td><td/><td/><td/><td/><td colspan=\"4\">Order Item Task</td><td/><td/><td>Order Item Task</td><td>Order Item Task</td></tr><tr><td/><td/><td>Opening</td><td/><td/><td/><td/><td/><td/><td colspan=\"2\">Opening</td><td/><td/><td/><td/><td/><td>Contact!Info</td><td>Opening</td><td>Contact!Info</td><td>Opening</td><td>Contact!Info</td><td>Shipping!Address</td></tr><tr><td>Hello</td><td colspan=\"3\">Request(MakeOrder) Ack</td><td colspan=\"3\">Ack</td><td colspan=\"5\">Hello Request(MakeOrder)</td><td>Ack</td><td colspan=\"2\">Ack</td><td/><td colspan=\"2\">Hello Request(MakeOrder) Ack</td><td>Ack</td><td>Hello</td><td>Request(MakeOrder) Ack</td><td>Ack</td></tr><tr><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td/><td>can i have your</td><td>can i have your</td></tr><tr><td colspan=\"2\">thank you for calling XYZ catalog</td><td>yes i would like to place an order</td><td colspan=\"2\">yes one second please</td><td colspan=\"2\">you thank</td><td colspan=\"2\">thank you for calling XYZ catalog</td><td colspan=\"3\">to place an order yes i would like</td><td colspan=\"2\">second please yes one</td><td>you thank</td><td/><td>thank you for calling XYZ catalog</td><td>please to place an order second yes one yes i would like</td><td>thank you</td><td>number with area code home telephone</td><td>thank you for calling XYZ catalog</td><td>yes i would like to place an order</td><td>yes one second please</td><td>you thank</td><td>home telephone number with area code ......</td></tr><tr><td colspan=\"2\">this is mary</td><td/><td/><td/><td/><td/><td colspan=\"2\">this is mary</td><td/><td/><td/><td/><td/><td/><td/><td>this is mary</td><td>this is mary</td></tr><tr><td colspan=\"2\">how may I</td><td/><td/><td/><td/><td/><td colspan=\"2\">how may I</td><td/><td/><td/><td/><td/><td/><td/><td>how may I</td><td>how may I</td></tr><tr><td>help you</td><td/><td/><td/><td/><td/><td/><td colspan=\"2\">help you</td><td/><td/><td/><td/><td/><td/><td/><td>help you</td><td>help you</td></tr><tr><td/><td/><td/><td/><td/><td/><td/><td colspan=\"4\">Order Item Task</td><td/><td/><td/><td/><td/><td colspan=\"2\">.........</td><td>Order Item Task</td></tr><tr><td/><td/><td/><td/><td/><td/><td colspan=\"2\">Opening</td><td/><td/><td/><td colspan=\"4\">Contact!Info</td><td colspan=\"2\">Shipping!Address</td><td>Opening</td><td>Contact!Info</td><td>Shipping!Address</td></tr><tr><td/><td/><td colspan=\"7\">Hello Request(MakeOrder) Ack</td><td colspan=\"2\">Ack</td><td/><td/><td/><td/><td/><td/><td>Hello</td><td>Request(MakeOrder) Ack</td><td>Ack</td></tr><tr><td/><td/><td colspan=\"4\">thank you for calling XYZ catalog help you how may I this is mary</td><td colspan=\"2\">to place an order yes i would like</td><td colspan=\"2\">yes one second please</td><td>you thank</td><td colspan=\"5\">can i have your home telephone number with area code ......</td><td colspan=\"2\">yes please ...... may we deliver this order to your home</td><td>help you how may I this is mary thank you for calling XYZ catalog</td><td>yes i would like to place an order</td><td>yes one please second</td><td>thank you</td><td>can i have your home telephone number with area code ......</td><td>may we deliver this order to your home ...... yes please</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "text": ", we construct ...... Closing .........", |
| "html": null |
| }, |
| "TABREF4": { |
| "content": "<table><tr><td>Type</td><td>Subtype</td></tr><tr><td>Ask</td><td>Info</td></tr><tr><td>Explain</td><td>Catalog, CC Related, Discount, Order Info</td></tr><tr><td/><td>Order Problem, Payment Rel, Product Info</td></tr><tr><td/><td>Promotions, Related Offer, Shipping</td></tr><tr><td colspan=\"2\">Convers-Ack, Goodbye, Hello, Help, Hold,</td></tr><tr><td>-ational</td><td>YoureWelcome, Thanks, Yes, No, Ack,</td></tr><tr><td/><td>Repeat, Not(Information)</td></tr><tr><td>Request</td><td>Code, Order Problem, Address, Catalog,</td></tr><tr><td/><td>CC Related, Change Order, Conf, Credit,</td></tr><tr><td/><td>Customer Info, Info, Make Order, Name,</td></tr><tr><td/><td>Order Info, Order Status, Payment Rel,</td></tr><tr><td/><td>Phone Number, Product Info, Promotions,</td></tr><tr><td/><td>Shipping, Store Info</td></tr><tr><td>YNQ</td><td>Address, Email, Info, Order Info,</td></tr><tr><td/><td>Order Status,Promotions, Related Offer</td></tr></table>", |
| "type_str": "table", |
| "num": null, |
| "text": "Task/subtask labels in CHILD", |
| "html": null |
| }, |
| "TABREF5": { |
| "content": "<table/>", |
| "type_str": "table", |
| "num": null, |
| "text": "", |
| "html": null |
| } |
| } |
| } |
| } |