ACL-OCL / Base_JSON /prefixP /json /P01 /P01-1047.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "P01-1047",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T09:29:48.158016Z"
},
"title": "Extending Lambek grammars: a logical account of minimalist grammars",
"authors": [
{
"first": "Alain",
"middle": [],
"last": "Lecomte",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Universit\u00e9 Pierre Mend\u00e8s-France",
"location": {
"addrLine": "BSHM -1251 Avenue Centrale, Domaine Universitaire de St Martin d'H\u00e8res BP 47 -38040",
"settlement": "GRENOBLE cedex 9",
"country": "France"
}
},
"email": "alain.lecomte@upmf-grenoble.fr"
},
{
"first": "Christian",
"middle": [],
"last": "Retor\u00e9",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Universit\u00e9 de Nantes",
"location": {
"addrLine": "2, rue de la Houssini\u00e8re",
"postBox": "BP 92208",
"postCode": "44322",
"settlement": "Nantes cedex 03",
"country": "France"
}
},
"email": "retore@irisa.fr"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "We provide a logical definition of Minimalist grammars, that are Stabler's formalization of Chomsky's minimalist program. Our logical definition leads to a neat relation to categorial grammar, (yielding a treatment of Montague semantics), a parsing-asdeduction in a resource sensitive logic, and a learning algorithm from structured data (based on a typing-algorithm and type-unification). Here we emphasize the connection to Montague semantics which can be viewed as a formal computation of the logical form.",
"pdf_parse": {
"paper_id": "P01-1047",
"_pdf_hash": "",
"abstract": [
{
"text": "We provide a logical definition of Minimalist grammars, that are Stabler's formalization of Chomsky's minimalist program. Our logical definition leads to a neat relation to categorial grammar, (yielding a treatment of Montague semantics), a parsing-asdeduction in a resource sensitive logic, and a learning algorithm from structured data (based on a typing-algorithm and type-unification). Here we emphasize the connection to Montague semantics which can be viewed as a formal computation of the logical form.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "The connection between categorial grammars (especially in their logical setting) and minimalist grammars, which has already been observed and discussed (Retor\u00e9 and Stabler, 1999) , deserve a further study: although they both are lexicalized, and resource consumption (or feature checking) is their common base, they differ in various respects. On the one hand, traditional categorial grammar has no move operation, and usually have a poor generative capacity unless the good properties of a logical system are damaged, and on the other hand minimalist grammars even though they were provided with a precise formal definition (Stabler, 1997) , still lack some computational properties that are crucial both from a theoretical and a practical viewpoint. Regarding applications, one needs parsing, generation or learning algorithms, and, considering more conceptual aspects, such algorithms are needed too to validate or invalidate linguistic claims regarding economy or efficiency. Our claim is that a logical treatment of these grammars leads to a simpler description and well defined computational properties. Of course among these aspects the relation to semantics or logical form is quite important; it is claimed to be a central notion in minimalism, but logical forms are rather obscure, and no computational process from syntax to semantics is suggested. Our logical presentation of minimalist grammar is a first step in this direction: to provide a description of minimalist grammar in a logical setting immediately set up the computational framework regarding parsing, generation and even learning, but also yields some good hints on the computational connection with logical forms.",
"cite_spans": [
{
"start": 152,
"end": 178,
"text": "(Retor\u00e9 and Stabler, 1999)",
"ref_id": "BIBREF4"
},
{
"start": 625,
"end": 640,
"text": "(Stabler, 1997)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Presentation",
"sec_num": "1"
},
{
"text": "The logical system we use, a slight extension of (de Groote, 1996) , is quite similar to the famous Lambek calculus (Lambek, 1958) , which is known to be a neat logical system. This logic has recently shown to have good logical properties like the subformula property which are relevant both to linguistics and computing theory (e.g. for modeling concurrent processes). The logic under consideration is a super-imposition of the Lambek calculus (a non commutative logic) and of intuitionistic multiplicative logic (also known as Lambek calculus with permutation). The context, that is the set of current hypotheses, are endowed with an order, and this order is crucial for obtaining the expected order on pronounced and interpreted features but it can also be relaxed when necessary: that is when its effects have already been recorded (in the labels) and the corresponding hypotheses can therefore be discharged.",
"cite_spans": [
{
"start": 53,
"end": 66,
"text": "Groote, 1996)",
"ref_id": "BIBREF1"
},
{
"start": 116,
"end": 130,
"text": "(Lambek, 1958)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Presentation",
"sec_num": "1"
},
{
"text": "Having this logical description of syntactic analyses allows to reduce parsing (and production) to deduction, and to extract logical forms from the proof; we thus obtain a close connection between syntax and semantics as the one between Lambek-style analyses and Montague semantics.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Presentation",
"sec_num": "1"
},
{
"text": "The general picture of these logical grammars is as follows. A lexicon maps words (or, more generally, items) onto a logical formula, called the (syntactic) type of the word. Types are defined from syntactic of formal features (which are propositional variables from the logical viewpoint):",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The grammatical architecture",
"sec_num": "2"
},
{
"text": "\u00a1 categorial features (categories) involved in merge: BASE \u00a2 \u00a4 \u00a3 \u00a6 \u00a5 \u00a7 \u00a9 \u00a7 \u00a7 \u00a7 \u00a7 \" ! \u00a1 functional features involved in move: FUN \u00a2 \u00a4 \u00a3# \u00a7$ \u00a7% ' & \u00a7 ( !",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The grammatical architecture",
"sec_num": "2"
},
{
"text": "The connectives in the logic for constructing formulae are the Lambek implications (or slashes) ) \u00a7 1 0 together with the commutative product of linear logic",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The grammatical architecture",
"sec_num": "2"
},
{
"text": ". 1 Once an array of items has been selected, a sentence (or any phrase) is a deduction of IP (or of the phrasal category) under the assumptions provided by the syntactic types of the involved items. This first step works exactly as Lambek grammars, except that the logic and the formulae are richer. Now, in order to compute word order, we proceed by labeling each formula in the proof. These labels, that are called phonological and semantic features in the transformational tradition, are computed from the proofs and consist of two parts that can be superimposed: a phonological label, denoted by 0 4 3 6 5 8 7 \u00a6 9 @ 0",
"cite_spans": [
{
"start": 2,
"end": 3,
"text": "1",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "2",
"sec_num": null
},
{
"text": ", and a semantic label 2 denoted by A B 3 6 5 8 7 \u00a6 9 @ C -the super-imposition of both 1 The logical system also contains a commutative implication,",
"cite_spans": [
{
"start": 88,
"end": 89,
"text": "1",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "2",
"sec_num": null
},
{
"text": ", and a non commutative product F b ut they do not appear in the lexicon, and because of the subformula property, they are not needed for the proofs we use. 2 We prefer semantic label to logical form not to confuse logical forms with the logical formulae present at each node of the proof. label being denoted by 3 6 5 8 7 \u00a6 9",
"cite_spans": [
{
"start": 157,
"end": 158,
"text": "2",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": ". The reason for having such a double labeling, is that, as usual in minimalism, semantic and phonological features can move separately. It should be observed that the labels are not some extraneous information; indeed the whole information is encoded in the proof, and the labeling is just a way to extract the phonological form and the logical form from the proof.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "We rather use chains or copy theory than movements and traces: once a label or one aspect (semantic or phonological) has been met it should be ignored when it is met again. For instance a label",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "G I H Q P R H 7 S A U T \u00a4 V W 7 Y X C b a c 5 8 d H 4 e T f V ' 7 Y X",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "corresponds to a semantic label",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "A G I H Q P R H 7 g C h A U T f V ' 7 Y X i C p A c a c 5 8 d H C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "and to the phonological form",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "0 G I H P H 7 W 0 q 0 8 a c 5 4 d H r e 0 s 0 t T f V ' 7 Y X i 0 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "D \u00a6 E",
"sec_num": null
},
{
"text": "Because of the sub-formula property we need not present all the rules of the system, but only the ones that can be used according to the types that appear in the lexicon. Further more, up to now there is no need to use introduction rules (called hypothetical reasoning in the Lambek calculus): so our system looks more like Combinatory Categorial Grammars or classical ABgrammars. Nevertheless some hypotheses can be cancelled during the derivation by the productelimination rule. This is essential since this rule is the one representing chains or movements. We also have to specify how the labels are carried out by the rules. At this point some non logical properties can be taken into account, for instance the strength of the features, if we wish to take them into account. They are denoted by lower-case variables. The rules of this system in a Natural Deduction format are: x . In the figure 1, the reader is provided with an example of a lexicon and of a derivation. The resulting label is ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "u w v y x ' 0 8 v X 0 8 h u v x X W v X u w v y x ) ) p u v X x W u A U y Q C v H P 7 Y 5 1 d e X u A U y Q \u00a7 f Q C v u w v g h g 2 i \u00a7 x W \u00a7 X \u00a7 f j v k l @ m 2 s p \u00a7 u \u00a7 f j v n k g 0 W \u00a3 x \u00a7 X ! @ m",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "A c V \u00a1 5 Y 5 \u00a3 \u00a2 C 7 H V @ 9 e V \u00a4 5 8 5 \u00a3 \u00a2 phonologi- cal form is 0 4 7 H V ' 9 e 0 p 0 8 V \u00a5 5 8 5 \u00a3 \u00a2 0 while the resulting logical form is A c V \u00a6 5 Y 5 \u00a3 \u00a2 C A B 7 H V",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "V \u00a5 5 Y 5 \u00a3 \u00a2 S 7 H V ' 9 e V \u00a5 5 8 5 \u00a3 \u00a2 which yields the phonolog- ical from 0 8 V \u00a7 1 5 Y 5 \u00a3 \u00a2 0 s 0 4 7 H V",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "A c V \u00a6 5 Y 5 \u00a3 \u00a2 C A B 7 H V @ 9 e C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": ". Observe that although entropy which suppresses some order has been used, the labels consist in ordered sequences of phonological and logical forms. It is so because when using [/ E] and [ ) E], we necessarily order the labels, and this order is then recorded inside the label and is never suppressed, even when using the entropy rule: at this moment, it is only the order on hypotheses which is relaxed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "In order to represent the minimalist grammars of (Stabler, 1997) , the above subsystem of partially commutative intuitionistic linear logic (de Groote, 1996) is enough and the types appearing in the lexicon also are a strict subset of all possible types: gives a co-indexation of two nodes that we can see as a move step. For instance in a tree presentation of natural deduction, we shall only keep the coindexation (corresponding to the cancellation of and : this is harmless since the conclusion is not modified, and makes our natural deduction T-markers).",
"cite_spans": [
{
"start": 49,
"end": 64,
"text": "(Stabler, 1997)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "A A ) A ) \" A ) A S t 2 \u00a6 s r 2 2 ! # \" p 2 C C C C 0 $ t C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "Such lexical entries, when processed wit\u1e27 \u00a9 -rules encompass Stabler minimalist grammars; this system nevertheless overgenerates, because some minimalist principles are not yet satisfied: they correspond to constraints on derivations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logico-grammatical rules for merge and phrasal movement",
"sec_num": "3"
},
{
"text": "The restriction which is still lacking concerns the way the proofs are built. Observe that this is an algorithmic advantage, since it reduces the search space.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conditions on derivations",
"sec_num": "3.1"
},
{
"text": "The simplest of these restriction is the following: the attractor F in the label L of the target % locates the closest F' in its domain. This simply corresponds to the following restriction. is correct.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conditions on derivations",
"sec_num": "3.1"
},
{
"text": "V ' 9 e \u00a2 v 7 H V @ 9 e I A A# ) \u00a1 C 0 8 C V \u00a2 v V A A c p 2 # C 0 Q C 1 5 Y 5 \u00a3 \u00a2 \u00a2 v 5 Y 5\u00a2 v V A A c h 2 # i C 0 Q C v 5 Y 5 \u00a3 \u00a2 0 8 p v V \u00a6 5 Y 5 \u00a3 \u00a2 p 2 # X # v X # v 7 H V @ 9 e S A A# ) \u00a2 C 0 8 C x v y x 0 8 h x v 7 H V ' 9 e x A# ) \u00a2 e C ) p X # x v X p 7 H V @ 9 e x \u00a1 H Q P 7 \u00a6 5 1 d e X ' X # \u00a7 x v X p 7 H V @ 9 e x \u00a1 2 p p v A c V 5 Y 5 \u00a3 \u00a2 C 7 H V @ 9 e V ! 1 5 Y 5 \u00a3 \u00a2 \u00a2 1. \u00a3 \u00a5 \u00a4 ... \u00a3 \u00a6 \u00a4 ... \u00a3 Q ... v C is forbidden 2. \u00a1 if there is a sequent ... \u00a3 ... v \u00a3 j ) C \u00a1 if there is a type \u00a3 j s uch that u w v \u00a3 j2 \u00a7 \u00a3",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conditions on derivations",
"sec_num": "3.1"
},
{
"text": "We have seen above that we are able to account for SVO and SOV orders quite easily. Nevertheless we could not handle this way VSO language. Indeed this order requires head-movement.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Extension to head-movement",
"sec_num": "3.2"
},
{
"text": "In order to handle head-movement, we shall also use the product 2 but between functor types. As a first example, let us take the very simple example of: peter loves mary. Starting from the following lexicon in figure 3 we can build the tree given in the same figure; it represents a natural deduction in our system, hence a syntactic analysis. The resulting phonological form is",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Extension to head-movement",
"sec_num": "3.2"
},
{
"text": "0 G I H Q P R H 7 W 0 \u00a6 0 8 a B 5 8 d H r e 0 \u00a6 0 t T f V ' 7 Y X i 0",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Extension to head-movement",
"sec_num": "3.2"
},
{
"text": "while the resulting logical form is ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Extension to head-movement",
"sec_num": "3.2"
},
{
"text": "A G I H Q P R H 7 g C t A U T \u00a4 V ' 7 8 X C t A c",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Extension to head-movement",
"sec_num": "3.2"
},
{
"text": "In categorial grammar (Moortgat, 1996) , the production of logical forms is essentially based on the association of pairs e r P 7 \u00a5 \u00a7 P X 4 d H with lambda terms representing the logical form of the items, and on the application of the Curry-Howard homomorphism: each (0 or ) )elimination rule translates into application and each introduction step into abstraction. Compositionality assumes that each step in a derivation is associated with a semantical operation.",
"cite_spans": [
{
"start": 22,
"end": 38,
"text": "(Moortgat, 1996)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The interface between syntax and semantics",
"sec_num": "4"
},
{
"text": "In generative grammar (Chomsky, 1995) , the production of logical forms is in last part of the derivation, performed after the so-called Spell Out point, and consists in movements of the semantical features only. Once this is done, two forms can be extracted from the result of the derivation: a phonological form and a logical one.",
"cite_spans": [
{
"start": 22,
"end": 37,
"text": "(Chomsky, 1995)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The interface between syntax and semantics",
"sec_num": "4"
},
{
"text": "These two approaches are therefore very differ- ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The interface between syntax and semantics",
"sec_num": "4"
},
{
"text": "Figure 2: Complex NP constraint X g # A A# ) \u00a9 C 0 4 \u00a1 C x \u00a4 # I 2 i \u00a1 X #",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The interface between syntax and semantics",
"sec_num": "4"
},
{
"text": "\u00a3 \u00a2 C 0 4 \u00a1 C b 2 A A# ) A c ) \u00a1 C C 0 8 i C d H Q P R H 7 \u00a2 v d H P H 7 # I 2 i \u00a9 V ' 7 Y X \u00a2 v \u00a9 V W 7 Y X # h 2 \u00a2 peter # A# ) \u00a3 \u00a2 e C loves\u00a4 A A# ) \u00a3 \u00a2 e C 0 4 \u00a2 C \u00a1 A c ) \u00a2 C (mary) # A# ) A c ) \u00a2 C C (to love) A A# ) A c ) \u00a1 C C 0 8 C mary",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The interface between syntax and semantics",
"sec_num": "4"
},
{
"text": "ent, but we can try to make them closer by replacing semantic features by lambda-terms and using some canonical transformations on the derivation trees. Instead of converting directly the derivation tree obtained by composition of types, something which is not possible in our translation of minimalist grammars, we extract a logical tree from the previous, and use the operations of Curry-Howard on this extracted tree. Actually, this extracted tree is also a deduction tree: it represents the proof we could obtain in the semantic component, by combining the semantic types associated with the syntactic ones (by a homomorphism to specify). Such a proof is in fact a proof in implicational intuitionistic linear logic.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The interface between syntax and semantics",
"sec_num": "4"
},
{
"text": "Coindexed nodes refer to ancient hypotheses which have been discharged simultaneously, thus resulting in phonological features and semantical ones at their right place 3 . By extracting the subtree the leaves of which are full of semantic content, we obtain a structure that can be easily seen as a composition:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logical form for example 3",
"sec_num": "4.1"
},
{
"text": "(peter)((mary)(to love)) If we replace these \"semantic features\" by \u00a1 terms, we have:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logical form for example 3",
"sec_num": "4.1"
},
{
"text": "A \u00a2 \u00a1 \u00a4 \u00a3 \u00a5 \u00a3 Ad H Q P R H 7 W C 1 \u00a7 r A \u00a2 \u00a1 \u00a4 \u00a3 \u00a5 \u00a3 A\u00a9 V ' 7 8 X C 1 \u00a7 \u00a6 \u00a1 x \u00a7 \u00a1 X a c 5 8 d H A B X \u00a7 x C C C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logical form for example 3",
"sec_num": "4.1"
},
{
"text": "This shows that necessarily raised constituants in the structure are not only \"syntactically\" raised but also \"semantically\" lifted, in the sense that",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logical form for example 3",
"sec_num": "4.1"
},
{
"text": "\u00a1 \u00a4 \u00a3 \u00a5 \u00a3 Ad H Q P R H 7 W C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logical form for example 3",
"sec_num": "4.1"
},
{
"text": "is the high order representation of the individual peter 4 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Logical form for example 3",
"sec_num": "4.1"
},
{
"text": "Let us look at now the example: mary seems to work From the lexicon in figure 4 we obtain the deduction tree given in the same figure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "3 For the time being, we make abstraction of the representation of time, mode, aspect... that would be supported by the inflection category. 4 It is important to notice that if we consider\u00a9",
"cite_spans": [
{
"start": 141,
"end": 142,
"text": "4",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "\u00a9 \u00a6 ! # \"",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "a typed lambda term, we must only assume it is of some type freely raised from $ , something we can represent by",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "% $ ' & ) ( 0 \" 1 & ) ( 0 \"",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": ", where X is a type-variable, here X = $ 2 & 4 3",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "5 \"",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "because6 \u00a8 ! 7 8 @ 9 B A C D ! F E G 6 7 \" has type $ 2 & H $ 2 & 4 3 I \" % \"",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "This time, it is not so easy to obtain the logical representation : e 8 H r H \u00a9 A P 5 3 6 5 8 7\u00a2 A\u00a9 ",
"cite_spans": [],
"ref_spans": [
{
"start": 66,
"end": 113,
"text": ": e 8 H r H \u00a9 A P 5 3 6 5 8 7\u00a2 A\u00a9",
"ref_id": null
}
],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "V W 7 Y X C C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "The best way to handle this situation consists in assuming that: just before the semantic content of the specifier (here the nominative position, occupied by",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "\u00a1 \u00a4 \u00a3 \u00a5 \u00a3 A\u00a9 V ' 7 8 X C ) applies.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "This shows that the semantic tree we want to extract from the derivation tree in types logic is not simply the subtree the leaves of which are semantically full. We need in fact some transformation which is simply the stretching of some nodes. These stretchings correspond to P -introduction steps in a Natural deduction tree. They are allowed each time a variable has been used before, which is not yet discharged and they necessarily occur just before a semantically full content of a specifier node (that means in fact a node labelled by a functional feature) applies. Actually, if we say that the tree so obtained represents a deduction in a natural deduction format, we have to specify which formulae it uses and what is the conclusion formula. We must therefore define a homomorphism between syntactic and semantic types. Let be this homomorphism.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "We shall assume:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "\u00a1 ( \u00a2 )=t, ( \u00a1 )Q \u00a3 t ,A G R S P U T 8 C f ! , ( )=e, \u00a1 w A \u00a2 V ) W C = w A W 0 F V ' C = A H A \u00a2 V @ C X P HA W C C , \u00a1 \u1ef2 a HA a C b Q \u00a3 ' A A G R c P e d C f P ) d C 1 \u00a7 r A G d g P ) d C f ! 5 5",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "X is a variable of type. This may appear as nondeterminism but the instantiation of X is always unique. Moreover, when",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "h is of type i p & ) i q \"",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": ", it is in fact endowed with the identity function, something which happens everytime h is linked by a chain to a higher node. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "C 0 4 \u00a1 C b 2 A B \u00a2 0 4 \u00a1 e C \u00a9 V ' 7 Y X \u00a2 v \u00a9 V W 7 Y X s 2 # P 5 3 q 5 4 7\u00a2 \u00a2 v y P 5 3 6 5 8 7\u00a2 A c ) \u00a1 C \u00a2 mary # A# ) \u00a3 \u00a2 C seems A A# ) \u00a3 \u00a2 C 0 4 \u00a2 e C \u00a1 (to seem) A B \u00a1 0 4 \u00a1 C \u00a1 to work A c ) \u00a1 C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "With this homomorphism of labels, the transformation of trees consisting in stretching \"intermediary projection nodes\" and erasing leaves without semantic content, we obtain from the derivation tree of the second example, the following \"semantic\" tree:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "seem(to work(mary)) T\u00a8 \u00a9 \u00a9 0 B ! # \" A A G R S P U T 8 C X P U T 8 C 6 \u00a1 I C 5 C I \u00a2 9 \u00a3 9 \u00a6 \u00a5 \u00a4 6 F \" % \" A G R c P U T 8 C \u1e97 A \u00a1 I C B C I A \" AT 0 P ) T 8 C to work(x) T\u00a8 ! F \u00a2 9 \u00a3 9 B \u00a5 \u00a4 ! # \" A G R c P ) T 8 C x R",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "where coindexed nodes are linked by the discharging relation. Let us notice that the characteristic weak or strong of the features may often be encoded in the lexical entries. For instance, Head-movement from V to I is expressed by the fact that tensed verbs are such that: Unfortunately, such rigid assignment does not always work. For instance, for phrasal movement (say of a to a # ) that depends of course on the particular # -node in the tree (for instance the situation is not necessary the same for nominative and for accusative case). In such cases, we may assume that multisets are associated with lexical entries instead of vectors.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Subject raising",
"sec_num": "4.2"
},
{
"text": "Let us try now to enrich this lexicon by considering other phenomena, like reflexive pronouns. . We obtain for paul shaves himself as the syntactical tree something similar to the tree obtained for our first little example (peter loves mary), and the semantic tree is given in figure 5.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Reflexives",
"sec_num": "4.3"
},
{
"text": "In our setting, parsing is reduced to proof search, it is even optimized proof-search: indeed the re- ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Remarks on parsing and learning",
"sec_num": "5"
},
{
"text": "\u00a6 \u00a9 F 8 \" A A G R c P ) T Y C X P U T 8 C # \u00a1 # A C D # E D \" A G R S P U T 8 C shave(z,z) T z R # \u00a1 A C E \u00a9 D \" A G R S P U T Y C\u00a8 \u00a9 \u00a8 # \u00a9 E D \" A A G R P A G R P U T Y C C P A G R S P U T 8 C C 6 \u00a8 ! F \u00a1 # A C D ! F E G 6 F \" A G R P A G R P U T Y C \u00a4 C\u00a8 ! F \u00a1 # A C D ! F E G 6 F \" A G R S P U T 8 C\u00a8 6 ! F \u00a1 A C D ! 7 E 6 7 \" A G R c P A G R c P ) T 8 C C 6 R \u00a4",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Remarks on parsing and learning",
"sec_num": "5"
},
{
"text": "striction on types, and on the structure of proof imposed by the shortest move principle and the absence of introduction rules considerably reduce the search space, and yields a polynomial algorithm. Nevertheless this is so when traces are known: otherwise one has to explore the possible places of theses traces.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Remarks on parsing and learning",
"sec_num": "5"
},
{
"text": "Here we did focus on the interface with semantics. Another excellent property of categorial grammars is that they allow -especially when there are no introduction rules -for learning algorithms, which are quite efficient when applied to structured data. This kind of algorithm applies here as well when the input of the algorithm are derivations.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Remarks on parsing and learning",
"sec_num": "5"
},
{
"text": "In this paper, we have tried to bridge a gap between minimalist program and the logical view of categorial grammar. We thus obtained a description of minimalist grammars which is quite formal and allows for a better interface with semantics, and some usual algorithms for parsing and learning.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6"
},
{
"text": "as long we don't take a semantical representation of tense and aspect in consideration.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "The minimalist program",
"authors": [
{
"first": "Noam",
"middle": [],
"last": "Chomsky",
"suffix": ""
}
],
"year": 1995,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Noam Chomsky. 1995. The minimalist program. MIT Press, Cambridge, MA.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Partially commutative linear logic",
"authors": [
{
"first": "Philippe",
"middle": [],
"last": "De",
"suffix": ""
},
{
"first": "Groote",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": 1996,
"venue": "Third Roma Workshop: Proofs and Linguistics Categories",
"volume": "",
"issue": "",
"pages": "199--208",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Philippe de Groote. 1996. Partially commutative lin- ear logic. In M. Abrusci and C. Casadio, editors, Third Roma Workshop: Proofs and Linguistics Cat- egories, pages 199-208. Bologna:CLUEB.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "The mathematics of sentence structure",
"authors": [
{
"first": "Joachim",
"middle": [],
"last": "Lambek",
"suffix": ""
}
],
"year": 1958,
"venue": "American mathematical monthly",
"volume": "65",
"issue": "",
"pages": "154--169",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Joachim Lambek. 1958. The mathematics of sen- tence structure. American mathematical monthly, 65:154-169.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Categorial type logic",
"authors": [
{
"first": "Michael",
"middle": [],
"last": "Moortgat",
"suffix": ""
}
],
"year": 1996,
"venue": "Handbook of Logic and Language",
"volume": "",
"issue": "",
"pages": "93--177",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Michael Moortgat. 1996. Categorial type logic. In J. van Benthem and A. ter Meulen, editors, Hand- book of Logic and Language, chapter 2, pages 93- 177. North-Holland Elsevier, Amsterdam.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Resource logics and minimalist grammars: introduction to the ESSLLI workshop. To appear in Language and Computation",
"authors": [
{
"first": "Christian",
"middle": [],
"last": "Retor\u00e9",
"suffix": ""
},
{
"first": "Edward",
"middle": [],
"last": "Stabler",
"suffix": ""
}
],
"year": 1999,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Christian Retor\u00e9 and Edward Stabler. 1999. Re- source logics and minimalist grammars: intro- duction to the ESSLLI workshop. To ap- pear in Language and Computation RR-3780",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Derivational minimalism",
"authors": [
{
"first": "Edward",
"middle": [],
"last": "Stabler",
"suffix": ""
}
],
"year": 1997,
"venue": "LNCS/LNAI",
"volume": "1328",
"issue": "",
"pages": "68--95",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Edward Stabler. 1997. Derivational minimalism. In Christian Retor\u00e9, editor, LACL'96, volume 1328 of LNCS/LNAI, pages 68-95. Springer-Verlag.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"text": "Definition 1\u00a9 -proofs contain only three kinds of steps: \u00a1 implication steps (elimination rules for / and ) ) \u00a1 tensor steps (elimination rule for 2 ) \u00a1 entropy steps (entropy rule) Definition 2 A lexical entry consists in an axiom v 3 where is a type:",
"type_str": "figure",
"num": null,
"uris": null
},
"FIGREF1": {
"text": "Shortest Move) : A\u00a9 -proof is said to respect the shortest move condition if it is such that: \u00a1 the same formula never occurs twice as a hypothesis of any sequent \u00a1 every active hypothesis during the proof process is discharged as soon as possibleThe consequences of this definition are the following:",
"type_str": "figure",
"num": null,
"uris": null
},
"FIGREF2": {
"text": "the verbal infinitive head (here to work) the main verb (here to seem) applies to the result, in order to obtain",
"type_str": "figure",
"num": null,
"uris": null
},
"FIGREF3": {
"text": "Figure 4: Mary seems to work",
"type_str": "figure",
"num": null,
"uris": null
},
"FIGREF4": {
"text": "the full phonology is associated with the inflection component, \u00a1 the empty phonology and the semantics are associated with the second one, \u00a1 the empty semantics occupies the first one 6",
"type_str": "figure",
"num": null,
"uris": null
},
"FIGREF5": {
"text": "The assignment for himself is given in figure 5 -where the semantical type of himself is assumed to be",
"type_str": "figure",
"num": null,
"uris": null
},
"TABREF0": {
"content": "<table><tr><td>This later rule encodes movement and deserves special attention. The label k g 0 W \u00a3 x \u00a7 X ! means the substitution of g to the unordered set \u00a3 x , X ! that is the simultaneous substitution of g for both x and X , no matter the order between x</td></tr><tr><td>versus strong case cide that only the semantic part that is $ ) , it is possible to de-A g i s sub-C stituted with</td></tr></table>",
"text": "Here some non logical but linguistically motivated distinction can be made. For instance according to the strength of a feature (e.g. weak case #",
"type_str": "table",
"html": null,
"num": null
}
}
}
}