ACL-OCL / Base_JSON /prefixP /json /P84 /P84-1022.json
Benjamin Aw
Add updated pkl file v3
6fa4bc9
{
"paper_id": "P84-1022",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T08:20:50.586742Z"
},
"title": "A PARSING ARCHITECTURE BASED ON DISTRIBUTED MEMORY MACHINES",
"authors": [
{
"first": "Jon",
"middle": [
"M"
],
"last": "Slack",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Open University Milton",
"location": {
"postCode": "MK7 6AA",
"settlement": "Keynes",
"country": "ENGLAND"
}
},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "The paper begins by defining a class of distributed memory machines which have useful properties as retrieval and filtering devices. These memory mechanisms store large numbers of associations on a single composite vector. They provide a natural format for encoding the syntactic and semantic constraints associated with linguistic elements. A computational architecture for parsing natural language is proposed which utillses the retrieval and associative features of these devices. The parsing mechanism is based on the principles of Lexlcal Functional Grammar and the paper demonstrates how these principles can be derived from the properties of the memory mechanisms.",
"pdf_parse": {
"paper_id": "P84-1022",
"_pdf_hash": "",
"abstract": [
{
"text": "The paper begins by defining a class of distributed memory machines which have useful properties as retrieval and filtering devices. These memory mechanisms store large numbers of associations on a single composite vector. They provide a natural format for encoding the syntactic and semantic constraints associated with linguistic elements. A computational architecture for parsing natural language is proposed which utillses the retrieval and associative features of these devices. The parsing mechanism is based on the principles of Lexlcal Functional Grammar and the paper demonstrates how these principles can be derived from the properties of the memory mechanisms.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Recently, interest has focussed on computational architectures employing massively parallel processing lip2].",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "I INTRODUCTION",
"sec_num": null
},
{
"text": "Some of these systems have used a distributed form of knowledge representation [3] . This type of representation encodes an item of knowledge in terms of the relationships among a collection of elementary processing units, and such assemblages can encode large numbers of items. Representational similarity and the ability to generalize are the principal features of such memory systems.",
"cite_spans": [
{
"start": 79,
"end": 82,
"text": "[3]",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "I INTRODUCTION",
"sec_num": null
},
{
"text": "The next section defines a distributed memory machine which incorporates some of the computational advantages of distributed representations within a traditional yon Neumann architecture.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "I INTRODUCTION",
"sec_num": null
},
{
"text": "The rest of the paper explores the properties of such machines as the basis for natural language parsing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "I INTRODUCTION",
"sec_num": null
},
{
"text": "Distributed memory machines (DMM) can be represented formally by the septuple DMM=(V,X,Y,Q,qo,p,A) , where V is a finite set denoting the total vocabulary; X is a finite set of inputs, and XGV; Y is a finite set of acceptable outputs and Y~V; Q is a set of internal states;",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "II DISTRIBUTED MEMORY MACHINES",
"sec_num": null
},
{
"text": "q0 is a distinguished initial state; ~.QxX-->Q, the retrieval function; A:Q-->Qxy, the output function. Further, where Y\" denotes the set of all finite concatenations of the elements of the set Y, Q~Y', and therefore QgV'.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "II DISTRIBUTED MEMORY MACHINES",
"sec_num": null
},
{
"text": "This statement represents the notion that internal states of DMMs can encode multiple outputs or hypotheses. The vocabulary, V, can be represented by the space I k, where I is some interval range defined within a chosen number system, N; IoN. The elements of X, Y and Q are encoded as k-element vectors, referred to as memory vectozs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "II DISTRIBUTED MEMORY MACHINES",
"sec_num": null
},
{
"text": "One form of DMM is the holographic associative memory [4, 5, 6] which encodes large numbers of associations on a single composite vector. This method of association formation is fundamental to the concept of holographic memory and the resulting associative trace is denoted <A>*<B>.",
"cite_spans": [
{
"start": 54,
"end": 57,
"text": "[4,",
"ref_id": "BIBREF4"
},
{
"start": 58,
"end": 60,
"text": "5,",
"ref_id": "BIBREF5"
},
{
"start": 61,
"end": 63,
"text": "6]",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "The operation of convolution is define by the equation (<A>*<B>~=.~AIB~. i and has the following propertles* [7] :",
"cite_spans": [
{
"start": 109,
"end": 112,
"text": "[7]",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "Commutative: <A>*<B> = <B>*<A>, Associative:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "<A>*(<B>*<C>) = (<A>*<B>)*<C>.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "Further, where a delta vector, denoted ~, is defined as a vector that has values of zero on all features except the central feature, which has a value of one, then <A>* ~ffi <A>. Moreover, <A>*0 ffi 0, where 0 is a zero vector in which all feature values are zero.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "Convolving an item wlth an attenuated delta vector (i.e., a vector with values of zero on all features except the central one, which has a value between 0 and i) produces the original item with a strength that is equal to the value of the central feature of the attenuated delta vector.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "The initial state, qo, encodes all the associations stored in the machine.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "In this model, associative traces are concatenated (+) through the operations of vector addition and normalization to produce a single vector. Overlapping associative items produce composite vectors which represent both the range of items stored and the central tendency of the those items.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "This form of prototype generation is a basic property of distributed memories.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "The retrieval function,@ , is simulated by the operation of correlation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "If the state, q~, encodes the association <A>*<B>, then presenting say <A> as an input, or retrieval key, produces a new state, q{~, which encodes the item <B>', a noisy version of <B>, under the operation of correlation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "This operation is defined by the equation (<A>#<B>~=~A%Bm,%and has the following properties: % An item correlated with itself, autocorrelation, produces an approximation to a delta vector.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "If two similar memory vectors are correlated, the central feature of the resulting vector will be equal to their similarity, or dot product, producing an attenuated delta vector.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "If the two items are completely independent, correlation produces a zero vector.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Holographic Associative Memory",
"sec_num": null
},
{
"text": "<A>~(<A>*<B>) = (<A>~<A>)*<B> + (<A>~<B>)*<A> + noise ...(I) where the noise component results from some of the less significant cross products.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "Assuming that <A> and <B> are unrelated, Equation (I) becomes:",
"cite_spans": [
{
"start": 50,
"end": 53,
"text": "(I)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "<AMI(<A>*<B>) = ~*<B> + 0*<A> + noise -<B> + 0 + noise Extending these results to a composite trace, suppose that q encodes two associated pairs of four unrelated items forming the vector (<A>*<B> + <C>*<D>).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "When <A> is given as the retrieval cue, the reconstruction can be characterized as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "<A>~(<A>*<B> + <C>*<D>) = (<A>~t<A>)*<B> + (<A>~<B>)*<A> + noise + (<A>~<C>)*<D> + (<A>@<D>)*<C> + noise = ~ *<B>+0*<A>+noise+O*<D>+O*<C>+noise -<B> + noise + noise When the additional unrelated items are added to the memory trace their affect on retrieval is to add noise to the reconstructed item <B>, which was associated with the retrieval cue.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "In a situation in which the encoded items are related to each other, the composite trace causes all of the related items to contribute to the reconstructed pattern, in addition to producing noise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "The amount of noise added to a retrieved item is a function of both the amount of information held on the composite memory vector and the size of the vector.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The relation between convolution and correlation is given by",
"sec_num": null
},
{
"text": "The computational properties of distributed memory machines (DMM) make them natural mechanisms for case-frame parsing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "Consider a DMM which encodes case-frame structures of the following form: <Pred>*(<Cl>*<Pl> + <C2>*<P2> + ...+ <Cn>*<Pn>) where <Pred> is the vector representing the predicate associated with the verb of an input clause;",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "<C1> to <Cn> are the case vectors such as <agent>, <instrument>, etc., and <PI> to <Pn> are vectors representing prototype concepts which can fill the associated cases. These structures can be made more complex by including tagging vectors which indicate such features as obligatory case, as shown in the case-frame vector for the predicate BREAK:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "(<agent>*<anlobJ+natforce> + <obJect>*<physobJ> *<obllg> + <instrument>*<physobJ>) In this example, the object case has a prototype covering the category of physical objects, and is tagged as obligatory.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "The initial state of the DMM, qo, encodes the concatenation of the set of case-frame vectors stored by the parser.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "The system receives two types of inputs, noun concept vectors representing noun phrases, and predicate vectors representing the verb components.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "If the system is in state qo only a predicate vector input produces a significant new state representing the case-frame structure associated with it. Once in this state, noun vector inputs identify the case slots they can potentially fill as illustrated in the following example:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "In parsing the sentence Fred broke the window with e stone, the input vector encodin E broke will retrieve the case-frame structure for break given above. The input of <Fred> now gives <Fred>~q<agent>*<Pa>+<obJ>*<Po>+<instr>*<Pi>) \" <Fred>g<agent>*<Pa>+<Fred>~<Pa>*<agent> + ... -0*<Pa>+ee*<agent> O*<Po>+e@*<obJ> + O*<Pi>+e%*<instr> : e~agent> + e~obJ> + es<instr> where ej is a measure of the similarity between the vectors, and underlying concepts, <Fred> and the case prototype <Pj>.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "In this example, <Fred> would be identified as the agent because e 0 and e~ would be low relative to ee. The vector is \"cleaned-up\" by a threshold function which is a component of the output function,)%. This process is repeated for the other noun concepts in the sentence, linking <window> and <stone> with the object and instrument cases, respectively.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "However, the parser requires additional machinery to handle the large set of sentences in which the case assignment is ambiguous using semantic knowledge alone.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "A. Case-Frame Parsing",
"sec_num": null
},
{
"text": "Unambiguous case assignment can only be achieved through the integration of syntactic and semantic processing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "Moreover, an adequate parser should generate an encoding of the grammatical relations between sentential elements in addition to a semantic representation. The rest of the paper demonstrates how the properties of DMMs can be combined with the ideas embodied in the theory of Lextcal-functional CTammar (LFG) [8] in a parser which builds both types of relational structure.",
"cite_spans": [
{
"start": 308,
"end": 311,
"text": "[8]",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "In LFG the mapping between grammatical and semantic relations is represented directly in the semantic form of the lexlcal entries for verbs. For example, the lexlcal entry for the verb hands is given by hands: V, #participle) = NONE #tense) = PRESENT (tsubJ hum) = SO ~pred)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "= HAND[@subJ)#obj2)@obJ)] where the arguments of the predicate HAND are ordered such that they map directly onto the arguments of the semantic predicate-argument structure. The order and value of the arguments in a lexical entry are transformed by lexlcal rules, such as the passive, to produce new lexical entries, e.g., HAND[#byobJ)~subJ)(~oobJ)]. The direct mapping between lexical predicates and case-frame structures is encoded on the case-frame DMM by augmenting the vectors as follows:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "Hands:-<HAND>*(<agent>*<Pa>*<subJ> + <obJect>*<Po>*<obJ2>+<goal>*<Pg>*<obJ>) When the SUBJ component has been identified through syntactic processing the resulting association vector, for example <subJ>*<John> for the sentence John handed Mary the book, will retrieve <agent> on input to the CF-DMM, according to the principles specified above. The multiple lexical entries produced by lexical rules have corresponding multiple case-frame vectors which are tagged by the appropriate grammatical vector. The CF-DMM encodes multiple case-frame entries for verbs, and the grammatical vector tags, such as <PASSIVE>, generated by the syntactic component, are input to the CF-DMM to retrieve the appropriate case-frame for the verb. This compatibility between f-structures and memory vectors is the basis for an efficient procedure for deriving f-structures from input strings. In LFG f-structures are generated in three steps. First, a context-free grammar (CFG) is used to derive an input string's constituent structure (C-structure). The grammar is augmented so that it generates a phrase structure tree which includes statements about the properties of the string's f-structure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "In the next step, this structure is condensed to derive a series of equations, called the functional description of the string. Finally, the f-structure is derived from the f-description.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "The properties of DMMs enable a simple procedure to be written which derives f-structures from augmented phrase structure trees, obviating the need for an f-descrlptlon. Consider the tree in figure 1 generated for our example sentence: The f-structure, encoded as a memory vector, can be derived from this tree by the following procedure. First, all the grammatical functions, features and semantic forms must be encoded as vectors. The~-variables, f,-f#, have no values at this point; they are derived by the procedure. All the vectors dominated by a node are concatenated to produce a single vector at that node. The symbol '=\" is interpreted as the association operator ,*. Applying this interpretation to the tree from the bottom up produces a memory vector for the value of f! which encodes the f-structure for the string, as given above. Accordingly, f~ takes the value (<TNUM>*<SG>+<TPRED>*<JOHN>); applying the rule specified at the node, (f, SUBJ)=f~ gives <tSUBJ>*(<tNUM>*<SG>+<TPRED>*<JOHN>) as a component of f,. The other components of fl are derived in the same way. The front-end CFG can be veiwed as generating the control structure for the derivation of a memory vector which represents the input string's f-structure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "~SUBJ) -& St& ~ENSE)-PAST \\ @FRED) =HAND[ ..] ~ \\ (I'NUM)-SO [ ~OBJ)-~, ~PRED)=JOHN) I (~qUM) =SG ~PRED)=MARY #OBJ2)=& / [ b",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "The properties of memory vectors also enable the procedure to automatically determine the consistency Df the structure. For example, in deriving the value of f& the concatenation operator merges the (%NUM)~SG features for A and book to form a single component of the f~vector, (<SPEC>*<A>+<NUM>*<SG>+<PRED>*<MARY>). .owever, if the two features had not matched, producing the vector component <NU}~*(<SG>+<PL>) for example, the vectors encoding the incompatible feature values are set such that their concatenation produces a special control vector which signals the mismatch. As elements of the input string enter the lexicon their associated entries are retrieved. The syntactic category of the element is passed onto the CFG, and the lexical schemata {e.g., ~PRED)='JOHN'}, encoded as memory vectors, are passed to the f-structure working memory. The lexical entry associated with the verb is passed to the case-frame memory to retrieve the appropriate set of structures. The partial results of the CFG control the formation of memory vectors in the f-structure memory, as indicated by the broad arrow. The CFG also generates grammatical vectors as inputs for case-frame memory to select the appropriate structure from the multiple encodings associated with each verb. The partial f-structure encoding can then be used as input to the case-frame memory to assign the semantic forms of grammatical functions to case slots. When the end of the string is reached both the case-frame instantiation and the f-structure should be complete.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "B. Encodin~ Syntactic Knowledge",
"sec_num": null
},
{
"text": "This paper attempts to demonstrate the value of distributed memory machines as components of a parsing system which generates both semantic and grammatical relational structures. The ideas presented are similar to those being developed within the connectlonist paradigm [I] . Small, and his colleagues [9] , have proposed a parsing model based directly on connectionist principles. The computational architecture consists of a large number of appropriately connected computing units communicating through weighted levels of excitation and inhibition. The ideas presented here differ from those embodied in the connectionist parser in that they emphasise distributed information storage and retrieval, rather than distributed parallel processing. Retrieval and filtering are achieved through simple computable functions operating on k-element arrays, in contrast to the complex interactions of the independent units in connectlonist models. In figure 2, although the network of machines requires heterarchical control, the architecture can be considered to be at the lower end of the family of parallel processing machines [i0].",
"cite_spans": [
{
"start": 270,
"end": 273,
"text": "[I]",
"ref_id": null
},
{
"start": 302,
"end": 305,
"text": "[9]",
"ref_id": "BIBREF9"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "IV CONCLUSIONS",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Connectionist models and their properties",
"authors": [
{
"first": "J",
"middle": [
"A"
],
"last": "Feldman",
"suffix": ""
},
{
"first": "D",
"middle": [
"H"
],
"last": "Ballard",
"suffix": ""
}
],
"year": 1982,
"venue": "Cognitive Science",
"volume": "6",
"issue": "",
"pages": "205--254",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Feldman, J.A. and Ballard, D.H. Connection- ist models and their properties. Cognitive Science, 1982, 6, 205-254.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Parallel Models of Associative Memory",
"authors": [
{
"first": "G",
"middle": [
"E"
],
"last": "Hinton",
"suffix": ""
},
{
"first": "Anderson",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hinton, G.E. and Anderson, J.A. (Eds) Parallel Models of Associative Memory.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Shape representation in parallel systems",
"authors": [
{
"first": "G",
"middle": [
"E"
],
"last": "Hinton",
"suffix": ""
}
],
"year": 1981,
"venue": "Proceedinss of the Seventh International Joint Conference on Artificial Intelli~ence",
"volume": "2",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hinton, G.E. Shape representation in parallel systems. In Proceedinss of the Seventh International Joint Conference on Artificial Intelli~ence, Vol. 2, Vancouver BC, Canada, August, 1981.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Theories of associative recall",
"authors": [
{
"first": "H",
"middle": [
"C"
],
"last": "Longuet-Higgins",
"suffix": ""
},
{
"first": "D",
"middle": [
"J"
],
"last": "Willshaw",
"suffix": ""
},
{
"first": "O",
"middle": [
"P"
],
"last": "Bunemann",
"suffix": ""
}
],
"year": 1970,
"venue": "~uarterly Reviews of Biophysics",
"volume": "3",
"issue": "",
"pages": "223--244",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Longuet-Higgins, H.C., Willshaw, D.J., and Bunemann, O.P. Theories of associative recall. ~uarterly Reviews of Biophysics, 1970, 3, 223-244.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "A theory for the storage and retrieval of item and associative information",
"authors": [
{
"first": "B",
"middle": [
"B"
],
"last": "Murdock",
"suffix": ""
}
],
"year": 1982,
"venue": "Psychological Review",
"volume": "89",
"issue": "",
"pages": "609--627",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Murdock, B.B. A theory for the storage and retrieval of item and associative information. Psychological Review, 1982, 89, 609-627.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Associative memory~ systemtheoretical approach",
"authors": [
{
"first": "T",
"middle": [],
"last": "Kohonen",
"suffix": ""
}
],
"year": 1977,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kohonen, T. Associative memory~ system- theoretical approach. Berlin: Springer- Verlag, 1977.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Convolution and Correlation algebras",
"authors": [
{
"first": "A",
"middle": [],
"last": "Borsellino",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "Poggio",
"suffix": ""
}
],
"year": 1973,
"venue": "Kybernetik",
"volume": "13",
"issue": "",
"pages": "113--122",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Borsellino, A., and Poggio, T. Convolution and Correlation algebras. Kybernetik, 1973, 13, 113-122.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Lexical-Functional Grammar: A formal system for grammatical representation",
"authors": [
{
"first": "R",
"middle": [],
"last": "Kaplan",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Bresnan",
"suffix": ""
}
],
"year": 1982,
"venue": "The Mental 9~presentation of Grammatical Relations",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kaplan, R., and Bresnan, J. Lexical-Functional Grammar: A formal system for grammatical representation. In J. Bresnan (ed.), The Mental 9~presentation of Grammatical Relations. Cambridge, Mass.:MIT Press, 1982.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "Toward connectlonlst parsing",
"authors": [
{
"first": "S",
"middle": [
"L"
],
"last": "Small",
"suffix": ""
},
{
"first": "G",
"middle": [
"W"
],
"last": "Cottre11",
"suffix": ""
},
{
"first": "L",
"middle": [],
"last": "Shastri",
"suffix": ""
}
],
"year": 1982,
"venue": "Proceedings of the National Conference on Artificial Intelligence",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Small, S.L., Cottre11, G.W., and Shastri, L. Toward connectlonlst parsing. In Proceedings of the National Conference on Artificial Intelligence, Pittsburgh, P~nsylvanla, 1982.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Massively parallel architectures for AI: NETL, THISTLE, and BOLTZMANNmachines",
"authors": [
{
"first": "S",
"middle": [],
"last": "Fahlman",
"suffix": ""
},
{
"first": "E",
"middle": [],
"last": "Hinton",
"suffix": ""
},
{
"first": "G",
"middle": [
"E"
],
"last": "Sejnowski",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "",
"suffix": ""
}
],
"year": null,
"venue": "Proceedings of the National Conference on Artificial Intelli~enc~e",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Fahlman, S,E., Hinton, G.E., and Sejnowski, T. Massively parallel architectures for AI: NETL, THISTLE, and BOLTZMANNmachines. In Proceed- ings of the National Conference on Artificial Intelli~enc~e, Washington D.C., I~3o",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"num": null,
"text": "Items of information are encoded as k-element zero-centred vectors over an interval such as [-I,+I]; <X>=(...x.t,x0,x~t,...). Two items, <A> and <B> (angular brackets denote memory vectors), are associated in memory through the operation of convolution.",
"uris": null,
"type_str": "figure"
},
"FIGREF1": {
"num": null,
"text": "The grammatical relations Between the sententlal elements are represented in the form of functional structure (f-structures) as in LFG. These structures correspond to embedded lists of attrlbute-value pairs, and because of the Uniqueness criterion which governs their format they are efficiently encoded as memory vectors. As an example, the grammatical relations for the sentence John handed Mary a book are encoded in the f-structure below:The lists of grammatical functions and features are encoded as single vectors under the + operator, and the embedded structure is preserved by the associative operator, *. The f-structure is encoded by the vector (<SUBJ>*(<NUM>*<SG>+<PRED>*<JOHN>) + <TENSE> *<PAST> + <PRED>*(<HAND>*(<#SUBJ>*<TOBJ2>* <TOBJ>)) + <OBJ>*(<NUM>*<SG>+<PRED>*<MARY>)+ <OBJ2>*(<SPEC>*<A>+<NUM>*<SG>+<PRED>*<BOOK>))",
"uris": null,
"type_str": "figure"
},
"FIGREF2": {
"num": null,
"text": "Figure I. Augmented Phrase Structure Tree",
"uris": null,
"type_str": "figure"
},
"FIGREF3": {
"num": null,
"text": "C. A Parsing Architecture The ideas outlined above are combined in the design of a tentative parsing architecture shown in figure 2. The diamonds denote DMMs, and the r Parsing Architecture ellipse denotes a form of DMM functioning as a working memory for encoding temporary f-structures.",
"uris": null,
"type_str": "figure"
}
}
}
}