| { |
| "paper_id": "E91-1039", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T10:37:51.489492Z" |
| }, |
| "title": "SEMANTIC CLASSIFICATION OF VERBS FROM THEIR SYNTACTIC CONTEXTS: AN IMPLEMENTED CLASSIFIER FOR STATIVITY", |
| "authors": [ |
| { |
| "first": "Michael", |
| "middle": [ |
| "R" |
| ], |
| "last": "Brent", |
| "suffix": "", |
| "affiliation": {}, |
| "email": "michael@ai.mit.edu" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "This paper discusses an implemented program that automatically classifies verbs into those that ~ describe only states of the world, such as to know, and those that describe events, such as to look. It works by exploiting the con, straint between the syntactic environments in which a verb can occur and its meaning. The only input is on-line text. This demonstrates an important new technique for the automatic generation of lexical databases.", |
| "pdf_parse": { |
| "paper_id": "E91-1039", |
| "_pdf_hash": "", |
| "abstract": [ |
| { |
| "text": "This paper discusses an implemented program that automatically classifies verbs into those that ~ describe only states of the world, such as to know, and those that describe events, such as to look. It works by exploiting the con, straint between the syntactic environments in which a verb can occur and its meaning. The only input is on-line text. This demonstrates an important new technique for the automatic generation of lexical databases.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Abstract", |
| "sec_num": null |
| } |
| ], |
| "body_text": [ |
| { |
| "text": "Young children and natural language processing programs face a common problem: everyone else knows a lot more about words. Children, it is hypothesized, catch up by observing the linguistic and non-linguistic contexts in which words are used. This paper focuses on the value and accessibility of the linguistic context. It argues that linguistic context by itself can provide useful cues about verb meaning to an artificial learner. This is demonstrated by a program that exploits two particular cues from the linguistic context to classify verbs automatically into those whose sole sense is one describing a state, and those that have a sense describing an event. 1 The approach described here accounts for a certain degree of noise in the input due both to mis-apprehension of input sentences and to their occasional real.formation. This work shows that the two cues are available and are reliable given the statistical methods applied.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "Language users, whether natural or artificial, need detailed semantic and syntactic classifications of words. Ultimately, any artificial language IThe input sentences are those compiled in the Lancaster/Oslo/Bergen (LOB) Corpus, a balanced corpus of one million words of British English. The LOB consists primarily of edited prose. user must be able to add new words to its lexicon, if only to accommodate the many neologisms it will encounter. And our lexicographic needs grow with our understanding of language. A number of current approaches to satisfying the lexical requirements for artificial devices do not involve unsupervised learning from examples. Boguraev and Briscoe (1987) discusses interpreting the information published in on-line dictionaries, while Zernik a/~d Dyer (1987) discuss tutored learning in a con. trolled environment. But any method that requires explicit human intervention --be it that of lexicographers, knowledge engineers, or \"tutors\" --will lag behind both the growth of vocabulary and the growth of linguistics, and even with the lag their maintenance will be expensive. By contrast, dictionaries constructed by automated learners from real sentences will not lag behind vocabulary growth; examples of current language use are free and nearly infinite. These observations have led ~everal researchers, including Hindle (1990) and Smadja and McKeown (1990) , to begin investigating automatic acquisition of semantics. Hindle and Smadja and McKeown rely purely on the ability of one particular word to statistically predict the occurrence of another in a particular position. In contrast, the approach described here is targeted at particular semantic classes that are revealed by specific linguistic constructions.", |
| "cite_spans": [ |
| { |
| "start": 672, |
| "end": 686, |
| "text": "Briscoe (1987)", |
| "ref_id": "BIBREF0" |
| }, |
| { |
| "start": 779, |
| "end": 790, |
| "text": "Dyer (1987)", |
| "ref_id": "BIBREF3" |
| }, |
| { |
| "start": 1348, |
| "end": 1361, |
| "text": "Hindle (1990)", |
| "ref_id": null |
| }, |
| { |
| "start": 1366, |
| "end": 1391, |
| "text": "Smadja and McKeown (1990)", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Introduction", |
| "sec_num": "1" |
| }, |
| { |
| "text": "The Questions This section discusses work on two linguistic cues that reveal the availability of non-stative senses for verbs. This work attempts to determine the difficulty of using the cues to classify verbs: into those describing states and those describing events. To that end, it focuses on two questions:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2", |
| "sec_num": null |
| }, |
| { |
| "text": "1. Is it possible to reliably detect the two cues using only a simple syntactic mechanism and minimal syntactic knowledge? How simple can the syntax be? (The less knowledge required to learn using a given technique, the more useful the technique will be.) 2. Assuming minimal syntactic power, how reliable are our two cues in real text, which is subject to performance limitations? Are there learning strategies under which their reliability is adequate?", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2", |
| "sec_num": null |
| }, |
| { |
| "text": "Section 2.1 describes syntactic constructions studied and demonstrates their relation to the stative semantic class. Sections 2.2 answers questions 1 in the affirmative. Section 2.4 answers question 2 in the affirmative, discusses the statistical method used for noise reduction, and demonstrates the program that learns the state-event distinction.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2", |
| "sec_num": null |
| }, |
| { |
| "text": "The differences between verbs describing states (statives) and those describing events (nonstatives) has been studied by linguists at least since Lakoff (1965) . (For a more precise semantic characterization of stativ\u00a2s see Dowty, 1979.) Classic examples of stative verbs are know, believe, desire, and love. A number of syntactic tests have been proposed to distinguish between statives and non-statives (again see Dowry, 1979) . For example, stative verbs are anomalous when used in the progressive aspect and when 'modified by rate adverbs such as quickly and slowly:", |
| "cite_spans": [ |
| { |
| "start": 146, |
| "end": 159, |
| "text": "Lakoff (1965)", |
| "ref_id": "BIBREF1" |
| }, |
| { |
| "start": 224, |
| "end": 237, |
| "text": "Dowty, 1979.)", |
| "ref_id": null |
| }, |
| { |
| "start": 416, |
| "end": 428, |
| "text": "Dowry, 1979)", |
| "ref_id": null |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "(1) a.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "* Jon is knowing calculus b.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "* Jon knows calculus quickly Perception verbs like see and hear share with st~rives a strong resistance to the progressive aspect, but not to rate adverbs:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "(2) a. * Jon is seeing the car b. OK Jon quickly saw the car Agentive verbs describing attempts to gain perceptions, like look and listen, do not share either property:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "(3) a. OK Jon is looking at a car b. OK Jon quickly looked at his watch", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "The classification program relies primarily on the progressive cue, but uses evidence from the rate adverb cue when it is available.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "2.2 Syntactic Requirements for Cue Detection Consider first how much syntactic analysis is needed to detect the progressive and rate adverb constructions. Initially, suppose that the availability of a non-stative sense is aii intrinsic property of a verb 2 not affected by its syntactic context. To detect progressives one need only parse a trivial part of the auxiliary system, which is known to 2This is not true in general, as shown by the f&ct that think that.., is stative while think about.., is not. be finite-state. Detecting the rate aclverb cue requires determining what the adverb modifies, and that can be trickier. For example, adverbs may appear after the direct object, (4a), and this must not be confused with the case where they appear after the subject of an embedded clause, (4b).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "(4) a. ,Ion fixed the robot quickly b. ,Ion knew his hostess rapidly lost intereat in such things", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "Using simple, finite-state machinery one would be forced to deal with (4b) by recognizing the position of the adverb as ambiguous and rejecting the example. Or one could deploy more sophist icated syntax to try determining the boundaries of embedded sentences. But even the best syntactic parser will fail on truly ambiguous cases like the following:", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "(5) a. Jon fixed the robot that had spoken slowly b. Jon believed the robot that had spoken slowly The data on rate adverbs were collected using the parsing approach, which required a substantial amount of machinery, but a finite-state approach might do almost as well. (See Brent and Berwick, 1991, for automatic iexical acquisition using simple finite-state parsing.)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "l'teveaHng Constructions", |
| "sec_num": "2.1" |
| }, |
| { |
| "text": "To test the power of the two proposed cues, the LOB corpus was automatically processed to determine what percentage of each verb's occurrences were in the progressive, and what percent. age were modified by rate adverbs. Sampling error was handled by calculating the probability distribution of the true percentage for each verb assuming that the sentences in the corpus were drawn at random from some infinitely large corpus. The overall frequency of the progressive construction was substantially higher than that of the rate adverb construction and so provided more significant data. Figure 1 shows a histogram constructed by summing these distributions of true frequency in the progressive over each of the 38 most common verbs in the corpus. 3 Figure 1 shows that, at least for these most common verbs, there are three and perhaps four distinct populations. In other words, these verbs do not vary continuously in their frequency of occurrence in the progressive, but rather show a marked tendency to cluster around certain values. As will be shown in the next section, the 3Histograms that include less frequent verbs have the same general character, but the second local maximum gets somewhst blurred by the many verbs whose true frequency in the progressive is poorly localized due to insufficient sample size. Figure 1 : A histogram constructed by smnming the probability distributions of true frequency in the progressive over each of the 38 most common verbs in the corpus stative verbs fall in the first population, to the left of the slight discontinuity at 1.35% of occurrence in the progressive.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 587, |
| "end": 595, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 749, |
| "end": 757, |
| "text": "Figure 1", |
| "ref_id": null |
| }, |
| { |
| "start": 1319, |
| "end": 1327, |
| "text": "Figure 1", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "Data on Cues from the Corpus", |
| "sec_num": "2.3" |
| }, |
| { |
| "text": "1 implemented a program that attempts to classify verbs into those with event senses, and those whose only meaning describes a state rather than an event. It does this by first detecting occurrences of the progressive and rate adverb constructions in the LOB corpus, and then computing confidence intervals on the true frequency of occurrence of each verb in an arbitrarily large corpus of the same composition. The program classifies the verbs according to bounds, which are for the moment supplied by the researcher, on the confidence intervals. For example, on the run shown in Figure 2 , the program classifies verbs which occur at least .1% of the time either in the progressive or modified by a rate adverb, as having an event (non-stative) sense. The classifier acts on .1% bound only if the sample-size is large enough to guarantee the bound with 95% confidence. Accuracy in ascribing non-stative senses according with this technique is excellent --no purely sta. tire verbs are ntis-classified as having non-stative senses. In fact, this result is not very sensitive to raising the minimum progressive frequency from .1% to as high as .6% or .7%, since most verbs with non-stative senses yield observed frequencies of at least two or three percent. Now consider the other side of the problem, classifying verbs as purely stative. Here the program takes verbs that fail the test for having a non-stative sense, and in addition whose true frequency in the progressive falls below a given upper bound with sufficient confidence. The rate-adverb construction is not used, except insofar as the verbs must fail the . 1% lower bound, because this construction turns out to be so rare that only a few of the most frequent verbs provide sufficiently tight bounds. The results for identifying pure sta- Figure 2 : One run of the stative/non-stative classification program on verbs occurring at least 100 times in the LOB corpus tives are also quite good, although more sensitive to the precise bounds than were the results for identifying non-statives. If the upper bound on progressive frequency is set at 1.35%, as in Figure 2, then eleven verbs are identified as purely stative, of the 204 distinct verbs occurring at least 100 times each in the corpus. Two of these, hear and agree, have relatively rare non-stative senses, meaning to imagine one hears (\"I am hearing a ringing in my ears\") and to communicate agreement (\"Rachel was already agreeing when Jon interrupted her with yet another tirade\"). If the upper bound on progressive frequency is tightened to 1.20% then hear and agree drop into the \"indeterminate\" category of verbs that pass neither test. So, too, do three pure statives, mean, require, and understand.", |
| "cite_spans": [], |
| "ref_spans": [ |
| { |
| "start": 581, |
| "end": 589, |
| "text": "Figure 2", |
| "ref_id": null |
| }, |
| { |
| "start": 1803, |
| "end": 1811, |
| "text": "Figure 2", |
| "ref_id": null |
| }, |
| { |
| "start": 2120, |
| "end": 2126, |
| "text": "Figure", |
| "ref_id": null |
| } |
| ], |
| "eq_spans": [], |
| "section": "2,4 The Classification Program", |
| "sec_num": null |
| }, |
| { |
| "text": "It is worth noting the importance of using some sort of noise reduction technique, such as the confidence intervals used here. There are two sources of noise in the linguistic input. First speakers do utter anomalous sentences. For ex. ample, the stative verb mean occurred one time out of 450 in the progressive. The sentence, \"It's a stroke, that was what he was meaning\" is clearly anomalous. The second source of noise is failure of the learner to detect the cue accurately. The accuracy of our automatic cue detection detection is described in the following section.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "2,4 The Classification Program", |
| "sec_num": null |
| }, |
| { |
| "text": "Section 2.2 discussed how much structure must be imposed on sentences if the progressive and rate-adverb constructions are to be detected. Section 2.3 showed that the progressive and rateadverb constructions are indeed reliable cues for the availability of a non-stative sense. This section discusses tile accuracy with which these cues can be detected. It is not practical to check manually every verb occurrence that our program judged tO be progressive. Instead, I checked 300 such sentences selected at random from among the most commonly occurring verbs. This check revealed only one sentence that did not truly describe a progressive event. That sentence is shown in (6a).", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Accuracy of Cue Detection", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "(6) a. go: What that means in this case is go.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Accuracy of Cue Detection", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "ing back to the war years... b. see: The task was solely to see how speedily it could be met... c. compare: ...the purchasing power of the underdeveloped countries in the commonwealth will rise slowly compared with that of Europe.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Accuracy of Cue Detection", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "It is not clear how to automatically determine that (6a) does not describe an event of going in progress. Rate adverbs are infrequent enough that it was possible to verify manually all 281 cases the program found. In four of those cases the rate adverb actually modified a verb other than the one that the program chose. Three of these four cases had the structure of (6b), where a wh-relative is not recognized as signaling the beginning of a new clause. This reflects an oversight in the grammar that should be easily correctable. The one remaining case of a mis-attributed rate adverb, (6c), would again require some work, and perhaps substantial syntactic knowledge, to correct. The rate of false positives in cue detection, then can be estimated at about one serious hazard in 300 for both \"t'~sts.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Accuracy of Cue Detection", |
| "sec_num": "2.5" |
| }, |
| { |
| "text": "This work demonstrates a promising approach to automatic semantic classification of verbs based only on their immediate linguistic contexts. Some sort of statistical smoothing is essential to avoid being permanently mislead by anomalous and misunderstood utterances, and this work demonstrated the sufficiency of an-approach based on binomial confidence-intervals. These methods, in combination with pure collocational methods like those of [Hindle, 1990] and [Smadja and McKeown, 1990] , may eventually yield substantial progress toward automatic acquisition of word meaning, or some aspects thereof, by language using devices.", |
| "cite_spans": [ |
| { |
| "start": 441, |
| "end": 455, |
| "text": "[Hindle, 1990]", |
| "ref_id": null |
| }, |
| { |
| "start": 460, |
| "end": 486, |
| "text": "[Smadja and McKeown, 1990]", |
| "ref_id": "BIBREF1" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "3" |
| }, |
| { |
| "text": "The initial results described here suggest many more experiments, some of which are already u~nderway (see Brent and Berwick, 1991) . These include attempting to take into account the ability of local syntactic context to influence a verb's meaning as well as to reveal it. For example, think that is stative while think about and think of are not. Separating these two senses automatically could add substantial power to our classifier. Next, there are many more linguistic cues to verb meaning to be detected and exploited. For example, the ability to take both adirect object and a propositional complement, as in \"tell him that he's a fool\", reveal verbs of communication. While the progressive cue is not available in Romance languages, the ability to take a direct object and a propositional complement seems to be diagnostic of communication verbs in Romance as well as in English. It would be valuable to demonstrate cues like this on non-English text. It would also be valuable to apply these techniques to a greater variety of input sentences, including transcriptions of mother's speech to their children. Finally, substantially larger corpora should be used in order to enlarge the number of verbs classified. All of these planned extensions serve the goal of automatically classifying thousands of verbs by dozens of different syntactic criteria, and thereby yielding a valuable, adaptable lexicon for natural language processing and artificial intelligence.", |
| "cite_spans": [ |
| { |
| "start": 107, |
| "end": 131, |
| "text": "Brent and Berwick, 1991)", |
| "ref_id": "BIBREF0" |
| } |
| ], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "Conclusions", |
| "sec_num": "3" |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "Large Lexicons for Natural Language Processing: Utilising the Grammar Coding System of LDOCE", |
| "authors": [ |
| { |
| "first": ";", |
| "middle": [ |
| "B" |
| ], |
| "last": "Briscoe", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Boguraev", |
| "suffix": "" |
| }, |
| { |
| "first": ";", |
| "middle": [ |
| "M" |
| ], |
| "last": "Briscoe", |
| "suffix": "" |
| }, |
| { |
| "first": "R", |
| "middle": [], |
| "last": "Brent", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Berwick", |
| "suffix": "" |
| } |
| ], |
| "year": 1979, |
| "venue": "Proceedings of the 4th Darpa Speech and Natural Language Workshop. Defense Advanced Research Projects Agency", |
| "volume": "13", |
| "issue": "", |
| "pages": "268--275", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "and Briscoe, 1987] B. Boguraev and T. Briscoe. Large Lexicons for Natural Lan- guage Processing: Utilising the Grammar Cod- ing System of LDOCE. Comp. Ling., 13(3), 1987. [Brent and Berwick, 1991] M. Brent and R. Berwick. Automatic Acquisition of Subcate- gorization Frames From Free Text Corpora. In Proceedings of the 4th Darpa Speech and Natu- ral Language Workshop. Defense Advanced Re- search Projects Agency, Arlington, VA, USA, 1991. [Dowty, 1979] D. Dowty. Word Meaning and Montague Grammar. Synthese Language Li- brary. D. Reidel, Boston, 1979. [Hindle, 1990] D. Hindle. Noun cla-qsification from predicate argument structures. In Proceedings of the ~Sth Annual Meeting of the ACL, pages 268-275. ACL, 1990.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "Automatically extracting and representing collocations for language generation", |
| "authors": [ |
| { |
| "first": ";", |
| "middle": [ |
| "G" |
| ], |
| "last": "Lakoff", |
| "suffix": "" |
| }, |
| { |
| "first": "; F", |
| "middle": [], |
| "last": "Lakoff", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Smadja", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Mckeown", |
| "suffix": "" |
| } |
| ], |
| "year": 1965, |
| "venue": "~8th Annual Meeting of the Association for Comp. Ling", |
| "volume": "", |
| "issue": "", |
| "pages": "252--259", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Lakoff, 1965] G. Lakoff. On the Nature of Syntac. tic Irregularity. PhD thesis, Indiana University, 1965. Published by Holt, Rinhard, and Winston as Irregularity in Syntax, 1970. [Smadja and McKeown, 1990] F. Smadja and K. McKeown. Automatically extracting and representing collocations for lan- guage generation. In ~8th Annual Meeting of the Association for Comp. Ling., pages 252-259.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "The self-extending phrasal lexicon", |
| "authors": [ |
| { |
| "first": "]", |
| "middle": [ |
| "U" |
| ], |
| "last": "Dyer", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Zernik", |
| "suffix": "" |
| }, |
| { |
| "first": "", |
| "middle": [], |
| "last": "Dyer", |
| "suffix": "" |
| } |
| ], |
| "year": null, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "and DYer, 1987] U. Zernik and M. Dyer. The self-extending phrasal lexicon.", |
| "links": null |
| } |
| }, |
| "ref_entries": { |
| "FIGREF0": { |
| "text": "S :wax-index 200 :scale tOO0) NIL", |
| "uris": null, |
| "num": null, |
| "type_str": "figure" |
| }, |
| "TABREF1": { |
| "text": "KHOU SEER LIKE RELIEVE URHT OERRIN HERR HERR REOUIRE UNDEGSTRHO RCGEE HRS-UOH-STRTIVE-SEHSE: UEflR URZT 00 TRLK SROU SO \u00a3TRY TRY VISIT LISTEN LIE SIT PREPRRE FRIL SEEK HONOEG FIGHT UORK STOOY B(GZH ORIVE URTCH OERL OCT EHJOY SETTLE SflILE PLRY OZE LIVE HUH HOVE 8TRRO HOPPER HOLK CRERTE PROVE CRUSE OR(OK DROP LOOK CRRRY FRCE RTTEND EXPECT FRLL (HO OEVELOP OFFER ERT URIT[ OEO0 CLOSE RECOil( P OIHT OORU RETURN RISE BUILD CHRHCE DIN COnE LERRN PUBLISH PICK RSSOCIRTE GET PRODUCE REPLY PRY LERO ZRTRODUC[ COflPLETE REFUGE SRV( NOTICE PULL OVOID RECEIVE SERVE PDES\u00a3RT STOP OPEN ENTER SET SPEHO ~IGH FORGET HOTE RGSURE PLRCE IHCRERSE OCCUR COHPRRE COHSIOER SUGGEST COVER DISCOVER SELL THINK OEGOGO RFF~ CT KEEP HOLO FOLLOU PUT HEET RCCEPT 8EHO HELP REVERL BOISG ORIGE OPPERR PGOVIO[ TONE GEH\u00a3HOEO SPEEK FINISH TURN DSK RERCH LET TELL RRK[ FELL RHSUER FIND LERV[ RCHIEVE FEEL CRLL 6HON USE (XPLRIN GIVE", |
| "num": null, |
| "content": "<table><tr><td>(\u00a9lassiry-statlve-non-statlve</td><td>:nax-proor-for-statlve</td><td>.OlOS</td></tr><tr><td/><td colspan=\"2\">~nln-rlte-for-non-statlve :nln-proor-for-non-statlve .00|) .OOI</td></tr><tr><td/><td colspan=\"2\">LRCKS-NRN-STRTIVE-SERSE:</td></tr><tr><td>OGTRIN OECIOE DRY SEE</td><td/><td/></tr><tr><td/><td colspan=\"2\">IHOETEGHIRATE:</td></tr><tr><td colspan=\"3\">THROU REFER BRSE CHOOSE CRrCH ORRIVE KRRRGE SUPPORT EXIST BELONG OHISE BERD REEO CUT IHTEHO IRRCIUE C</td></tr><tr><td colspan=\"3\">LRIH FQRfl BUY HE STRTE gILL RPPLY REISQVE RERLISE HOPE RRINTRIN JOIN HEHTIGN FILL OOnI! OEPEHO REPORT</td></tr><tr><td colspan=\"3\">OLLOU flROOY ESTRRLISR IHOICRTE LRY LOOSE SPERK SUPPOSE REDUCE REPRESENT LOVE ZHUOLUE COHTRIH OO0 STORY</td></tr><tr><td colspan=\"2\">8R IHCLUOE COHCERH COHTIHU[ OE$CRIBE</td><td/></tr><tr><td>HIL</td><td/><td/></tr><tr><td>Oynarnic Lisp Listener f</td><td/><td/></tr></table>", |
| "html": null, |
| "type_str": "table" |
| } |
| } |
| } |
| } |