File size: 86,055 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
{
    "paper_id": "P89-1010",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T08:14:50.598386Z"
    },
    "title": "Word Association Norms, Mutual Information, and Lexicography",
    "authors": [
        {
            "first": "Kenneth",
            "middle": [
                "Ward"
            ],
            "last": "Church",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "N.J. Patrick Hanks CoLlins Publishers Glasgow",
                "location": {
                    "country": "Scotland"
                }
            },
            "email": ""
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "The term word assaciation is used in a very particular sense in the psycholinguistic literature. (Generally speaking, subjects respond quicker than normal to the word \"nurse\" if it follows a highly associated word such as \"doctor.\") We wilt extend the term to provide the basis for a statistical description of a variety of interesting linguistic phenomena, ranging from semantic relations of the doctor/nurse type (content word/content word) to lexico-syntactic co-occurrence constraints between verbs and prepositions (content word/function word). This paper will propose a new objective measure based on the information theoretic notion of mutual information, for estimating word association norms from computer readable corpora. (The standard method of obtaining word association norms, testing a few thousand subjects on a few hundred words, is both costly and unreliable.) The , proposed measure, the association ratio, estimates word association norms directly from computer readable corpora, waki,~g it possible to estimate norms for tens of thousands of words.",
    "pdf_parse": {
        "paper_id": "P89-1010",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "The term word assaciation is used in a very particular sense in the psycholinguistic literature. (Generally speaking, subjects respond quicker than normal to the word \"nurse\" if it follows a highly associated word such as \"doctor.\") We wilt extend the term to provide the basis for a statistical description of a variety of interesting linguistic phenomena, ranging from semantic relations of the doctor/nurse type (content word/content word) to lexico-syntactic co-occurrence constraints between verbs and prepositions (content word/function word). This paper will propose a new objective measure based on the information theoretic notion of mutual information, for estimating word association norms from computer readable corpora. (The standard method of obtaining word association norms, testing a few thousand subjects on a few hundred words, is both costly and unreliable.) The , proposed measure, the association ratio, estimates word association norms directly from computer readable corpora, waki,~g it possible to estimate norms for tens of thousands of words.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "The search for increasingly delicate word classes is not new. In lexicography, for example, it goes back at least to the \"verb patterns\" described in Hornby's Advanced Learner's Dictionary (first edition 1948) . What is new is that facilities for the computational storage and analysis of large bodies of natural language have developed significantly in recent years, so that it is now becoming possible to test and apply informal assertions of this kind in a more 76 rigorous way, and to see what company our words do keep.",
                "cite_spans": [
                    {
                        "start": 150,
                        "end": 209,
                        "text": "Hornby's Advanced Learner's Dictionary (first edition 1948)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "The proposed statistical description has a large number of potentially important applications, including: (a) constraining the language model both for speech recognition and optical character recognition (OCR), (b) providing disambiguation cues for parsing highly ambiguous syntactic structures such as noun compounds, conjunctions, and prepositional phrases, (c) retrieving texts from large databases (e.g., newspapers, patents), (d) enhancing the productivity of computational linguists in compiling lexicons of lexico-syntactic facts, and (e) enhancing the productivity of lexicographers in identifying normal and conventional usage.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Practical Applications",
                "sec_num": "2."
            },
            {
                "text": "Consider the optical character recognizer (OCR) application. Suppose that we have an OCR device such as [Kahan, Pavlidis, Baird (1987) ], and it has assigned about equal probability to having recognized \"farm\" and \"form,\" where the context is either: (1) \"federal t credit\" or (2) \"some of.\" The proposed association measure can make use of the fact that \"farm\" is much more likely in the first context and \"form\" is much more likely in the second to resolve the ambiguity. Note that alternative disambiguation methods based on syntactic constraints such as part of speech are unlikely to help in this case since both \"form\" and \"farm\" are commonly used as nouns.",
                "cite_spans": [
                    {
                        "start": 104,
                        "end": 134,
                        "text": "[Kahan, Pavlidis, Baird (1987)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Practical Applications",
                "sec_num": "2."
            },
            {
                "text": "Word association norms are well known to be an important factor in psycholinguistic research, especially in the area of lexical retrieval. Generally speaking, subjects respond quicker than normal to the word \"nurse\" if it follows a highly associated word such as \"doctor.\" \"Some resuhs and impl~tfions ere summarized from rexcfion-fime .experiments in which subjects either (a) ~as~f'mi successive strings of lenen as words and nonwords, c~ (b) pronounced the sUnriSe. Both types of response to words (e.g., BUTTER) were consistently fester when preceded by associated words (e.g., BREAD) rather than unassociated words (e.g, NURSE).\" [Meyer, Schvaneveldt and Ruddy (1975), p. 98] Much of this psycholinguistic research is based on empirical estimates of word association norms such as [Palermo and Jenkins (1964) ], perhaps the most influential study of its kind, though extremely small and somewhat dated. This study measured 200 words by asking a few thousand subjects to write down a word after each of the 200 words to be measured. Results are reported in tabular form, indicating which words were written down, and by how many subjects, factored by grade level and sex. The word \"doctor,\" for example, is reported on pp. 98-100, to be most often associated with \"nurse,\" followed by \"sick,\" \"health,\" \"medicine,\" \"hospital,\" \"man,\" \"sickness,\" \"lawyer,\" and about 70 more words.",
                "cite_spans": [
                    {
                        "start": 374,
                        "end": 377,
                        "text": "(a)",
                        "ref_id": null
                    },
                    {
                        "start": 635,
                        "end": 680,
                        "text": "[Meyer, Schvaneveldt and Ruddy (1975), p. 98]",
                        "ref_id": null
                    },
                    {
                        "start": 786,
                        "end": 813,
                        "text": "[Palermo and Jenkins (1964)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Word Association and Psycholingui~tics",
                "sec_num": "3."
            },
            {
                "text": "We propose an alternative measure, the association ratio, for measuring word association norms, based on the information theoretic concept of mutual information. The proposed measure is more objective and less costly than the subjective method employed in [Palermo and Jenkins (1964) ]. The association ratio can be scaled up to provide robust estimates of word association norms for a large portion of the language. Using the association ratio measure, the five most associated words are (in order): \"dentists,\" \"nurses,\" \"treating,\" \"treat,\" and \"hospitals.\"",
                "cite_spans": [
                    {
                        "start": 256,
                        "end": 283,
                        "text": "[Palermo and Jenkins (1964)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "What is \"mutual information\"? According to [Fano (1961) , p. 28], if two points (words), x and y, have probabilities P(x) and P(y), then their mutual information, l(x,y), is defined to be",
                "cite_spans": [
                    {
                        "start": 43,
                        "end": 55,
                        "text": "[Fano (1961)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "l(x,y) -Io- P(x,y) s2 P(x) P(y)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "Informally, mutual information compares the probability of observing x and y together (the joint probability) with the probabilities of observing x and y independently (chance). If there is a genuine association between x and y, then the joint probability P(x,y) will be much larger than chance P(x) P(y), and consequently l(x,y) >> 0. If there is no interesting relationship between x and y, then P(x,y) ~ P(x) P(y), and thus, I(x,y) ~-0.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "If x and y are in complementary distribution, then P(x,y) will be much less than P(x) P(y), forcing l(x,y) << O.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "In our application, word probabilities, P(x) and P(y), are estimated by counting the number of observations of x and y in a corpus, f(x) and f(y), and normalizing by N, the size of the corpus. (Our examples use a number of different corpora with different sizes: 15 million words for the 1987 AP corpus, 36 million words for the 1988 AP corpus, and 8.6 million tokens for the tagged corpus.) Joint probabilities, P(x,y), are estimated by counting the number of times that x is followed by y in a window of w words,f,,(x,y) , and normalizing by N.",
                "cite_spans": [
                    {
                        "start": 508,
                        "end": 522,
                        "text": "words,f,,(x,y)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "The window size parameter allows us to look at different scales. Smaller window sizes will identify fixed expressions (idioms) and other relations that hold over short ranges; larger window sizes will highlight semantic concepts and other relationships that hold over larger scales. For the remainder of this paper, the window size, w, will be set to 5 words as a compromise; this setting is large enough to show some of the constraints between verbs and arguments, but not so large that it would wash out constraints that make use of strict adjacency.1",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "Since the association ratio becomes unstable when the counts are very small, we will not discuss word pairs with f(x,y) $ 5. An improvement would make use of t-scores, and throw out pairs that were not significant. Unfortunately, this requffes an estimate of the variance of f(x,y), which goes beyond the scope of this paper. For the remainder of this paper, we will adopt the simple but arbitrary threshold, and ignore pairs with small counts.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "Technically, the association ratio is different from mutual information in two respects. First, joint probabilities are supposed to be symmetric: P(x,y) = P(y,x), and thus, mutual information is also symmetric: l(x,y)=l (y,x) . However, the association ratio is not symmetric, since f(x,y) encodes linear precedence. (Recall that f(x,y) denotes the number of times that word x appears before y in the window of w words, not the number of times the two words appear in either order.) Although we could fix this problem by redefining f(x,y) to be symmetric (by averaging the matrix with its transpose), we have decided not to do so, since order information appears to be very interesting. Notice the asymmetry in the pairs below (computed from 36 million words of 1988 AP text), illustrating a wide variety of biases ranging w-I (which has the consequence of subtracting Iog2(w-l) --2 from our association ratio scores). This adjustment has the additional benefit of assuring that",
                "cite_spans": [
                    {
                        "start": 220,
                        "end": 225,
                        "text": "(y,x)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "~ f(x,y) ffi ~ f(x) ffi ~ f(y)ffi N.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "When l(x,y) is large, the association ratio produces very credible results not unlike those reported in ~alermo and Jenkins (1964)], as illustrated in the tabl~ below. In contrast, when l(x,y) ~ 0, the pairs less interesting. (As a very rough rule of thumb, we have observed that pairs with l(x,y) > 3 tend to be interesting, and pairs with smaller l(x,y) are generally not. One can make this statement precise by calibrating the measure with subjective measures. Alternatively, one could make estimates of the variance and then make statements about confidence levels, e.g., with 95% confidence, P(x,y) >",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "An Information Theoretic Measure",
                "sec_num": "4."
            },
            {
                "text": "Some Interesting Associations with \"Doctor\" in the 1987 AP Corpus (N = 15 minion) corpora are too small (and our measurement techniques are too crude). Suppose, for example, that both x and y appear about i0 times per million words of text. Then, P(x)=P(y)=iO -s and chance is P(x)P(x)ffi tO -l\u00b0. Thus, to say that l(x,y) is much less than 0, we need to say that P(x,y) is much less than 10-~\u00b0 a statement that is hard to make with much confidence given the size of presently available corpora. In fact, we cannot (easily) observe a probability less than 1/N = 10 -7, and therefore, it is hard to know ff l(x,y) is much less than chance or not, unless chance is very large. (In fact, the pair (a, doctors)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "P(x) P(y).)",
                "sec_num": null
            },
            {
                "text": "I(x, y) fix, y) fix) x fly)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "P(x) P(y).)",
                "sec_num": null
            },
            {
                "text": "above, appears significantly less often than chance. But to justify this statement, we need to compensate for the window size (which shifts the score downward by 2.0, e.g. from 0.96 down to -1.04) and we need to estimate the standard deviation, using a method such as [Good (1953) ].)",
                "cite_spans": [
                    {
                        "start": 268,
                        "end": 280,
                        "text": "[Good (1953)",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "P(x) P(y).)",
                "sec_num": null
            },
            {
                "text": "Although the psycholinguistic literature documents the significance of noun/noun word associations such as doctor/nurse in considerable detail, relatively little is said about associations among verbs, function words, adjectives, and other non-nouns. In addition to identifying semantic relations of the doctor/nurse variety, we believe the association ratio can also be used to search for interesting lexico-syntactic relationships between verbs and typical arguments/adjuncts. The proposed association ratio That is enough to show its main patterning and it suggests that in currently-held corpora there will be found sufficient evidence for the desc~'iption of a substantial collection of phrases... [Sinclair (1987)\u00a2. pp. 151-152] It happens that set ... offwas found 177 times in the 1987 AP Corpus of approximately 15 million words, about the same number of occurrences per million as Sinclair found in his (mainly British) corpus. Quantitatively, l(set,off) = 5.9982, indicating that the probability of set ... off is almost 64 times greater than chance. This association is relatively strong; the other particles that Sincliir mentions have association ratios of: about (1.4), in (2.9), up (6.9), out (4.5), on (3.3) in the 1987 AP Corpus.",
                "cite_spans": [
                    {
                        "start": 703,
                        "end": 734,
                        "text": "[Sinclair (1987)\u00a2. pp. 151-152]",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexico-$yntactic Regularities",
                "sec_num": "5."
            },
            {
                "text": "As Sinclair suggests, the approach is well suited for identifying phrasal verbs. However, phrasal verbs involving the preposition to raise an interesting problem because of the possible confusion with the infinitive marker to. We have found that if we first tag every word in the corpus with a part of speech using a method such as [Church (1988) ], and then measure associations between tagged words, we can identify interesting contrasts between verbs associated with a following preposition to~in and verbs associated with a following infinitive marker to~to. (Part of speech notation is borrowed from [Francis and Kucera (1982) ]; in = preposition; to = infinitive marker; vb = bare verb; vbg = verb + ins; vbd = verb + ed; vbz = verb + s; vbn = verb + en.) The association ratio identifies quite a number of verbs associated in an interesting way with to; restricting our attention to pairs with a score of 3.0 or more, there are 768 verbs associated with the preposition to~in and 551 verbs with the infinitive marker to~to. The ten verbs found to be most associated before to~in are:",
                "cite_spans": [
                    {
                        "start": 332,
                        "end": 346,
                        "text": "[Church (1988)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 618,
                        "end": 631,
                        "text": "Kucera (1982)",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexico-$yntactic Regularities",
                "sec_num": "5."
            },
            {
                "text": "\u2022 to~in: alluding/vbg, adhere/vb, amounted/vbn, relating/vbg, amounting/vbg, revert/vb, reverted/vbn, resorting/vbg, relegated/vbn",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexico-$yntactic Regularities",
                "sec_num": "5."
            },
            {
                "text": "\u2022 to~to: obligated/vbn, trying/vbg, compened/vbn, enables/vbz, supposed/vbn, intends/vbz, vowing/vbg, tried/vbd, enabling/vbg, tends/vbz, tend/vb, intend/vb, tries/vbz Thus, we see there is considerable leverage to be gained by preprocessing the corpus and manipulating the inventory of tokens. For measuring syntactic constraints, it may be useful to include some part of speech information and to exclude much of the internal structure of noun phrases. For other purposes, it may be helpful to tag items and/or phrases with semantic libels such as *person*, *place*, *time*, *body-part*, *bad*, etc. Hindle (personal communication) has found it helpful to preprocess the input with the Fidditch parser ~I-. Iindle (1983a,b) ] in order to identify associations between verbs and arguments, and postulate semantic classes for nouns on this basis.",
                "cite_spans": [
                    {
                        "start": 709,
                        "end": 725,
                        "text": "Iindle (1983a,b)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexico-$yntactic Regularities",
                "sec_num": "5."
            },
            {
                "text": "Large machine-readable corpora are only just now becoming available to lexicographers. Up to now, lexicographers have been reliant either on citations collected by human readers, which introduced an element of selectivity and so inevitably distortion (rare words and uses were collected but common uses of common words were not), or on small corpora of only a million words or so, which are reliably informative for only the most common uses of the few most frequent words of English. (A million-word corpus such as the Brown Corpus is reliable, roughly, for only some uses of only some of the forms of around 4000 dictionary entries. But standard dictionaries typically contain twenty times this number of entries.)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "The computational tools available for studying machine-readable corpora are at present still rather primitive. There are concordancing programs (see Figure 1 at the end of this paper), which are basically KWIC (key word in context [Aho, Kernighan, and Weinberger (1988) , p. 122]) indexes with additional features such as the ability to extend the context, sort leftwards as well as rightwards, and so on. There is very little interactive software. In a typical skuation in the lexicography of the 1980s, a lexicographer is given the concordances for a word, marks up the printout with colored pens in order to identify the salient senses, and then writes syntactic descriptions and definitions.",
                "cite_spans": [
                    {
                        "start": 231,
                        "end": 269,
                        "text": "[Aho, Kernighan, and Weinberger (1988)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [
                    {
                        "start": 149,
                        "end": 157,
                        "text": "Figure 1",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "Although this technology is a great improvement on using human readers to collect boxes of citation index cards (the method Murray used in constructing the Oxford English Dictionary a century ago), it works well if there are no more than a few dozen concordance lines for a word, and only two or three main sense divisions. In analyzing a complex word such as \"take\", \"save\", or \"from\", the lexicographer is trying to pick out significant patterns and subtle distinctions that are buried in literally thousands of concordance lines: pages and pages of computer printout. The unaided human mind simply cannot discover all the significant patterns, let alone group them and rank in order of importance.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "The AP 1987 concordance to \"save\" is many pages long; there are 666 lines for the base form alone, and many more for the inflected forms \"saved,\" \"saves,\" \"saving,\" and \"savings.\" In the discussion that follows, we shall, for the sake of simplicity, not analyze the inflected forms and we shall only look at the patterns to the right of \"save\". Figure 1 that the word \"to\" often comes before \"save\" and the word \"the\" often comes after \"save,\" it is hard to say from examination of a concordance alone whether either or both of these co-occurrences have any significance.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 345,
                        "end": 353,
                        "text": "Figure 1",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "Two examples will be illustrate how the association ratio measure helps make the analysis both quicker and more accurate.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "6.1 F.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "xamp/e 1: \"save ... from\"",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "The association ratios (above) show that association norms apply to function words as well as content words. For example, one of the words significantly associated with \"save\" is \"from\". Many dictionaries, for example Merriam-Webster's Ninth, make no explicit mention of \"from\" in the entry for \"save\", although British learners' dictionaries do make specific mention of \"from\" in connection with \"save\". These learners' dictionaries pay more attention to language structure and collocation than do American collegiate dictionaries, and lexicographers trained in the British tradition are often fairly skilled at spotting these generalizations. However, teasing out such facts, and distinguishing true intuitions from false intuitions takes a lot of time and hard work, and there is a high probability of inconsistencies and omissions.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "Which other verbs typically associate with \"from,\" and where does \"save\" rank in such a list? The association ratio identified 1530 words that are associated with \"from\"; 911 of them were tagged as verbs. The first I00 verbs are:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "refi 'aJn/vb, gleaned/vii, stems/vbz, stemmed/vbd, stemmins/vbg, renging/vbg, stemmed/vii, ranged/vii, derived/vii, reng~/vbd, extort/vb, gradu|ted/vbd, barred/vii, benefltiag/vbg, benefmect/vii, benefited/vii, m'hing/vbg, range/vb, exempts/vbz, suffers/vbz, exemptingtvbg, benefited/vbd, In.evented/vbd (7.0), seepins/vbs, btrted/vbd, tnevents/vbz, suffering/vbs, exe.laded/vii, mtrks/vbz, pmfitin~vbs, recoverins/vbg, discharged/vii, reboundins/vbg, vary/vb, exempted/vbn, ~te/vb, blmished/vii, withdrawing/vbg, ferry/vb, prevented/vii, pmfit/vb, bar/vb, excused/vii, bars/vbz, benefit/vb, emerget/vbz, em~se/vb, vm'tes/vbz, differ/vb, removed/vim, exemln/vb, expened/vbn, withdraw/vb, stem/vb, separated/vii, judging/vbg, adapted/vbn, escapins/vbs, inherited/vii, differed/vbd, emerged/vbd, withheld/vbd, kaked/vbn, strip/vb, i~mlting/vbs, discouruge/vb, I~'event/vb, withdrew/vbd, pmhibits/vbz, borrowing/vbg , preventing/vbg, prohibit/vb, resulted/vbd (6.0), predude/vb, divert/vb, distin~hh/vb, pulled/vbn, fell/vbn, varied/vbn, emerging/vbs, suHe~r/vb, prohibiting/vbg, extract/vb, remverA, b, paralyzed/vii, stole/vbd, departing/vbs, escaped/vii, l~ohibited/vbn, forbid/vb, evacuated/vii, reap/vb, barring/vbg, removing/vbg, stolen/vii, receives/vbz. \"Save ... from\" is a good example for illustrating the advantages of the association ratio. Save is ranked 319th in this list, indicating that the association is modest, strong enough to be important (21 times more likely than chance), but not so strong that it would pop out at us in a concordance, or that it would be one of the first things to come to mind.",
                "cite_spans": [
                    {
                        "start": 5,
                        "end": 13,
                        "text": "'aJn/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 14,
                        "end": 26,
                        "text": "gleaned/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 27,
                        "end": 37,
                        "text": "stems/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 38,
                        "end": 50,
                        "text": "stemmed/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 51,
                        "end": 64,
                        "text": "stemmins/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 65,
                        "end": 77,
                        "text": "renging/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 78,
                        "end": 90,
                        "text": "stemmed/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 91,
                        "end": 102,
                        "text": "ranged/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 103,
                        "end": 115,
                        "text": "derived/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 116,
                        "end": 126,
                        "text": "reng~/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 127,
                        "end": 137,
                        "text": "extort/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 138,
                        "end": 152,
                        "text": "gradu|ted/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 153,
                        "end": 164,
                        "text": "barred/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 165,
                        "end": 180,
                        "text": "benefltiag/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 181,
                        "end": 195,
                        "text": "benefmect/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 196,
                        "end": 210,
                        "text": "benefited/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 211,
                        "end": 222,
                        "text": "m'hing/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 223,
                        "end": 232,
                        "text": "range/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 233,
                        "end": 245,
                        "text": "exempts/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 246,
                        "end": 258,
                        "text": "suffers/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 259,
                        "end": 273,
                        "text": "exemptingtvbg,",
                        "ref_id": null
                    },
                    {
                        "start": 274,
                        "end": 288,
                        "text": "benefited/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 289,
                        "end": 310,
                        "text": "In.evented/vbd (7.0),",
                        "ref_id": null
                    },
                    {
                        "start": 311,
                        "end": 323,
                        "text": "seepins/vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 324,
                        "end": 335,
                        "text": "btrted/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 336,
                        "end": 349,
                        "text": "tnevents/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 350,
                        "end": 364,
                        "text": "suffering/vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 365,
                        "end": 379,
                        "text": "exe.laded/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 380,
                        "end": 390,
                        "text": "mtrks/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 391,
                        "end": 403,
                        "text": "pmfitin~vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 404,
                        "end": 419,
                        "text": "recoverins/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 420,
                        "end": 435,
                        "text": "discharged/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 436,
                        "end": 451,
                        "text": "reboundins/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 452,
                        "end": 460,
                        "text": "vary/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 461,
                        "end": 474,
                        "text": "exempted/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 475,
                        "end": 482,
                        "text": "~te/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 483,
                        "end": 496,
                        "text": "blmished/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 497,
                        "end": 513,
                        "text": "withdrawing/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 514,
                        "end": 523,
                        "text": "ferry/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 524,
                        "end": 538,
                        "text": "prevented/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 539,
                        "end": 548,
                        "text": "pmfit/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 549,
                        "end": 556,
                        "text": "bar/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 557,
                        "end": 569,
                        "text": "excused/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 570,
                        "end": 579,
                        "text": "bars/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 580,
                        "end": 591,
                        "text": "benefit/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 592,
                        "end": 604,
                        "text": "emerget/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 605,
                        "end": 614,
                        "text": "em~se/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 615,
                        "end": 626,
                        "text": "vm'tes/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 627,
                        "end": 637,
                        "text": "differ/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 638,
                        "end": 650,
                        "text": "removed/vim,",
                        "ref_id": null
                    },
                    {
                        "start": 651,
                        "end": 661,
                        "text": "exemln/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 662,
                        "end": 674,
                        "text": "expened/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 675,
                        "end": 687,
                        "text": "withdraw/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 688,
                        "end": 696,
                        "text": "stem/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 697,
                        "end": 711,
                        "text": "separated/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 712,
                        "end": 724,
                        "text": "judging/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 725,
                        "end": 737,
                        "text": "adapted/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 738,
                        "end": 751,
                        "text": "escapins/vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 752,
                        "end": 766,
                        "text": "inherited/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 767,
                        "end": 780,
                        "text": "differed/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 781,
                        "end": 793,
                        "text": "emerged/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 794,
                        "end": 807,
                        "text": "withheld/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 808,
                        "end": 818,
                        "text": "kaked/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 819,
                        "end": 828,
                        "text": "strip/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 829,
                        "end": 842,
                        "text": "i~mlting/vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 843,
                        "end": 857,
                        "text": "discouruge/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 858,
                        "end": 870,
                        "text": "I~'event/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 871,
                        "end": 884,
                        "text": "withdrew/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 885,
                        "end": 898,
                        "text": "pmhibits/vbz,",
                        "ref_id": null
                    },
                    {
                        "start": 899,
                        "end": 914,
                        "text": "borrowing/vbg ,",
                        "ref_id": null
                    },
                    {
                        "start": 915,
                        "end": 930,
                        "text": "preventing/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 931,
                        "end": 943,
                        "text": "prohibit/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 944,
                        "end": 963,
                        "text": "resulted/vbd (6.0),",
                        "ref_id": null
                    },
                    {
                        "start": 964,
                        "end": 975,
                        "text": "predude/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 976,
                        "end": 986,
                        "text": "divert/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 987,
                        "end": 1000,
                        "text": "distin~hh/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 1001,
                        "end": 1012,
                        "text": "pulled/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 1013,
                        "end": 1022,
                        "text": "fell/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 1023,
                        "end": 1034,
                        "text": "varied/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 1035,
                        "end": 1048,
                        "text": "emerging/vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 1049,
                        "end": 1059,
                        "text": "suHe~r/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 1060,
                        "end": 1076,
                        "text": "prohibiting/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 1077,
                        "end": 1088,
                        "text": "extract/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 1089,
                        "end": 1097,
                        "text": "remverA,",
                        "ref_id": null
                    },
                    {
                        "start": 1098,
                        "end": 1100,
                        "text": "b,",
                        "ref_id": null
                    },
                    {
                        "start": 1101,
                        "end": 1115,
                        "text": "paralyzed/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 1116,
                        "end": 1126,
                        "text": "stole/vbd,",
                        "ref_id": null
                    },
                    {
                        "start": 1127,
                        "end": 1141,
                        "text": "departing/vbs,",
                        "ref_id": null
                    },
                    {
                        "start": 1142,
                        "end": 1154,
                        "text": "escaped/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 1155,
                        "end": 1170,
                        "text": "l~ohibited/vbn,",
                        "ref_id": null
                    },
                    {
                        "start": 1171,
                        "end": 1181,
                        "text": "forbid/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 1182,
                        "end": 1196,
                        "text": "evacuated/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 1197,
                        "end": 1205,
                        "text": "reap/vb,",
                        "ref_id": null
                    },
                    {
                        "start": 1206,
                        "end": 1218,
                        "text": "barring/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 1219,
                        "end": 1232,
                        "text": "removing/vbg,",
                        "ref_id": null
                    },
                    {
                        "start": 1233,
                        "end": 1244,
                        "text": "stolen/vii,",
                        "ref_id": null
                    },
                    {
                        "start": 1245,
                        "end": 1258,
                        "text": "receives/vbz.",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "If the dictionary is going to list \"save ... from,\" then, for consistency's sake, it ought to consider listing all of the more important associations as well. Of the 27 bare verbs (tagged 'vb3 in the list above, all but 7 are listed in the Cobuild dictionary as occurring with \"from\". However, this dictionary does not note that vary, ferry, strip, divert, forbid, and reap occur with \"from.\" If the Cobuild lexicographers had had access to the proposed measure, they could possibly have obtained better coverage at less cost.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Applications in Lexicography",
                "sec_num": "6."
            },
            {
                "text": "Having established the relative importance of \"save ... from\", and having noted that the two words are rarely adjacent, we would now like to speed up the labor-intensive task of categorizing the concordance lines. Ideally, we would like to develop a set of semi-automatic tools that would help a lexicographer produce something like Figure 2 , which provides an annotated summary of the 65 concordance lines for \"save ... from. ''a The \"save ... from\" pattern occurs in about 10% of the 666 concordance lines for \"save.\"",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 333,
                        "end": 341,
                        "text": "Figure 2",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Example 2: Identifying Semantic Classes",
                "sec_num": "6.2"
            },
            {
                "text": "Traditionally, semantic categories have been only vaguely recognized, and to date little effort has been devoted to a systematic classification of a large corpus.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Example 2: Identifying Semantic Classes",
                "sec_num": "6.2"
            },
            {
                "text": "Lexicographers have tended to use concordances impressionistically; semantic theorist, AI-ers, and others have concentrated on a few interesting examples, e.g., '*bachelor,\" and have not given much thought to how the results might be scaled up.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Example 2: Identifying Semantic Classes",
                "sec_num": "6.2"
            },
            {
                "text": "With this concern in mind, it seems reasonable to ask how well these 65 lines for \"save ... from\" fit in with all other uses of \"save\"?. A laborious concordance analysis was undertaken to answer this question. When it was nearing completion, we noticed that the tags that we were inventing to capture the generalizations could in most cases have been suggested by looking at the lexical items listed in the association ratio table for \"save\". For example, we had failed to notice the significance of time adverbials in our analysis of \"save,\" and no 2. The last unclassifaat line, \"...save shoppers anywhere from $S0...\" raises imeres~g problems. Syntactic \"chunking\" shows that, in spite of its ~o-coearreaoe of \"from\" with \"save\", this line does ant belong hm'e. An intriguing exerciw, given the lookup table we are trying to construct, is how to guard against false inferences such u that since \"shoppm's\" is tagged [PERSON], \"$$0 to 5500\" must here count u either BAD m\" a LOCATION. Accidental coincidmlces of this kind do not have a significant effect on the measure, however, although they do secve as a reminder of the probabilistic nature of the findings. dictionary records this. Yet it should be clear from the association ratio table above that \"annually\" and \"month ''3 are commonly found with \"save\". More detailed inspection shows that the time adverbials correlate interestingly with just one group of \"save\" objects, namely those tagged [MONEY] . The AP wire is fuU of discussions of \"saving $1.2 billion per month\"; computational lexicography should measure and record such patterns ff they are general, even when traditional dictionaries do not. If we had looked at the association ratio tables before labeling the 65 lines for \"save ... from,\" we might have noticed the very large value for \"save ... forests,\" suggesting that there may be an important pattern here.",
                "cite_spans": [
                    {
                        "start": 1453,
                        "end": 1460,
                        "text": "[MONEY]",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Example 2: Identifying Semantic Classes",
                "sec_num": "6.2"
            },
            {
                "text": "In fact, this pattern probably subsumes most of the occurrences of the \"save [ANIMAL]\" pattern noticed in Figure 2 . Thus, tables do not provide semantic tags, but they provide a powerful set of suggestions to the lexicographer for what needs to be accounted for in choosing a set of semantic tags.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 106,
                        "end": 114,
                        "text": "Figure 2",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Example 2: Identifying Semantic Classes",
                "sec_num": "6.2"
            },
            {
                "text": "It may be that everything said here about \"save\" and other words is true only of 1987 American journalese. Intuitively, however, many of the patterns discovered seem to be good candidates for conventions of general English. A future step would be to examine other more balanced corpora and test how well the patterns hold up.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Example 2: Identifying Semantic Classes",
                "sec_num": "6.2"
            },
            {
                "text": "We began this paper with the psycholinguistic notion \u2022 of word association norm, and extended that concept toward the information theoretic def'mition of mutual information. This provided a precise statistical calculation that could be applied to a very 3. The word \"time\" itself also occurs significantly in the table, but on clco~ examination it is clear that this use of \"time\" (e.g., \"to save time\") counts as something like a commodity or resource, not as part of a time adjunct. Such are the pitfalls of lexicography (obvious when they are pointed out).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "large corpus of text in order to produce a table of associations for tens of thousands of words, We were then able to show that the table encoded a number of very interesting patterns ranging from doctor ... nurse to save ... from. We finally concluded by showing how the patterns in the association ratio table might help a lexicographer organize a concordance.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "In point of fact, we actually developed these resuks in basically the reverse order. Concordance analysis is stilt extremely labor-intensive, and prone to errors of omission. The ways that concordances are sorted don't adequately support current lexicographic practice. Despite the fact that a concordance is indexed by a single word, often lexicographers actually use a second word such as \"from\" or an equally common semantic concept such as a time adverbial to decide how to categorize concordance lines. In other words, they use two words to triangulate in on a word sense. This triangulation approach clusters concordance Lines together into word senses based primarily on usage (distributional evidence), as opposed to intuitive notions of meaning. Thus, the question of what is a word sense can be addressed with syntactic methods (symbol pushing), and need not address semantics (interpretation), even though the inventory of tags may appear to have semantic values.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "The triangulation approach requires \"art.\" How does the lexicographer decide which potential cut points are \"interesting\" and which are merely due to chance? The proposed association ratio score provides a practical and objective measure which is often a fairly good approximation to the \"art.\" Since the proposed measure is objective, it can be applied in a systematic way over a large body of material, steadily improving consistency and productivity.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "But on the other hand, the objective score can be misleading. The score takes only distributional evidence into account. For example, the measure favors \"set ... for\" over \"set ... down\"; it doesn't know that the former is less interesting because its semantics are compositional. In addition, the measure is extremely superficial; it cannot cluster words into appropriate syntactic classes without an explicit preprocess such as Church's parts program \"or Hindle's parser. Neither of these preprocesses, though, can help highlight the \"natural\" similarity between nouns such as \"picture\" and \"photograph.\" Although one might imagine a preprocess that would help in this particular case, there will probably always be a class of generalizations that are obvious to an intelligent lexicographer, but lie hopelessly beyond the objectivity of a computer.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "Despite these problems, the association ratio could be an important tool to aid the lexicographer, rather like an index to the concordances, It can help us decide what to look for; it provides a quick summary of what company our words do keep. \u2022 out-of*work steelworker. \" because that doesn't \" We suspend reality when we say we']] scientists has won the first round in an effort to about three children in a mining town who plot to GM executives say the shutdowns will rtmant as receiver, instructed officials to try to",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "The package, which is to newly elshanced image as the moderate who moved to million offer from chairman Victor Posner to help after telling a delivery-room do~or not to try to h birthday Tuesday. cheered by those who fought to at he had formed an ellianco with Moslem rebels to \" Basically we could",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "We worked for a year to their expensive rob'mrs, just like in wartime, to ard of many who risked their own lives in order to",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            },
            {
                "text": "We must inct~tse the amount Americans save China from poverty. save enormous sums of money in contracting out individual c save enough for a down payment on 8 home. save jobs, that costs jobs. \" save money by spending $10,000 in wages for a public works save one of Egypt's great treasures, the decaying tomb of R save the \"pit ponies \"doomed to be slaughtered. save the automak~r $$00 milfion a year in operating costs a save the company rather than liquidate it and then declared save the counU3, nearly $2 billion, also includes a program save the country. save the fmanclaliy troubled company, but said Posner sail save the infant by inserting a tube in its throat to help i save the majestic Beaux Arts architectural masterpie~,e. save the nation from communism. save the operating costs of the Pershings and ground-launch save the site at enormous expense to us. \" said Leveiilee. save them from drunken Yankee brawlers, \"Tass said. save those who were passengers. \" save. \" ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "ConcluMom",
                "sec_num": "7."
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "A Stochastic Pans Program and Noun Phrase Parser for Unrestricted Text",
                "authors": [
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Church",
                        "suffix": ""
                    }
                ],
                "year": 1988,
                "venue": "Second Conference on AppU~ Natural Language Processing",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Church, K., (1988), \"A Stochastic Pans Program and Noun Phrase Parser for Unrestricted Text,\" Second Conference on AppU~ Natural Language Processing, Austin, Texas.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Tranamlx~n of Information",
                "authors": [
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Fano",
                        "suffix": ""
                    }
                ],
                "year": 1961,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Fano, R., (1961), Tranamlx~n of Information, MIT Press, Cambridge, Massechusens.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "A Synopsis of Linguistic Theory 1930-1955",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Firth",
                        "suffix": ""
                    }
                ],
                "year": 1957,
                "venue": "Smdiea in l.AnguLvd\u00a2 Analysis",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Firth, J., (1957), \"A Synopsis of Linguistic Theory 1930-1955\" in Smdiea in l.AnguLvd\u00a2 Analysis, Philological Society, Oxford;",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Frequency AnalysiJ of EnglhOt U,~&e",
                "authors": [
                    {
                        "first": "W",
                        "middle": [],
                        "last": "Pranch",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Kucera",
                        "suffix": ""
                    }
                ],
                "year": 1982,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Pranch, W., and Kucera, H., (1982), Frequency AnalysiJ of EnglhOt U,~&e, Houghton Mifflin Company, Boston.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "The Population Frequemctea of Species and the F..tttnmrlan of Population Parametera",
                "authors": [
                    {
                        "first": "I",
                        "middle": [
                            "J"
                        ],
                        "last": "Good",
                        "suffix": ""
                    }
                ],
                "year": 1953,
                "venue": "Biomelxika",
                "volume": "40",
                "issue": "",
                "pages": "237--264",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Good, I. J., (1953), The Population Frequemctea of Species and the F..tttnmrlan of Population Parametera, Biomelxika, Vol. 40, pp, 237-264.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Definitions and Explanations",
                "authors": [
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Hanks",
                        "suffix": ""
                    }
                ],
                "year": 1987,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hanks, P. (198\"0, \"Definitions and Explanations,\" in Sinclair (1987b).",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Deterministic Parsing of Syntactic Nonfluancks",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Hindle",
                        "suffix": ""
                    }
                ],
                "year": 1983,
                "venue": "ACL Proceedings",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hindle, D., (1983a), \"Deterministic Parsing of Syntactic Non- fluancks,\" ACL Proceedings.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "User manual for Fidditch, a deterministic parser",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Hindle",
                        "suffix": ""
                    }
                ],
                "year": 1983,
                "venue": "Naval Research Laboratory Technical Memorandum",
                "volume": "",
                "issue": "",
                "pages": "7590--142",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hindle, D., (1983b), \"User manual for Fidditch, a deterministic parser,\" Naval Research Laboratory Technical Memorandum \u00a27590-142",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "The Advanced Learner's D/cn'onary",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Hornby",
                        "suffix": ""
                    }
                ],
                "year": 1948,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hornby, A., (1948), The Advanced Learner's D/cn'onary, Oxford Univenity Press.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "On the Recognition of Printed Characters of any Font or She",
                "authors": [
                    {
                        "first": "$",
                        "middle": [],
                        "last": "Kahaa",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Pavlidis",
                        "suffix": ""
                    },
                    {
                        "first": "H",
                        "middle": [],
                        "last": "Baird",
                        "suffix": ""
                    }
                ],
                "year": 1987,
                "venue": "IEEE Transections PAMI",
                "volume": "",
                "issue": "",
                "pages": "274--287",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Kahaa, $., Pavlidis, T., and Baird, H., (1987) \"On the Recognition of Printed Characters of any Font or She,\" IEEE Transections PAMI, pp. 274-287.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Loci of Contextual Effects on Visual Word-Reoognition",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Meyer",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Schvaneveldt",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Ruddy",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Rabbin",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Domic",
                        "suffix": ""
                    }
                ],
                "year": 1975,
                "venue": "Attention and Performance V",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Meyer, D., Schvaneveldt, R.. and Ruddy, M., (1975), \"Loci of Contextual Effects on Visual Word-Reoognition,\" in Rabbin, P., and Domic, S., (ads.), Attention and Performance V, Academic Press, London, New York, San PrantAwo.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Word Asr,~:iation Norms",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Pakn-Mo",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Jenkins",
                        "suffix": ""
                    }
                ],
                "year": 1964,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Pakn-mo, D,, and Jenkins, J., (1964) \"Word Asr,~:iation Norms,\" University of Minnesota Press, Minn~po~.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "CoUtma Cobulld Engllah Language DlcrlanaW",
                "authors": [
                    {
                        "first": "",
                        "middle": [],
                        "last": "Sine",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Lair",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Hanks",
                        "suffix": ""
                    },
                    {
                        "first": "G",
                        "middle": [],
                        "last": "Fox",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Moon",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Stock",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sine.lair, J., Hanks, P., Fox, G., Moon, R., Stock, P. (ads), (1997a), CoUtma Cobulld Engllah Language DlcrlanaW, Collins, London and Glasgow.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Looking Up: an account of the COBUILD Project in lexical co",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Sinclair",
                        "suffix": ""
                    }
                ],
                "year": null,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Sinclair, J., (lgSTo), \"The Nature of the Evidence,\" in Sinclair, J. (ed.), Looking Up: an account of the COBUILD Project in lexical co.orang, Collins, London and Glasgow.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "type_str": "figure",
                "num": null,
                "text": "can be viewed as a formalization of Sinciair's argument: \"How common are the phrasal verbs with set7 Set is particularly rich in making combinations with words like about, in, up, out, on, off, and these words are themselves very common. How likely is set off to occur? Both are frequent words; [set occurs approximately 250 times in a million words and] off occurs approximately 556 times in a million words... IT]he question we are asking can be roughly rephrased as follows: how Likely is off to occur immediately after set? ... This is 0.00025x0.00055 [P(x) P(y)], which gives us the tiny figure of 0.0000001375 ... The assumption behind this calculation is that the words are distributed at random in a text [at chance, in our terminology]. It is obvious to a linguist that this is not so, and a cough measure of how much set and off attract each other is to cumpare the probability with what actually happens... $~ off o~urs nearly 70 times in the 7.3 million word corpus [P(x,y)-70/(7.3 106) >> P(x) P(y)].",
                "uris": null
            },
            "FIGREF1": {
                "type_str": "figure",
                "num": null,
                "text": "Short Sample of the Concordance to \"Save\" from the AP 1987 Corpus rs Sunday, ~aIlins for greater economic reforms to mmts.qion af~efted that \" the Postai Servi~ COUld Then, she said. the family hopes to",
                "uris": null
            },
            "FIGREF2": {
                "type_str": "figure",
                "num": null,
                "text": "Some AP 1987 Concordance lines to 'save ... from,' roughly sorted into categories save X from Y (6S concordance lines) 1 save PERSON from Y (23 concordance lanes)",
                "uris": null
            },
            "TABREF0": {
                "num": null,
                "text": "",
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td>1. This definition fw(x,y) uses \u2022 rectangular window. It might bc interesting to consider alternatives (e.g., \u2022 triangular window or \u2022 decaying exponential) that would weight words less and less as they are separated from sexism to syntax. Asymmetry in 1988 AP Corpus ('N ffi 36 million) x y fix,y) fly, x) doctors nurses 81 10 man woman 209 42 doctors lawyers 25 16 bread butter 14 0 save life 106 8 save money 155 8 save from 144 16 supposed to 982 21 Secondly, one might expect f(x,y)&lt;-f(x) and f(x,y) ~f(y), but the way we have been counting, this needn't be the case if x and y happen to appear several times in the window. For example, given the sentence, \"Library workers were prohibited from saving books from this heap of ruins,\" which appeared in an AP story on April l, 1988, f(prohibited) ffi 1 and f(prohibited, from) ffi 2. This problem can he fixed by dividing f(x,y) by</td></tr></table>"
            },
            "TABREF4": {
                "num": null,
                "text": "save PERSON &ore (BAD) LOC(ATION) (4 concordance lines) rescoers who helped save the toddler[pERSON] from an abandoned weli['LOC] will be feted with a parade while attempting to save two drowning boys[PERSON] from a turbulent[BAD] creek[LOC] in Ohio[LOCI 2. save INSTtTFUTION) &ore (ECON) BAD (27 concordance lines) membe~ states to help save the BEC[INST] from possible bankrnptcy[BCONJ[BAD] this year. should be sought \"to save the company[CORP[lNST]] from bankruptey(ECON][BAD] . law was necessary to save the cuuntry[NATION[INST]] from disast~[BAD] . operation \" to save the nafion[NATION[INST]] from Communism[BAD]~q3LITICAL] , were not needed to save the system from bankrnptcy[ECON][BAD] . his efforts to save the world[IN'ST] from the likes of Lothar and the Spider Woman",
                "html": null,
                "type_str": "table",
                "content": "<table><tr><td colspan=\"2\">1.1 save PERSON from BAD (19 concordance lines)</td></tr><tr><td>( Robert DeNiro ) to</td><td>save Indian Iribes[PERSON] from se~ocide[DESTRUCT[BAD]] at the hands of</td></tr><tr><td>'~ We wanted to</td><td>save him[PERSON] from undue uouble[BAD] and loti[BAD] of money, \"</td></tr><tr><td>Murphy WLV sacriflcod to</td><td>save more powerful Democrats[PERsoN] from harm[BAD] .</td></tr><tr><td>\"God sent this man to</td><td>save my five children[PERsoN] from being burned to death[DESTRUCT[BAD]] and</td></tr><tr><td>Pope John Paul H to \"</td><td>save us[PERSON] from sin[BAD] . \"</td></tr><tr><td colspan=\"2\">1.2 3. save ANIMAL ~'om DESTRUCT(ION) (5 concordance lines)</td></tr><tr><td>sire them the money to</td><td>rove the dogs[ANIMAL] from being des~'oyed[DESTRUCT] ,</td></tr><tr><td>pmgrem intended to</td><td>save the slant birds(ANIMAL] from extinction[DESTRUCT] ,</td></tr><tr><td>UNCLASSIFIED (10</td><td>concordance lines)</td></tr><tr><td>wainut and ash trees to</td><td/></tr><tr><td>after the attack to,</td><td/></tr><tr><td>~.n'~t~ttes that would</td><td/></tr></table>"
            }
        }
    }
}