File size: 66,393 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
{
    "paper_id": "P99-1002",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T09:32:05.344665Z"
    },
    "title": "AUTOMATIC SPEECH RECOGNITION AND ITS APPLICATION TO INFORMATION EXTRACTION",
    "authors": [
        {
            "first": "Sadaoki",
            "middle": [],
            "last": "Furui",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "Tokyo institute of Technology",
                "location": {
                    "addrLine": "2-12-1, Meguro-ku",
                    "postCode": "152-8552",
                    "settlement": "Ookayama, Tokyo",
                    "country": "Japan"
                }
            },
            "email": "furui@cs.titech.ac.jp"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "This paper describes recent progress and the author's perspectives of speech recognition technology. Applications of speech recognition technology can be classified into two main areas, dictation and human-computer dialogue systems.",
    "pdf_parse": {
        "paper_id": "P99-1002",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "This paper describes recent progress and the author's perspectives of speech recognition technology. Applications of speech recognition technology can be classified into two main areas, dictation and human-computer dialogue systems.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "In the dictation domain, the automatic broadcast news transcription is now actively investigated, especially under the DARPA project. The broadcast news dictation technology has recently been integrated with information extraction and retrieval technology and many application systems, such as automatic voice document indexing and retrieval systems, are under development. In the human-computer interaction domain, a variety of experimental systems for information retrieval through spoken dialogue are being investigated. In spite of the remarkable recent progress, we are still behind our ultimate goal of understanding free conversational speech uttered by any speaker under any environment. This paper also describes the most important research issues that we should attack in order to advance to our ultimate goal of fluent speech recognition.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "pattern recognition paradigm, a data-driven approach which makes use of a rich set of speech utterances from a large population of speakers, the use of stochastic acoustic and language modeling, and the use of dynamic programmingbased search methods.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "A series of (D)ARPA projects have been a major driving force of the recent progress in research on large-vocabulary, continuous-speech recognition. Specifically, dictation of speech reading newspapers, such as north America business newspapers including the Wall Street Journal (WSJ), and conversational speech recognition using an Air Travel Information System (ATIS) task were actively investigated. More recent DARPA programs are the broadcast news dictation and natural conversational speech recognition using Switchboard and Call Home tasks. Research on human-computer dialogue systems, the Communicator program, has also started [ 1 ] . Various other systems have been actively investigated in US, Europe and Japan stimulated by DARPA projects. Most of them can be classified into either dictation systems or human-computer dialogue systems.",
                "cite_spans": [
                    {
                        "start": 635,
                        "end": 640,
                        "text": "[ 1 ]",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "The field of automatic speech recognition has witnessed a number of significant advances in the past 5 -10 years, spurred on by advances in signal processing, algorithms, computational architectures, and hardware. These advances include the widespread adoption of a statistical Figure 1 shows a mechanism of state-of-the-art speech recognizers [2] . Common features of these systems are the use of cepstral parameters and their regression coefficients as speech features, triphone HMMs as acoustic models, vocabularies of several thousand or several ten thousand entries, and stochastic language models such as bigrams and trigrams. Such methods have been applied not only to English but also to French, German, Italian, Spanish, Chinese and Japanese. Although there are several languagespecific characteristics, similar recognition results have been obtained. The remainder of this paper is organized as follows. Section 2 describes recent progress in broadcast news dictation and its application to information extraction, and Section 3 describes human-computer dialogue systems. In spite of the remarkable recent progress, we are still far behind our ultimate goal of understanding free conversational speech uttered by any speaker under any environment. Section 4 describes how to increase the robustness of speech recognition, and Section 5 describes perspectives of linguistic modeling for spontaneous speech recognition/ understanding. Section 6 concludes the paper.",
                "cite_spans": [
                    {
                        "start": 344,
                        "end": 347,
                        "text": "[2]",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 278,
                        "end": 286,
                        "text": "Figure 1",
                        "ref_id": "FIGREF0"
                    }
                ],
                "eq_spans": [],
                "section": "INTRODUCTION",
                "sec_num": "1."
            },
            {
                "text": "With the introduction of the broadcast news test bed to the DARPA project in 1995, the research effort took a profound step forward. Many of the deficiencies of the WSJ domain were resolved in the broadcast news domain [3] . Most importantly, the fact that broadcast news is a real-",
                "cite_spans": [
                    {
                        "start": 219,
                        "end": 222,
                        "text": "[3]",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "DARPA Broadcast News Dictation Project",
                "sec_num": "2.1"
            },
            {
                "text": "We have been developing a largevocabulary continuous-speech recognition (LVCSR) system for Japanese broadcast-news speech transcription [4] [5] . This is a part of a joint research with the NHK broadcast company whose goal is the closed-captioning of TV programs. The broadcast-news manuscripts that were used for constructing the language models were taken from the period between July 1992 \u2022 and May 1996, and comprised roughly 500k sentences and 22M words. To calculate word ngram language models, we segmented the broadcast-news manuscripts into words by using a morphological analyzer since Japanese sentences are written without spaces between words. A word-frequency list was derived for the news manuscripts, and the 20k most frequently used words were selected as vocabulary words. This 20k vocabulary covers about 98% of the words in the broadcast-news manuscripts. We calculated bigrams and trigrams and estimated unseen n-grams using Katz's back-off smoothing method.",
                "cite_spans": [
                    {
                        "start": 136,
                        "end": 139,
                        "text": "[4]",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 140,
                        "end": 143,
                        "text": "[5]",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Japanese Broadcast News Dictation System",
                "sec_num": "2.2"
            },
            {
                "text": "Japanese text is written by a mixture of three kinds of characters: Chinese characters (Kanji) and two kinds of Japanese characters (Hira-gana and Kata-kana). Most Kanji have multiple readings, and correct readings can only be decided according to context. Conventional language models usually assign equal probability to all possible readings of each word. This causes recognition errors because the assigned probability is sometimes very different from the true probability. We therefore constructed a language model that depends on the readings of words in order to take into account the frequency and context-dependency of the readings. Broadcast news speech includes filled pauses at the beginning and in the middle of sentences, which cause recognition errors in our language models that use news manuscripts written prior to broadcasting. To cope with this problem, we introduced filled-pause modeling into the language model. News speech data, from TV broadcasts in July 1996, were divided into two parts, a clean part and a noisy part, and were separately evaluated. The clean part consisted of utterances with no background noise, and the noisy part consisted of utterances with background noise. The noisy part included spontaneous speech such as reports by correspondents. We extracted 50 male utterances and 50 female utterances for each part, yielding four evaluation sets; male-clean (m/c), male-noisy (m/n), female-clean (f/c), femalenoisy (fin). Each set included utterances by five or six speakers. All utterances were manually segmented into sentences. Table 1 shows the experimental results for the baseline language model (LM 1) and the new language models. LM2",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 1572,
                        "end": 1579,
                        "text": "Table 1",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Japanese Broadcast News Dictation System",
                "sec_num": "2.2"
            },
            {
                "text": "is the reading-dependent language model, and LM3 is a modification of LM2 by filled-pause modeling. For clean speech, LM2 reduced the word error rate by 4.7 % relative to LM1, and LM3 model reduced the word error rate by 10.9 % relative to LM2 on average.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Japanese Broadcast News Dictation System",
                "sec_num": "2.2"
            },
            {
                "text": "News is filled with events, people, and organizations and all manner of relations among them. The great richness of material and the naturally evolving content in broadcast news has leveraged its value into areas of research well beyond speech recognition. In the DARPA project, the Spoken Document ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Information Extraction in the DARPA Project",
                "sec_num": "2.3"
            },
            {
                "text": "Summarizing transcribed news speech is useful for retrieving or indexing broadcast news. We investigated a method for extracting topic words from nouns in the speech recognition results on the basis of a significance measure [4] [5] . The extracted topic words were compared with \"true\" topic words, which were given by three human subjects. The results are shown in Figure 2 .",
                "cite_spans": [
                    {
                        "start": 225,
                        "end": 228,
                        "text": "[4]",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 229,
                        "end": 232,
                        "text": "[5]",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 367,
                        "end": 375,
                        "text": "Figure 2",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Information Extraction from Japanese Broadcast News",
                "sec_num": "2.4"
            },
            {
                "text": "When the top five topic words were chosen (recall=13%), 87% of them were correct on average. ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Information Extraction from Japanese Broadcast News",
                "sec_num": "2.4"
            },
            {
                "text": "Recently a number of sites have been working on human-computer dialogue systems. The followings are typical examples.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Typical Systems in US and Europe",
                "sec_num": "3.1"
            },
            {
                "text": "The University of Karlsruhe focuses its speech research on a content-addressable multimedia information retrieval system, under a multi-lingual environment, where queries and multimedia documents may appear in multiple languages [7] . The system is called \"View4You\" and their research is conducted in cooperation with the Informedia project at CMU [6] . In the View4You system, German and Servocroatian public newscasts are recorded daily. The newscasts are automatically segmented and an index is created for each of the segments by means of automatic speech recognition. The user can query the system in natural language by keyboard or through a speech utterance. The system returns a list of segments which is sorted by relevance with respect to the user query. By selecting a segment, the user can watch the corresponding part of the news show on his/her computer screen.",
                "cite_spans": [
                    {
                        "start": 229,
                        "end": 232,
                        "text": "[7]",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 349,
                        "end": 352,
                        "text": "[6]",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "(a) The View4You system at the University of Karksruhe",
                "sec_num": null
            },
            {
                "text": "The system overview is shown in Fig. 3 . segments the speech archive and generates transcripts, (2) an information-retrieval engine which indexes the transcriptions and formulates hypotheses regarding document relevance to user-submitted queries and (3) a graphical-userinterface which supports search and local contextual navigation based on the machinegenerated transcripts and graphical representations of query-keyword distribution in the retrieved speech transcripts. The speech recognition component of SCAN includes an intonational phrase boundary detection module and a classification module, These subcomponents preprocess the speech data before passing the speech to the recognizer itself.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 32,
                        "end": 38,
                        "text": "Fig. 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "(a) The View4You system at the University of Karksruhe",
                "sec_num": null
            },
            {
                "text": "The GALAXY-II conversational system at MIT Galaxy is a clientserver architecture developed at MIT for accessing online information using spoken dialogue [9] . Ithas served as the testbed for developing human language Phone technology at MIT for several years. Recently, they have initiated a significant redesign of the GALAXY architecture to make it easier for researchers to develop their own applications, using either exclusively their own servers or intermixing them with servers developed by others. This redesign was done in part due to the fact that GALAXY has been designed as the first reference architecture for the new DARPA Communicator program. The resulting configuration of the GALAXY-II architecture is shown in Fig. 5 . The boxes in this figure represent various human language technology servers as well as information and domain servers. The label in italics next to each box identifies the corresponding MIT system component. Interactions between servers are mediated by the hub and managed in the hub script. A particular dialogue session is initiated by a user either through interaction with a graphical interface at a Web site, through direct telephone dialup, or through a desktop agent.",
                "cite_spans": [
                    {
                        "start": 153,
                        "end": 156,
                        "text": "[9]",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 729,
                        "end": 735,
                        "text": "Fig. 5",
                        "ref_id": "FIGREF3"
                    }
                ],
                "eq_spans": [],
                "section": "(c)",
                "sec_num": null
            },
            {
                "text": "Text-to-speech ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "DECTALK & ENVOICE",
                "sec_num": null
            },
            {
                "text": "The ARISE (Automatic Railway Information Systems for Europe) projects aims developing prototype telephone information services for train travel information in several European countries [ 10] . In collaboration with the Vecsys company and with the SNCF (the French Railways), LIMSI has developed a prototype telephone service providing timetables, simulated fares and reservations, and information on reductions and services for the main French intercity connections. A prototype French/English service for the high speed trains between Paris and London is also under development. The system is based on the spoken language systems developed for the RailTel project [11] and the ESPRIT Mask project [12] . Compared to the RailTel system, the main advances in ARISE are in dialogue management, confidence measures, inclusion of optional spell mode for ci, ty/station names, and barge-in capability to allow more natural interaction between the user and the machine.",
                "cite_spans": [
                    {
                        "start": 186,
                        "end": 191,
                        "text": "[ 10]",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 666,
                        "end": 670,
                        "text": "[11]",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 699,
                        "end": 703,
                        "text": "[12]",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "(d) The ARISE train travel information system at LIMSI",
                "sec_num": null
            },
            {
                "text": "We have recently investigated a paradigm for designing multimodal dialogue systems [ 13] . An example task of the system was to retrieve particular information about different shops in the Tokyo Metropolitan area, such as their names, addresses and phone numbers. The system accepted speech and screen touching as input, and presented retrieved information on a screen display or by synthesized speech as shown in Fig.  6 . The speech recognition part was modeled by the FSN (finite state network) consisting of keywords and fillers, both of which were implemented by the DAWG (directed acyclic word-graph) structure. The number ofkeywords was 306, consisting of district names and business names. hypothesized word, and if the score is below an empirically determined threshold, the hypothesized word is marked as uncertain. The uncertain words are ignored by the understanding component or used by the dialogue manager to start clarification subdialogues.",
                "cite_spans": [
                    {
                        "start": 83,
                        "end": 88,
                        "text": "[ 13]",
                        "ref_id": "BIBREF11"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 414,
                        "end": 421,
                        "text": "Fig.  6",
                        "ref_id": "FIGREF4"
                    }
                ],
                "eq_spans": [],
                "section": "Designing a Multimodal Dialogue System for Information Retrieval",
                "sec_num": "3.2"
            },
            {
                "text": "was calculated for each strategy, and the best strategy was selected according to the keyword recognition accuracy.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Designing a Multimodal Dialogue System for Information Retrieval",
                "sec_num": "3.2"
            },
            {
                "text": "Ultimately, speech recognition systems should be capable of f robust, speakerindependent or speakeradaptive, continuous speech recognition\u2022 Figure 7 shows main causes of acoustic variation in speech [14] . ~. It is crucial to establish methods that are robust against voice variation due to individuality, the physical and psychological condition of the speaker, telephone sets, microphones, network characteristics, additive background noise, speaking styles, and so on. Figure 8 shows main methods for making speech recognition systems robust against voice variation. It is also important for the systems to impose few restrictions on tasks and vocabulary. To solve these problems, it is essential to develop automatic adaptation techniques.",
                "cite_spans": [
                    {
                        "start": 199,
                        "end": 203,
                        "text": "[14]",
                        "ref_id": "BIBREF12"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 140,
                        "end": 148,
                        "text": "Figure 7",
                        "ref_id": null
                    },
                    {
                        "start": 472,
                        "end": 480,
                        "text": "Figure 8",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Automatic adaptation",
                "sec_num": "4.1"
            },
            {
                "text": "Extraction and normalization of. (adaptation to) voice individuality is one of the most important issues [ 14] . A small percentage of people occasionally cause systems to produce exceptionally low recognition rates\u2022 This is an example of the \"sheep and goats\" phenomenon. instantaneous/incremental adaptation is ideal, since the system works as if it were a speakerindependent system, and it performs increasingly better as it is used. However, since we have to adapt many phonemes using a limited size of utterances including only a limited number of phonemes, it is crucial to use reasonable modeling of speaker-to-speaker variablity or constraints. Modeling of the mechanism of speech production is expected to provide a useful modeling of speaker-to-speaker variability.",
                "cite_spans": [
                    {
                        "start": 105,
                        "end": 110,
                        "text": "[ 14]",
                        "ref_id": "BIBREF12"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Automatic adaptation",
                "sec_num": "4.1"
            },
            {
                "text": "Since, in broadcast news, each speaker utters several sentences in succession, the recognition error rate can be reduced by adapting acoustic models incrementally within a segment that contains only one speaker. We applied on-line, unsupervised, instantaneous and incremental speaker adaptation combined with automatic detection of speaker changes [4] . The MLLR [ 15] -MAP [ 16] and VFS (vector-field smoothing) [17] methods were instantaneously and incrementally carried out for each utterance. The adaptation process is as follows. For the first input utterance, the speaker-independ\u00a2nt model is used for both recognition and adaptation, and the first speaker-adapted model is created. For the second input utterance, the likelihood value of the utterance given the speaker-independent model and that given the speaker-adapted model are calculated and compared. If the former value is larger, the utterance is considered to be the beginning of a new speaker, and another speakeradapted model is created. Otherwise, the existing speaker-adapted model is incrementally adapted. For the succeeding input utterances, speaker changes are detected in the same way by comparing the acoustic likelihood values of each utterance obtained from the speaker-independent model and some speaker-adapted models. If the speaker-independent model yields a larger likelihood than any of the speaker-adapted models, a speaker change is detected and a new speaker-adapted model is constructed. Experimental results show that the adaptation reduced the word error rate by 11.8 % relative to the speaker-independent models.",
                "cite_spans": [
                    {
                        "start": 348,
                        "end": 351,
                        "text": "[4]",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 363,
                        "end": 368,
                        "text": "[ 15]",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 374,
                        "end": 379,
                        "text": "[ 16]",
                        "ref_id": "BIBREF14"
                    },
                    {
                        "start": 413,
                        "end": 417,
                        "text": "[17]",
                        "ref_id": "BIBREF15"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "On-line speaker adaptation in broadcast news dictation",
                "sec_num": "4.2"
            },
            {
                "text": "One of the most important issues for speech recognition is how to create language models (rules) for spontaneous speech. When recognizing spontaneous speech in dialogues, it is necessary to deal with variations that are not encountered when recognizing speech that is read from texts. These variations include extraneous words, out-of-vocabulary words, ungrammatical sentences, disfluency, partial words, repairs, hesitations, and repetitions. It is crucial to develop robust and flexible parsing algorithms that match the characteristics of spontaneous speech. A paradigm shift from the present transcription-based approach to a detection-based approach will be important to solve such problems [2] . How to extract contextual information, predict users' responses, and focus on key words are very important issues.",
                "cite_spans": [
                    {
                        "start": 696,
                        "end": 699,
                        "text": "[2]",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Language modeling for spontaneous speech recognition",
                "sec_num": "5.1"
            },
            {
                "text": "Stochastic language modeling, such as bigrams and trigrams, has been a very powerful tool, so it would be very effective to extend its utility by incorporating semantic knowledge. It would also be useful to integrate unification grammars and context-free grammars for efficient word prediction. Style shifting is also an important problem in spontaneous speech recognition. In typical laboratory experiments, speakers are reading lists of words rather than trying to accomplish a real task. Users actually trying to accomplish a task, however, use a different linguistic style. Adaptation of linguistic models according to tasks, topics and speaking styles is a very important issue, since collecting a large linguistic database for every new task is difficult and costly.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Language modeling for spontaneous speech recognition",
                "sec_num": "5.1"
            },
            {
                "text": "automatic speech recognition will continue to find applications, such as meeting/conference summarization, automatic closed captioning, and interpreting telephony. It is expected that speech recognizer will become the main input device of the \"wearable\" computers that are now actively investigated. In order to materialize these applications, we have to solve many problems. The most important issue is how to make the speech recognition systems robust against acoustic and lingustic variation in speech. In this context, a paradigm shitt from speech recognition to understanding where underlying messages of the speaker, that is, meaning/context that the speaker intended to convey are extracted, instead of transcribing all the spoken words, will be indispensable.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Language modeling for spontaneous speech recognition",
                "sec_num": "5.1"
            }
        ],
        "back_matter": [
            {
                "text": "State-of-the-art automatic speech recognition systems employ the criterion of maximizing P(/4,qX), where W is a word sequence, and X is an acoustic observation sequence. This criterion is reasonable for dictating read speech. However, the ultimate goal of automatic speech recognition is to extract the underlying messages of the speaker from the speech signals. Hence we need to model the process of speech generation and recognition as shown in Fig. 9 [ 18] , where M is the message (content) that a speaker intended to convey. models in the same way as in usual recognition processes. We assume that P(M) has a uniform probability for all M. Therefore, we only need to consider further the term P(~M). We assume that P(~M) can be expressed as follows. where ~, 0<-/1.<1, is a weighting factor. P(W), the first term of the right hand side, represents a part of P(~M) that is independent of Mand can be given by a general statistical language model. P'(WIM), the second term of the right hand side, represents the part ofP(WIA D that depends on M. We consider that M is represented by a co-occurrence of words based on the distributional hypothesis by Harris [ 19] . Since this approach formulates P'(WIM) without explicitly representing M, it can use information about the speaker's message M without being affected by the quantization problem of topic classes. This new formulation of speech recognition was applied to the Japanese broadcast news dictation, and it was found that word error rates for the clean set were slightly reduced by this method.",
                "cite_spans": [
                    {
                        "start": 454,
                        "end": 459,
                        "text": "[ 18]",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 1160,
                        "end": 1165,
                        "text": "[ 19]",
                        "ref_id": "BIBREF17"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 447,
                        "end": 453,
                        "text": "Fig. 9",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Message-Driven Speech Recognition",
                "sec_num": "5.2"
            },
            {
                "text": "( Based on the progress, various application systems have been developed using dictation and spoken dialogue technology. One of the most important applications is information extraction and retrieval. Using the speech recognition technology, broadcast news can be automatically indexed, producing a wide range of capabilities for browsing news archives interactively. Since speech is the most natural and efficient communication method between humans,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "maxP(MIX) = max]~ P(MIW)P(WIX).",
                "sec_num": null
            }
        ],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Future directions in speech information processing",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Furui",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proc. 16th ICA and 135th Meeting ASA",
                "volume": "",
                "issue": "",
                "pages": "1--4",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Furui: \"Future directions in speech information processing\", Proc. 16th ICA and 135th Meeting ASA, Seattle, pp. 1-4 (1998)",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Broadcast news is good news",
                "authors": [
                    {
                        "first": "F",
                        "middle": [],
                        "last": "Kubala",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "DARPA Broadcast News Workshop",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "F. Kubala: \"Broadcast news is good news\", DARPA Broadcast News Workshop, Virginia (1999)",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Improvements in Japanese broadcast news transcription",
                "authors": [
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Ohtsuki",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Furui",
                        "suffix": ""
                    },
                    {
                        "first": "N",
                        "middle": [],
                        "last": "Sakurai",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Iwasaki",
                        "suffix": ""
                    },
                    {
                        "first": "Z.-P",
                        "middle": [],
                        "last": "Zhang",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K. Ohtsuki, S. Furui, N. Sakurai, A. Iwasaki and Z.-P. Zhang: \"Improvements in Japanese broadcast news transcription\", DARPA Broadcast News Workshop, Virginia (1999)",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "~lessage-driven speech recognition and topicword extraction",
                "authors": [
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Ohtsuki",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Furui",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Iwasaki",
                        "suffix": ""
                    },
                    {
                        "first": "N",
                        "middle": [],
                        "last": "Sakurai",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "Proc. IEEE Int. Conf. Acoust., Speech, Signal Process",
                "volume": "",
                "issue": "",
                "pages": "625--628",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K. Ohtsuki, S. Furui, A. Iwasaki and N. Sakurai: \"~lessage-driven speech recognition and topic- word extraction\", Proc. IEEE Int. Conf. Acoust., Speech, Signal Process., Phoenix, pp. 625-628 (1999)",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Speech recognition and information retrieval: Experiments in retrieving spoken documents",
                "authors": [
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Witbrock",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [
                            "G"
                        ],
                        "last": "Hauptmann",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "Proc. DARPA Speech Recognition Workshop",
                "volume": "",
                "issue": "",
                "pages": "160--164",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "M. Witbrock and A. G. Hauptmann: \"Speech recognition and information retrieval: Experiments in retrieving spoken documents\", Proc. DARPA Speech Recognition Workshop, Virginia, pp. 160-164 (1997). See also http:// www.informedia.cs.cmu.edu/",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "The interactive systems labs View4You video indexing system",
                "authors": [
                    {
                        "first": "T",
                        "middle": [],
                        "last": "Kemp",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Geutner",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Schmidt",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Tomaz",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Weber",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Westphal",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Waibel",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proc. Int. Conf. Spoken Language Processing",
                "volume": "",
                "issue": "",
                "pages": "1639--1642",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "T. Kemp, P. Geutner, M. Schmidt, B. Tomaz, M. Weber, M. Westphal and A. Waibel: \"The interactive systems labs View4You video indexing system\", Proc. Int. Conf. Spoken Language Processing, Sydney, pp. 1639-1642 (1998)",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "SCAN -speech content based audio navigator: a systems overview",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Choi",
                        "suffix": ""
                    },
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Hindle",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Hirschberg",
                        "suffix": ""
                    },
                    {
                        "first": "I",
                        "middle": [],
                        "last": "Magrin-Chagnolleau",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Nakatani",
                        "suffix": ""
                    },
                    {
                        "first": "F",
                        "middle": [],
                        "last": "Pereira",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Singhal",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Whittaker",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proc. Int. Conf. Spoken Language Processing",
                "volume": "",
                "issue": "",
                "pages": "2867--2870",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. Choi, D. Hindle, J. Hirschberg, I. Magrin- Chagnolleau, C. Nakatani, F. Pereira, A. Singhal and S. Whittaker: \"SCAN -speech content based audio navigator: a systems overview\", Proc. Int. Conf. Spoken Language Processing, Sydney, pp. 2867-2870 (1998)",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "GALAXY-II: a reference architecture for conversational system development",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Seneff",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Hurley",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Lau",
                        "suffix": ""
                    },
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Pao",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Schmid",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Zue",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proc. Int. Conf. Spoken Language Processing",
                "volume": "",
                "issue": "",
                "pages": "931--934",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Seneff, E. Hurley, R. Lau, C. Pao, P. Schmid and V. Zue: \"GALAXY-II: a reference architecture for conversational system development\", Proc. Int. Conf. Spoken Language Processing, Sydney, pp. 931-934 (1998)",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Proc. IEEE Int. Conf. Acoust., Speech, Signal Process",
                "authors": [
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Lamel",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Rosset",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "L"
                        ],
                        "last": "Gauvain",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Bennacef",
                        "suffix": ""
                    }
                ],
                "year": 1999,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "501--504",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "L. Lamel, S. Rosset, J. L. Gauvain and S. Bennacef: \"The LIMSI ARISE system for train travel information\", Proc. IEEE Int. Conf. Acoust., Speech, Signal Process., Phoenix, pp. 501-504 (1999)",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "The LIMSI RailTel system: Field trial of a telephone service for rail travel information",
                "authors": [
                    {
                        "first": "L",
                        "middle": [
                            "F"
                        ],
                        "last": "Lamel",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [
                            "K"
                        ],
                        "last": "Bennacef",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Rosset",
                        "suffix": ""
                    },
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Devillers",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Foukia",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "J"
                        ],
                        "last": "Gangolf",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "L"
                        ],
                        "last": "Gauvain",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "",
                "volume": "23",
                "issue": "",
                "pages": "67--82",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "L. F. Lamel, S. K. Bennacef, S. Rosset, L. Devillers, S. Foukia, J. J. Gangolf and J. L. Gauvain: \"The LIMSI RailTel system: Field trial of a telephone service for rail travel information\", Speech Communication, 23, pp. 67-82 (1997)",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Speech recognition for an information Kiosk",
                "authors": [
                    {
                        "first": "J",
                        "middle": [
                            "L"
                        ],
                        "last": "Gauvain",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "J"
                        ],
                        "last": "Gangolf",
                        "suffix": ""
                    },
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Lamel",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proc. Int. Conf. Spoken Language Processing",
                "volume": "",
                "issue": "",
                "pages": "849--852",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. L. Gauvain, J. J. Gangolf and L. Lamel: \"Speech recognition for an information Kiosk\", Proc. Int. Conf. Spoken Language Processing, Philadelphia, pp. 849-852 (1998)",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Designing a multimodal dialogue system for information retrieval",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Furui",
                        "suffix": ""
                    },
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Yamaguchi",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "Proc. Int. Conf. Spoken Language Processing",
                "volume": "",
                "issue": "",
                "pages": "1191--1194",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Furui and K. Yamaguchi: \"Designing a multimodal dialogue system for information retrieval\", Proc. Int. Conf. Spoken Language Processing, Sydney, pp. 1191-1194 (1998)",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Recent advances in robust speech recognition",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Furui",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "Proc. ESCA-NATO Workshop on Robust Speech Recognition for Unknown Communication Channels",
                "volume": "",
                "issue": "",
                "pages": "11--20",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "S. Furui: \"Recent advances in robust speech recognition\", Proc. ESCA-NATO Workshop on Robust Speech Recognition for Unknown Communication Channels, Pont-a-Mousson, France, pp. 11-20 (1997)",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Maximum likelihood linear regression for speaker adaptation of continuous density hidden Markov models",
                "authors": [
                    {
                        "first": "C",
                        "middle": [
                            "J"
                        ],
                        "last": "Leggetter",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [
                            "C"
                        ],
                        "last": "Woodland",
                        "suffix": ""
                    }
                ],
                "year": 1995,
                "venue": "Computer Speech and Language",
                "volume": "",
                "issue": "",
                "pages": "171--185",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "C. J. Leggetter and P. C. Woodland: \"Maximum likelihood linear regression for speaker adaptation of continuous density hidden Markov models\", Computer Speech and Language, pp. 171-185 (1995).",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Maximum a posteriori estimation for multivariate Gaussian mixture observations of Markov chains",
                "authors": [
                    {
                        "first": "J. -L",
                        "middle": [],
                        "last": "Gauvain",
                        "suffix": ""
                    },
                    {
                        "first": "C.-H",
                        "middle": [],
                        "last": "Lee",
                        "suffix": ""
                    }
                ],
                "year": 1994,
                "venue": "IEEE Trans. on Speech and Audio Processing",
                "volume": "2",
                "issue": "2",
                "pages": "291--298",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. -L. Gauvain and C.-H. Lee: \"Maximum a posteriori estimation for multivariate Gaussian mixture observations of Markov chains\" IEEE Trans. on Speech and Audio Processing, 2, 2, pp. 291-298 (1994).",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "Speaker adaptation based on transfer vector field smoothing with continuous mixture density HMMs",
                "authors": [
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Ohkura",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Sugiyama",
                        "suffix": ""
                    },
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Sagayama",
                        "suffix": ""
                    }
                ],
                "year": 1992,
                "venue": "Proc. Int. Conf. Spoken Language Processing",
                "volume": "",
                "issue": "",
                "pages": "369--372",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "K. Ohkura, M. Sugiyama and S. Sagayama: \"Speaker adaptation based on transfer vector field smoothing with continuous mixture density HMMs\", Proc. Int. Conf. Spoken Language Processing, Banff, pp. 369-372 (1992)",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "Automatic speech recognition: Problems, progress & prospects",
                "authors": [
                    {
                        "first": "B.-H",
                        "middle": [],
                        "last": "Juang",
                        "suffix": ""
                    }
                ],
                "year": 1996,
                "venue": "IEEE Workshop on Neural Networks for Signal Processing",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "B.-H. Juang: \"Automatic speech recognition: Problems, progress & prospects\", IEEE Workshop on Neural Networks for Signal Processing (1996)",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "Co-occurrence and transformation in linguistic structure",
                "authors": [
                    {
                        "first": "Z",
                        "middle": [
                            "S"
                        ],
                        "last": "Harris",
                        "suffix": ""
                    }
                ],
                "year": 1957,
                "venue": "",
                "volume": "33",
                "issue": "",
                "pages": "283--340",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Z. S. Harris: \"Co-occurrence and transformation in linguistic structure\", Language, 33, pp. 283- 340 (1957)",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "num": null,
                "text": "Mechanism of state-of-the-art speech recognizers.",
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF1": {
                "num": null,
                "text": "Topic word extraction results.",
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF2": {
                "num": null,
                "text": "The SCAN-speech content based audio navigator at AT&T Labs SCAN (Speech Content based Audio Navigator) is a spoken document retrieval system developed at AT&T Labs integrating speaker-independent, large-vocabulary speech recognition with information-retrieval to support query-based retrieval of information from speech archives[8]. Initial development focused on the application of SCAN to the broadcast news domain. An overview of the system architecture is provided inFig. 4. The system consists of three components: (1) a speaker-independent largevocabulary speech recognition engine which (System overview of the View4You system. Overview of the SCAN spoken document system architecture.",
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF3": {
                "num": null,
                "text": "Architecture of GALAXY-II.",
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF4": {
                "num": null,
                "text": "The fillers accepted roughly 100,000 non-keywords/phrases occuring in spontaneous speech. A variety of dialogue strategies were designed and evaluated based on an objective cost function having a set of actions and states as parameters. Expected dialogue costThe speech recognizer uses n-gram backoff language models estimated on the transcriptions of spoken queries. Since the amount of language model training data is small, some grammatical classes, such as cities, days and months, are used to provide more robust estimates of the ngram probabilitiesMultimodal dialogue system structure for information retrieval.",
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF5": {
                "num": null,
                "text": "Main causes of acoustic variation in speech. [ ............... fClose-talking microphone / (Microphone array Microphone \u2022 fAuditory models Analysis and feature extraction ..... ~(EIH, SMC, PLP) filtering venture-level normmizatiorv/ 1 (n,~t'.tr,'jl .... inn ada t tion r'--x ~'v ......... vv...~ de) composition(PMC) ........................... \"~ Model transformation(MLLR) Model-level t ...... I, Bayesian adaptive learning normalization/Linguisti c processing t .... Language model adaptation Main methods to cope with voice variation in speech recognition.",
                "uris": null,
                "type_str": "figure"
            },
            "TABREF0": {
                "type_str": "table",
                "html": null,
                "content": "<table><tr><td/><td/><td/><td colspan=\"2\">world domain of totally unconstrained, broadcast news</td></tr><tr><td/><td/><td/><td/><td>is a superb stress test that requires new</td></tr><tr><td/><td>Acoustic analysis I ~XI'..X T</td><td/><td/><td>algorithms to work across widely varying conditions. Algorithms need to solve a specific problem without degrading any other condition.</td></tr><tr><td>I</td><td>Gl\u00b0bal search:</td><td colspan=\"2\">~'-P(xr\"xTIwr\"wk) Ph\u00b0nemeinvent\u00b0ryl I</td><td>Another advantage of this domain is</td></tr><tr><td colspan=\"3\">| IP( xr.. xT IWr..wt).P(wr..wt )l maximize</td><td>Pronunciation lexicon[</td><td>that news is easy to collect and the supply of data is boundless. The data</td></tr><tr><td/><td>\u00b0ver Wl'\" wt</td><td colspan=\"2\">J,,P(wl\"\"wk) tLanguagemodel [</td></tr><tr><td/><td>1 Recognized word sequence</td><td/><td/></tr></table>",
                "num": null,
                "text": "obvious value has lead to rapid technology transfer of speech recognition into other research areas and applications. Since the variations in speaking style and accent as well as in channel and environment conditions are"
            },
            "TABREF1": {
                "type_str": "table",
                "html": null,
                "content": "<table><tr><td>Language</td><td/><td colspan=\"2\">Evaluation sets</td><td/></tr><tr><td>model</td><td>m/c</td><td>m/n</td><td>f/c</td><td>f/n</td></tr><tr><td>LM1</td><td>17.6</td><td>37.2</td><td>14.3</td><td>41.2</td></tr><tr><td>LM2</td><td>16.8</td><td>35.9</td><td>13.6</td><td>39.3</td></tr><tr><td>LM3</td><td>14.2</td><td>33.1</td><td>12.9</td><td>38.1</td></tr></table>",
                "num": null,
                "text": "Experimental results of Japanese broadcast news dictation with various language models (word error rate [%])"
            }
        }
    }
}