File size: 66,157 Bytes
6fa4bc9 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 | {
"paper_id": "P86-1024",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T09:12:52.083367Z"
},
"title": "A SENTENCE ANALYSIS METHOD FOR A JAPANESE BOOK READING MACHINE FOR THE BLIND",
"authors": [
{
"first": "Yutaka",
"middle": [],
"last": "Ohyama",
"suffix": "",
"affiliation": {
"laboratory": "C&C Systems Research Laboratories",
"institution": "NEC Corporation",
"location": {
"addrLine": "1-1, Miyazaki 4-chome, Miyamae-ku",
"postCode": "213",
"settlement": "Kawasaki-city, Kanagawa",
"country": "Japan"
}
},
"email": ""
},
{
"first": "Toshikazu",
"middle": [],
"last": "Fukushima",
"suffix": "",
"affiliation": {
"laboratory": "C&C Systems Research Laboratories",
"institution": "NEC Corporation",
"location": {
"addrLine": "1-1, Miyazaki 4-chome, Miyamae-ku",
"postCode": "213",
"settlement": "Kawasaki-city, Kanagawa",
"country": "Japan"
}
},
"email": ""
},
{
"first": "Tomoki",
"middle": [],
"last": "Shutoh",
"suffix": "",
"affiliation": {
"laboratory": "C&C Systems Research Laboratories",
"institution": "NEC Corporation",
"location": {
"addrLine": "1-1, Miyazaki 4-chome, Miyamae-ku",
"postCode": "213",
"settlement": "Kawasaki-city, Kanagawa",
"country": "Japan"
}
},
"email": ""
},
{
"first": "Masamichi",
"middle": [],
"last": "Shutoh",
"suffix": "",
"affiliation": {
"laboratory": "C&C Systems Research Laboratories",
"institution": "NEC Corporation",
"location": {
"addrLine": "1-1, Miyazaki 4-chome, Miyamae-ku",
"postCode": "213",
"settlement": "Kawasaki-city, Kanagawa",
"country": "Japan"
}
},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "The following proposal is for a Japanese sentence analysis method to be used in a Japanese book reading machine. This method is designed to allow for several candidates in case of ambiguous characters. Each sentence is analyzed to compose a data structure by defining the relationship between words and phrases. This structure (named network structure) involves all possible combinations of syntactically collect phrases. After network structure has been completed, heuristic rules are applied in order to determine the most probable way to arrange the phrases and thus organize the best sentence. All information about each sentence ~ the pronunciation of each word with its accent and the structure of phrases ~ will be used during speech synthesis. Experiment results reveal: 99.1% of all characters were given their correct pronunciation. Using several recognized character candidates is more efficient than only using first ranked characters as the input for sentence analysis. Also this facility increases the efficiency of the book reading machine in that it enables the user to select other ways to organize sentences.",
"pdf_parse": {
"paper_id": "P86-1024",
"_pdf_hash": "",
"abstract": [
{
"text": "The following proposal is for a Japanese sentence analysis method to be used in a Japanese book reading machine. This method is designed to allow for several candidates in case of ambiguous characters. Each sentence is analyzed to compose a data structure by defining the relationship between words and phrases. This structure (named network structure) involves all possible combinations of syntactically collect phrases. After network structure has been completed, heuristic rules are applied in order to determine the most probable way to arrange the phrases and thus organize the best sentence. All information about each sentence ~ the pronunciation of each word with its accent and the structure of phrases ~ will be used during speech synthesis. Experiment results reveal: 99.1% of all characters were given their correct pronunciation. Using several recognized character candidates is more efficient than only using first ranked characters as the input for sentence analysis. Also this facility increases the efficiency of the book reading machine in that it enables the user to select other ways to organize sentences.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "English text-to-speech conversion technology has substantially progressed through massive research ( e.g., Allen 1973 Allen , 1976 Allen , 1986 Klatt 1982 Klatt , 1986 . A book reading machine for the blind is a typical use for text-tospeech technology in the welfare field ( Allen 1973 ) . According to the Kurzweil Reading Machine Update ( 1985 ) , the Machine is in use by thousands of people in over 500 locations worldwide.",
"cite_spans": [
{
"start": 107,
"end": 117,
"text": "Allen 1973",
"ref_id": "BIBREF3"
},
{
"start": 118,
"end": 130,
"text": "Allen , 1976",
"ref_id": "BIBREF2"
},
{
"start": 131,
"end": 143,
"text": "Allen , 1986",
"ref_id": "BIBREF0"
},
{
"start": 144,
"end": 154,
"text": "Klatt 1982",
"ref_id": "BIBREF6"
},
{
"start": 155,
"end": 167,
"text": "Klatt , 1986",
"ref_id": "BIBREF5"
},
{
"start": 274,
"end": 288,
"text": "( Allen 1973 )",
"ref_id": "BIBREF3"
},
{
"start": 340,
"end": 348,
"text": "( 1985 )",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "I. Introduction",
"sec_num": null
},
{
"text": "In the case of Japanese, however, due to the complexities of the language, Japanese text-to-speech conversion technology hasn't progressed as fast as that of English. Recently a Japanese text-to-speech synthesizer has been introduced ( Kabeya et al. 1985 ) . However, this synthesizer accepts only Japanese character code strings and doesn't include the character recognition facility.",
"cite_spans": [
{
"start": 234,
"end": 256,
"text": "( Kabeya et al. 1985 )",
"ref_id": "BIBREF4"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "I. Introduction",
"sec_num": null
},
{
"text": "Since 1982, the authors have been engaged in the research and development of a Japanese sentence analysis method to be used in a book reading machine for the blind. The first version of the Japanese book reading machine, which is aimed to exarnine algorithms and its performance, has developed in 1984 ( Tsuji and Asai 1985; Tsukurno and Asai 1985; Fukushima et al. 1985; Fushikida 1985, 1986 ) . Figure 1 shows the book reading process of the machine. A pocket-size book is first scanned, then each character on the page is detected and recognized. Sentence analysis ( parsing ) is accomplished by using character recognition result. Finally, synthesized speech is generated. The speech can be recorded for future use. The pages will turn automatically. The Japanese sentence analysis method that the authors have developed has two functions: One, to choose an appropriate character among several input character candidates when the character recognition result is ambiguous. Two, to convert the written character strings into phonetic symbols. The written character strings are made up Kanji ( Chinese } characters and kana ( Japanese consonant-vowel combination ) characters. These phonetic symbols depict both the pronunciation and accent of each word. The structure of the phrases is also obtained in order to determine the pause positions and intonation.",
"cite_spans": [
{
"start": 302,
"end": 324,
"text": "( Tsuji and Asai 1985;",
"ref_id": "BIBREF11"
},
{
"start": 325,
"end": 348,
"text": "Tsukurno and Asai 1985;",
"ref_id": null
},
{
"start": 349,
"end": 371,
"text": "Fukushima et al. 1985;",
"ref_id": "BIBREF9"
},
{
"start": 372,
"end": 394,
"text": "Fushikida 1985, 1986 )",
"ref_id": null
}
],
"ref_spans": [
{
"start": 397,
"end": 405,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "I. Introduction",
"sec_num": null
},
{
"text": "After briefly describing the difficulty of Japanese sentence analysis technology compared to that of English, this paper will outline the Japanese sentence analysis method, as well as experimental results.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "I. Introduction",
"sec_num": null
},
{
"text": "In this section, the difficulty of Japanese sentence analysis is described by comparing with that of English.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Comparison of Japanese and English as Input for a Book Reading Machine",
"sec_num": "2."
},
{
"text": "Characters to Phonetic Symbols",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conversion from Written",
"sec_num": "2.1"
},
{
"text": "In English, text-to-speech conversion can be achieved by applying general rules. For exceptional words which are outside the rules, an exceptional word dictionary is used. Accentuation can be also achieved by rules and an exceptional dictionary.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conversion from Written",
"sec_num": "2.1"
},
{
"text": "Roughly speaking, Japanese text-to-speech conversion is similar to that of English. However, in case of Japanese, more diligent analysis is required. Japanese sentences are written by using Kanji characters and kana characters. Thousands of kinds of Kanji characters are generally used in Japanese sentences. And, most of the Kanji characters have several readings ( Figure 2 (a) ). On the other hand, the number of kana characters is less than one hundred. Each kana character corresponds to certain monosyllable. Therefore, in the conversion of kana characters, kana-to-phoneme conversion rules seem to be successfully applied. However, in two cases, kana characters l~ and ~', are used as Kaku-Joshi, Japanese preposition which follows a noun to form a noun phrase, then the pronunciation changes ( Figure 2 (b) }. Subsequently the reading of numerical words also changes ( Figure 2 (c)).",
"cite_spans": [],
"ref_spans": [
{
"start": 365,
"end": 379,
"text": "( Figure 2 (a)",
"ref_id": null
},
{
"start": 800,
"end": 810,
"text": "( Figure 2",
"ref_id": null
},
{
"start": 875,
"end": 885,
"text": "( Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "Conversion from Written",
"sec_num": "2.1"
},
{
"text": "As described above, the pronunciation of each character in Japanese sentences is determined by a neighbor character which combines to form a word. There are too many exceptions in Japanese to create general rules. Therefore, a large size word dictionary which covers all commonly used words is generally used to analyze Japanese sentences.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conversion from Written",
"sec_num": "2.1"
},
{
"text": "In English sentences, the boundaries between words are indicated by spaces and punctuation marks. This is quite helpful in detecting phrase structure, which is used to determinate pause positions and intonation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Required Sentence Analysis Level",
"sec_num": "2.2"
},
{
"text": "On the contrary, Japanese sentences only have punctuation marks. They don't have any spaces which indicate word boundaries, Therefore, more precise analysis is required in order to detect word boundaries at first. The structure of the sentence will be analyzed after the word detection. Examples of Japanese Word.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Required Sentence Analysis Level",
"sec_num": "2.2"
},
{
"text": "English sentences consist of twenty-six alphabet characters and other characters, such as numbers and punctuations. Because of the fewer number of the English alphabet characters, characters can be recognized accurately.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Character Recognition Accuracy",
"sec_num": "2.3"
},
{
"text": "Japanese sentences consist of thousands of Kanji characters, more than one hundred different kana characters ( two kana character sets ~ Hiragana and Katakana are used in Japanese sentences ) and alphanumeric characters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Character Recognition Accuracy",
"sec_num": "2.3"
},
{
"text": "Because of the variety of characters, even when using a well-established character recognition method, the result is sometimes ambiguous.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Character Recognition Accuracy",
"sec_num": "2.3"
},
{
"text": "The Japanese sentence analysis method has the following characteristics.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Characteristics of Sentence Analysis Method",
"sec_num": "3."
},
{
"text": "I. The mixed Kanji-kana strings are analyzed both through word extraction and syntactical examination. An internal data structure ( named network structure in this paper ), which defines the relationship of all possible words and phrases, is composed through word extraction and syntactical examination.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Characteristics of Sentence Analysis Method",
"sec_num": "3."
},
{
"text": "After network structure has been completed, heuristic rules are applied in order to determine the most probable way to arrange the phrases and thus organize a sentence.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Characteristics of Sentence Analysis Method",
"sec_num": "3."
},
{
"text": "ambiguous, several candidates per character are accepted. Unsuitable character candidates are eliminated through sentence analysis.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "When an obtained character recognition result is",
"sec_num": "2."
},
{
"text": "Sentence analysis of Japanese reads back to front between punctuation marks. For example, the analysis starts from the position of the first punctuation mark and works to the beginning of the sentence. Thus, word dictionaries and their indexes have been organized so they can be used through this sequence.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each punctuation mark is used as a delimiter.",
"sec_num": "3."
},
{
"text": "computing time to analyze unrestricted Japanese text. Therefore, it has been designed not to analyze deep sentence structure, such as semantic or pragmatic correlates.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The sentence analysis method is required for short",
"sec_num": "4."
},
{
"text": "5. By the user's request, the book reading machine can read the same sentence again and again. If the user wants to change the way of reading ( e.g. in the case that there are homographs ), the machine can also crest other ways of reading. In order to achieve this goal, several pages of sentence analysis result is kept while the machine is in use.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The sentence analysis method is required for short",
"sec_num": "4."
},
{
"text": "As shown in Figure 3 , the Japanese sentence analysis system consists of two subsystems and word dictionaries. Two subsystems are named \"network structure composition subsystem\" and \"speech information organization subsystem\", respectively. These subsystems work asynchronously. Figure 3 . Sentence Analysis System Outline.",
"cite_spans": [],
"ref_spans": [
{
"start": 12,
"end": 20,
"text": "Figure 3",
"ref_id": null
},
{
"start": 279,
"end": 287,
"text": "Figure 3",
"ref_id": null
}
],
"eq_spans": [],
"section": "Outline of Sentence Analysis System",
"sec_num": "4."
},
{
"text": "As the input, the network structure composition subsystem receives character recognition results. When the character recognition result is ambiguous, several character candidates appear. During the character recognition, the probability of each character candidate is also obtained. Figure 4 is an example of character recognition result. Figure 4 describes: The first character of the sentence as having three character candidates. The fifth and seventh characters as having two candidates. Except the fifth character, all of the first ranking character candidates are correct. However, the fifth character proves an exception with the second ranking character candidate as the desired character.",
"cite_spans": [],
"ref_spans": [
{
"start": 283,
"end": 291,
"text": "Figure 4",
"ref_id": "FIGREF1"
},
{
"start": 339,
"end": 347,
"text": "Figure 4",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "With the recognized result, the network structure composition subsystem is activated. Figure 5 describes how the recognition result ( shown in Figure 4 ) is analyzed.",
"cite_spans": [],
"ref_spans": [
{
"start": 86,
"end": 94,
"text": "Figure 5",
"ref_id": "FIGREF2"
},
{
"start": 143,
"end": 151,
"text": "Figure 4",
"ref_id": "FIGREF1"
}
],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "Through the detection of punctuation marks in the input sentence ( recognition result ), the subsystem determines the region to be analyzed. After one region has been analyzed, the next punctuation mark which determines the next region is detected. In case of Figure 5 , for example, whole data will be analyzed at once, because the first punctuation mark is located at the end of the sentence.",
"cite_spans": [],
"ref_spans": [
{
"start": 260,
"end": 269,
"text": "Figure 5",
"ref_id": "FIGREF2"
}
],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "Characters in the region are analyzed from the detected punctuation to the beginning of the sentence. The analysis is accomplished by both word extraction ;~nd syntactical examination.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "Words in dictionaries are extracted by using character strings which are obtained by combining character candidates. The type of the characters ( kana, Kanji etc. ) determines which index for the dictionaries will be used. After extracting the words, phrases are composed by combining the words.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "Using syntactical rules ( i.e. conjugation rules ), only syntactically correct phrases are composed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "Finally, by using these phrases, network structure is composed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "Network structure obtained through the analysis described in Figure 5 is shown in Figure 6 . This structure involves the following information.",
"cite_spans": [],
"ref_spans": [
{
"start": 61,
"end": 69,
"text": "Figure 5",
"ref_id": "FIGREF2"
},
{
"start": 82,
"end": 90,
"text": "Figure 6",
"ref_id": "FIGREF4"
}
],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "\u2022 hierarchical relationship between sentence, phrases and words",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "\u2022 syntactical meaning of each word",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "\u2022 pointers to the pronunciation and accent information of for each word in dictionaries",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "\u2022 pointers between phrases which are used when the user selects other ways of reading Some features of Japanese language are utilized in the network structure composition subsystem. Some examples of them are as follow.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "1. In general, a Japanese phrase consists of both an independent word and dependent words. The prefix word and/or the suffix word are sometimes adjoined. The number of dependent words is not so many as compared with independent words. It seems to be efficient to analyze dependent words first. Thus, the analysis is accomplished from the end of the region to the beginning.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Network Structure Composition Subsystem",
"sec_num": "4.1"
},
{
"text": "3.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "2.",
"sec_num": null
},
{
"text": "Independent words mostly include non-kana characters, alternately, dependent words are written in kana characters. Therefore, higher priority is given both to independent words which include a non-kana characters and to dependent words which consist of only kana characters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "2.",
"sec_num": null
},
{
"text": "The number of Kanji characters is far greater than that of kana characters. Therefore, it seems efficient to use a Kanji character as the search key to scan the dictionary indexes. These indexes are designed so that the search key must be a non-kana character in cases where there is one or more non-kana character.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "2.",
"sec_num": null
},
{
"text": "With the user's request for speech synthesis, the speech information organization subsystem is activated. This subsystem determines the best sentence ( a combination of phrases ) by examining the phrases in network structure.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speech Information Organization Subsystem",
"sec_num": "4.2"
},
{
"text": "After organizing the sentence, the information for speech synthesis is then organized. The pronunciation and accent of each word are determined by using the dictionaries. The structure of the sentence is obtained by analyzing the relationship between phrases. In case of numerical words, such as 1,234..56, a special procedure is activated to generate the reading. In case the user requests other ways of reading the sentence, the subsystem chooses other phrases in network structure, thus organizing the speech synthesis information. In order to determine the most probable phrase combination in network structure, heuristic rules axe applied.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speech Information Organization Subsystem",
"sec_num": "4.2"
},
{
"text": "The rules have been obtained mainly by experiments. Some of them are as follow.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Words",
"sec_num": null
},
{
"text": "The sentence which contains the least number of phrases will be given the highest priority.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "[11 Number of Phrases in a Sentence",
"sec_num": null
},
{
"text": "i21 Probabilities of Characters The phrase which contains more probable character candidates will be given higher priority. This probability is obtained as the result of character recognition.",
"cite_spans": [
{
"start": 4,
"end": 31,
"text": "Probabilities of Characters",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "[11 Number of Phrases in a Sentence",
"sec_num": null
},
{
"text": "Independent words written in kana characters will be given lower priority. Independent words written in one character will be also given lower priority.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "!3] Written Format of Words",
"sec_num": null
},
{
"text": "The frequently used syntactical combination will be given higher priority. ( e.g. noun-preposition combination )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "14! Syntactical Combination Appearance Frequency",
"sec_num": null
},
{
"text": "!51 Selected Phrases The phrase which once has been selected by a user will be given higher priority.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "14! Syntactical Combination Appearance Frequency",
"sec_num": null
},
{
"text": "In the case of Figure 3 , the best way of arranging phrases is determined by applying the heuristic rule [1].",
"cite_spans": [],
"ref_spans": [
{
"start": 15,
"end": 23,
"text": "Figure 3",
"ref_id": null
}
],
"eq_spans": [],
"section": "14! Syntactical Combination Appearance Frequency",
"sec_num": null
},
{
"text": "Dictionaries used in this system are the following.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Dictionaries",
"sec_num": "4.3"
},
{
"text": "(1) Independent Word Dictionary Nouns, Verbs, Adjectives, Adverbs, Conjunctions etc. 65,850 words",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Dictionaries",
"sec_num": "4.3"
},
{
"text": "(2) Each word stored in these dictionaries has the following information. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Word Dictionaries",
"sec_num": "4.3"
},
{
"text": "Some experiments have achieved in order to evaluate the sentence analysis method.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Results",
"sec_num": "5."
},
{
"text": "In this section, these experimental results are described.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Results",
"sec_num": "5."
},
{
"text": "The accuracy of pronunciation has been evaluated by counting correctly pronounced characters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "In this experiment, character code strings were used as the input data. The following two whole books are analyzed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "\u2022 Tetsugaku Annai ( Introduction to Philosophy ) by Tetsuzo Tanikawa ( an essay )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "\u2022 Touzoku Gaisha ( The Thief Company ) by Shin-ichi Hoshi ( a collection of short stories )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "As shown in Table I , 99.1% of all characters have been given their correct pronunciation. The major cases for mispronunciation are as follows.",
"cite_spans": [],
"ref_spans": [
{
"start": 12,
"end": 19,
"text": "Table I",
"ref_id": null
}
],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "(1) Unregistered words in dictionaries (l-a) uncommon words (l-b) proper nouns (l-c) uncommon written style",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "(2) Pronunciation changes in the case of compound words",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "(3) Homographs The efficiency as the postprocessing roll for character recognition has been evaluated by comparing the characters used for speech synthesis with the character recognition result. Twelve pages of character recognition results ( four pages of three books ) have been analyzed. The books used as the input data are as follow.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "\u2022 Tetsugaku Annai ( Introduction to Philosophy )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "by Tetsuzo Tanikawa ( an essay ) \u2022 Touzoku Gaisha ( The Thief Company ) by Shin-ichi Hoshi ( a collection of short stories }",
"cite_spans": [],
"ref_spans": [
{
"start": 11,
"end": 32,
"text": "Tanikawa ( an essay )",
"ref_id": null
},
{
"start": 50,
"end": 71,
"text": "( The Thief Company )",
"ref_id": null
}
],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "\u2022 Yujo ( The friendship ) by Saneatsu Mushanokouji ( a novel ) Table 2 shows scores for the character recognition result. Table 3 shows the score for characters which are' chosen as correct characters by the sentence analysis method, as well as the score for correctly pronounced characters. As shown in Tables 2 and 3, the score for correct characters obtained after the sentence analysis was 99.7%, while the score for the 1st ranking chaxacters obtained in the character recognition result was 99.5%. This experimental result reveals that the sentence analysis method is effective as a postprocessing roll of character recognition.",
"cite_spans": [],
"ref_spans": [
{
"start": 7,
"end": 25,
"text": "( The friendship )",
"ref_id": null
},
{
"start": 63,
"end": 70,
"text": "Table 2",
"ref_id": "TABREF5"
},
{
"start": 122,
"end": 129,
"text": "Table 3",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "The state of errors found during the experiment is shown in Table 4 . The difference between (b') and (b3) in Table 4 indicates the effectiveness of the sentence analysis method. The score 99.0% in Table 3 indicates the efficiency of the sentence analysis method in the book reading machine. ",
"cite_spans": [],
"ref_spans": [
{
"start": 60,
"end": 67,
"text": "Table 4",
"ref_id": "TABREF7"
},
{
"start": 110,
"end": 117,
"text": "Table 4",
"ref_id": "TABREF7"
},
{
"start": 198,
"end": 205,
"text": "Table 3",
"ref_id": "TABREF6"
}
],
"eq_spans": [],
"section": "Pronunciation Accuracy",
"sec_num": "5.1"
},
{
"text": "To examine the efficiency, an experiment has been conducted where sentences have been read both automatically and with the help of manual manipulation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Efficiency of Selection by Manual",
"sec_num": "5.3"
},
{
"text": "The same text used in Section 5.2 was used in this experiment. Table 5 shows scores for the correctly pronounced characters. As shown in Table 5 , 99.9% and 99.8~ of all characters were given correct pronunciation after the manual selection, while 99.3% and 99.0e~ of all characters had been given their correct pronunciation before the manual selection, respectively. These scores reveal that most mispronunciation could be recovered by manual selection so that nearly all accurately pronounced reading can be taped. ",
"cite_spans": [],
"ref_spans": [
{
"start": 63,
"end": 70,
"text": "Table 5",
"ref_id": "TABREF8"
},
{
"start": 137,
"end": 144,
"text": "Table 5",
"ref_id": "TABREF8"
}
],
"eq_spans": [],
"section": "Efficiency of Selection by Manual",
"sec_num": "5.3"
},
{
"text": "A sentence analysis method used in a Japanese book reading machine has been described. Input sentences, where each character is allowed to have other candidates, are analyzed by using several word dictionaries, as well as employing syntactical examinations. After generating network structure, heuristic rules are applied in order to determine the most desirable sentence used for speech information generation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6."
},
{
"text": "The results of experiments reveal: 99.1% of all characters used in two whole books have been correctly converted to their pronunciation. Even when the character recognition result is ambiguous, correct characters can often be chosen by the sentence analysis method. By manual selection, most incorrect characters can be corrected.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6."
},
{
"text": "Currently, the authors are improving the sentence analysis method including 'the heuristic rules and the contents of dictionaries through book reading experiments and data examinations. This work is, needless to say, aimed in offering better quality speech to the blind users in a short.computing time. Authors are expecting that their efforts will contribute to the welfare field.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "6."
}
],
"back_matter": [
{
"text": "The authors would like to express their appreciation to Mr. S. Hanaki for his constant encouragement and effective advice. The authors would also like to express their appreciation to Ms. A. Ohtake for her enthusiasm and cooperation throughout the research.This research has been accomplished as the research project \"Book-Reader for the Blind', which is one project of The National Research and Development Program for Medical and Welfare Apparatus, Agency of Industrial Science and Technology, Ministry of International Trade and Industry.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "ACKNOWLEDGEMENTS",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "From Text to Speech: The MITalk System",
"authors": [
{
"first": "J",
"middle": [],
"last": "Allen",
"suffix": ""
}
],
"year": 1986,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Allen, J., ed., 1986 From Text to Speech: The MITalk System. Cambridge University Press.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Speech Synthesis from Unrestricted Text",
"authors": [
{
"first": "J",
"middle": [],
"last": "Allen",
"suffix": ""
}
],
"year": 1985,
"venue": "Computer Speech Processing",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Allen, J. 1985 Speech Synthesis from Unrestricted Text. In Fallside, F. and Woods, W.A., eds., Computer Speech Processing. Prentice-Hall.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Synthesis of Speech from Unrestricted Text",
"authors": [
{
"first": "J",
"middle": [],
"last": "Allen",
"suffix": ""
}
],
"year": 1976,
"venue": "Proc. IEEE",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Allen, J. 1976 Synthesis of Speech from Unrestricted Text. Proc. IEEE, 64.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Reading Machine for the Blind: The Technical Problems and the Methods Adopted for Their Solution",
"authors": [
{
"first": "J",
"middle": [],
"last": "Allen",
"suffix": ""
}
],
"year": 1973,
"venue": "IEEE Trans., AU",
"volume": "",
"issue": "3",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Allen, J. 1973 Reading Machine for the Blind: The Technical Problems and the Methods Adopted for Their Solution. IEEE Trans., AU-21(3).",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "A Japanese Text-To-Speech Synthesizer",
"authors": [
{
"first": "K",
"middle": [],
"last": "Kabeya",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Hakoda",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Ishikawa",
"suffix": ""
}
],
"year": 1985,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kabeya, K.; Hakoda, K.; and Ishikawa, K. 1985 A Japanese Text-To-Speech Synthesizer. Proe. A VIOS '85.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Text to Speech: Present and Future",
"authors": [
{
"first": "D",
"middle": [
"H"
],
"last": "Klatt",
"suffix": ""
}
],
"year": 1986,
"venue": "Proe. Speech Tech",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Klatt, D.H. 1986 Text to Speech: Present and Future. Proe. Speech Tech '86.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "The Klattalk Text-to-Speech System. Proe. ICASSP '8Z",
"authors": [
{
"first": "D",
"middle": [
"H"
],
"last": "Klatt",
"suffix": ""
}
],
"year": 1982,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Klatt, D.H. 1982 The Klattalk Text-to-Speech System. Proe. ICASSP '8Z.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Japanese Speech Synthesis System in a Book Reader for the Blind",
"authors": [
{
"first": "",
"middle": [
"Y"
],
"last": "Mitome",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Fushikida",
"suffix": ""
}
],
"year": 1986,
"venue": "Proc. ICASSP '86",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mitome. Y. and Fushikida, K. 1986 Japanese Speech Synthesis System in a Book Reader for the Blind. Proc. ICASSP '86.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Kurzweil Reading Machine Update",
"authors": [],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kurzweil Reading Machine Update.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "A sentence analysis method for Japanese text-to-speech conversion in the Japanese book reading machine for the 51ind",
"authors": [
{
"first": "Kurzweil",
"middle": [],
"last": "Computer Products",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "Fukushima",
"suffix": ""
},
{
"first": "Y",
"middle": [],
"last": "Ohyama",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Ohtake",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "Shutoh",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Shutoh",
"suffix": ""
}
],
"year": 1985,
"venue": "Inf. Process. Soc. Jpn",
"volume": "",
"issue": "",
"pages": "2--4",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kurzweil Computer Products. << in Japanese >> Fukushima, T.; Ohyama, Y.; Ohtake, A.; Shutoh, T; and Shutoh, M. 1985 A sentence analysis method for Japanese text-to-speech conversion in the Japanese book reading machine for the 51ind. WG preprint, Inf. Process. Soc. Jpn., WGJDP 2-4.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Japanese Speech Synthesis by Rule using Formant-CV, Speech Compilation Method",
"authors": [
{
"first": "Y",
"middle": [],
"last": "Mitome",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Fushikida",
"suffix": ""
}
],
"year": 1985,
"venue": "Trans. Committee on Speech Res., Acoust. Soc. Jpn",
"volume": "",
"issue": "",
"pages": "85--116",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mitome, Y. and Fushikida, K. 1985 Japanese Speech Synthesis by Rule using Formant-CV, Speech Compilation Method. Trans. Committee on Speech Res., Acoust. Soc. Jpn., $85-31.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Document Image Analysis, based upon Split Detection Method",
"authors": [
{
"first": "Y",
"middle": [],
"last": "Tsuji",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Asai",
"suffix": ""
}
],
"year": 1985,
"venue": "IECE Jpn",
"volume": "",
"issue": "",
"pages": "85--102",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tsuji, Y. and Asai, K. 1985 Document Image Analysis, based upon Split Detection Method. Tech. Rep., IECE Jpn., PRL85-17.",
"links": null
},
"BIBREF12": {
"ref_id": "b12",
"title": "Machine Printed Chinese Character Recognition by Improved Loci Features",
"authors": [
{
"first": "J",
"middle": [],
"last": "Tsukumo",
"suffix": ""
},
{
"first": "K",
"middle": [],
"last": "Asai",
"suffix": ""
}
],
"year": 1985,
"venue": "IECE Jpn",
"volume": "",
"issue": "",
"pages": "85--102",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Tsukumo, J. and Asai, K. 1985 Machine Printed Chinese Character Recognition by Improved Loci Features. Tech. Rcp., IECE Jpn., PRL85-17.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"uris": null,
"type_str": "figure",
"text": "I. The Book Reading Machine Outline.",
"num": null
},
"FIGREF1": {
"uris": null,
"type_str": "figure",
"text": "Character Recognition Result Example.",
"num": null
},
"FIGREF2": {
"uris": null,
"type_str": "figure",
"text": "Sentence Analysis Example.",
"num": null
},
"FIGREF4": {
"uris": null,
"type_str": "figure",
"text": "Network Structure Example.",
"num": null
},
"FIGREF5": {
"uris": null,
"type_str": "figure",
"text": "(a) written mixed Kanji-kana string (first-choice) (b) syntactical meaning (c) pronunciation (d) accent position Items (a) and (b) of all words are gathered to form the following four indexes. * Kana Independent Word Index * Kana Dependent Words and Kana Suffix Word Index * Non-Kana Word Index * Prefix Word Index These indexes are used by the network structure composition subsystem. Items (c) and (d) are used by the speech information organization subsystem.",
"num": null
},
"FIGREF6": {
"uris": null,
"type_str": "figure",
"text": "as the Postprocessing Roll for Character Recognition",
"num": null
},
"TABREF1": {
"content": "<table><tr><td/><td/><td>Speech Information</td><td>,Speech</td></tr><tr><td/><td/><td>Organization Subsystem</td><td>Information</td></tr><tr><td/><td>Network</td><td/></tr><tr><td/><td>Structure</td><td/></tr><tr><td>I</td><td>Indexes</td><td>Contents</td></tr><tr><td/><td colspan=\"2\">Word Dictionaries</td></tr></table>",
"num": null,
"text": "",
"type_str": "table",
"html": null
},
"TABREF4": {
"content": "<table><tr><td>Total Characters</td><td>128,289 (100%)</td></tr><tr><td>Correct Characters</td><td>127,108 (99.1%)</td></tr></table>",
"num": null,
"text": "Score for Correct Pronunciation.",
"type_str": "table",
"html": null
},
"TABREF5": {
"content": "<table><tr><td>Total Characters</td><td>6,793 (100%)</td></tr><tr><td>Correct Characters</td><td>6,757 (99.5%)</td></tr><tr><td>( at 1st Ranking )</td><td/></tr><tr><td>Correct Characters</td><td>6,7s3 (99.9%)</td></tr><tr><td>( in 1st to 5th Ranking )</td><td/></tr></table>",
"num": null,
"text": "Character Recognition Result.",
"type_str": "table",
"html": null
},
"TABREF6": {
"content": "<table><tr><td>Total Characters</td><td>6,793 (100%)</td></tr><tr><td>Characters Treated as</td><td>6,772 (99.7%)</td></tr><tr><td>Correct Characters</td><td/></tr><tr><td>Characters Correctly</td><td>6,72s (99.0%)</td></tr><tr><td>Pronounced</td><td/></tr></table>",
"num": null,
"text": "Scores after Sentence Analysis.",
"type_str": "table",
"html": null
},
"TABREF7": {
"content": "<table><tr><td colspan=\"2\"><< Character Recognition Error >></td><td/><td/></tr><tr><td>Ca)</td><td>1st Ranking Chars are Incorrect</td><td>36</td><td/></tr><tr><td>(al)</td><td>Correct Chars in 2nd-5th</td><td/><td>26</td></tr><tr><td>(a2)</td><td>Not among Candidates</td><td/><td>10</td></tr><tr><td colspan=\"2\"><< Sentence Analysis Error >></td><td/><td/></tr><tr><td>(b) (bl) (b2) (b3)</td><td>Total Incorrect Char Incorrect Chars among (al) Incorrect Chars among (a2) Incorrect Chars While Char</td><td>21</td><td>4 10 7</td></tr><tr><td/><td>Recognition was Correct</td><td/><td/></tr><tr><td>(b')</td><td>Correct Chars While the 1st</td><td>22</td><td/></tr><tr><td/><td>Ranking Chars were Incorrect</td><td/><td/></tr><tr><td/><td>( b' = al -bl</td><td/><td/></tr></table>",
"num": null,
"text": "State of Errors.",
"type_str": "table",
"html": null
},
"TABREF8": {
"content": "<table><tr><td>Total Characters</td><td>6,793 (100\u00b0~)</td></tr><tr><td colspan=\"2\"><< Input Data is Correct Characters >></td></tr><tr><td>Before Selection</td><td>6,745 (99.3%)</td></tr><tr><td>After Selection</td><td>6,787 (99.9%)</td></tr><tr><td colspan=\"2\"><< Input Data is Recognized Characters >></td></tr><tr><td>Before Selection</td><td>6,728 (99.0\u00b0~)</td></tr><tr><td>After Selection</td><td>6,777 (99.8\u00b0~)</td></tr></table>",
"num": null,
"text": "Scores for Characters.",
"type_str": "table",
"html": null
}
}
}
} |